Deploy a model and extract entities from text using the runtime API.

Once you are satisfied with how your model performs, it is ready to be deployed, and used to recognize entities in text. You can only send entity recognition tasks through the API, not from Language Studio.

Prerequisites

See the application development lifecycle for more information.

Deploy your model

Deploying a model hosts it, and makes it available for predictions through an endpoint.

When a model is deployed, you will be able to test the model directly in the portal or by calling the API associated with it.

Note

You can only have ten deployment names

  1. Go to your project in Language studio.

  2. From the left panel, select Deploy model.

  3. Click on Add deployment to submit a new deployment job.

    A screenshot showing the deployment button

  4. In the window that appears, you can create a new deployment name or override an existing one. Then, you can add a trained model to this deployment name.

    A screenshot showing the deployment screen

Delete deployment

To delete a deployment, select the deployment you want to delete and select Delete deployment

Tip

You can test your model in Language Studio by sending samples of text for it to classify.

  1. Select Test model from the menu on the left side of your project in Language Studio.
  2. Select the model you want to test.
  3. Add your text to the textbox, you can also upload a .txt file.
  4. Click on Run the test.
  5. In the Result tab, you can see the extracted entities from your text. You can also view the JSON response under the JSON tab.

Send an entity recognition request to your model

Using Language studio

  1. After the deployment is completed, select the model you want to use and from the top menu click on Get prediction URL and copy the URL and body.

    run-inference

  2. In the window that appears, under the Submit pivot, copy the sample request into your command line

  3. Replace <YOUR_DOCUMENT_HERE> with the actual text you want to classify.

    run-inference-2

  4. Submit the request

  5. In the response header you receive extract jobId from operation-location, which has the format: {YOUR-ENDPOINT}/text/analytics/v3.2-preview.2/analyze/jobs/<jobId}>

  6. Copy the retrieve request and replace jobId and submit the request.

    run-inference-3

Retrieve the results of your job

  1. Select Retrieve from the same window you got the example request you got earlier and copy the sample request into a text editor.

    Screenshot showing the prediction retrieval request and URL

  2. Replace <OPERATION_ID> with the jobId from the previous step.

  3. Submit the GET cURL request in your terminal or command prompt. You'll receive a 202 response with the API results if the request was successful.