Skip to content

Commit f9a3b80

Browse files
authored
Update README.md - added note for MOE LLM deployments compatibility with ai inference sdk (#39038)
* Update README.md added note for MOE LLM deployments. * Fix typo in Managed Compute Endpoints description
1 parent e989390 commit f9a3b80

File tree

1 file changed

+1
-0
lines changed

1 file changed

+1
-0
lines changed

sdk/ai/azure-ai-inference/samples/README.md

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -42,6 +42,7 @@ See [Prerequisites](https://github.com/Azure/azure-sdk-for-python/blob/main/sdk/
4242
To construct any of the clients, you will need to pass in the endpoint URL. If you are using key authentication, you also need to pass in the key associated with your deployed AI model.
4343

4444
* For Serverless API and Managed Compute endpoints, the endpoint URL has the form `https://your-unique-resouce-name.your-azure-region.models.ai.azure.com`, where `your-unique-resource-name` is your globally unique Azure resource name and `your-azure-region` is the Azure region where the model is deployed (e.g. `eastus2`).
45+
* For Managed Compute Endpoints, do not include the inference path (e.g. `/score`) in endpoint URL.
4546

4647
* For Azure OpenAI endpoints, the endpoint URL has the form `https://your-unique-resouce-name.openai.azure.com/openai/deployments/your-deployment-name`, where `your-unique-resource-name` is your globally unique Azure OpenAI resource name, and `your-deployment-name` is your AI Model deployment name.
4748

0 commit comments

Comments
 (0)