Skip to content

Refresh AITK Model Catalog Page #8462

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
wants to merge 4 commits into
base: main
Choose a base branch
from
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
3 changes: 3 additions & 0 deletions docs/intelligentapps/images/models/custom_1.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
3 changes: 3 additions & 0 deletions docs/intelligentapps/images/models/custom_2.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
3 changes: 3 additions & 0 deletions docs/intelligentapps/images/models/custom_3.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
3 changes: 3 additions & 0 deletions docs/intelligentapps/images/models/foundry_models.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
3 changes: 3 additions & 0 deletions docs/intelligentapps/images/models/model_operation.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
3 changes: 3 additions & 0 deletions docs/intelligentapps/images/models/models.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
148 changes: 106 additions & 42 deletions docs/intelligentapps/models.md
Original file line number Diff line number Diff line change
@@ -1,89 +1,106 @@
---
ContentId: 52ad40fe-f352-4e16-a075-7a9606c5df3b
DateApproved: 12/11/2024
DateApproved: 06/14/2025
MetaDescription: Find a popular generative AI model by publisher and source. Bring your own model that is hosted with a URL, or select an Ollama model.
---
# Models in AI Toolkit
# Explore models in AI Toolkit

AI Toolkit supports a broad range of generative AI models. Both Small Language Models (SLM) and Large Language Models (LLM) are supported.
AI Toolkit provides comprehensive support for a wide variety of generative AI models, including both Small Language Models (SLMs) and Large Language Models (LLMs).

In the model catalog, you can access models from various sources:
Within the model catalog, you can explore and utilize models from multiple hosting sources:

- GitHub-hosted models (Llama3, Phi-3, Mistral models)
- Publisher-hosted models (OpenAI ChatGPT models, Anthropic Claude, Google Gemini)
- Locally downloaded models, for example from HuggingFace
- Locally running Ollama models
- Connect to Bring-Your-Own-Models
- Models hosted on GitHub, such as Llama3, Phi-3, and Mistral.
- Models from Azure AI Foundry, such as OpenAI's GPT-4, Grok by xAI, and many open-source models.
- Models provided directly by publishers, including OpenAI's ChatGPT, Anthropic's Claude, and Google's Gemini.
- Models downloaded locally from repositories like Ollama and ONNX.
- Custom self-hosted or externally deployed models accessible via Bring-Your-Own-Model (BYOM) integration.

![AI Toolkit model catalog displaying various generative AI models](./images/models/models.png)

## Find a model

To find a model in the model catalog:

1. Select the AI Toolkit view in the Activity Bar

1. Select **CATALOG** > **Models** to open the model catalog

![Select model in model catalog](./images/models/model_catalog.png)

Select a model card in the model catalog to view more details of the selected model.

1. Select **MODELS** > **Catalog** to open the model catalog
1. Use the filters to reduce the list of available models

- **Hosted by**: AI Toolkit supports GitHub, ONNX, OpenAI, Anthropic, Google as model hosting sources.

- **Publisher**: The publisher for AI models, such as Microsoft, Meta, Google, OpenAI, Anthropic, Mistral AI, and more.

- **Tasks**: Currently, only `Text Generation` is supported.
- **Feature**: Supported features of the model, such as `Text Attachment`, `Image Attachment`, `Web Search`, `Structured Outputs`, and more.

- **Model type**: Filter models that can run remotely or locally on CPU, GPU, or NPU. This filter depends on the local availability.

- **Fine-tuning Support**: Show models that can be used to run fine-tuning.
1. Browse the models in different categories, such as:
- **Popular Models** is a curated list of widely used models across various tasks and domains.
- **Azure AI Foundry Hosted Models** includes models hosted on Azure AI Foundry, such as OpenAI's GPT-4, Grok by xAI, and many open-source models.
- **GitHub Models** provide easy access to popular models hosted on GitHub. It's best for fast prototyping and experimentation.
- **ONNX Models** are optimized for local execution and can run on CPU, GPU, or NPU.
- **Ollama Models** are popular models that can run locally with Ollama, supporting CPU via GGUF quantization.
1. Alternatively, use the search box to find a specific model by name or description

To reference a self-hosted model or locally-running Ollama model:
## Add a model from the catalog
To add a model from the model catalog:
1. Locate the model you want to add in the model catalog
1. Select the **Add** button on the model card
1. The flow for adding models will be slightly different based on the providers:

1. Select **+ Add model** in the model catalog
- **GitHub**: AI Toolkit will ask for your GitHub credentials to access the model repository. Once authenticated, the model will be added directly into AI Toolkit.
- **Azure AI Foundry**: AI Toolkit will prompt you to sign in with your Azure account and set a default project in Azure AI Foundry. Once authenticated, you can deploy a model to your Azure AI Foundry project and use it in AI Toolkit.
- **ONNX**: The model is downloaded from ONNX and added to AI Toolkit.
- **Ollama**: The model is downloaded from Ollama and added to AI Toolkit.

1. Choose between Ollama or a custom model in the model Quick Pick
> [!TIP]
> You can edit the API key later by right clicking the model and selecting **Edit** and view the excrypted value in `${HOME}/.aikt/models/my-models/yml` file.

1. Provide details to add the model
![AI Toolkit interface showing a model card with options Try in Playground, Download, and Load in Playground.](./images/models/model_operation.png)

## License and sign-in
- **OpenAI**, **Anthropic**, and **Google**: AI Toolkit will prompt you to enter the API Key.
- **Custom models**: Refer to the [Add a custom model](#add-a-custom-model) section for detailed instructions.

Some models require a publisher or hosting-service license and account to sign-in. In that case, before you can run the model in the [model playground](/docs/intelligentapps/playground.md), you are prompted to provide this information.
Once added, the model will appear under **MY MODELS** in the tree view, and you can use it in the [**Playground**](/docs/intelligentapps/playground.md) or [**Agent Builder**](/docs/intelligentapps/agentbuilder.md).

## Select a model for testing
## Deploy a model to Azure AI Foundry
If you want to deploy a model to Azure AI Foundry, you can do so directly from the model catalog. This allows you to run the model in your Azure AI Foundry project and use it in AI Toolkit.
1. Locate the model you want to deploy in the model catalog
1. Select the **Deploy to Azure** button on the model card
1. AI Toolkit will prompt you to sign in with your Azure account and set a default project in Azure AI Foundry
1. Once authenticated, you can deploy the model to your Azure AI Foundry project
1. After deployment, the model will be available in your Azure AI Foundry project and can be used in AI Toolkit

AI Toolkit enables you to test run a model in the playground for chat completions. You have different options, available through the actions on the model card in the model catalog.
![AI Toolkit interface showing a model card with options Deploy to Azure, Try in Playground, and Load in Playground.](./images/models/foundry_models.png)

- **Try in Playground**: load the selected model for testing in the playground without downloading it
- **Download**: download the model from a source like Hugging Face
- **Load in Playground**: load a downloaded model into the playground for chat
## Add a custom model
You can also add your own models that are hosted externally or run locally. There are several options available:
- Add Ollama models from the Ollama library or custom Ollama endpoints.
- Add custom models that has an OpenAI compatible endpoint, such as a self-hosted model or a model running on a cloud service.
- Add custom ONNX models, such as those from Hugging Face, using AI Toolkit's [model conversion tool](/docs/intelligentapps/modelconversion.md).

## Bring your own models
There are several entrypoints to add models to AI Toolkit:
- From **MY MODELS** in the tree view, hover over it and select the `+` icon.
![AI Toolkit interface showing the Model Catalog toolbar with the + Add model button highlighted, indicating where users can click to add a new custom model.](./images/models/custom_1.png)

AI Toolkit's playground also supports remote models. If you have a self-hosted or deployed model that is accessible from the internet, you can add it to AI Toolkit and use it in the playground.
- From the **Model Catalog**, select the **+ Add model** button from the tool bar.
![AI Toolkit interface showing the Model Catalog toolbar with the + Add model button highlighted. The toolbar is located at the top of the catalog view, and the + Add model button is emphasized to indicate where users can click to add a new custom model.](./images/models/custom_2.png)

1. Hover over **MY MODELS** in the tree view, and select the `+` icon to add a remote model into AI Toolkit.
1. Fill in the requested information, such as model name, display name, model hosting URL, and optional auth string.
- From the **Add Custom Models** section in the model catalog, select **+ Add Your Own Model**.
![AI Toolkit interface showing the Custom Models section in the model catalog. The + Add model button is highlighted, indicating where users can click to add a new custom model.](./images/models/custom_3.png)

![Bring Your Own Models](./images/models/byom.png)

## Add Ollama models
### Add Ollama models

Ollama enables many popular genAI models to run locally with CPU via GGUF quantization. If you have Ollama installed on your local machine with downloaded Ollama models, you can add them to AI Toolkit for use in the model playground.

### Prerequisites
Prerequisites for using Ollama models in AI Toolkit:

- AI Toolkit v0.6.2 or newer.
- [Ollama](https://ollama.com/download) (Tested on Ollama v0.4.1)

### Add local Ollama into AI Toolkit

1. Hover over **MY MODELS** in the tree view and select the "+" icon to add a model
To add local Ollama into AI Toolkit

Alternatively, select the **+ Add model** button in the model catalog or playground.

1. Select **Add an Ollama model**
1. From one of the entrypoints mentioned above, select **Add Ollama Model**.

![Select model type to add](./images/models/select-type.png)

Expand All @@ -100,3 +117,50 @@ Ollama enables many popular genAI models to run locally with CPU via GGUF quanti

> [!NOTE]
> Attachment is not support yet for Ollama models. Since we connect to Ollama using its [OpenAI compatible endpoint](https://github.com/ollama/ollama/blob/main/docs/openai.md) and it doesn't support attachments yet.

### Add a custom model with OpenAI compatible endpoint

If you have a self-hosted or deployed model that is accessible from the internet with an OpenAI compatible endpoint, you can add it to AI Toolkit and use it in the playground.

1. From one of the entry points above, select **Add Custom Model**.
1. Enter the OpenAI compatible endpoint URL and the required information.

To add a self-hosted or locally running Ollama model:

1. Select **+ Add model** in the model catalog.
1. In the model Quick Pick, choose **Ollama** or **Custom model**.
1. Enter the required details to add the model.

### Add a custom ONNX model

To add a custom ONNX model, first convert it to the AI Toolkit model format using the [model conversion tool](/docs/intelligentapps/modelconversion.md). After conversion, add the model to AI Toolkit.

## Select a model for testing

You can test a model in the playground for chat completions.

Use the actions on the model card in the model catalog:

- **Try in Playground**: Load the selected model for testing in the [Playground](/docs/intelligentapps/playground.md).
- **Try in Agent Builder**: Load the selected model in the [Agent Builder](/docs/intelligentapps/agentbuilder.md) to build AI agents.

## Manage models
You can manage your models in the **MY MODELS** section of the AI Toolkit view. Here you can:
- View the list of models you have added to AI Toolkit.
- Right-click on a model to access options such as:
- **Load in Playground**: Load the model in the [Playground](/docs/intelligentapps/playground.md) for testing.
- **Copy Model Name**: Copy the model name to the clipboard for use in other contexts, such as your code integration.
- **Refresh**: Refresh the model configuration to ensure you have the latest settings.
- **Edit**: Modify the model settings, such as the API key or endpoint.
- **Delete**: Remove the model from AI Toolkit.
- **About this Model**: View detailed information about the model, including its publisher, source, and supported features.

- Right-click on `ONNX` section title to access options such as:
- **Start Server**: Start the ONNX server to run ONNX models locally.
- **Stop Server**: Stop the ONNX server if it is running.
- **Copy Endpoint**: Copy the ONNX server endpoint to the clipboard for use in other contexts, such as your code integration.

## License and sign-in

Some models require a publisher or hosting-service license and account to sign-in. In that case, before you can run the model in the [model playground](/docs/intelligentapps/playground.md), you are prompted to provide this information.