You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: doc/source/models/builtin/llm/index.rst
+8-1Lines changed: 8 additions & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -366,6 +366,11 @@ The following is a list of built-in LLM in Xinference:
366
366
- 32768
367
367
- MiniCPM4 series are highly efficient large language models (LLMs) designed explicitly for end-side devices, which achieves this efficiency through systematic innovation in four key dimensions: model architecture, training data, training algorithms, and inference systems.
368
368
369
+
* - :ref:`minimax-m2 <models_llm_minimax-m2>`
370
+
- chat, tools, reasoning
371
+
- 196608
372
+
- MiniMax-M2, a Mini model built for Max coding & agentic workflows.
@@ -534,7 +539,7 @@ The following is a list of built-in LLM in Xinference:
534
539
* - :ref:`qwen3 <models_llm_qwen3>`
535
540
- chat, reasoning, hybrid, tools
536
541
- 40960
537
-
- Qwen3 is the latest generation of large language models in Qwen series, offering a comprehensive suite of dense and mixture-of-experts (MoE) models. Built upon extensive training, Qwen3 delivers groundbreaking advancements in reasoning, instruction-following, agent capabilities, and multilingual support
542
+
- Qwen3 is the latest generation of large language models in Qwen series, offering a comprehensive suite of dense and mixture-of-experts (MoE) models. Built upon extensive training, Qwen3 delivers groundbreaking advancements in reasoning, instruction-following, agent capabilities, and multilingual support.
538
543
539
544
* - :ref:`qwen3-coder <models_llm_qwen3-coder>`
540
545
- chat, tools
@@ -846,6 +851,8 @@ The following is a list of built-in LLM in Xinference:
0 commit comments