Skip to content

Commit 22cebd1

Browse files
qinxuyezhcn000000
authored andcommitted
FEAT: [model] support GLM-4.5 series (xorbitsai#3882)
1 parent 7bb2f3f commit 22cebd1

File tree

14 files changed

+328
-39
lines changed

14 files changed

+328
-39
lines changed

doc/source/models/builtin/llm/codegeex4.rst

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -21,8 +21,8 @@ Model Spec 1 (pytorch, 9 Billion)
2121
- **Model Size (in billions):** 9
2222
- **Quantizations:** none
2323
- **Engines**: vLLM, Transformers
24-
- **Model ID:** THUDM/codegeex4-all-9b
25-
- **Model Hubs**: `Hugging Face <https://huggingface.co/THUDM/codegeex4-all-9b>`__, `ModelScope <https://modelscope.cn/models/ZhipuAI/codegeex4-all-9b>`__
24+
- **Model ID:** zai-org/codegeex4-all-9b
25+
- **Model Hubs**: `Hugging Face <https://huggingface.co/zai-org/codegeex4-all-9b>`__, `ModelScope <https://modelscope.cn/models/ZhipuAI/codegeex4-all-9b>`__
2626

2727
Execute the following command to launch the model, remember to replace ``${quantization}`` with your
2828
chosen quantization method from the options listed above::
@@ -37,8 +37,8 @@ Model Spec 2 (ggufv2, 9 Billion)
3737
- **Model Size (in billions):** 9
3838
- **Quantizations:** IQ2_M, IQ3_M, Q4_K_M, Q5_K_M, Q6_K_L, Q8_0
3939
- **Engines**: vLLM, llama.cpp
40-
- **Model ID:** THUDM/codegeex4-all-9b-GGUF
41-
- **Model Hubs**: `Hugging Face <https://huggingface.co/THUDM/codegeex4-all-9b-GGUF>`__, `ModelScope <https://modelscope.cn/models/ZhipuAI/codegeex4-all-9b-GGUF>`__
40+
- **Model ID:** zai-org/codegeex4-all-9b-GGUF
41+
- **Model Hubs**: `Hugging Face <https://huggingface.co/zai-org/codegeex4-all-9b-GGUF>`__, `ModelScope <https://modelscope.cn/models/ZhipuAI/codegeex4-all-9b-GGUF>`__
4242

4343
Execute the following command to launch the model, remember to replace ``${quantization}`` with your
4444
chosen quantization method from the options listed above::

doc/source/models/builtin/llm/cogagent.rst

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -21,8 +21,8 @@ Model Spec 1 (pytorch, 9 Billion)
2121
- **Model Size (in billions):** 9
2222
- **Quantizations:** none
2323
- **Engines**: Transformers
24-
- **Model ID:** THUDM/cogagent-9b-20241220
25-
- **Model Hubs**: `Hugging Face <https://huggingface.co/THUDM/cogagent-9b-20241220>`__, `ModelScope <https://modelscope.cn/models/ZhipuAI/cogagent-9b-20241220>`__
24+
- **Model ID:** zai-org/cogagent-9b-20241220
25+
- **Model Hubs**: `Hugging Face <https://huggingface.co/zai-org/cogagent-9b-20241220>`__, `ModelScope <https://modelscope.cn/models/ZhipuAI/cogagent-9b-20241220>`__
2626

2727
Execute the following command to launch the model, remember to replace ``${quantization}`` with your
2828
chosen quantization method from the options listed above::

doc/source/models/builtin/llm/deepseek-v3-0324.rst

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -45,6 +45,7 @@ chosen quantization method from the options listed above::
4545

4646
xinference launch --model-engine ${engine} --model-name deepseek-v3-0324 --size-in-billions 671 --model-format awq --quantization ${quantization}
4747

48+
4849
Model Spec 3 (mlx, 671 Billion)
4950
++++++++++++++++++++++++++++++++++++++++
5051

@@ -59,3 +60,4 @@ Execute the following command to launch the model, remember to replace ``${quant
5960
chosen quantization method from the options listed above::
6061

6162
xinference launch --model-engine ${engine} --model-name deepseek-v3-0324 --size-in-billions 671 --model-format mlx --quantization ${quantization}
63+

doc/source/models/builtin/llm/glm-4.1v-thinking.rst

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -21,8 +21,8 @@ Model Spec 1 (pytorch, 9 Billion)
2121
- **Model Size (in billions):** 9
2222
- **Quantizations:** none
2323
- **Engines**: vLLM, Transformers
24-
- **Model ID:** THUDM/GLM-4.1V-9B-Thinking
25-
- **Model Hubs**: `Hugging Face <https://huggingface.co/THUDM/GLM-4.1V-9B-Thinking>`__, `ModelScope <https://modelscope.cn/models/ZhipuAI/GLM-4.1V-9B-Thinking>`__
24+
- **Model ID:** zai-org/GLM-4.1V-9B-Thinking
25+
- **Model Hubs**: `Hugging Face <https://huggingface.co/zai-org/GLM-4.1V-9B-Thinking>`__, `ModelScope <https://modelscope.cn/models/ZhipuAI/GLM-4.1V-9B-Thinking>`__
2626

2727
Execute the following command to launch the model, remember to replace ``${quantization}`` with your
2828
chosen quantization method from the options listed above::
@@ -37,8 +37,8 @@ Model Spec 2 (awq, 9 Billion)
3737
- **Model Size (in billions):** 9
3838
- **Quantizations:** Int4
3939
- **Engines**: vLLM, Transformers
40-
- **Model ID:** dengcao/GLM-4.1V-9B-Thinking-AWQ
41-
- **Model Hubs**: `Hugging Face <https://huggingface.co/dengcao/GLM-4.1V-9B-Thinking-AWQ>`__, `ModelScope <https://modelscope.cn/models/dengcao/GLM-4.1V-9B-Thinking-AWQ>`__
40+
- **Model ID:** QuantTrio/GLM-4.1V-9B-Thinking-AWQ
41+
- **Model Hubs**: `Hugging Face <https://huggingface.co/QuantTrio/GLM-4.1V-9B-Thinking-AWQ>`__, `ModelScope <https://modelscope.cn/models/tclf90/GLM-4.1V-9B-Thinking-AWQ>`__
4242

4343
Execute the following command to launch the model, remember to replace ``${quantization}`` with your
4444
chosen quantization method from the options listed above::
@@ -53,8 +53,8 @@ Model Spec 3 (gptq, 9 Billion)
5353
- **Model Size (in billions):** 9
5454
- **Quantizations:** Int4-Int8Mix
5555
- **Engines**: vLLM, Transformers
56-
- **Model ID:** dengcao/GLM-4.1V-9B-Thinking-GPTQ-Int4-Int8Mix
57-
- **Model Hubs**: `Hugging Face <https://huggingface.co/dengcao/GLM-4.1V-9B-Thinking-GPTQ-Int4-Int8Mix>`__, `ModelScope <https://modelscope.cn/models/dengcao/GLM-4.1V-9B-Thinking-GPTQ-Int4-Int8Mix>`__
56+
- **Model ID:** QuantTrio/GLM-4.1V-9B-Thinking-GPTQ-Int4-Int8Mix
57+
- **Model Hubs**: `Hugging Face <https://huggingface.co/QuantTrio/GLM-4.1V-9B-Thinking-GPTQ-Int4-Int8Mix>`__, `ModelScope <https://modelscope.cn/models/tclf90/GLM-4.1V-9B-Thinking-GPTQ-Int4-Int8Mix>`__
5858

5959
Execute the following command to launch the model, remember to replace ``${quantization}`` with your
6060
chosen quantization method from the options listed above::
Lines changed: 111 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,111 @@
1+
.. _models_llm_glm-4.5:
2+
3+
========================================
4+
glm-4.5
5+
========================================
6+
7+
- **Context Length:** 65536
8+
- **Model Name:** glm-4.5
9+
- **Languages:** en, zh
10+
- **Abilities:** chat, reasoning
11+
- **Description:** The GLM-4.5 series models are foundation models designed for intelligent agents.
12+
13+
Specifications
14+
^^^^^^^^^^^^^^
15+
16+
17+
Model Spec 1 (pytorch, 355 Billion)
18+
++++++++++++++++++++++++++++++++++++++++
19+
20+
- **Model Format:** pytorch
21+
- **Model Size (in billions):** 355
22+
- **Quantizations:** none
23+
- **Engines**: Transformers
24+
- **Model ID:** zai-org/GLM-4.5
25+
- **Model Hubs**: `Hugging Face <https://huggingface.co/zai-org/GLM-4.5>`__, `ModelScope <https://modelscope.cn/models/ZhipuAI/GLM-4.5>`__
26+
27+
Execute the following command to launch the model, remember to replace ``${quantization}`` with your
28+
chosen quantization method from the options listed above::
29+
30+
xinference launch --model-engine ${engine} --model-name glm-4.5 --size-in-billions 355 --model-format pytorch --quantization ${quantization}
31+
32+
33+
Model Spec 2 (fp8, 355 Billion)
34+
++++++++++++++++++++++++++++++++++++++++
35+
36+
- **Model Format:** fp8
37+
- **Model Size (in billions):** 355
38+
- **Quantizations:** FP8
39+
- **Engines**:
40+
- **Model ID:** zai-org/GLM-4.5-FP8
41+
- **Model Hubs**: `Hugging Face <https://huggingface.co/zai-org/GLM-4.5-FP8>`__, `ModelScope <https://modelscope.cn/models/ZhipuAI/GLM-4.5-FP8>`__
42+
43+
Execute the following command to launch the model, remember to replace ``${quantization}`` with your
44+
chosen quantization method from the options listed above::
45+
46+
xinference launch --model-engine ${engine} --model-name glm-4.5 --size-in-billions 355 --model-format fp8 --quantization ${quantization}
47+
48+
49+
Model Spec 3 (mlx, 355 Billion)
50+
++++++++++++++++++++++++++++++++++++++++
51+
52+
- **Model Format:** mlx
53+
- **Model Size (in billions):** 355
54+
- **Quantizations:** 4bit
55+
- **Engines**: MLX
56+
- **Model ID:** mlx-community/GLM-4.5-{quantization}
57+
- **Model Hubs**: `Hugging Face <https://huggingface.co/mlx-community/GLM-4.5-{quantization}>`__, `ModelScope <https://modelscope.cn/models/mlx-community/GLM-4.5-{quantization}>`__
58+
59+
Execute the following command to launch the model, remember to replace ``${quantization}`` with your
60+
chosen quantization method from the options listed above::
61+
62+
xinference launch --model-engine ${engine} --model-name glm-4.5 --size-in-billions 355 --model-format mlx --quantization ${quantization}
63+
64+
65+
Model Spec 4 (pytorch, 106 Billion)
66+
++++++++++++++++++++++++++++++++++++++++
67+
68+
- **Model Format:** pytorch
69+
- **Model Size (in billions):** 106
70+
- **Quantizations:** none
71+
- **Engines**: Transformers
72+
- **Model ID:** zai-org/GLM-4.5-Air
73+
- **Model Hubs**: `Hugging Face <https://huggingface.co/zai-org/GLM-4.5-Air>`__, `ModelScope <https://modelscope.cn/models/ZhipuAI/GLM-4.5-Air>`__
74+
75+
Execute the following command to launch the model, remember to replace ``${quantization}`` with your
76+
chosen quantization method from the options listed above::
77+
78+
xinference launch --model-engine ${engine} --model-name glm-4.5 --size-in-billions 106 --model-format pytorch --quantization ${quantization}
79+
80+
81+
Model Spec 5 (fp8, 106 Billion)
82+
++++++++++++++++++++++++++++++++++++++++
83+
84+
- **Model Format:** fp8
85+
- **Model Size (in billions):** 106
86+
- **Quantizations:** FP8
87+
- **Engines**:
88+
- **Model ID:** zai-org/GLM-4.5-Air-FP8
89+
- **Model Hubs**: `Hugging Face <https://huggingface.co/zai-org/GLM-4.5-Air-FP8>`__, `ModelScope <https://modelscope.cn/models/ZhipuAI/GLM-4.5-Air-FP8>`__
90+
91+
Execute the following command to launch the model, remember to replace ``${quantization}`` with your
92+
chosen quantization method from the options listed above::
93+
94+
xinference launch --model-engine ${engine} --model-name glm-4.5 --size-in-billions 106 --model-format fp8 --quantization ${quantization}
95+
96+
97+
Model Spec 6 (mlx, 106 Billion)
98+
++++++++++++++++++++++++++++++++++++++++
99+
100+
- **Model Format:** mlx
101+
- **Model Size (in billions):** 106
102+
- **Quantizations:** 2bit, 3bit, 4bit, 5bit, 8bit
103+
- **Engines**: MLX
104+
- **Model ID:** mlx-community/GLM-4.5-Air-{quantization}
105+
- **Model Hubs**: `Hugging Face <https://huggingface.co/mlx-community/GLM-4.5-Air-{quantization}>`__, `ModelScope <https://modelscope.cn/models/mlx-community/GLM-4.5-Air-{quantization}>`__
106+
107+
Execute the following command to launch the model, remember to replace ``${quantization}`` with your
108+
chosen quantization method from the options listed above::
109+
110+
xinference launch --model-engine ${engine} --model-name glm-4.5 --size-in-billions 106 --model-format mlx --quantization ${quantization}
111+

doc/source/models/builtin/llm/glm-4v.rst

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -21,8 +21,8 @@ Model Spec 1 (pytorch, 9 Billion)
2121
- **Model Size (in billions):** 9
2222
- **Quantizations:** none
2323
- **Engines**: Transformers
24-
- **Model ID:** THUDM/glm-4v-9b
25-
- **Model Hubs**: `Hugging Face <https://huggingface.co/THUDM/glm-4v-9b>`__, `ModelScope <https://modelscope.cn/models/ZhipuAI/glm-4v-9b>`__
24+
- **Model ID:** zai-org/glm-4v-9b
25+
- **Model Hubs**: `Hugging Face <https://huggingface.co/zai-org/glm-4v-9b>`__, `ModelScope <https://modelscope.cn/models/ZhipuAI/glm-4v-9b>`__
2626

2727
Execute the following command to launch the model, remember to replace ``${quantization}`` with your
2828
chosen quantization method from the options listed above::

doc/source/models/builtin/llm/glm-edge-chat.rst

Lines changed: 12 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -21,8 +21,8 @@ Model Spec 1 (pytorch, 1_5 Billion)
2121
- **Model Size (in billions):** 1_5
2222
- **Quantizations:** none
2323
- **Engines**: Transformers
24-
- **Model ID:** THUDM/glm-edge-1.5b-chat
25-
- **Model Hubs**: `Hugging Face <https://huggingface.co/THUDM/glm-edge-1.5b-chat>`__, `ModelScope <https://modelscope.cn/models/ZhipuAI/glm-edge-1.5b-chat>`__
24+
- **Model ID:** zai-org/glm-edge-1.5b-chat
25+
- **Model Hubs**: `Hugging Face <https://huggingface.co/zai-org/glm-edge-1.5b-chat>`__, `ModelScope <https://modelscope.cn/models/ZhipuAI/glm-edge-1.5b-chat>`__
2626

2727
Execute the following command to launch the model, remember to replace ``${quantization}`` with your
2828
chosen quantization method from the options listed above::
@@ -37,8 +37,8 @@ Model Spec 2 (pytorch, 4 Billion)
3737
- **Model Size (in billions):** 4
3838
- **Quantizations:** none
3939
- **Engines**: Transformers
40-
- **Model ID:** THUDM/glm-edge-4b-chat
41-
- **Model Hubs**: `Hugging Face <https://huggingface.co/THUDM/glm-edge-4b-chat>`__, `ModelScope <https://modelscope.cn/models/ZhipuAI/glm-edge-4b-chat>`__
40+
- **Model ID:** zai-org/glm-edge-4b-chat
41+
- **Model Hubs**: `Hugging Face <https://huggingface.co/zai-org/glm-edge-4b-chat>`__, `ModelScope <https://modelscope.cn/models/ZhipuAI/glm-edge-4b-chat>`__
4242

4343
Execute the following command to launch the model, remember to replace ``${quantization}`` with your
4444
chosen quantization method from the options listed above::
@@ -53,8 +53,8 @@ Model Spec 3 (ggufv2, 1_5 Billion)
5353
- **Model Size (in billions):** 1_5
5454
- **Quantizations:** Q4_0, Q4_1, Q4_K, Q4_K_M, Q4_K_S, Q5_0, Q5_1, Q5_K, Q5_K_M, Q5_K_S, Q6_K, Q8_0
5555
- **Engines**: llama.cpp
56-
- **Model ID:** THUDM/glm-edge-1.5b-chat-gguf
57-
- **Model Hubs**: `Hugging Face <https://huggingface.co/THUDM/glm-edge-1.5b-chat-gguf>`__, `ModelScope <https://modelscope.cn/models/ZhipuAI/glm-edge-1.5b-chat-gguf>`__
56+
- **Model ID:** zai-org/glm-edge-1.5b-chat-gguf
57+
- **Model Hubs**: `Hugging Face <https://huggingface.co/zai-org/glm-edge-1.5b-chat-gguf>`__, `ModelScope <https://modelscope.cn/models/ZhipuAI/glm-edge-1.5b-chat-gguf>`__
5858

5959
Execute the following command to launch the model, remember to replace ``${quantization}`` with your
6060
chosen quantization method from the options listed above::
@@ -69,8 +69,8 @@ Model Spec 4 (ggufv2, 1_5 Billion)
6969
- **Model Size (in billions):** 1_5
7070
- **Quantizations:** F16
7171
- **Engines**: llama.cpp
72-
- **Model ID:** THUDM/glm-edge-1.5b-chat-gguf
73-
- **Model Hubs**: `Hugging Face <https://huggingface.co/THUDM/glm-edge-1.5b-chat-gguf>`__, `ModelScope <https://modelscope.cn/models/ZhipuAI/glm-edge-1.5b-chat-gguf>`__
72+
- **Model ID:** zai-org/glm-edge-1.5b-chat-gguf
73+
- **Model Hubs**: `Hugging Face <https://huggingface.co/zai-org/glm-edge-1.5b-chat-gguf>`__, `ModelScope <https://modelscope.cn/models/ZhipuAI/glm-edge-1.5b-chat-gguf>`__
7474

7575
Execute the following command to launch the model, remember to replace ``${quantization}`` with your
7676
chosen quantization method from the options listed above::
@@ -85,8 +85,8 @@ Model Spec 5 (ggufv2, 4 Billion)
8585
- **Model Size (in billions):** 4
8686
- **Quantizations:** Q4_0, Q4_1, Q4_K, Q4_K_M, Q4_K_S, Q5_0, Q5_1, Q5_K, Q5_K_M, Q5_K_S, Q6_K, Q8_0
8787
- **Engines**: llama.cpp
88-
- **Model ID:** THUDM/glm-edge-4b-chat-gguf
89-
- **Model Hubs**: `Hugging Face <https://huggingface.co/THUDM/glm-edge-4b-chat-gguf>`__, `ModelScope <https://modelscope.cn/models/ZhipuAI/glm-edge-4b-chat-gguf>`__
88+
- **Model ID:** zai-org/glm-edge-4b-chat-gguf
89+
- **Model Hubs**: `Hugging Face <https://huggingface.co/zai-org/glm-edge-4b-chat-gguf>`__, `ModelScope <https://modelscope.cn/models/ZhipuAI/glm-edge-4b-chat-gguf>`__
9090

9191
Execute the following command to launch the model, remember to replace ``${quantization}`` with your
9292
chosen quantization method from the options listed above::
@@ -101,8 +101,8 @@ Model Spec 6 (ggufv2, 4 Billion)
101101
- **Model Size (in billions):** 4
102102
- **Quantizations:** F16
103103
- **Engines**: llama.cpp
104-
- **Model ID:** THUDM/glm-edge-4b-chat-gguf
105-
- **Model Hubs**: `Hugging Face <https://huggingface.co/THUDM/glm-edge-4b-chat-gguf>`__, `ModelScope <https://modelscope.cn/models/ZhipuAI/glm-edge-4b-chat-gguf>`__
104+
- **Model ID:** zai-org/glm-edge-4b-chat-gguf
105+
- **Model Hubs**: `Hugging Face <https://huggingface.co/zai-org/glm-edge-4b-chat-gguf>`__, `ModelScope <https://modelscope.cn/models/ZhipuAI/glm-edge-4b-chat-gguf>`__
106106

107107
Execute the following command to launch the model, remember to replace ``${quantization}`` with your
108108
chosen quantization method from the options listed above::

doc/source/models/builtin/llm/glm4-0414.rst

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -21,8 +21,8 @@ Model Spec 1 (pytorch, 9 Billion)
2121
- **Model Size (in billions):** 9
2222
- **Quantizations:** none
2323
- **Engines**: vLLM, Transformers
24-
- **Model ID:** THUDM/GLM-4-9B-0414
25-
- **Model Hubs**: `Hugging Face <https://huggingface.co/THUDM/GLM-4-9B-0414>`__, `ModelScope <https://modelscope.cn/models/ZhipuAI/GLM-4-9B-0414>`__
24+
- **Model ID:** zai-org/GLM-4-9B-0414
25+
- **Model Hubs**: `Hugging Face <https://huggingface.co/zai-org/GLM-4-9B-0414>`__, `ModelScope <https://modelscope.cn/models/ZhipuAI/GLM-4-9B-0414>`__
2626

2727
Execute the following command to launch the model, remember to replace ``${quantization}`` with your
2828
chosen quantization method from the options listed above::
@@ -37,8 +37,8 @@ Model Spec 2 (pytorch, 32 Billion)
3737
- **Model Size (in billions):** 32
3838
- **Quantizations:** none
3939
- **Engines**: vLLM, Transformers
40-
- **Model ID:** THUDM/GLM-4-32B-0414
41-
- **Model Hubs**: `Hugging Face <https://huggingface.co/THUDM/GLM-4-32B-0414>`__, `ModelScope <https://modelscope.cn/models/ZhipuAI/GLM-4-32B-0414>`__
40+
- **Model ID:** zai-org/GLM-4-32B-0414
41+
- **Model Hubs**: `Hugging Face <https://huggingface.co/zai-org/GLM-4-32B-0414>`__, `ModelScope <https://modelscope.cn/models/ZhipuAI/GLM-4-32B-0414>`__
4242

4343
Execute the following command to launch the model, remember to replace ``${quantization}`` with your
4444
chosen quantization method from the options listed above::

doc/source/models/builtin/llm/glm4-chat-1m.rst

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -21,8 +21,8 @@ Model Spec 1 (pytorch, 9 Billion)
2121
- **Model Size (in billions):** 9
2222
- **Quantizations:** none
2323
- **Engines**: vLLM, Transformers
24-
- **Model ID:** THUDM/glm-4-9b-chat-1m-hf
25-
- **Model Hubs**: `Hugging Face <https://huggingface.co/THUDM/glm-4-9b-chat-1m-hf>`__, `ModelScope <https://modelscope.cn/models/ZhipuAI/glm-4-9b-chat-1m-hf>`__
24+
- **Model ID:** zai-org/glm-4-9b-chat-1m-hf
25+
- **Model Hubs**: `Hugging Face <https://huggingface.co/zai-org/glm-4-9b-chat-1m-hf>`__, `ModelScope <https://modelscope.cn/models/ZhipuAI/glm-4-9b-chat-1m-hf>`__
2626

2727
Execute the following command to launch the model, remember to replace ``${quantization}`` with your
2828
chosen quantization method from the options listed above::

doc/source/models/builtin/llm/glm4-chat.rst

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -21,8 +21,8 @@ Model Spec 1 (pytorch, 9 Billion)
2121
- **Model Size (in billions):** 9
2222
- **Quantizations:** none
2323
- **Engines**: vLLM, Transformers
24-
- **Model ID:** THUDM/glm-4-9b-chat-hf
25-
- **Model Hubs**: `Hugging Face <https://huggingface.co/THUDM/glm-4-9b-chat-hf>`__, `ModelScope <https://modelscope.cn/models/ZhipuAI/glm-4-9b-chat-hf>`__
24+
- **Model ID:** zai-org/glm-4-9b-chat-hf
25+
- **Model Hubs**: `Hugging Face <https://huggingface.co/zai-org/glm-4-9b-chat-hf>`__, `ModelScope <https://modelscope.cn/models/ZhipuAI/glm-4-9b-chat-hf>`__
2626

2727
Execute the following command to launch the model, remember to replace ``${quantization}`` with your
2828
chosen quantization method from the options listed above::

0 commit comments

Comments
 (0)