49
49
<img src =" https://raw.githubusercontent.com/imoneoi/openchat/master/assets/benchmarks-openchat-3.6-20240522.svg " style =" width : 95% ;" >
50
50
</div >
51
51
52
-
53
52
<details >
54
53
<summary >Reproducing benchmarks</summary >
55
54
@@ -77,6 +76,14 @@ HumanEval is run using the official [EvalPlus repository](https://github.com/eva
77
76
| Open-source SOTA** | 13B-70B | 61.4 | 7.71 | 73.2 | 49.7 | 41.7 | 62.3 | 63.7 | 82.3 | 41.4 |
78
77
| | | | WizardLM 70B | WizardCoder 34B | Orca 13B | Orca 13B | Platypus2 70B | WizardLM 70B | MetaMath 70B | Flan-T5 11B |
79
78
79
+ 🔥 OpenChat-3.5-0106 (7B) now outperforms Grok-0 (33B) on ** all 4 benchmarks** and Grok-1 (314B) on average and ** 3/4 benchmarks** .
80
+
81
+ | | License | # Param | Average | MMLU | HumanEval | MATH | GSM8k |
82
+ | -----------------------| -------------| ---------| ----------| --------| -----------| ----------| ----------|
83
+ | ** OpenChat-3.5-0106** | Apache-2.0 | ** 7B** | ** 61.0** | 65.8 | ** 71.3** | ** 29.3** | ** 77.4** |
84
+ | Grok-0 | Proprietary | 33B | 44.5 | 65.7 | 39.7 | 15.7 | 56.8 |
85
+ | Grok-1 | Proprietary | 314B | 55.8 | ** 73** | 63.2 | 23.9 | 62.9 |
86
+
80
87
<details >
81
88
<summary >Evaluation details</summary >
82
89
* : ChatGPT (March) results are from GPT-4 Technical Report, Chain-of-Thought Hub, and our evaluation.
@@ -130,19 +137,9 @@ python gen_judgment.py --model-list openchat-3.5-0106 --parallel 8 --mode single
130
137
131
138
</details >
132
139
133
- ## 🎇 Comparison with [ X.AI Grok] ( https://x.ai/ )
134
-
135
- 🔥 OpenChat-3.5-0106 (7B) now outperforms Grok-0 (33B) on ** all 4 benchmarks** and Grok-1 (???B) on average and ** 3/4 benchmarks** .
136
-
137
- | | License | # Param | Average | MMLU | HumanEval | MATH | GSM8k |
138
- | -----------------------| -------------| ---------| ----------| --------| -----------| ----------| ----------|
139
- | ** OpenChat-3.5-0106** | Apache-2.0 | ** 7B** | ** 61.0** | 65.8 | ** 71.3** | ** 29.3** | ** 77.4** |
140
- | Grok-0 | Proprietary | 33B | 44.5 | 65.7 | 39.7 | 15.7 | 56.8 |
141
- | Grok-1 | Proprietary | ???B | 55.8 | ** 73** | 63.2 | 23.9 | 62.9 |
142
-
143
140
# ⬇️ Installation
144
141
> [ !NOTE]
145
- > Need [ ` pytorch ` ] ( https://pytorch.org/get-started/locally/#start-locally ) to run OpenChat
142
+ > Need [ ` pytorch ` ] ( https://pytorch.org/get-started/locally/#start-locally ) and [ CUDA ] ( https://developer.nvidia.com/cuda-toolkit-archive ) to run OpenChat
146
143
147
144
## pip
148
145
@@ -199,22 +196,29 @@ pip3 install -e . # Editable mode, you can make changes in this cloned repo
199
196
200
197
📎 Note: For 20 series or older GPUs that do not support ` bfloat16 ` , add ` --dtype float16 ` to the server args.
201
198
199
+ ### List of currently supported models
200
+
201
+ | MODEL_TYPE | MODEL_REPO | License |
202
+ | --------------| -----------------------------------------------------------------------------------------------| ------------|
203
+ | openchat_3.6 | [ openchat/openchat-3.6-8b-20240522] ( https://huggingface.co/openchat/openchat-3.6-8b-20240522 ) | Llama 3 |
204
+ | openchat_3.5 | [ openchat/openchat-3.5-0106] ( https://huggingface.co/openchat/openchat-3.5-0106 ) | Apache 2.0 |
205
+
202
206
### For a single GPU (e.g. RTX 3090, 4090)
203
207
204
208
``` bash
205
- python -m ochat.serving.openai_api_server --model openchat/openchat-3.5-0106
209
+ python -m ochat.serving.openai_api_server --model MODEL_REPO
206
210
```
207
211
208
212
### For multiple GPUs (tensor parallel)
209
213
210
214
``` bash
211
215
# N is the number of tensor parallel GPUs
212
- python -m ochat.serving.openai_api_server --model openchat/openchat-3.5-0106 --engine-use-ray --worker-use-ray --tensor-parallel-size N
216
+ python -m ochat.serving.openai_api_server --model MODEL_REPO --engine-use-ray --worker-use-ray --tensor-parallel-size N
213
217
```
214
218
215
219
use ` -h ` to see more settings
216
220
``` bash
217
- python -m ochat.serving.openai_api_server --model openchat/openchat-3.5-0106 -h
221
+ python -m ochat.serving.openai_api_server --model MODEL_REPO -h
218
222
```
219
223
220
224
<details >
@@ -234,7 +238,7 @@ Once started, the server listens at `localhost:18888` for requests and is compat
234
238
curl http://localhost:18888/v1/chat/completions \
235
239
-H " Content-Type: application/json" \
236
240
-d ' {
237
- "model": "openchat_3.5 ",
241
+ "model": "MODEL_TYPE ",
238
242
"messages": [{"role": "user", "content": "You are a large language model named OpenChat. Write a poem to describe yourself"}]
239
243
}'
240
244
```
@@ -245,7 +249,7 @@ curl http://localhost:18888/v1/chat/completions \
245
249
curl http://localhost:18888/v1/chat/completions \
246
250
-H " Content-Type: application/json" \
247
251
-d ' {
248
- "model": "openchat_3.5 ",
252
+ "model": "MODEL_TYPE ",
249
253
"condition": "Math Correct",
250
254
"messages": [{"role": "user", "content": "10.3 − 7988.8133 = "}]
251
255
}'
0 commit comments