Skip to content

Commit 104359e

Browse files
authored
Update README.md
1 parent d27140c commit 104359e

File tree

1 file changed

+21
-17
lines changed

1 file changed

+21
-17
lines changed

README.md

Lines changed: 21 additions & 17 deletions
Original file line numberDiff line numberDiff line change
@@ -49,7 +49,6 @@
4949
<img src="https://raw.githubusercontent.com/imoneoi/openchat/master/assets/benchmarks-openchat-3.6-20240522.svg" style="width: 95%;">
5050
</div>
5151

52-
5352
<details>
5453
<summary>Reproducing benchmarks</summary>
5554

@@ -77,6 +76,14 @@ HumanEval is run using the official [EvalPlus repository](https://github.com/eva
7776
| Open-source SOTA** | 13B-70B | 61.4 | 7.71 | 73.2 | 49.7 | 41.7 | 62.3 | 63.7 | 82.3 | 41.4 |
7877
| | | | WizardLM 70B | WizardCoder 34B | Orca 13B | Orca 13B | Platypus2 70B | WizardLM 70B | MetaMath 70B | Flan-T5 11B |
7978

79+
🔥 OpenChat-3.5-0106 (7B) now outperforms Grok-0 (33B) on **all 4 benchmarks** and Grok-1 (314B) on average and **3/4 benchmarks**.
80+
81+
| | License | # Param | Average | MMLU | HumanEval | MATH | GSM8k |
82+
|-----------------------|-------------|---------|----------|--------|-----------|----------|----------|
83+
| **OpenChat-3.5-0106** | Apache-2.0 | **7B** | **61.0** | 65.8 | **71.3** | **29.3** | **77.4** |
84+
| Grok-0 | Proprietary | 33B | 44.5 | 65.7 | 39.7 | 15.7 | 56.8 |
85+
| Grok-1 | Proprietary | 314B | 55.8 | **73** | 63.2 | 23.9 | 62.9 |
86+
8087
<details>
8188
<summary>Evaluation details</summary>
8289
*: ChatGPT (March) results are from GPT-4 Technical Report, Chain-of-Thought Hub, and our evaluation.
@@ -130,19 +137,9 @@ python gen_judgment.py --model-list openchat-3.5-0106 --parallel 8 --mode single
130137

131138
</details>
132139

133-
## 🎇 Comparison with [X.AI Grok](https://x.ai/)
134-
135-
🔥 OpenChat-3.5-0106 (7B) now outperforms Grok-0 (33B) on **all 4 benchmarks** and Grok-1 (???B) on average and **3/4 benchmarks**.
136-
137-
| | License | # Param | Average | MMLU | HumanEval | MATH | GSM8k |
138-
|-----------------------|-------------|---------|----------|--------|-----------|----------|----------|
139-
| **OpenChat-3.5-0106** | Apache-2.0 | **7B** | **61.0** | 65.8 | **71.3** | **29.3** | **77.4** |
140-
| Grok-0 | Proprietary | 33B | 44.5 | 65.7 | 39.7 | 15.7 | 56.8 |
141-
| Grok-1 | Proprietary | ???B | 55.8 | **73** | 63.2 | 23.9 | 62.9 |
142-
143140
# ⬇️ Installation
144141
> [!NOTE]
145-
> Need [`pytorch`](https://pytorch.org/get-started/locally/#start-locally) to run OpenChat
142+
> Need [`pytorch`](https://pytorch.org/get-started/locally/#start-locally) and [CUDA](https://developer.nvidia.com/cuda-toolkit-archive) to run OpenChat
146143
147144
## pip
148145

@@ -199,22 +196,29 @@ pip3 install -e . # Editable mode, you can make changes in this cloned repo
199196

200197
📎 Note: For 20 series or older GPUs that do not support `bfloat16`, add `--dtype float16` to the server args.
201198

199+
### List of currently supported models
200+
201+
| MODEL_TYPE | MODEL_REPO | License |
202+
|--------------|-----------------------------------------------------------------------------------------------|------------|
203+
| openchat_3.6 | [openchat/openchat-3.6-8b-20240522](https://huggingface.co/openchat/openchat-3.6-8b-20240522) | Llama 3 |
204+
| openchat_3.5 | [openchat/openchat-3.5-0106](https://huggingface.co/openchat/openchat-3.5-0106) | Apache 2.0 |
205+
202206
### For a single GPU (e.g. RTX 3090, 4090)
203207

204208
```bash
205-
python -m ochat.serving.openai_api_server --model openchat/openchat-3.5-0106
209+
python -m ochat.serving.openai_api_server --model MODEL_REPO
206210
```
207211

208212
### For multiple GPUs (tensor parallel)
209213

210214
```bash
211215
# N is the number of tensor parallel GPUs
212-
python -m ochat.serving.openai_api_server --model openchat/openchat-3.5-0106 --engine-use-ray --worker-use-ray --tensor-parallel-size N
216+
python -m ochat.serving.openai_api_server --model MODEL_REPO --engine-use-ray --worker-use-ray --tensor-parallel-size N
213217
```
214218

215219
use `-h` to see more settings
216220
```bash
217-
python -m ochat.serving.openai_api_server --model openchat/openchat-3.5-0106 -h
221+
python -m ochat.serving.openai_api_server --model MODEL_REPO -h
218222
```
219223

220224
<details>
@@ -234,7 +238,7 @@ Once started, the server listens at `localhost:18888` for requests and is compat
234238
curl http://localhost:18888/v1/chat/completions \
235239
-H "Content-Type: application/json" \
236240
-d '{
237-
"model": "openchat_3.5",
241+
"model": "MODEL_TYPE",
238242
"messages": [{"role": "user", "content": "You are a large language model named OpenChat. Write a poem to describe yourself"}]
239243
}'
240244
```
@@ -245,7 +249,7 @@ curl http://localhost:18888/v1/chat/completions \
245249
curl http://localhost:18888/v1/chat/completions \
246250
-H "Content-Type: application/json" \
247251
-d '{
248-
"model": "openchat_3.5",
252+
"model": "MODEL_TYPE",
249253
"condition": "Math Correct",
250254
"messages": [{"role": "user", "content": "10.3 − 7988.8133 = "}]
251255
}'

0 commit comments

Comments
 (0)