Skip to content

Commit 959ac67

Browse files
authored
update readme (#949)
1 parent 97512a4 commit 959ac67

File tree

2 files changed

+3
-2
lines changed

2 files changed

+3
-2
lines changed

README.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -27,6 +27,8 @@ and [fbaldassarri](https://huggingface.co/fbaldassarri). For usage instructions,
2727

2828

2929
## 🆕 What's New
30+
[2025/10] We enhanced the RTN mode (--iters 0) to significantly reduce quantization cost compared to the default tuning mode. Check out [this doc](./docs/opt_rtn.md) for some accuracy results. If you don’t have sufficient resources, you can use this mode for 4-bit quantization.
31+
3032
[2025/10] We proposed a fast algorithm to generate **mixed bits/datatypes** schemes in minutes. Please
3133
refer to the documentation for accuracy [results](./docs/auto_scheme_acc.md) and [this guide](https://github.com/intel/auto-round/blob/main/docs/step_by_step.md#autoscheme) for usage instructions.
3234

docs/auto_scheme_acc.md

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -6,8 +6,7 @@ to stabilize accuracy during evaluation. All other settings follow the default c
66
We ignore the scale and zp bits in the tables below. The accuracy may change a little as we modified a little of the
77
implementation. We will rerun all the experiments.
88

9-
For mxfp experiment, we use fake model while for weight only model we use real model. **No tuning is applied unless explicit stated.
10-
**
9+
For mxfp experiment, we use fake model while for weight only model we use real model. **No tuning is applied unless explicit stated.**
1110

1211
*Average accuracy across `lambada_openai`, `hellaswag`, `piqa`, `winogrande`, and `mmlu`.*
1312

0 commit comments

Comments
 (0)