-
Notifications
You must be signed in to change notification settings - Fork 2.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Inference] Update fakequant script #9054
[Inference] Update fakequant script #9054
Conversation
2. add llama3.1 and qwen2 ptq config 3. update quantization.md
…nto add_new_fakequant_type
Thanks for your contribution! |
Codecov ReportAll modified and coverable lines are covered by tests ✅
Additional details and impacted files@@ Coverage Diff @@
## develop #9054 +/- ##
===========================================
+ Coverage 53.76% 53.82% +0.06%
===========================================
Files 652 652
Lines 104507 104529 +22
===========================================
+ Hits 56190 56265 +75
+ Misses 48317 48264 -53 ☔ View full report in Codecov by Sentry. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
* 1. add a8w8(fp8) a8w8c8(int8) quant_type support 2. add llama3.1 and qwen2 ptq config 3. update quantization.md * fix load_quant_model bug * fix load quant bug * update ll/README.md * remove useless code * update quant observer config * resolve wrong modify * fix prepare_qconfig * remove unuse files * update quantization.md * reformat quantization.md and argument.py * update prepare data method for ceval ptq
PR types
Others
PR changes
Docs
Description
update quantization.md