-
Notifications
You must be signed in to change notification settings - Fork 2.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[LLM Inference] Support qwen2 #8893
Conversation
Thanks for your contribution! |
Codecov ReportAttention: Patch coverage is
Additional details and impacted files@@ Coverage Diff @@
## develop #8893 +/- ##
===========================================
- Coverage 55.29% 55.18% -0.11%
===========================================
Files 632 634 +2
Lines 98910 99345 +435
===========================================
+ Hits 54690 54827 +137
- Misses 44220 44518 +298 ☔ View full report in Codecov by Sentry. |
7844fb7
to
a6bde28
Compare
@@ -0,0 +1,21 @@ | |||
{ |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Qwen2属于比较重要的模型,加上单测?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
如线下沟通结果,单测后续统一补上~
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
PR types
New features
PR changes
Models
Description
支持qwen2 bf16/wint8高性能推理