-
Notifications
You must be signed in to change notification settings - Fork 1.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
add ipex moe implementation to support Mixtral and PhiMoe #2707
Conversation
Signed-off-by: Wang, Yi A <yi.a.wang@intel.com>
need xpu ipex to upgrade to 2.5 as well. or else it will block tgi-xpu |
This is still marked as draft, should we review/merge or not ? The code change looks Ok. |
Signed-off-by: Wang, Yi A <yi.a.wang@intel.com>
It also need the upgrade of xpu ipex too. once ready, I will mark it ready for review. thanks |
Signed-off-by: Wang, Yi A <yi.a.wang@intel.com>
@Narsil please help review the PR. thanks |
Signed-off-by: Wang, Yi A <yi.a.wang@intel.com>
@Narsil could you review it. the PR also upgrade xpu ipex which means to support AWQ and GPTQ |
@danieldk could you help? |
Looking at it now! |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks, it's nice to have fused MoE on ipex. I added some comments.
Co-authored-by: Daniël de Kok <me@github.danieldk.eu>
@OlivierDehaene OR @Narsil