-
-
Notifications
You must be signed in to change notification settings - Fork 4.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Mixtral-8x7B-Instruct-v0.1-GPTQ weight loading error #2202
Comments
Yep, I'm facing the same issue. 2x A30 -- On Runpod Command:
|
None of the params in params_dict (mixtral.py:406)
matches the weights loaded from the safetensors file from Seems like a silly mistake, a goofy fault, indeed. Pleaze fix ASAP. |
@jbohnslav @Sirri69 Thanks for reporting the bug! I only tested the Mixtral GPTQ model with a single GPU. I reproduced the bug and fixed it in #2208 |
Thanks so much @WoosukKwon! I'll check it out now. |
Confirm fixed! |
Command:
Result:
I think there's a mismatch between the keys of all the HF quantized models and the default model uploaded by Mistral. Maybe we just need a lookup table? Or do the instruct versions have different params?
The text was updated successfully, but these errors were encountered: