Replies: 1 comment
-
|
@xdd130 Does your model have any weight of shape 10304? Since 10304=64*161, you should set -gs=64 in the command. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I try used the AutoGPTQ tool to quantize the Qwen2.5-3B-Instruct model to 3 bits. I successfully obtained the model in GPTQ format, but when I compiled the script using T-MAC:
I got the following error:
Does this indicate that there is a problem with my quantization step or that t-mac does not support the direct use of 3-bit models? What additional operations do I need to do?
Beta Was this translation helpful? Give feedback.
All reactions