-
-
Notifications
You must be signed in to change notification settings - Fork 4.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Support for chatglm-6b #231
Comments
mark |
2 similar comments
mark |
mark |
If anyone is familiar with chatGLM model architecture, feel free to help on #625. I am new to transformer architecture and not sure if my changes is correct.. |
Vllm can support chatglm 6b now? |
If anyone have bandwidth to help us implement ChatGLM3 support, please leave a comment and coordinate here: #1552 |
ChatGLM supported in #1261 |
SUMMARY: * switch build phase to GCP k8s cluster TEST PLAN: runs on remote push Co-authored-by: andy-neuma <andy@neuralmagic.com>
Removes unnecessary mark step from MoE OP loop to speed up computation
* add fp8 for dbrx * linting
It would be great if you could support chatglm-6b,It's a popular chinese model。
https://huggingface.co/THUDM/chatglm-6b
The text was updated successfully, but these errors were encountered: