forked from ggerganov/llama.cpp
-
Notifications
You must be signed in to change notification settings - Fork 360
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Support for QWEN models #549
Labels
enhancement
New feature or request
Comments
FYI here is the log from a successful llama.cpp run
|
+1 |
Should be added in the next version |
Hi all, please try 1.52 and let me know if it works for you. |
Great Job !!! Yes this is now working. Many thanks for the update. |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Hi, thanks for your amazing work on this software. I am tring to run some of the latest QWEN models that are topping the leader boards and on paper currently the best base model. Specifically QWEN-72b. This currently works correctly in llama.cpp with the latest release #https://github.com/ggerganov/llama.cpp/releases/tag/b1610. I beleive this functionality was added in ggerganov#4281 (3 days ago)
When I run this in Kobold.1.51.1 and try to load the model from #https://huggingface.co/CausalLM/72B-preview-GGUF I get the following error on load. Maybe this pull has not been incorperated in the latest release yet?
Many thanks in advance
The text was updated successfully, but these errors were encountered: