-
Notifications
You must be signed in to change notification settings - Fork 9.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Requesting Qwen-7B Support #2528
Comments
Qwen is similar to Llama model. I wonder if we can convert tiktoken vocab format to sentencepiece directly. |
https://huggingface.co/JosephusCheung/Qwen-LLaMAfied-7B-Chat I think this model is still slightly different from the original Qwen especially in QwenAttentionBlock, Qwen applied log(n) to query. |
Sadly I couldn't convert it, always getting cryptic python errors. |
I found this script for converting from Tiktoken vocab to HF: https://gist.github.com/xenova/a452a6474428de0182b17605a98631ee (didn't test it, but it looks reasonable and seems to be from a HF person.) To actually use it, you'll also need to use #3743 since there wasn't already support for loading from @TheBloke too since I assume you're looking to support these Qwen models. |
Is there a blocker to support these models in |
https://github.com/QwenLM/qwen.cpp |
We should first support QKVO bias in llama. |
There was an implementation here but failed, can anyone figure out what's wrong with this? |
|
should this be closed by #5037 ? |
https://huggingface.co/Qwen/Qwen-7B
https://huggingface.co/Qwen/Qwen-7B-Chat
These two models are performing better than 13B models. and on C-eval beating ChatGPT.
Requesting model support for Llama.cpp.
The text was updated successfully, but these errors were encountered: