-
Notifications
You must be signed in to change notification settings - Fork 9.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Please add support for kfkas llama-2-ko-7b-chat #2877
Comments
I suggested creating this but it looks like it's probably an issue with |
To follow up on this, right now llama.cpp itself doesn't fully handle unicode for BPE tokenizing. That means regardless of what the convert script does (and that probably also still has issues) models like this one aren't going to work. From what I've found out, the reason is being implementing full unicode support for BPE would require a large amount of complex code. In the short term, it doesn't seem like there's a definite plan to do this. If you need to run this model (or similar BPE non-Latin character based models) then you should probably look for another solution, at least for now. Note: Speaking as an individual here with information I believe to be correct. I don't have any authority to say what the |
Thank you for writing your feedback. While searching on the Internet, I found a case of translating into gguf on the hugging face (https://huggingface.co/StarFox7/Llama-2-ko-7B-chat-gguf) ). I'm going to ask this person how he converted it. I will leave a reply on this page again if there is any progress. Thank you for always caring! |
I'd be interested in knowing that too! Maybe there's a relatively easy way to get these models working that I missed. |
I am the developer of kfkas/Llama-2-ko-7b-Chat. Having trouble? |
Hello, developer. Here's how I've tried. 1. 1. convert-llama-hf-to-gguf.Attempt to convert to py▷ DOS prompt:
→ The GGUF file is created, but the following message occurs during model loading. ▷ log : llama_model_loader: - type f16: 226 tensors
2. Create vocab.json and attempt to convert to convert.py▷ DOS prompt:
→ The execution went well, but there was a problem that the string was not created. ▷ Execution Log:
While searching on the Internet, I found a case of translating into gguf on the hugging face (https://huggingface.co/StarFox7/Llama-2-ko-7B-chat-gguf). I wonder how you converted it into a GGUF file. |
I am also in the process of converting 'kfkas/Llama-2-ko-7b-Chat' to the gguf format. However, when using the 'bpe' tokenizer, I encountered the following error message:
Upon investigating, I discovered that the developer of the pre-trained model for this model used the sentencepiece tokenizer. Therefore, it seems that the tokenizer configuration should be set to 'spm' (default) instead of 'bpe'. I proceeded to obtain the tokenizer.model from original llama-2 model and added_tokens.json from the pre-trained model (beomi/llama-2-ko-7b) accordingly and performed the conversion.
Upon running the conversion, I encountered the following message in both gguf converted and quantized after gguf converted:
Error message is :
The same results are produced at local and google colab. full message:
I don't know how to fix it |
This issue was closed because it has been inactive for 14 days since being marked as stale. |
Failed to convert model https://huggingface.co/kfkas/Llama-2-ko-7b-Chat in hf to gguf. Please add support to convert to gguf.
Please refer to the link below for the context of the failed issue.
Converting kfkas Llama-2-ko-7b-Chat to GGUF fails.
The text was updated successfully, but these errors were encountered: