Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

feat: support bloom-1b4-zh #543

Closed

Conversation

xingchensong
Copy link

@xingchensong xingchensong commented Sep 28, 2023

fp16:
image
image

q4_0
image
image

@xingchensong
Copy link
Author

xingchensong commented Sep 28, 2023

PPL test on single sentence:

torch fp16 ggml fp16 ggml q8_0 ggml q5_0 ggml q4_0
1.9083 1.90837552 1.91213162 1.95716960 2.21781030

the code I used to test ppl is from this PR: #521

@xingchensong xingchensong changed the title bloom: support bloom-1b4-zh feat: support bloom-1b4-zh Sep 28, 2023
@ggerganov
Copy link
Owner

ggerganov commented Oct 6, 2023

Thanks for this contribution!

Are you interested in trying to add support for this model straight into llama.cpp?
Recently, we have started to support many different model architectures and people will help out with testing and keeping up-to-date.

@xingchensong
Copy link
Author

Thanks for this contribution!

Are you interested in trying to add support for this model straight into llama.cpp?
Recently, we have started to support many different model architectures and people will help out with testing and keeping up-to-date.

sure, will do it ASAP

@ggerganov
Copy link
Owner

Great! Take a look at this PR to get an idea of what is needed:

ggerganov/llama.cpp#3187

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants