Replies: 2 comments
-
Do you know if llama.cpp can load MiniCPM models? |
Beta Was this translation helpful? Give feedback.
0 replies
-
Yes, they do. Supported models are: Inference with llama.cpp MiniCPM-V 2.6 can run with llama.cpp now! See our fork of llama.cpp for more detail. This implementation supports smooth inference of 16~18 token/s on iPad (test environment:iPad Pro + M4). |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Is there a way to add support or somehow be able to load MiniCPM model like this one here?
https://huggingface.co/openbmb/MiniCPM-V-2_6-gguf
Thank you in advance!
Beta Was this translation helpful? Give feedback.
All reactions