Replies: 2 comments
-
Hey community Here we have a good explanation to help you run the python version locally |
Beta Was this translation helpful? Give feedback.
-
This is newly relevant with the release of Gemini Diffusion yesterday: https://simonwillison.net/2025/May/21/gemini-diffusion/ I might be interested in exploring this. Wrapping up a quantization PR for Iwan's fork (ikawrakow/ik_llama.cpp#441) and am interested in what it would take to add support for a whole new architecture to llama.cpp (although I can't guarantee immediate bandwidth). |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Mercury Coder is a diffusion-model based LLM which its creators claim offers huge efficiency improvements over standard LLMs. Is there any possibility that llama.cpp will add support for it?
https://huggingface.co/spaces/multimodalart/LLaDA
https://ml-gsai.github.io/LLaDA-demo/
https://github.com/ML-GSAI/LLaDA
Beta Was this translation helpful? Give feedback.
All reactions