Skip to content

Commit 4bb625b

Browse files
ryan-mangenogabe-l-hartngxsonCISCcompilade
authored
Smoldocling support (#14597)
* support for smoldocling * fixed merge conflicts * Update gguf-py/gguf/tensor_mapping.py Co-authored-by: Gabe Goodhart <gabe.l.hart@gmail.com> * Update gguf-py/gguf/tensor_mapping.py Co-authored-by: Gabe Goodhart <gabe.l.hart@gmail.com> * merge conflicts * pre tokenizer merge fix * convert : fix smollm3 jinja template (#14586) Signed-off-by: ryan-mangeno <ryanmangeno@gmail.com> * support for smoldocling Signed-off-by: ryan-mangeno <ryanmangeno@gmail.com> * fixed merge conflicts Signed-off-by: ryan-mangeno <ryanmangeno@gmail.com> * Update src/llama-vocab.cpp Co-authored-by: Sigbjørn Skjæret <sigbjorn.skjaeret@scala.com> * Update gguf-py/gguf/tensor_mapping.py Co-authored-by: Sigbjørn Skjæret <sigbjorn.skjaeret@scala.com> * Update gguf-py/gguf/tensor_mapping.py Co-authored-by: Sigbjørn Skjæret <sigbjorn.skjaeret@scala.com> * Update src/llama-model.h Co-authored-by: Sigbjørn Skjæret <sigbjorn.skjaeret@scala.com> * safetensors tensor mapping Signed-off-by: ryan-mangeno <ryanmangeno@gmail.com> * added back accidental removal of clean spaces for hunyuan * Update src/llama-vocab.cpp Co-authored-by: Sigbjørn Skjæret <sigbjorn.skjaeret@scala.com> * updated hash and reordererd model list * Update gguf-py/gguf/tensor_mapping.py Co-authored-by: Sigbjørn Skjæret <sigbjorn.skjaeret@scala.com> * Update src/llama-vocab.cpp Co-authored-by: Sigbjørn Skjæret <sigbjorn.skjaeret@scala.com> * Update include/llama.h Co-authored-by: Sigbjørn Skjæret <sigbjorn.skjaeret@scala.com> * Update convert_hf_to_gguf.py Co-authored-by: Sigbjørn Skjæret <sigbjorn.skjaeret@scala.com> * Update convert_hf_to_gguf_update.py Co-authored-by: Sigbjørn Skjæret <sigbjorn.skjaeret@scala.com> * Update src/llama-vocab.cpp Co-authored-by: Sigbjørn Skjæret <sigbjorn.skjaeret@scala.com> * removed old tensor name * removed tensor mappings -> handled by smolvlm * Update gguf-py/gguf/tensor_mapping.py Co-authored-by: Sigbjørn Skjæret <sigbjorn.skjaeret@scala.com> * Update gguf-py/gguf/tensor_mapping.py Co-authored-by: Sigbjørn Skjæret <sigbjorn.skjaeret@scala.com> * Update gguf-py/gguf/tensor_mapping.py Co-authored-by: Sigbjørn Skjæret <sigbjorn.skjaeret@scala.com> --------- Signed-off-by: ryan-mangeno <ryanmangeno@gmail.com> Co-authored-by: Gabe Goodhart <gabe.l.hart@gmail.com> Co-authored-by: Xuan-Son Nguyen <son@huggingface.co> Co-authored-by: Sigbjørn Skjæret <sigbjorn.skjaeret@scala.com> Co-authored-by: compilade <git@compilade.net>
1 parent 11ee0fe commit 4bb625b

File tree

3 files changed

+5
-0
lines changed

3 files changed

+5
-0
lines changed

src/llama-model.cpp

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -40,6 +40,7 @@ const char * llm_type_name(llm_type type) {
4040
case LLM_TYPE_190M: return "190M";
4141
case LLM_TYPE_220M: return "220M";
4242
case LLM_TYPE_250M: return "250M";
43+
case LLM_TYPE_256M: return "256M";
4344
case LLM_TYPE_270M: return "270M";
4445
case LLM_TYPE_335M: return "335M";
4546
case LLM_TYPE_410M: return "410M";
@@ -581,6 +582,7 @@ void llama_model::load_hparams(llama_model_loader & ml) {
581582
case 22: type = LLM_TYPE_1B; break;
582583
case 26: type = LLM_TYPE_3B; break;
583584
case 28: type = LLM_TYPE_3B; break; // Llama 3.2 3B
585+
case 30: type = LLM_TYPE_256M; break; // smoldocling 256M
584586
// granite uses a vocab with len 49152
585587
case 32: type = n_vocab == 49152 ? LLM_TYPE_3B : (n_vocab < 40000 ? LLM_TYPE_7B : LLM_TYPE_8B); break;
586588
case 36: type = LLM_TYPE_8B; break; // granite

src/llama-model.h

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -32,6 +32,7 @@ enum llm_type {
3232
LLM_TYPE_190M,
3333
LLM_TYPE_220M,
3434
LLM_TYPE_250M,
35+
LLM_TYPE_256M,
3536
LLM_TYPE_270M,
3637
LLM_TYPE_335M,
3738
LLM_TYPE_410M,

src/llama-vocab.cpp

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1846,6 +1846,7 @@ void llama_vocab::impl::load(llama_model_loader & ml, const LLM_KV & kv) {
18461846
|| t.first == "<EOT>"
18471847
|| t.first == "_<EOT>"
18481848
|| t.first == "<|end▁of▁sentence|>" // DeepSeek
1849+
|| t.first == "<end_of_utterance>" // smoldocling
18491850
) {
18501851
special_eot_id = t.second;
18511852
if ((id_to_token[t.second].attr & LLAMA_TOKEN_ATTR_CONTROL) == 0) {
@@ -2005,6 +2006,7 @@ void llama_vocab::impl::load(llama_model_loader & ml, const LLM_KV & kv) {
20052006
|| t.first == "<EOT>"
20062007
|| t.first == "_<EOT>"
20072008
|| t.first == "<|end_of_text|>"
2009+
|| t.first == "<end_of_utterance>" // smoldocling
20082010
) {
20092011
special_eog_ids.insert(t.second);
20102012
if ((id_to_token[t.second].attr & LLAMA_TOKEN_ATTR_CONTROL) == 0) {

0 commit comments

Comments
 (0)