Skip to content

convert_hf_to_gguf: add JoyAI-LLM-Flash tokenizer hash mapping to deepseek-v3#19651

Merged
CISC merged 5 commits intoggml-org:masterfrom
dranger003:joyai-llm-flash
Feb 16, 2026
Merged

convert_hf_to_gguf: add JoyAI-LLM-Flash tokenizer hash mapping to deepseek-v3#19651
CISC merged 5 commits intoggml-org:masterfrom
dranger003:joyai-llm-flash

Conversation

@dranger003
Copy link
Contributor

  • adding hash for jdopensource/JoyAI-LLM-Flash mapping to existing deepseek-v3
  • DeepseekV3ForCausalLM architecture already supported
  • moved GLM-4.7-Flash entry together with the other glm entries
  • conversion and quantization successful and output generation looks great

@CISC
Copy link
Collaborator

CISC commented Feb 16, 2026

Thanks, but convert_hf_to_gguf_update.py needs to be run to update convert_hf_to_gguf.py, which is why CI failed. :)

Anyway, I think I prefer that you create a new pre-tokenizer name for this in case there are any subtle differences discovered later on; move it out of pre_computed_hashes and add it in llama-vocab.cpp together with DEEPSEEK3_LLM and HUNYAN_DENSE.

@dranger003
Copy link
Contributor Author

Ah, of course I changed the order after running the script, thanks.

Copy link
Collaborator

@CISC CISC left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

You also need to add a section like this:

llama.cpp/src/llama-vocab.cpp

Lines 2039 to 2042 in 7af1cce

} else if (
tokenizer_pre == "hunyuan-dense") {
pre_type = LLAMA_VOCAB_PRE_TYPE_HUNYUAN_DENSE;
clean_spaces = false;

@CISC CISC merged commit 05fa625 into ggml-org:master Feb 16, 2026
54 of 79 checks passed
michaelneale added a commit to michaelneale/llama.cpp that referenced this pull request Feb 17, 2026
* upstream/master: (88 commits)
  ci : bump komac version (ggml-org#19682)
  build : link ws2_32 as PUBLIC on Windows (ggml-org#19666)
  build : cleanup library linking logic (ggml-org#19665)
  convert : add JoyAI-LLM-Flash (ggml-org#19651)
  perplexity: add proper batching (ggml-org#19661)
  common : inline functions (ggml-org#18639)
  ggml : make `ggml_is_view` as API (ggml-org#19539)
  model: Add support for Tiny Aya Models (ggml-org#19611)
  build : rework llama_option_depr to handle LLAMA_CURL (ggml-org#19658)
  Adjust workaround for ROCWMMA_FATTN/GFX9 to only newer ROCm veresions (ggml-org#19591)
  models : deduplicate delta-net graphs for Qwen family (ggml-org#19597)
  graph : fix KQ mask, lora, cvec reuse checks (ggml-org#19644)
  ggml: aarch64: Implement SVE in Gemm q4_k 8x8 q8_k Kernel  (ggml-org#19132)
  sync : ggml
  ggml : bump version to 0.9.7 (ggml/1425)
  ggml : bump version to 0.9.6 (ggml/1423)
  cuda: optimize iq2xxs/iq2xs/iq3xxs dequantization (ggml-org#19624)
  docs: update s390x build docs (ggml-org#19643)
  build : remove LLAMA_HTTPLIB option (ggml-org#19623)
  cmake : check if KleidiAI API has been fetched (ggml-org#19640)
  ...
liparetejas pushed a commit to liparetejas/llama.cpp that referenced this pull request Feb 23, 2026
* convert_hf_to_gguf: add JoyAI-LLM-Flash tokenizer hash mapping to deepseek-v3

* llama-vocab: create a new pre-tokenizer name for joyai-llm.

* add missing vocab type section

* Update convert_hf_to_gguf_update.py

Co-authored-by: Sigbjørn Skjæret <sigbjorn.skjaeret@scala.com>

* Update convert_hf_to_gguf.py

Co-authored-by: Sigbjørn Skjæret <sigbjorn.skjaeret@scala.com>

---------

Co-authored-by: Sigbjørn Skjæret <sigbjorn.skjaeret@scala.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

python python script changes

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants