Skip to content

Commit c3e195f

Browse files
amirai21pwilkin
authored andcommitted
convert : correctly handle LLaMA tokenizer for Jamba (ggml-org#16470)
* fix: convert_hf_to_gguf - change Jamba non-sentencepiece mode (tokenizer.json) vocab construction * fix: convert_hf_to_gguf - jamba non-sentencepiece tokenizer to use _set_vocab_llama_hf func * fix: convert_hf_to_gguf - removed get_vocab_base_pre from jamba
1 parent ebf89db commit c3e195f

File tree

1 file changed

+2
-10
lines changed

1 file changed

+2
-10
lines changed

convert_hf_to_gguf.py

Lines changed: 2 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -5996,20 +5996,12 @@ def modify_tensors(self, data_torch: Tensor, name: str, bid: int | None) -> Iter
59965996
class JambaModel(TextModel):
59975997
model_arch = gguf.MODEL_ARCH.JAMBA
59985998

5999-
def get_vocab_base_pre(self, tokenizer) -> str:
6000-
del tokenizer # unused
6001-
6002-
return "gpt-2"
6003-
60045999
def set_vocab(self):
60056000
if (self.dir_model / "tokenizer.model").is_file():
6006-
# Using Jamba's tokenizer.json causes errors on model load
6007-
# (something about "byte not found in vocab"),
6008-
# but there's a working tokenizer.model
60096001
self._set_vocab_sentencepiece()
60106002
else:
6011-
# Some Jamba models only have a tokenizer.json, which works.
6012-
self._set_vocab_gpt2()
6003+
self._set_vocab_llama_hf()
6004+
self.gguf_writer.add_add_space_prefix(False)
60136005

60146006
def set_gguf_parameters(self):
60156007
d_model = self.find_hparam(["hidden_size", "mamba_d_model"])

0 commit comments

Comments
 (0)