Open
Description
jetmoe-8b
model runs fine but for jetmoe-8b-chat
with even the latest transformers
and tokenizer
I get:
Traceback (most recent call last):
File "/home/cqrl/.local/lib/python3.11/site-packages/transformers/models/auto/configuration_auto.py", line 932, in from_pretrained
config_class = CONFIG_MAPPING[config_dict["model_type"]]
~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/cqrl/.local/lib/python3.11/site-packages/transformers/models/auto/configuration_auto.py", line 634, in __getitem__
raise KeyError(key)
KeyError: 'jetmoe'
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/cqrl/jetmoe/chat-jet.py", line 6, in <module>
model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=torch.bfloat16, attn_implementation="eager", trust_remote_code=True)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/cqrl/.local/lib/python3.11/site-packages/transformers/models/auto/auto_factory.py", line 523, in from_pretrained
config, kwargs = AutoConfig.from_pretrained(
^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/cqrl/.local/lib/python3.11/site-packages/transformers/models/auto/configuration_auto.py", line 934, in from_pretrained
raise ValueError(
ValueError: The checkpoint you are trying to load has model type `jetmoe` but Transformers does not recognize this architecture. This could be because of an issue with the checkpoint, or because your version of Transformers is out of date.
Metadata
Metadata
Assignees
Labels
No labels