Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Converter not working for NLLB models #1711

Open
marcelgoya opened this issue May 27, 2024 · 9 comments
Open

Converter not working for NLLB models #1711

marcelgoya opened this issue May 27, 2024 · 9 comments

Comments

@marcelgoya
Copy link

When I run the following converter script:

ct2-transformers-converter --model facebook/nllb-200-distilled-1.3B --quantization float16 --output_dir nllb-200-distilled-1.3B-ct2-float16

I now get the following error:

config.json: 100%|████████████████████████████████████████████████████████████████████| 808/808 [00:00<00:00, 4.01MB/s] pytorch_model.bin: 100%|███████████████████████████████████████████████████████████| 5.48G/5.48G [00:22<00:00, 248MB/s] generation_config.json: 100%|█████████████████████████████████████████████████████████| 189/189 [00:00<00:00, 1.27MB/s] tokenizer_config.json: 100%|██████████████████████████████████████████████████████████| 564/564 [00:00<00:00, 3.51MB/s] sentencepiece.bpe.model: 100%|█████████████████████████████████████████████████████| 4.85M/4.85M [00:00<00:00, 139MB/s] tokenizer.json: 100%|██████████████████████████████████████████████████████████████| 17.3M/17.3M [00:00<00:00, 335MB/s] special_tokens_map.json: 100%|████████████████████████████████████████████████████| 3.55k/3.55k [00:00<00:00, 20.6MB/s] Traceback (most recent call last): File "/usr/local/bin/ct2-transformers-converter", line 8, in <module> sys.exit(main()) File "/usr/local/lib/python3.10/dist-packages/ctranslate2/converters/transformers.py", line 2234, in main converter.convert_from_args(args) File "/usr/local/lib/python3.10/dist-packages/ctranslate2/converters/converter.py", line 50, in convert_from_args return self.convert( File "/usr/local/lib/python3.10/dist-packages/ctranslate2/converters/converter.py", line 89, in convert model_spec = self._load() File "/usr/local/lib/python3.10/dist-packages/ctranslate2/converters/transformers.py", line 142, in _load spec = loader(model, tokenizer) File "/usr/local/lib/python3.10/dist-packages/ctranslate2/converters/transformers.py", line 194, in __call__ spec = self.get_model_spec(model) File "/usr/local/lib/python3.10/dist-packages/ctranslate2/converters/transformers.py", line 429, in get_model_spec return super().get_model_spec(model) File "/usr/local/lib/python3.10/dist-packages/ctranslate2/converters/transformers.py", line 257, in get_model_spec self.set_encoder(spec.encoder, model.model.encoder) File "/usr/local/lib/python3.10/dist-packages/ctranslate2/converters/transformers.py", line 286, in set_encoder self.set_common_layers(spec, encoder) File "/usr/local/lib/python3.10/dist-packages/ctranslate2/converters/transformers.py", line 347, in set_common_layers spec.scale_embeddings = module.embed_scale File "/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py", line 1709, in __getattr__ raise AttributeError(f"'{type(self).__name__}' object has no attribute '{name}'") AttributeError: 'M2M100Encoder' object has no attribute 'embed_scale'

@minhthuc2502
Copy link
Collaborator

Which ctranslate2's version that you used? I don't have any issue like this.

@marcelgoya
Copy link
Author

marcelgoya commented May 27, 2024

@minhthuc2502 I am using the latest version, 4.3.0. I am also using the latest Transformers and Torch versions

@minhthuc2502
Copy link
Collaborator

Try to reinstall Transformers. M2M100Encoder has attribute embed_scale.

@jorirsan
Copy link

Can confirm that the conversion script when using Transformers >=4.41.0 is broken. Looking at the commit history of modeling_m2m_100.py in the Tranformers repo, it seems that huggingface/transformers#30410 may be the cause

@btseytlin
Copy link

Having the same problem

@thewh1teagle
Copy link

Can confirm that the conversion script when using Transformers >=4.41.0 is broken

Using

pip install transformers==4.40.2

Fixed the issue for now

@thangld201
Copy link

thangld201 commented Jul 9, 2024

For new transformers version (v.4.41+), embed_scale is no longer a class attribute. I think we would need to change converter script (python/ctranslate2/converters/transformers.py)
e.g. from

    def set_common_layers(self, spec, module):
        spec.scale_embeddings = module.embed_scale

to

    def set_common_layers(self, spec, module):
        import math
        if not hasattr(module,'embed_scale'): embed_scale = math.sqrt(module.config.d_model) if module.config.scale_embedding else 1.0
        else: embed_scale = module.embed_scale
        spec.scale_embeddings = embed_scale

@mayowaosibodu
Copy link

For new transformers version (v.4.41+), embed_scale is no longer a class attribute. I think we would need to change converter script (python/ctranslate2/converters/transformers.py) e.g. from

    def set_common_layers(self, spec, module):
        spec.scale_embeddings = module.embed_scale

to

    def set_common_layers(self, spec, module):
        import math
        if not hasattr(module,'embed_scale'): embed_scale = math.sqrt(module.config.d_model) if module.config.scale_embedding else 1.0
        else: embed_scale = module.embed_scale
        spec.scale_embeddings = embed_scale

Super helpful. Editing my ctranslate2/converters/transformers.py this way, helped me get the ct2 conversion working. Thanks.

@minhthuc2502
Copy link
Collaborator

Sorry for the late response.

For new transformers version (v.4.41+), embed_scale is no longer a class attribute. I think we would need to change converter script (python/ctranslate2/converters/transformers.py) e.g. from

Thank you for your investigation. I will update the script transformers.py.

homink pushed a commit to homink/CTranslate2 that referenced this issue Sep 13, 2024
minhthuc2502 pushed a commit that referenced this issue Sep 13, 2024
* Wav2Vec2 upgrade with Conv1D options

* refining scripts

* refining script again

* fix the formats

* fix the isort format

* refining the library

* update based on the suggestions

* update the variable name

* adding unk_token removal for the Python testing

* adding whitespace

* update Python format

* update variables

* update variables

* update variables

* update variables

* Wav2Vec2Bert ASR support

* sync with the main repository

* update missing parts

* update missing parts2

* update the logic for make_relative_positions

* update test case name

* separate the asymmetric relative positions

* clean empty lines

* update typo

* adding the version check for transformers

* udpate the format

* adding the version check for transformers2

* upgrade transformers 4.41 for python test

* patch from #1711

---------

Co-authored-by: hkwon <homin.kwon@sri.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

7 participants