Align EOS token ID between tokenizer and generation config #663
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
This PR addresses a nasty footgun in
transformers
, where a change to the base model's tokenizer must ALSO be propagated to the model's generation config. Without this change, thepipeline()
function produces unbounded generations because it relies on the base model's EOS token (e.g.<|endoftext|>
) instead of the one we set in the tokenizer config (e.g.<|im_end|>
)Note that this has no impact on
vllm
since there the EOS token is inferred from thetokenizer_config.json
file.cc @qgallouedec @kashif for viz as this might be necessary to include on the SFT script of TRL too.
I also took the opportunity to clean up some of the "demo" recipes as I think it's better to have a single source of truth for well-tested recipes.