-
Notifications
You must be signed in to change notification settings - Fork 1.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Let DPOTrainer Support padding_free #2422
Comments
Thanks @fzyzcjy! Can you elaborate a bit? What is this padding free method? |
Oh sorry I provided the wrong link, now the link is updated to point to the correct "padding_free" article |
Thanks for the pointer. This would be nice addition! Any contribution is welcome. I mark this one as good second issue |
The guideline is basically to:
|
Thank you! |
but why make `xxx_attention_mask' optional ?
does this make sense? |
I think it makes sense yes. |
@fzyzcjy @qgallouedec if no one is working on this I would like to help |
@dame-cell I do not have time recently for that, and your PR would be great and many thanks! |
Is it possible for PPO to support padding_free? |
Feature request
Hi thanks for the library! It seems that https://huggingface.co/blog/packing-with-FA2 introduces a way to avoid a lot of pad tokens in SFT, and makes training faster. Therefore, it would be great if the same thing can be used for DPO.
Motivation
(see above)
Your contribution
n/a
The text was updated successfully, but these errors were encountered: