-
Notifications
You must be signed in to change notification settings - Fork 24
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Wrap evaluation benchmark using HF-trainer #61
Comments
Hey @sbmaruf, thanks for the input! We haven't thought much about parallelizing large models yet (focusing on the baselines for now), but I totally agree this is something we should have in mind moving forward. Do you imagine the process would require a lot of modification to the codebase? Just wondering what a proof of concept implementation of this would look like. |
For a proof of concept implementation, These are some of the possible places that might get affected, *** (I'm pasting some example codes from one of my large codebase, there might be some redundant stuff) ***
I can work/review on related pull request if you want. |
This might sounds like a bit of re-structuring but for the sake of future compatibility, I propose the following,
huggingface
trainer: This will help the repo to automatically adapt todeepspeed
and all the exclusive features of transformers library.--
data_loader
--
DataCollator
--
compute_metrics
--
predictions
(if needed)finetune
our full model, we don't have to change a lot in the surface level.I would love to take some responsibility if needed. Let me know. @jaketae @tianjianjiang @wilsonyhlee
The text was updated successfully, but these errors were encountered: