Open
Description
This might sounds like a bit of re-structuring but for the sake of future compatibility, I propose the following,
- Move to
huggingface
trainer: This will help the repo to automatically adapt todeepspeed
and all the exclusive features of transformers library. - We don't have to re-invent the wheel. Given that we are using huggingface trainer, we only need to implement the following functions for a trainer for different tasks.
--data_loader
--DataCollator
--compute_metrics
--predictions
(if needed) - In case if we want to
finetune
our full model, we don't have to change a lot in the surface level.
I would love to take some responsibility if needed. Let me know. @jaketae @tianjianjiang @wilsonyhlee
Metadata
Metadata
Assignees
Labels
No labels