Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Performance in T5.1.1 #6

Open
toriving opened this issue Oct 25, 2022 · 4 comments
Open

Performance in T5.1.1 #6

toriving opened this issue Oct 25, 2022 · 4 comments

Comments

@toriving
Copy link

toriving commented Oct 25, 2022

Hi, I think this is a very cool study.
I have a question about this study. Have you tried the T5.1.1?
When I experimented, it was confirmed that the performance of T5.1.1 was lower than that of T5.
Not only that, there were cases where the performance was significantly lower than full fine-tuning.
Do you have any comments regarding this?

@ylsung
Copy link
Owner

ylsung commented Oct 28, 2022

Thanks for reporting this. I didn't try T5.1.1 before. I will look into this when I am not that swamped.... Probably some time at the end of November.

@wang-yiwei
Copy link

Hello Sung 😄
Do you have any plan to integrate your tuning algorithm into hugginface's peft lib?

@ylsung
Copy link
Owner

ylsung commented May 1, 2023

I would love to, and I do have a simpler implementation right now. But currently, I don't have a concrete timeline for when I can contribute to the peft lib, as I have some new projects at hand now.

@qincuiping
Copy link

I would love to, and I do have a simpler implementation right now. But currently, I don't have a concrete timeline for when I can contribute to the peft lib, as I have some new projects at hand now.

We want to try other models (like llama) on your tuning algorithm, so hope you could take some time out to integrating your tuning algorithm into hugginface's peft lib. 😄

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants