Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fine-tuning result gradually becoming noise #579

Open
5 tasks done
knabx opened this issue Sep 23, 2024 · 4 comments
Open
5 tasks done

Fine-tuning result gradually becoming noise #579

knabx opened this issue Sep 23, 2024 · 4 comments
Labels
enhancement New feature or request

Comments

@knabx
Copy link

knabx commented Sep 23, 2024

Self Checks

  • I have searched for existing issues search for existing issues, including closed ones.
  • I confirm that I am using English to submit this report (我已阅读并同意 Language Policy).
  • [FOR CHINESE USERS] 请务必使用英文提交 Issue,否则会被关闭。谢谢!:)
  • Please do not modify this template :) and fill in all the required fields.

1. Is this request related to a challenge you're experiencing? Tell me about your story.

yes, when I use fine-tuned model, result gradually becoming noise

2. Additional context or comments

No response

3. Can you help us with this feature?

  • I am interested in contributing to this feature.
@knabx knabx added the enhancement New feature or request label Sep 23, 2024
@knabx knabx changed the title Fine-tuning result gradually becoming noise, but top 5 accuracy is low Fine-tuning result gradually becoming noise Sep 23, 2024
@Stardust-minus
Copy link
Member

Can you share more information? Like how much data you use, the batch size, learning rate, and the step you train

@knabx
Copy link
Author

knabx commented Sep 23, 2024

Can you share more information? Like how much data you use, the batch size, learning rate, and the step you train

Thanks for your reply. I used 61 wav files, with an average length of about 3 minutes each. The batch size: 2, and accumulate_grad_batches: 4, learning rate: 1e-4, other parameters are the same as document. I trained for 1000 steps, saving checkpoint every 100 steps and validating the results . The loss gradually decreased, and the top 5 accuracy gradually increased, eventually reaching around 0.95. However, the generated audio by the checkpoints after 300 steps gradually declined, eventually become noise.

@PoTaTo-Mika
Copy link
Contributor

Can you share more information? Like how much data you use, the batch size, learning rate, and the step you train

Thanks for your reply. I used 61 wav files, with an average length of about 3 minutes each. The batch size: 2, and accumulate_grad_batches: 4, learning rate: 1e-4, other parameters are the same as document. I trained for 1000 steps, saving checkpoint every 100 steps and validating the results . The loss gradually decreased, and the top 5 accuracy gradually increased, eventually reaching around 0.95. However, the generated audio by the checkpoints after 300 steps gradually declined, eventually become noise.

Your learning rate is a bit too high, try 1e-5 to 5e-5. Also, the LLaMa part doesn't need much steps, about 100-300 is OK.

@knabx
Copy link
Author

knabx commented Sep 24, 2024

I have changed learning rate to 1e-5 and trained 300 steps, but loss and top 5 accuracy listed below. Does this look good? or should I train for more steps / increase learning rate?
train/loss=4.590, train/top_5_accuracy=0.350, val/loss=4.750, val/top_5_accuracy=0.354

BTW, there seems not much difference between audio generated by fine-tuned model and original model

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

3 participants