Replies: 1 comment
-
We tried further finetune the ReProver checkpoint, though the results were not promising. I believe it may have something to do with lean-dojo/LeanDojo#5. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi,
We're experimenting with fine-tuning ReProver on a slightly modified Mathlib dataset. We just used the training code in this repo with ReProver instead of ByT5 as the base model. It works, but I'm wondering if this is the right way. It perhaps results in a high initial learning rate and causes the model to forget too much in the first epochs.
Do you think it would be better to continue from ReProver's training checkpoint? Or should we just use a lower learning rate?
Thanks for any tips!
Beta Was this translation helpful? Give feedback.
All reactions