You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I would like to understand if I need to retrain or fine-tune the EncoderModel after adding these new words to the tokenizer. How will this modification affect the model's performance or behavior?
This question is related to the Imagen project, and I want to ensure that I am following the correct approach when incorporating new words into the tokenizer.
The text was updated successfully, but these errors were encountered:
When you add new tokens to the vocabulary (and add the entry in the embedding layer), you'll end up with randomly initialized values corresponding to the new token(s).
That I have added some new words to
t5.get_tokenizer()
as shown below:I would like to understand if I need to retrain or fine-tune the EncoderModel after adding these new words to the tokenizer. How will this modification affect the model's performance or behavior?
This question is related to the Imagen project, and I want to ensure that I am following the correct approach when incorporating new words into the tokenizer.
The text was updated successfully, but these errors were encountered: