Skip to content

fine_tune param #311

@JaimeJuanPerez

Description

@JaimeJuanPerez

Hello! I’ve run into an issue related to what seems like model “forgetfulness,” similar to what’s discussed in issue #163.

I fine-tune a model that performs very well on my custom dataset, but when I try to predict labels that were not present in that dataset, the results degrade significantly.

As an additional attempt to mitigate this, I tried using the fine_tune parameter, which is supposed to freeze the encoder. However, it doesn’t appear to have any effect—the number of trainable parameters remains unchanged. Am I misunderstanding how this parameter is intended to work? Should fine-tuning always update the full architecture, or is the encoder actually meant to be frozen when fine_tune is set?

Thank you in advance!

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions