You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
When using LLM to do NER task, there is a warning saying "This is a friendly reminder - the current text generation call will exceed the model's predefined maximum length (4096). Depending on the model, you may observe exceptions, performance degradation, or nothing at all."
How to change the maximum length of the LLM output?
Hi @yileitu, all model parameter are forwarded to transformers that handles the model. In most cases there is a max_length or max_new_tokens parameter you can set:
[components.llm.model]
@llm_models = "spacy.Llama2.v1"name = "Llama-2-7b-hf"max_length = 8192 # or any other value you want to set
When using LLM to do NER task, there is a warning saying "This is a friendly reminder - the current text generation call will exceed the model's predefined maximum length (4096). Depending on the model, you may observe exceptions, performance degradation, or nothing at all."
How to change the maximum length of the LLM output?
The text was updated successfully, but these errors were encountered: