-
-
Notifications
You must be signed in to change notification settings - Fork 4.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Finetuned Flan-T5 #434
Comments
Hi @patrafter1999, thanks for your interest in vLLM and good question!
Currently, we do not support encoder-decoder models like T5. It's on our roadmap.
You can use the same API for the fine-tuned models saved on your local disk (if the model architecture is supported). For example, the following should work: from vllm import LLM
llm = LLM(model="path/to/local/model") Note that if your model is fine-tuned by LoRA, you should combine the LoRA weights into the original model weights before using vLLM. Currently, we do not natively support inference with LoRA adapters. |
Hi @WoosukKwon, Thanks a lot for your kind answer. I have finetuned a Santacoder model, which is supported by vLLM. I tried the following in databricks.
And I get this error.
The finetuned parameter file and tokenizer file is in that model_dir folder. What could be the cause of this error? Thanks. |
Closing as duplicate of #187 |
Hi vllm team,
I know you guys are extremely busy with many action items. vLLM is now becoming a must when you run LLM.
I plan to use a finetuned FLAN-T5 model. My question is:
Thanks a lot for your kind answers.
The text was updated successfully, but these errors were encountered: