-
-
Notifications
You must be signed in to change notification settings - Fork 4.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Bug]: api_server.py: error: unrecognized arguments: --task embedding #9840
Comments
I think |
It looks like the "latest" Docker image there refers to the most recently released version, so it's not the one from main branch. |
Please follow these instructions for the latest code. |
Thanks @DarkLight1337 . Managed to run the server with the main branch. Unfortunately I didn't succeed in running bge-m3 as it seems that vllm doesn't support xlm-roberta architectures. |
Oh, I completely missed that you were trying to run bge-m3. Can you move over to #9847 ? |
Your current environment
I'm running in a Kubernetes environment vllm docker image vllm/vllm-openai:latest .
The latest version is supposed to support --task embedding flag. https://docs.vllm.ai/en/latest/serving/openai_compatible_server.html
Model Input Dumps
No response
🐛 Describe the bug
I start the server with the following command:
python3 -m vllm.entrypoints.openai.api_server
and arguments:--served-model-name BAAI/bge-m3 --enforce-eager --task embedding
.The pod fails to run. Error message:
api_server.py: error: unrecognized arguments: --task embedding
Any suggestions please?
Before submitting a new issue...
The text was updated successfully, but these errors were encountered: