Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Bug]: api_server.py: error: unrecognized arguments: --task embedding #9840

Closed
1 task done
javiplav opened this issue Oct 30, 2024 · 6 comments
Closed
1 task done
Labels
bug Something isn't working

Comments

@javiplav
Copy link

javiplav commented Oct 30, 2024

Your current environment

I'm running in a Kubernetes environment vllm docker image vllm/vllm-openai:latest .

The latest version is supposed to support --task embedding flag. https://docs.vllm.ai/en/latest/serving/openai_compatible_server.html

Model Input Dumps

No response

🐛 Describe the bug

I start the server with the following command: python3 -m vllm.entrypoints.openai.api_server and arguments: --served-model-name BAAI/bge-m3 --enforce-eager --task embedding .

The pod fails to run. Error message: api_server.py: error: unrecognized arguments: --task embedding

Any suggestions please?

Before submitting a new issue...

  • Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the documentation page, which can answer lots of frequently asked questions.
@javiplav javiplav added the bug Something isn't working label Oct 30, 2024
@DarkLight1337
Copy link
Member

DarkLight1337 commented Oct 30, 2024

I think --task embedding (#9424) has not been released outside of main branch yet. You'll have to clone the repository and build it from source to use it. Make sure that you have updated your Docker image to use the latest version (as in the latest, latest build).

@DarkLight1337
Copy link
Member

It looks like the "latest" Docker image there refers to the most recently released version, so it's not the one from main branch.

@DarkLight1337
Copy link
Member

Please follow these instructions for the latest code.

@javiplav
Copy link
Author

Thanks @DarkLight1337 . Managed to run the server with the main branch.

Unfortunately I didn't succeed in running bge-m3 as it seems that vllm doesn't support xlm-roberta architectures.

@DarkLight1337
Copy link
Member

Oh, I completely missed that you were trying to run bge-m3. Can you move over to #9847 ?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

2 participants