Skip to content

[vLLM] The ray serve using vLLM example on the website does not work. #50275

Closed
@power-puff-gg

Description

@power-puff-gg

What happened + What you expected to happen

On following all the steps mentioned in the tutorial: https://docs.ray.io/en/latest/serve/tutorials/vllm-example.html, I get the following error:

`(ServeReplica:default:VLLMDeployment pid=1328933)     return any(model.name == model_name for model in self.base_model_paths)
(ServeReplica:default:VLLMDeployment pid=1328933)            ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
(ServeReplica:default:VLLMDeployment pid=1328933)   File "/root/miniconda3/envs/venv/lib/python3.11/site-packages/vllm/entrypoints/openai/serving_engine.py", line 665, in <genexpr>

(ServeReplica:default:VLLMDeployment pid=1328933)     return any(model.name == model_name for model in self.base_model_paths)
(ServeReplica:default:VLLMDeployment pid=1328933)                ^^^^^^^^^^
(ServeReplica:default:VLLMDeployment pid=1328933) AttributeError: 'str' object has no attribute 'name'`

The model I was trying out was the one mentioned in the tutorial: NousResearch/Meta-Llama-3-8B-Instruct

Versions / Dependencies

vllm==0.6.6.post
ray==2.41.0

Reproduction script

Create llm.py and query.py as mentioned in the tutorial and run the following commands:

serve run llm:build_app model="NousResearch/Meta-Llama-3-8B-Instruct" tensor-parallel-size=2 accelerator="GPU"

python query.py

Issue Severity

High: It blocks me from completing my task.

Metadata

Metadata

Assignees

Labels

bugSomething that is supposed to be working; but isn'tllm

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions