Closed
Description
What happened + What you expected to happen
On following all the steps mentioned in the tutorial: https://docs.ray.io/en/latest/serve/tutorials/vllm-example.html, I get the following error:
`(ServeReplica:default:VLLMDeployment pid=1328933) return any(model.name == model_name for model in self.base_model_paths)
(ServeReplica:default:VLLMDeployment pid=1328933) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
(ServeReplica:default:VLLMDeployment pid=1328933) File "/root/miniconda3/envs/venv/lib/python3.11/site-packages/vllm/entrypoints/openai/serving_engine.py", line 665, in <genexpr>
(ServeReplica:default:VLLMDeployment pid=1328933) return any(model.name == model_name for model in self.base_model_paths)
(ServeReplica:default:VLLMDeployment pid=1328933) ^^^^^^^^^^
(ServeReplica:default:VLLMDeployment pid=1328933) AttributeError: 'str' object has no attribute 'name'`
The model I was trying out was the one mentioned in the tutorial: NousResearch/Meta-Llama-3-8B-Instruct
Versions / Dependencies
vllm==0.6.6.post
ray==2.41.0
Reproduction script
Create llm.py and query.py as mentioned in the tutorial and run the following commands:
serve run llm:build_app model="NousResearch/Meta-Llama-3-8B-Instruct" tensor-parallel-size=2 accelerator="GPU"
python query.py
Issue Severity
High: It blocks me from completing my task.