-
Notifications
You must be signed in to change notification settings - Fork 1.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Bump vllm to v0.4.2 #7198
Bump vllm to v0.4.2 #7198
Conversation
Hi @kebe7jun, thanks for submitting the PR. Can you elaborate on the specific models or features you're interested in that require this version upgrade? |
I need LLama3 optimizations from 0.4.1, as well as phi-3-mini support from 0.4.2. See: https://github.com/vllm-project/vllm/releases I have already signed the CLA, and I have had PR merged before. |
Looks like this might need a different version of |
@rmccorm4 , yesy, I was working on this. In latest vllm they've added installation of |
vLLM backend PR: triton-inference-server/vllm_backend#43 |
@kebe7jun, please rebase your branch with the latest main. Thank you. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM with @pskiran1 changes: triton-inference-server/vllm_backend#43
@pskiran1 , this branch has no conflicts, thus re-base is unnecessary IMO. Feel free to merge this PR with yours |
No description provided.