-
-
Notifications
You must be signed in to change notification settings - Fork 4.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Feature]: AssertionError: Speculative decoding not yet supported for RayGPU backend. #4358
Comments
I am having the same issue
|
running into this as well |
Running into this as well |
2 similar comments
Running into this as well |
Running into this as well |
This issue should have been resolved by #4840 |
This issue has been automatically marked as stale because it has not had any activity within 90 days. It will be automatically closed if no further activity occurs within 30 days. Leave a comment if you feel this issue should remain open. Thank you! |
🚀 The feature, motivation and pitch
Hi,
Do you guys have any workaround for the
Speculative decoding not yet supported for RayGPU backend.
error or idea when the RayGPU backend will support speculative decoding?I run vllm server with the following command:
However, I got
AssertionError: Speculative decoding not yet supported for RayGPU backend.
Alternatives
No response
Additional context
No response
The text was updated successfully, but these errors were encountered: