Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Feature]: Set RoPE scaling parameters dynamically #4334

Closed
LilianJim opened this issue Apr 24, 2024 · 0 comments · Fixed by #4638
Closed

[Feature]: Set RoPE scaling parameters dynamically #4334

LilianJim opened this issue Apr 24, 2024 · 0 comments · Fixed by #4638

Comments

@LilianJim
Copy link

LilianJim commented Apr 24, 2024

🚀 The feature, motivation and pitch

As it was implemented in #555, specifying RoPE parameters is only available from the model's config.json, and I haven't found a way to set it dynamically in my code. Is there currently a way of doing this?

Related to #910.

Alternatives

Right now, unless providing a modified config.json (which is very inconvenient in my setup), I haven't found an alternative. I've tried monkey patching vllm.transformers_utils.config.get_config function to no avail (Ray uses it in a way I don't understand).

Additional context

For context - specific to my setup - I'm using 2 GPUs to run a quantized Llama-3-70B (casperhansen's). Thus vLLM is using Ray.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging a pull request may close this issue.

1 participant