-
-
Notifications
You must be signed in to change notification settings - Fork 120
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add RoPE scaling arguments to engine #220
Comments
There are some models for long context tasks like storywriting that it'd be nice to use with a static RoPE scaling factor. +1 on this! |
Hi getting error for bigger context models like Microsoft Phi 3 medium with respect to rope scaling factors with exl2 format. |
It is something related to this I think, maybe not much needs to be done here, just implement this code , I will try to test if it doesn't breaks anything else , here is the git in vllm for this feature vllm-project/vllm#4638 |
vllm-project/vllm#4298 vllm has implemented rotatry scale embeddings like this |
Added in v0.6.0. |
Currently, we auto-scale using the
--max-model-len
argument. It may be more appropriate to have specific options for the scaling factor, etc.The text was updated successfully, but these errors were encountered: