-
-
Notifications
You must be signed in to change notification settings - Fork 11.1k
[FA/Chore] Bump vllm-flash-attention #25537
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[FA/Chore] Bump vllm-flash-attention #25537
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Code Review
This pull request updates the vllm-flash-attention dependency to a newer commit to pick up several bug fixes. The change is straightforward and correct. I've added one suggestion to improve maintainability by documenting the reason for pinning this specific commit hash.
| vllm-flash-attn | ||
| GIT_REPOSITORY https://github.com/vllm-project/flash-attention.git | ||
| GIT_TAG ee4d25bd84e0cbc7e0b9b9685085fd5db2dcb62a | ||
| GIT_TAG 4695e6bed5366c41e28c06cd86170166e4f43d00 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
For better maintainability, it's good practice to add a comment explaining what this Git commit hash corresponds to. This helps other developers understand why this specific commit is being used without having to look it up. You could mention the PRs this commit includes, as you did in the pull request description.
# Includes fixes for FA3 build on CUDA 12.3, 12.4, and 12.5.
# See https://github.com/vllm-project/flash-attention/pull/94
GIT_TAG 4695e6bed5366c41e28c06cd86170166e4f43d00
fd389d9 to
b67f5d3
Compare
|
@LucasWilkinson would you mind bumping directly to |
It takes a while to get it through CI (need to populate sccache) so ill try to create a fast follow that includes: vllm-project/flash-attention#96 👍 apologies for the the cumbersome process |
Signed-off-by: Lucas Wilkinson <lwilkins@redhat.com>
Signed-off-by: Lucas Wilkinson <lwilkins@redhat.com> Signed-off-by: yewentao256 <zhyanwentao@126.com>
Signed-off-by: Lucas Wilkinson <lwilkins@redhat.com> Signed-off-by: Tomer Asida <57313761+tomeras91@users.noreply.github.com>
Signed-off-by: Lucas Wilkinson <lwilkins@redhat.com>
Signed-off-by: Lucas Wilkinson <lwilkins@redhat.com> Signed-off-by: xuebwang-amd <xuebwang@amd.com>
Signed-off-by: Lucas Wilkinson <lwilkins@redhat.com>
Signed-off-by: Lucas Wilkinson <lwilkins@redhat.com>
Signed-off-by: Lucas Wilkinson <lwilkins@redhat.com> Signed-off-by: xuebwang-amd <xuebwang@amd.com>
Bump FA to pickup
vllm-project/flash-attention#94
vllm-project/flash-attention#91
vllm-project/flash-attention#87