Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Readme Update for FSDP #980

Merged
merged 4 commits into from
Jun 6, 2024
Merged

Readme Update for FSDP #980

merged 4 commits into from
Jun 6, 2024

Conversation

hlahkar
Copy link
Collaborator

@hlahkar hlahkar commented May 14, 2024

Updating Readme to indicate flash attention support for LLama2 70b with FSDP. This improves performance as well as memory footprint

What does this PR do?

Fixes # (issue)

Before submitting

  • This PR fixes a typo or improves the docs (you can dismiss the other checks if that's the case).
  • Did you make sure to update the documentation with your changes?
  • Did you write any new necessary tests?

Updating Readme to  indicate flash attention support for LLama2 70b with FSDP. This improves performance as well as memory footprint
@hlahkar hlahkar requested a review from vivekgoe May 14, 2024 04:14
@hlahkar hlahkar requested a review from regisss as a code owner May 14, 2024 04:14
@HuggingFaceDocBuilderDev

The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.

Copy link
Collaborator

@vivekgoe vivekgoe left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM.

@vivekgoe
Copy link
Collaborator

@regisss please review this readme change and merge this change when we transition to Synapse 1.16.0. We have tested this change on G2 with Synapse 1.16.0, this gives a small boost in performance.

@libinta
Copy link
Collaborator

libinta commented May 20, 2024

@hlahkar @vivekgoe can you also update the ci test and benchmark with the same parameters?

@regisss
Copy link
Collaborator

regisss commented May 21, 2024

Does this work with 1.15?

@hlahkar
Copy link
Collaborator Author

hlahkar commented May 23, 2024

@regisss This is for 1.16; it does not work properly for 1.15

@libinta libinta added synapse 1.16_dependency synapse 1.16 dependency and removed synapse1.16 labels May 23, 2024
Fused SDPA leads to a small drop in the perf for 7b model; so updating the test case. For 70b we see a perf improvement.
hsubramony added a commit that referenced this pull request May 29, 2024
@regisss regisss added the run-test Run CI for PRs from external contributors label May 30, 2024
hsubramony added a commit that referenced this pull request May 31, 2024
@hlahkar
Copy link
Collaborator Author

hlahkar commented Jun 6, 2024

@regisss Can we merge this now?

@regisss regisss merged commit acc7093 into huggingface:main Jun 6, 2024
11 of 12 checks passed
imangohari1 pushed a commit to imangohari1/optimum-habana that referenced this pull request Jun 13, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
run-test Run CI for PRs from external contributors synapse 1.16_dependency synapse 1.16 dependency
Projects
None yet
Development

Successfully merging this pull request may close these issues.

5 participants