Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add support for T5 for deepspeed-inference #2843

Closed
wants to merge 4 commits into from

Conversation

HeyangQin
Copy link
Contributor

@HeyangQin HeyangQin commented Feb 16, 2023

This is the PR to add support to T5. Currently it is still work in progress. A lot of the codes are adapted from #2451

@HeyangQin HeyangQin changed the title wrapper policy/container/model implementation for T5 Add support for T5 for deepspeed-inference Feb 16, 2023
@alexcoca
Copy link

alexcoca commented Mar 1, 2023

Hi @HeyangQin, thanks for your great work! Is there any update on the timeline of merging T5 support?

output_attentions=False,
return_dict=True,
): # Adapted from https://github.com/huggingface/transformers/blob/main/src/transformers/models/t5/modeling_t5.py
print("Using DS_T5Container.forward")

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

There is an extra log here

@baptistejamin
Copy link

For whatever reason, it seems it is requiring a lot more VRAM when this policy is injected.

@loadams loadams closed this Sep 7, 2023
@alexcoca
Copy link

alexcoca commented Sep 7, 2023

@HeyangQin @loadams I take it that this is not on your roadmap?

@HeyangQin
Copy link
Contributor Author

Hello @alexcoca. T5 support should already be live with #2962. Please feel free to let us know if you need more information

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants