Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Support for LLaMA #104

Closed
ustcwhy opened this issue Mar 31, 2023 · 1 comment
Closed

Support for LLaMA #104

ustcwhy opened this issue Mar 31, 2023 · 1 comment

Comments

@ustcwhy
Copy link

ustcwhy commented Mar 31, 2023

Thanks for your wonderful work!
Meta released their newest LLM, LLaMA. The checkpoint is available on Huggingface[1]. zphang has presented the code to use LLaMA based on the transformers repo. For FlexGen, could I directly replace OPT model with LLaMA to make inferences on a local card? Do you have any plan to support LLaMA in the future?

[1] https://huggingface.co/decapoda-research
[2] huggingface/transformers#21955

@BarfingLemurs
Copy link

(duplicate) #60

@ustcwhy ustcwhy closed this as completed Apr 10, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants