Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fine-tuning #23

Closed
alaaj27 opened this issue Feb 27, 2023 · 21 comments
Closed

Fine-tuning #23

alaaj27 opened this issue Feb 27, 2023 · 21 comments

Comments

@alaaj27
Copy link

alaaj27 commented Feb 27, 2023

Is it possible to Fine-tune LLaMA for downstream tasks? If so, how can we do that?

Edit: Reading the other opened issues, I realized that neither the training data nor the pre-trained weights were released. How the code is going to be useful anyway?

@archwolf118
Copy link

I have same question. Thank you for reply.

@shangdibufashi
Copy link

+1

6 similar comments
@gsanou
Copy link

gsanou commented Mar 2, 2023

+1

@tweedmann
Copy link

+1

@liaoxingjian
Copy link

+1

@aohenuo
Copy link

aohenuo commented Mar 3, 2023

+1

@chembiome
Copy link

+1

@comtomru
Copy link

comtomru commented Mar 3, 2023

+1

@turbo
Copy link

turbo commented Mar 4, 2023

There's an RLHF harness here: https://github.com/nebuly-ai/nebullvm/tree/main/apps/accelerate/chatllama

Instruction fine tuning should also be pretty easy.

@mvllwong
Copy link

mvllwong commented Mar 5, 2023

++

@sequoiar
Copy link

sequoiar commented Mar 7, 2023

+1

1 similar comment
@kev216
Copy link

kev216 commented Mar 7, 2023

+1

@ruian0
Copy link

ruian0 commented Mar 7, 2023

  • 1

@Chenzongchao
Copy link

+1

4 similar comments
@secain
Copy link

secain commented Mar 10, 2023

+1

@zanghu
Copy link

zanghu commented Mar 12, 2023

+1

@wallon-ai
Copy link

+1

@hujunchao
Copy link

+1

@ruian0
Copy link

ruian0 commented Mar 17, 2023

Found this work from Stanford Alpaca. They read and resave the model(so not dependent on FAIR) and it becomes trainable with the hugging face training framework.

https://crfm.stanford.edu/2023/03/13/alpaca.html
huggingface/transformers#21955

@dneupan1
Copy link

dneupan1 commented Aug 2, 2023

+1

@subramen
Copy link
Contributor

We recently shared scripts for finetuning and inference at https://github.com/facebookresearch/llama-recipes

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests