-
Notifications
You must be signed in to change notification settings - Fork 826
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Question] Reason behind removing lm_head
in modules
#13
Comments
Had the same thought. Have you figured it out? I didn't see anything in the paper either. If you want to add new tokens, you need to target the lm_head anyways. |
@artidoro @TimDettmers some insight on this would be greatly appreciated. |
Could someone share any observations or evaluations regarding the use of 'lm_head' as a target module? |
Not sure if this is related with this issue, but I found that when applying lora to Llama-2, and include |
Hello,
Thank you for the amazing repo. I was curious about this code below.
qlora/qlora.py
Lines 221 to 222 in e381744
Why is
lm_head
removed? What does it mean by for "needed for 16-bit"? Does it mean targeting this module for fp16 or so is incorrect?The text was updated successfully, but these errors were encountered: