Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

About GPU memory usage #8

Open
JY-CCK opened this issue Apr 1, 2024 · 1 comment
Open

About GPU memory usage #8

JY-CCK opened this issue Apr 1, 2024 · 1 comment

Comments

@JY-CCK
Copy link

JY-CCK commented Apr 1, 2024

Hello.
First of all, thanks for sharing a bitnet training code.

I have a question about GPU memory usage.
As I understanding, bitnet can reduce VRAM usage compared to fp16/bf16 precision.
However, by commenting code in the train_bitnet.py
model = apply_bitlinear(model, target_layers=target_layers) # comment this to train og llama
memory usage is reduced about 2G.
(with bitnet layer, it used 13G v.s. w/o bitnet layer, 11G)

Doesn't it make sense that using bitnet would actually result in lower memory usage?

Thanks.

@joey00072
Copy link
Owner

Oh is it, lmk

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants