Open
Description
Hi!
Could you let me know the minimum GPU resources needed to fine-tune the 32B model?
I trained the 32B model using 8 L40 GPUs, but the process was interrupted during model loading.
It seems like the model has not been loaded into the GPUs.
But it was okay when I trained the 3B model
I want to know whether I can train the 32B model with 8 L40 GPUs (8*40G)
Thanks!

Metadata
Metadata
Assignees
Labels
No labels