Skip to content

Conversation

feifeibear
Copy link
Contributor

@feifeibear feifeibear commented Jan 6, 2023

What's new

provide a memory efficient way to init model distributedly.
During ColoInitContext, it initializes 1/N parameter on each process.
Then, during tensor_parallelize, it changes the process group, distspec for tensor parallel.
So during the initialization phase, the overall memory requirement is just one copy of the model.

@feifeibear feifeibear merged commit 1aaeb59 into hpcaitech:main Jan 6, 2023
@feifeibear feifeibear deleted the dev0106_1 branch January 6, 2023 07:44
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant