Open
Description
openedon Apr 20, 2022
Hi, I'm adding deepspeed for my distributed model training framework.
When using pytorch native apis, everything is fine. For distributed training, originally I would wrapp model in an object of nn.parallel.DistributedDataParallel, and use model.no_sync() api to avoid unnecessary sync ops.
I cannot find the equivalent api in deepspeed. Can you offer some help?
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment