File tree Expand file tree Collapse file tree 2 files changed +2
-1
lines changed Expand file tree Collapse file tree 2 files changed +2
-1
lines changed Original file line number Diff line number Diff line change @@ -189,4 +189,4 @@ def _initialize(self, zero_config_dict):
189189 self .round_robin_gradients = get_scalar_param (
190190 zero_config_dict ,
191191 ZERO_OPTIMIZATION_ROUND_ROBIN_GRADIENTS ,
192- ZERO3_OPTIMIZATION_CONTIGUOUS_GRADIENTS_DEFAULT )
192+ ZERO_OPTIMIZATION_ROUND_ROBIN_GRADIENTS_DEFAULT )
Original file line number Diff line number Diff line change @@ -106,6 +106,7 @@ def __init__(self,
106106 logger .info (f"Reduce bucket size { reduce_bucket_size } " )
107107 logger .info (f"Allgather bucket size { allgather_bucket_size } " )
108108 logger .info (f"CPU Offload: { cpu_offload } " )
109+ logger .info (f'Round robin gradient partitioning: { round_robin_gradients } ' )
109110 # The fused optimizer does all the work. We need this layer for two reason:
110111 # 1. maintain same user API from apex.fp16_utils
111112 # 2. keep common stuff here in case we need to add ne552w fused optimizer later
You can’t perform that action at this time.
0 commit comments