Skip to content

Commit 97f7ed9

Browse files
authored
Use correct default for round robin gradients (#1258)
* Make round robin gradient partitioning configurable (default False) * Use the correct default * Log config setting
1 parent bfe7f0d commit 97f7ed9

File tree

2 files changed

+2
-1
lines changed

2 files changed

+2
-1
lines changed

deepspeed/runtime/zero/config.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -189,4 +189,4 @@ def _initialize(self, zero_config_dict):
189189
self.round_robin_gradients = get_scalar_param(
190190
zero_config_dict,
191191
ZERO_OPTIMIZATION_ROUND_ROBIN_GRADIENTS,
192-
ZERO3_OPTIMIZATION_CONTIGUOUS_GRADIENTS_DEFAULT)
192+
ZERO_OPTIMIZATION_ROUND_ROBIN_GRADIENTS_DEFAULT)

deepspeed/runtime/zero/stage2.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -106,6 +106,7 @@ def __init__(self,
106106
logger.info(f"Reduce bucket size {reduce_bucket_size}")
107107
logger.info(f"Allgather bucket size {allgather_bucket_size}")
108108
logger.info(f"CPU Offload: {cpu_offload}")
109+
logger.info(f'Round robin gradient partitioning: {round_robin_gradients}')
109110
# The fused optimizer does all the work. We need this layer for two reason:
110111
# 1. maintain same user API from apex.fp16_utils
111112
# 2. keep common stuff here in case we need to add ne552w fused optimizer later

0 commit comments

Comments
 (0)