Skip to content

Commit

Permalink
tweak parameters
Browse files Browse the repository at this point in the history
  • Loading branch information
Samuel Weinbach committed May 13, 2021
1 parent f61bb53 commit 130d9ad
Showing 1 changed file with 4 additions and 3 deletions.
7 changes: 4 additions & 3 deletions tests/model/test_model_train.py
Original file line number Diff line number Diff line change
Expand Up @@ -16,8 +16,9 @@
["rmsnorm", "none", "swish"]],
"pipe_parallel_size,model_parallel_size": [[0, 1], [1, 2], [0, 2]],
"no_weight_tying": binary,
"attention_config": [[[["global"], "all"]], [[["local"], "all"]], [[["sparse_variable"], "all"]],
[[["sparse_fixed"], "all"]]],
"num_layers": [4],
"attention_config": [[[["global"], "all"]], [[["local", "global"], "all"]], [[["sparse_variable", "global"], "all"]],
[[["sparse_fixed", "global"], "all"]]],
"scaled_upper_triang_masked_softmax_fusion,bias_gelu_fusion": [[True, False], [False, True]],
"checkpoint_activations": binary,
"log_gradient_noise_scale": [True],
Expand Down Expand Up @@ -56,7 +57,7 @@ def run_train_test(yaml_list=None, param_dict=None):
from megatron.training import train_step
from megatron.utils import Timers

max_steps = 256
max_steps = 64

model, optimizer, lr_scheduler, args_loaded = model_setup(yaml_list, param_dict)

Expand Down

0 comments on commit 130d9ad

Please sign in to comment.