Skip to content

Commit ada7a8c

Browse files
Revert "[CI] Add Compiled DDP and Compiled FSDP2 tests to test_inductor_distributed (pytorch#138178)"
This reverts commit 8cb9110. Reverted pytorch#138178 on behalf of https://github.com/yf225 due to because pytorch#138174 is reverted, we need to revert this too ([comment](pytorch#138178 (comment)))
1 parent 59158f6 commit ada7a8c

File tree

3 files changed

+4
-15
lines changed

3 files changed

+4
-15
lines changed

.ci/pytorch/test.sh

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -320,7 +320,6 @@ test_inductor_distributed() {
320320
python test/run_test.py -i distributed/test_c10d_functional_native.py --verbose
321321
python test/run_test.py -i distributed/_tensor/test_dtensor_compile.py --verbose
322322
python test/run_test.py -i distributed/tensor/parallel/test_micro_pipeline_tp.py --verbose
323-
python test/run_test.py -i distributed/_composable/test_replicate_with_compiler.py --verbose
324323
python test/run_test.py -i distributed/_composable/fsdp/test_fully_shard_comm.py --verbose
325324
python test/run_test.py -i distributed/_composable/fsdp/test_fully_shard_training.py -k test_train_parity_multi_group --verbose
326325
python test/run_test.py -i distributed/_composable/fsdp/test_fully_shard_training.py -k test_train_parity_with_activation_checkpointing --verbose
@@ -332,7 +331,6 @@ test_inductor_distributed() {
332331
python test/run_test.py -i distributed/_composable/fsdp/test_fully_shard_mixed_precision.py -k test_compute_dtype --verbose
333332
python test/run_test.py -i distributed/_composable/fsdp/test_fully_shard_mixed_precision.py -k test_reduce_dtype --verbose
334333
python test/run_test.py -i distributed/_composable/fsdp/test_fully_shard_clip_grad_norm_.py -k test_clip_grad_norm_2d --verbose
335-
python test/run_test.py -i distributed/_composable/fsdp/test_fully_shard_compile.py --verbose
336334
python test/run_test.py -i distributed/fsdp/test_fsdp_tp_integration.py -k test_fsdp_tp_integration --verbose
337335

338336
# this runs on both single-gpu and multi-gpu instance. It should be smart about skipping tests that aren't supported

test/distributed/_composable/fsdp/test_fully_shard_compile.py

Lines changed: 4 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -613,8 +613,7 @@ def input_creation_fn():
613613
@skipIfRocm
614614
@unittest.skipIf(not HAS_GPU, "Inductor+gpu needs triton and recent GPU arch")
615615
def test_nested_fully_shard_backend_aot_eager(self):
616-
# TODO: fix fwd_fullgraph=False case
617-
for fwd_fullgraph in [True]:
616+
for fwd_fullgraph in [True, False]:
618617
self._test_traceable_fsdp(
619618
*self._create_nested_fully_shard_factory_fns(
620619
fwd_fullgraph=fwd_fullgraph
@@ -626,8 +625,7 @@ def test_nested_fully_shard_backend_aot_eager(self):
626625
@skipIfRocm
627626
@unittest.skipIf(not HAS_GPU, "Inductor+gpu needs triton and recent GPU arch")
628627
def test_nested_fully_shard_backend_aot_eager_decomp_partition(self):
629-
# TODO: fix fwd_fullgraph=False case
630-
for fwd_fullgraph in [True]:
628+
for fwd_fullgraph in [True, False]:
631629
self._test_traceable_fsdp(
632630
*self._create_nested_fully_shard_factory_fns(
633631
fwd_fullgraph=fwd_fullgraph
@@ -732,7 +730,6 @@ def test_nested_fully_shard_backend_inductor_fullgraph_True(self):
732730
)
733731
file_check.run(bwd_code)
734732

735-
@unittest.skip("TODO: fix fwd_fullgraph=False case")
736733
@skipIfRocm
737734
@unittest.skipIf(not HAS_GPU, "Inductor+gpu needs triton and recent GPU arch")
738735
def test_nested_fully_shard_backend_inductor_fullgraph_False(self):
@@ -813,9 +810,8 @@ def _sdpa_with_graph_break(*args, **kwargs):
813810
@skipIfRocm
814811
@unittest.skipIf(not HAS_GPU, "Inductor+gpu needs triton and recent GPU arch")
815812
def test_transformer_backend_aot_eager(self):
816-
# TODO: fix fwd_fullgraph=False case
817813
for fwd_fullgraph, all_requires_grad in itertools.product(
818-
[True], [True, False]
814+
[True, False], [True, False]
819815
):
820816
with self._maybe_add_graph_break_to_sdpa(
821817
fwd_fullgraph
@@ -833,9 +829,8 @@ def test_transformer_backend_aot_eager(self):
833829
# TODO: native_dropout has worse accuracy after decomp, need to figure out why
834830
@torch._inductor.config.patch(fallback_random=True)
835831
def test_transformer_backend_aot_eager_decomp_partition(self):
836-
# TODO: fix fwd_fullgraph=False case
837832
for fwd_fullgraph, all_requires_grad in itertools.product(
838-
[True], [True, False]
833+
[True, False], [True, False]
839834
):
840835
with self._maybe_add_graph_break_to_sdpa(fwd_fullgraph):
841836
self._test_traceable_fsdp(
@@ -951,7 +946,6 @@ def test_transformer_backend_inductor_fullgraph_True(self):
951946
)
952947
file_check.run(bwd_code)
953948

954-
@unittest.skip("TODO: fix fwd_fullgraph=False case")
955949
@skipIfRocm
956950
@unittest.skipIf(not HAS_GPU, "Inductor+gpu needs triton and recent GPU arch")
957951
# TODO: native_dropout causes CUDA IMA error, need to figure out why

test/distributed/_composable/test_replicate_with_compiler.py

Lines changed: 0 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -385,9 +385,6 @@ def setUp(self):
385385
def tearDown(self):
386386
dist.destroy_process_group()
387387

388-
@unittest.skip(
389-
"Temporarily disabled due to SymInt error: `unhashable type: non-nested SymInt`"
390-
)
391388
@unittest.skipIf(not HAS_GPU, "Inductor+gpu needs triton and recent GPU arch")
392389
@skipIfRocm
393390
def test_ddp_tp(self):

0 commit comments

Comments
 (0)