Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
53 changes: 0 additions & 53 deletions src/transformers/testing_utils.py
Original file line number Diff line number Diff line change
Expand Up @@ -15,7 +15,6 @@
import ast
import collections
import contextlib
import copy
import doctest
import functools
import gc
Expand Down Expand Up @@ -1598,58 +1597,6 @@ def assert_screenout(out, what):
assert match_str != -1, f"expecting to find {what} in output: f{out_pr}"


def set_model_tester_for_less_flaky_test(test_case):
# NOTE: this function edits the config object, which may lead to hard-to-debug side-effects. Use with caution.
# Do not use in tests/models where objects behave very differently based on the config's hidden layer settings
# (e.g. KV caches, sliding window attention, ...)

# TODO (if possible): Avoid exceptional cases
exceptional_classes = [
"ZambaModelTester",
"Zamba2ModelTester",
"RwkvModelTester",
"AriaVisionText2TextModelTester",
"GPTNeoModelTester",
"DPTModelTester",
"Qwen3NextModelTester",
]
if test_case.model_tester.__class__.__name__ in exceptional_classes:
return

target_num_hidden_layers = 1
if hasattr(test_case.model_tester, "out_features") or hasattr(test_case.model_tester, "out_indices"):
target_num_hidden_layers = None

if hasattr(test_case.model_tester, "num_hidden_layers") and target_num_hidden_layers is not None:
test_case.model_tester.num_hidden_layers = target_num_hidden_layers
if (
hasattr(test_case.model_tester, "vision_config")
and "num_hidden_layers" in test_case.model_tester.vision_config
and target_num_hidden_layers is not None
):
test_case.model_tester.vision_config = copy.deepcopy(test_case.model_tester.vision_config)
if isinstance(test_case.model_tester.vision_config, dict):
test_case.model_tester.vision_config["num_hidden_layers"] = 1
else:
test_case.model_tester.vision_config.num_hidden_layers = 1
if (
hasattr(test_case.model_tester, "text_config")
and "num_hidden_layers" in test_case.model_tester.text_config
and target_num_hidden_layers is not None
):
test_case.model_tester.text_config = copy.deepcopy(test_case.model_tester.text_config)
if isinstance(test_case.model_tester.text_config, dict):
test_case.model_tester.text_config["num_hidden_layers"] = 1
else:
test_case.model_tester.text_config.num_hidden_layers = 1

# A few model class specific handling

# For Albert
if hasattr(test_case.model_tester, "num_hidden_groups"):
test_case.model_tester.num_hidden_groups = test_case.model_tester.num_hidden_layers


def set_config_for_less_flaky_test(config):
target_attrs = [
"rms_norm_eps",
Expand Down
3 changes: 0 additions & 3 deletions tests/models/efficientloftr/test_modeling_efficientloftr.py
Original file line number Diff line number Diff line change
Expand Up @@ -23,7 +23,6 @@
require_vision,
set_config_for_less_flaky_test,
set_model_for_less_flaky_test,
set_model_tester_for_less_flaky_test,
slow,
torch_device,
)
Expand Down Expand Up @@ -360,8 +359,6 @@ def recursive_check(batched_object, single_row_object, model_name, key):
msg += str(e)
raise AssertionError(msg)

set_model_tester_for_less_flaky_test(self)

config, batched_input = self.model_tester.prepare_config_and_inputs_for_common()
set_config_for_less_flaky_test(config)

Expand Down