Skip to content

Commit 02ac9de

Browse files
committed
fix tests after rebase
1 parent 38471f1 commit 02ac9de

File tree

2 files changed

+3
-0
lines changed

2 files changed

+3
-0
lines changed

test/test_utils/test_results_manager.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -352,6 +352,7 @@ def test_metric_results(metric, scores, ensemble_ends_later):
352352

353353

354354
def test_search_results_sprint_statistics():
355+
BaseTask.__abstractmethods__ = set()
355356
api = BaseTask()
356357
for method in ['get_search_results', 'sprint_statistics', 'get_incumbent_results']:
357358
with pytest.raises(RuntimeError):

test/test_utils/test_results_visualizer.py

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -146,6 +146,7 @@ def test_set_plot_args(params): # TODO
146146

147147
@pytest.mark.parametrize('metric_name', ('unknown', 'accuracy'))
148148
def test_raise_error_in_plot_perf_over_time_in_base_task(metric_name):
149+
BaseTask.__abstractmethods__ = set()
149150
api = BaseTask()
150151

151152
if metric_name == 'unknown':
@@ -159,6 +160,7 @@ def test_raise_error_in_plot_perf_over_time_in_base_task(metric_name):
159160
@pytest.mark.parametrize('metric_name', ('balanced_accuracy', 'accuracy'))
160161
def test_plot_perf_over_time(metric_name): # TODO
161162
dummy_history = [{'Timestamp': datetime(2022, 1, 1), 'train_accuracy': 1, 'test_accuracy': 1}]
163+
BaseTask.__abstractmethods__ = set()
162164
api = BaseTask()
163165
run_history_data = json.load(open(os.path.join(os.path.dirname(__file__),
164166
'runhistory.json'),

0 commit comments

Comments
 (0)