Skip to content

Commit

Permalink
Mark all flaky tests as integration_tests
Browse files Browse the repository at this point in the history
  • Loading branch information
eivindjahren authored Sep 25, 2024
1 parent 21af329 commit 5731434
Show file tree
Hide file tree
Showing 6 changed files with 7 additions and 0 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -214,6 +214,7 @@ def mock_send(msg):
assert len(lines) == 0, "expected 0 Job running messages"


@pytest.mark.integration_test
@pytest.mark.flaky(reruns=5)
@pytest.mark.skipif(
sys.platform.startswith("darwin"), reason="Performance can be flaky"
Expand Down
1 change: 1 addition & 0 deletions tests/ert/unit_tests/forward_model_runner/test_job.py
Original file line number Diff line number Diff line change
Expand Up @@ -142,6 +142,7 @@ def max_memory_per_subprocess_layer(layers: int) -> int:
assert max_seens[1] + memory_per_numbers_list < max_seens[2]


@pytest.mark.integration_test
@pytest.mark.flaky(reruns=3)
@pytest.mark.usefixtures("use_tmpdir")
def test_memory_profile_in_running_events():
Expand Down
1 change: 1 addition & 0 deletions tests/ert/unit_tests/scheduler/test_lsf_driver.py
Original file line number Diff line number Diff line change
Expand Up @@ -1119,6 +1119,7 @@ async def test_submit_with_resource_requirement_with_bsub_capture():
assert "hname" not in Path("captured_bsub_args").read_text(encoding="utf-8")


@pytest.mark.integration_test
@pytest.mark.usefixtures("use_tmpdir")
async def test_submit_with_num_cpu(pytestconfig, job_name):
if not pytestconfig.getoption("lsf"):
Expand Down
2 changes: 2 additions & 0 deletions tests/ert/unit_tests/scheduler/test_scheduler.py
Original file line number Diff line number Diff line change
Expand Up @@ -479,6 +479,7 @@ async def wait(iens):
assert killed_iens == [6, 7, 8, 9]


@pytest.mark.integration_test
@pytest.mark.flaky(reruns=5)
@pytest.mark.parametrize(
"submit_sleep, iens_stride, realization_runtime",
Expand Down Expand Up @@ -525,6 +526,7 @@ async def wait():
assert max(deltas) <= submit_sleep + 0.1


@pytest.mark.integration_test
@pytest.mark.flaky(reruns=5)
@pytest.mark.parametrize(
"submit_sleep, realization_max_runtime, max_running",
Expand Down
1 change: 1 addition & 0 deletions tests/ert/unit_tests/scheduler/test_slurm_driver.py
Original file line number Diff line number Diff line change
Expand Up @@ -350,6 +350,7 @@ async def test_submit_with_num_cpu(pytestconfig, job_name):
assert Path("test").read_text(encoding="utf-8") == "test\n"


@pytest.mark.integration_test
@pytest.mark.flaky(reruns=3)
async def test_kill_before_submit_is_finished(
tmp_path, monkeypatch, caplog, pytestconfig
Expand Down
1 change: 1 addition & 0 deletions tests/ert/unit_tests/simulator/test_batch_sim.py
Original file line number Diff line number Diff line change
Expand Up @@ -354,6 +354,7 @@ def test_batch_simulation_suffixes(batch_sim_example, storage):
assert act == pytest.approx(exp)


@pytest.mark.integration_test
@pytest.mark.flaky(reruns=3) # https://github.com/equinor/ert/issues/7309
@pytest.mark.timeout(10)
def test_stop_sim(copy_case, storage):
Expand Down

0 comments on commit 5731434

Please sign in to comment.