diff --git a/tests/test_modal.py b/tests/test_modal.py index 14531015..d22ef05b 100644 --- a/tests/test_modal.py +++ b/tests/test_modal.py @@ -50,7 +50,7 @@ def modal_deployment(project_root: Path): if result.returncode != 0: # if it fails simply because the environment does not exist, we can fix that - if "No such environment" in result.stderr: + if "No such environment" in result.stderr or "not found" in result.stderr: result = subprocess.run( ["modal", "environment", "create", modal_env], cwd=project_root / "src" / "runners", @@ -155,7 +155,6 @@ async def test_modal_launcher_python_script( # System info - test actual expected values assert gpu_type.name in result.system.gpu assert "Linux" in result.system.platform - assert result.system.torch.startswith("2.7") # update when the image changes # Test run structure assert "test" in result.runs @@ -184,6 +183,7 @@ async def test_modal_launcher_python_script( assert reporter.updates == ["✅ Waiting for modal run to finish... Done"] +@pytest.mark.skip(reason="Multi-GPU L4x4 NCCL issues on Modal infrastructure") @pytest.mark.integration @pytest.mark.asyncio @pytest.mark.parametrize("script, good", [("submission.py", True), ("wrong.py", False)]) @@ -236,6 +236,7 @@ async def test_modal_multi_gpu(modal_deployment, project_root: Path, script: str assert test_run.run.passed is good +@pytest.mark.skip(reason="Multi-GPU L4x4 NCCL issues on Modal infrastructure") @pytest.mark.integration @pytest.mark.asyncio @pytest.mark.parametrize("script, good", [("submission.py", True), ("wrong.py", False)])