From e480776343caab522e68822eaeed584776b312b1 Mon Sep 17 00:00:00 2001 From: Mark Saroufim Date: Tue, 26 Aug 2025 10:26:36 -0700 Subject: [PATCH] Revert "fix payload in workflows (#331)" This reverts commit 015607b6a77aafe9055e5c4244d94d7d929c7c7f. --- .github/workflows/amd_workflow.yml | 4 ++-- .github/workflows/nvidia_workflow.yml | 4 ++-- scripts/github_test_payload.json | 2 +- 3 files changed, 5 insertions(+), 5 deletions(-) diff --git a/.github/workflows/amd_workflow.yml b/.github/workflows/amd_workflow.yml index c1efdf4c..8a0014ef 100644 --- a/.github/workflows/amd_workflow.yml +++ b/.github/workflows/amd_workflow.yml @@ -39,8 +39,8 @@ jobs: # Apply mask to the extracted content echo "::add-mask::$PAYLOAD" - # Compress and base64 encode the payload to match what the Python script expects - echo "$PAYLOAD" | python3 -c "import sys, zlib, base64; print(base64.b64encode(zlib.compress(sys.stdin.read().encode('utf-8'))).decode('ascii'))" > payload.json + # Now write to file (won't be logged since it's masked) + echo "$PAYLOAD" > payload.json - name: Set venv directory based on runner run: | diff --git a/.github/workflows/nvidia_workflow.yml b/.github/workflows/nvidia_workflow.yml index 85cd25a1..e16cf4d5 100644 --- a/.github/workflows/nvidia_workflow.yml +++ b/.github/workflows/nvidia_workflow.yml @@ -46,8 +46,8 @@ jobs: # Apply mask to the extracted content echo "::add-mask::$PAYLOAD" - # Compress and base64 encode the payload to match what the Python script expects - echo "$PAYLOAD" | python3 -c "import sys, zlib, base64; print(base64.b64encode(zlib.compress(sys.stdin.read().encode('utf-8'))).decode('ascii'))" > payload.json + # Now write to file (won't be logged since it's masked) + echo "$PAYLOAD" > payload.json - name: Install uv uses: astral-sh/setup-uv@v3 diff --git a/scripts/github_test_payload.json b/scripts/github_test_payload.json index d762ad81..a2deff6e 100644 --- a/scripts/github_test_payload.json +++ b/scripts/github_test_payload.json @@ -4,5 +4,5 @@ "main.py": "import torch\n\nprint(f\"CUDA available: {torch.cuda.is_available()}\")\nprint(f\"PyTorch version: {torch.__version__}\")\n\nif torch.cuda.is_available():\n x = torch.randn(5, device='cuda')\n print(f\"Random tensor on GPU: {x}\")" }, "main": "main.py", - "mode": "test" + "mode": "script" }