Skip to content

Commit d6b17c2

Browse files
committed
trigger ci on A10
1 parent 0725cd6 commit d6b17c2

File tree

4 files changed

+21
-81
lines changed

4 files changed

+21
-81
lines changed

.github/workflows/model_jobs.yml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -107,9 +107,9 @@ jobs:
107107
run: |
108108
echo "${{ inputs.machine_type }}"
109109
110-
if [ "${{ inputs.machine_type }}" = "aws-g4dn-4xlarge-cache" ]; then
110+
if [ "${{ inputs.machine_type }}" = "aws-g5-4xlarge-cache" ]; then
111111
machine_type=single-gpu
112-
elif [ "${{ inputs.machine_type }}" = "aws-g4dn-12xlarge-cache" ]; then
112+
elif [ "${{ inputs.machine_type }}" = "aws-g5-12xlarge-cache" ]; then
113113
machine_type=multi-gpu
114114
else
115115
machine_type=${{ inputs.machine_type }}

.github/workflows/self-scheduled-caller.yml

Lines changed: 3 additions & 64 deletions
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@ on:
77
- cron: "17 2 * * *"
88
push:
99
branches:
10-
- run_scheduled_ci*
10+
- trigger_ci_on_a10
1111
workflow_dispatch:
1212
inputs:
1313
prev_workflow_run_id:
@@ -25,7 +25,7 @@ on:
2525
# Used for `push` to easily modiffy the target workflow runs to compare against
2626
env:
2727
prev_workflow_run_id: ""
28-
other_workflow_run_id: ""
28+
other_workflow_run_id: "15770139098"
2929

3030

3131
jobs:
@@ -50,70 +50,9 @@ jobs:
5050
uses: ./.github/workflows/self-scheduled.yml
5151
with:
5252
job: run_models_gpu
53-
slack_report_channel: "#transformers-ci-daily-models"
53+
slack_report_channel: "#transformers-ci-dummy"
5454
runner: daily-ci
5555
docker: huggingface/transformers-all-latest-gpu
5656
ci_event: Daily CI
5757
report_repo_id: hf-internal-testing/transformers_daily_ci
5858
secrets: inherit
59-
60-
torch-pipeline:
61-
name: Torch pipeline CI
62-
uses: ./.github/workflows/self-scheduled.yml
63-
with:
64-
job: run_pipelines_torch_gpu
65-
slack_report_channel: "#transformers-ci-daily-pipeline-torch"
66-
runner: daily-ci
67-
docker: huggingface/transformers-pytorch-gpu
68-
ci_event: Daily CI
69-
report_repo_id: hf-internal-testing/transformers_daily_ci
70-
secrets: inherit
71-
72-
example-ci:
73-
name: Example CI
74-
uses: ./.github/workflows/self-scheduled.yml
75-
with:
76-
job: run_examples_gpu
77-
slack_report_channel: "#transformers-ci-daily-examples"
78-
runner: daily-ci
79-
docker: huggingface/transformers-all-latest-gpu
80-
ci_event: Daily CI
81-
report_repo_id: hf-internal-testing/transformers_daily_ci
82-
secrets: inherit
83-
84-
trainer-fsdp-ci:
85-
name: Trainer/FSDP CI
86-
uses: ./.github/workflows/self-scheduled.yml
87-
with:
88-
job: run_trainer_and_fsdp_gpu
89-
slack_report_channel: "#transformers-ci-daily-training"
90-
runner: daily-ci
91-
docker: huggingface/transformers-all-latest-gpu
92-
ci_event: Daily CI
93-
report_repo_id: hf-internal-testing/transformers_daily_ci
94-
secrets: inherit
95-
96-
deepspeed-ci:
97-
name: DeepSpeed CI
98-
uses: ./.github/workflows/self-scheduled.yml
99-
with:
100-
job: run_torch_cuda_extensions_gpu
101-
slack_report_channel: "#transformers-ci-daily-training"
102-
runner: daily-ci
103-
docker: huggingface/transformers-pytorch-deepspeed-latest-gpu
104-
ci_event: Daily CI
105-
working-directory-prefix: /workspace
106-
report_repo_id: hf-internal-testing/transformers_daily_ci
107-
secrets: inherit
108-
109-
quantization-ci:
110-
name: Quantization CI
111-
uses: ./.github/workflows/self-scheduled.yml
112-
with:
113-
job: run_quantization_torch_gpu
114-
slack_report_channel: "#transformers-ci-daily-quantization"
115-
runner: daily-ci
116-
docker: huggingface/transformers-quantization-latest-gpu
117-
ci_event: Daily CI
118-
report_repo_id: hf-internal-testing/transformers_daily_ci
119-
secrets: inherit

.github/workflows/self-scheduled.yml

Lines changed: 15 additions & 15 deletions
Original file line numberDiff line numberDiff line change
@@ -53,7 +53,7 @@ jobs:
5353
name: Setup
5454
strategy:
5555
matrix:
56-
machine_type: [aws-g4dn-4xlarge-cache, aws-g4dn-12xlarge-cache]
56+
machine_type: [aws-g5-4xlarge-cache, aws-g5-12xlarge-cache]
5757
runs-on:
5858
group: '${{ matrix.machine_type }}'
5959
container:
@@ -111,7 +111,7 @@ jobs:
111111
strategy:
112112
fail-fast: false
113113
matrix:
114-
machine_type: [aws-g4dn-4xlarge-cache, aws-g4dn-12xlarge-cache]
114+
machine_type: [aws-g5-4xlarge-cache, aws-g5-12xlarge-cache]
115115
slice_id: ${{ fromJSON(needs.setup.outputs.slice_ids) }}
116116
uses: ./.github/workflows/model_jobs.yml
117117
with:
@@ -129,7 +129,7 @@ jobs:
129129
strategy:
130130
fail-fast: false
131131
matrix:
132-
machine_type: [aws-g4dn-4xlarge-cache, aws-g4dn-12xlarge-cache]
132+
machine_type: [aws-g5-4xlarge-cache, aws-g5-12xlarge-cache]
133133
slice_id: [0, 1]
134134
uses: ./.github/workflows/model_jobs.yml
135135
with:
@@ -147,7 +147,7 @@ jobs:
147147
strategy:
148148
fail-fast: false
149149
matrix:
150-
machine_type: [aws-g4dn-4xlarge-cache, aws-g4dn-12xlarge-cache]
150+
machine_type: [aws-g5-4xlarge-cache, aws-g5-12xlarge-cache]
151151
runs-on:
152152
group: '${{ matrix.machine_type }}'
153153
container:
@@ -181,9 +181,9 @@ jobs:
181181
run: |
182182
echo "${{ matrix.machine_type }}"
183183
184-
if [ "${{ matrix.machine_type }}" = "aws-g4dn-4xlarge-cache" ]; then
184+
if [ "${{ matrix.machine_type }}" = "aws-g5-4xlarge-cache" ]; then
185185
machine_type=single-gpu
186-
elif [ "${{ matrix.machine_type }}" = "aws-g4dn-12xlarge-cache" ]; then
186+
elif [ "${{ matrix.machine_type }}" = "aws-g5-12xlarge-cache" ]; then
187187
machine_type=multi-gpu
188188
else
189189
machine_type=${{ matrix.machine_type }}
@@ -215,7 +215,7 @@ jobs:
215215
strategy:
216216
fail-fast: false
217217
matrix:
218-
machine_type: [aws-g4dn-4xlarge-cache]
218+
machine_type: [aws-g5-4xlarge-cache]
219219
runs-on:
220220
group: '${{ matrix.machine_type }}'
221221
container:
@@ -249,9 +249,9 @@ jobs:
249249
run: |
250250
echo "${{ matrix.machine_type }}"
251251
252-
if [ "${{ matrix.machine_type }}" = "aws-g4dn-4xlarge-cache" ]; then
252+
if [ "${{ matrix.machine_type }}" = "aws-g5-4xlarge-cache" ]; then
253253
machine_type=single-gpu
254-
elif [ "${{ matrix.machine_type }}" = "aws-g4dn-12xlarge-cache" ]; then
254+
elif [ "${{ matrix.machine_type }}" = "aws-g5-12xlarge-cache" ]; then
255255
machine_type=multi-gpu
256256
else
257257
machine_type=${{ matrix.machine_type }}
@@ -284,7 +284,7 @@ jobs:
284284
strategy:
285285
fail-fast: false
286286
matrix:
287-
machine_type: [aws-g4dn-4xlarge-cache, aws-g4dn-12xlarge-cache]
287+
machine_type: [aws-g5-4xlarge-cache, aws-g5-12xlarge-cache]
288288
runs-on:
289289
group: '${{ matrix.machine_type }}'
290290
container:
@@ -346,9 +346,9 @@ jobs:
346346
run: |
347347
echo "${{ matrix.machine_type }}"
348348
349-
if [ "${{ matrix.machine_type }}" = "aws-g4dn-4xlarge-cache" ]; then
349+
if [ "${{ matrix.machine_type }}" = "aws-g5-4xlarge-cache" ]; then
350350
machine_type=single-gpu
351-
elif [ "${{ matrix.machine_type }}" = "aws-g4dn-12xlarge-cache" ]; then
351+
elif [ "${{ matrix.machine_type }}" = "aws-g5-12xlarge-cache" ]; then
352352
machine_type=multi-gpu
353353
else
354354
machine_type=${{ matrix.machine_type }}
@@ -383,7 +383,7 @@ jobs:
383383
fail-fast: false
384384
matrix:
385385
folders: ${{ fromJson(needs.setup.outputs.quantization_matrix) }}
386-
machine_type: [aws-g4dn-4xlarge-cache, aws-g4dn-12xlarge-cache]
386+
machine_type: [aws-g5-4xlarge-cache, aws-g5-12xlarge-cache]
387387
runs-on:
388388
group: '${{ matrix.machine_type }}'
389389
container:
@@ -426,9 +426,9 @@ jobs:
426426
run: |
427427
echo "${{ matrix.machine_type }}"
428428
429-
if [ "${{ matrix.machine_type }}" = "aws-g4dn-4xlarge-cache" ]; then
429+
if [ "${{ matrix.machine_type }}" = "aws-g5-4xlarge-cache" ]; then
430430
machine_type=single-gpu
431-
elif [ "${{ matrix.machine_type }}" = "aws-g4dn-12xlarge-cache" ]; then
431+
elif [ "${{ matrix.machine_type }}" = "aws-g5-12xlarge-cache" ]; then
432432
machine_type=multi-gpu
433433
else
434434
machine_type=${{ matrix.machine_type }}

utils/split_model_tests.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -62,4 +62,5 @@
6262
start = end
6363
end = start + num_jobs_per_splits + (1 if idx < num_jobs % args.num_splits else 0)
6464
model_splits.append(d[start:end])
65+
model_splits = [["models/vit"], ["models/clip"]]
6566
print(model_splits)

0 commit comments

Comments
 (0)