|
47 | 47 | machine_type: [single-gpu, multi-gpu] |
48 | 48 | runs-on: [self-hosted, docker-gpu, '${{ matrix.machine_type }}'] |
49 | 49 | container: |
50 | | - image: huggingface/transformers-all-latest-gpu |
| 50 | + image: huggingface/transformers-all-latest-gpu-push-ci |
51 | 51 | options: --gpus 0 --shm-size "16gb" --ipc host -v /mnt/cache/.cache/huggingface:/mnt/cache/ |
52 | 52 | steps: |
53 | 53 | - name: NVIDIA-SMI |
|
62 | 62 | machine_type: [single-gpu, multi-gpu] |
63 | 63 | runs-on: [self-hosted, docker-gpu, '${{ matrix.machine_type }}'] |
64 | 64 | container: |
65 | | - image: huggingface/transformers-all-latest-gpu |
| 65 | + image: huggingface/transformers-all-latest-gpu-push-ci |
66 | 66 | options: --gpus 0 --shm-size "16gb" --ipc host -v /mnt/cache/.cache/huggingface:/mnt/cache/ |
67 | 67 | outputs: |
68 | 68 | matrix: ${{ steps.set-matrix.outputs.matrix }} |
@@ -158,7 +158,7 @@ jobs: |
158 | 158 | machine_type: [single-gpu] |
159 | 159 | runs-on: [self-hosted, docker-gpu, '${{ matrix.machine_type }}'] |
160 | 160 | container: |
161 | | - image: huggingface/transformers-all-latest-gpu |
| 161 | + image: huggingface/transformers-all-latest-gpu-push-ci |
162 | 162 | options: --gpus 0 --shm-size "16gb" --ipc host -v /mnt/cache/.cache/huggingface:/mnt/cache/ |
163 | 163 | steps: |
164 | 164 | # Necessary to get the correct branch name and commit SHA for `workflow_run` event |
@@ -243,7 +243,7 @@ jobs: |
243 | 243 | machine_type: [multi-gpu] |
244 | 244 | runs-on: [self-hosted, docker-gpu, '${{ matrix.machine_type }}'] |
245 | 245 | container: |
246 | | - image: huggingface/transformers-all-latest-gpu |
| 246 | + image: huggingface/transformers-all-latest-gpu-push-ci |
247 | 247 | options: --gpus all --shm-size "16gb" --ipc host -v /mnt/cache/.cache/huggingface:/mnt/cache/ |
248 | 248 | steps: |
249 | 249 | # Necessary to get the correct branch name and commit SHA for `workflow_run` event |
@@ -328,7 +328,7 @@ jobs: |
328 | 328 | machine_type: [single-gpu] |
329 | 329 | runs-on: [self-hosted, docker-gpu, '${{ matrix.machine_type }}'] |
330 | 330 | container: |
331 | | - image: huggingface/transformers-pytorch-deepspeed-latest-gpu |
| 331 | + image: huggingface/transformers-pytorch-deepspeed-latest-gpu-push-ci |
332 | 332 | options: --gpus 0 --shm-size "16gb" --ipc host -v /mnt/cache/.cache/huggingface:/mnt/cache/ |
333 | 333 | steps: |
334 | 334 | # Necessary to get the correct branch name and commit SHA for `workflow_run` event |
@@ -410,7 +410,7 @@ jobs: |
410 | 410 | machine_type: [multi-gpu] |
411 | 411 | runs-on: [self-hosted, docker-gpu, '${{ matrix.machine_type }}'] |
412 | 412 | container: |
413 | | - image: huggingface/transformers-pytorch-deepspeed-latest-gpu |
| 413 | + image: huggingface/transformers-pytorch-deepspeed-latest-gpu-push-ci |
414 | 414 | options: --gpus all --shm-size "16gb" --ipc host -v /mnt/cache/.cache/huggingface:/mnt/cache/ |
415 | 415 | steps: |
416 | 416 | # Necessary to get the correct branch name and commit SHA for `workflow_run` event |
|
0 commit comments