6161 group : aws-g4dn-2xlarge
6262 container :
6363 image : diffusers/diffusers-pytorch-cuda
64- options : --shm-size "16gb" --ipc host --gpus 0
64+ options : --shm-size "16gb" --ipc host --gpus all
6565 steps :
6666 - name : Checkout diffusers
6767 uses : actions/checkout@v3
@@ -107,7 +107,7 @@ jobs:
107107 group : aws-g4dn-2xlarge
108108 container :
109109 image : diffusers/diffusers-pytorch-cuda
110- options : --shm-size "16gb" --ipc host --gpus 0
110+ options : --shm-size "16gb" --ipc host --gpus all
111111 defaults :
112112 run :
113113 shell : bash
@@ -178,7 +178,7 @@ jobs:
178178
179179 container :
180180 image : diffusers/diffusers-pytorch-cuda
181- options : --gpus 0 --shm-size "16gb" --ipc host
181+ options : --gpus all --shm-size "16gb" --ipc host
182182
183183 steps :
184184 - name : Checkout diffusers
@@ -222,7 +222,7 @@ jobs:
222222 group : aws-g6e-xlarge-plus
223223 container :
224224 image : diffusers/diffusers-pytorch-cuda
225- options : --shm-size "16gb" --ipc host --gpus 0
225+ options : --shm-size "16gb" --ipc host --gpus all
226226 steps :
227227 - name : Checkout diffusers
228228 uses : actions/checkout@v3
@@ -248,7 +248,7 @@ jobs:
248248 BIG_GPU_MEMORY : 40
249249 run : |
250250 python -m pytest -n 1 --max-worker-restart=0 --dist=loadfile \
251- -m "big_gpu_with_torch_cuda " \
251+ -m "big_accelerator " \
252252 --make-reports=tests_big_gpu_torch_cuda \
253253 --report-log=tests_big_gpu_torch_cuda.log \
254254 tests/
@@ -270,7 +270,7 @@ jobs:
270270 group : aws-g4dn-2xlarge
271271 container :
272272 image : diffusers/diffusers-pytorch-minimum-cuda
273- options : --shm-size "16gb" --ipc host --gpus 0
273+ options : --shm-size "16gb" --ipc host --gpus all
274274 defaults :
275275 run :
276276 shell : bash
@@ -321,55 +321,6 @@ jobs:
321321 name : torch_minimum_version_cuda_test_reports
322322 path : reports
323323
324- run_nightly_onnx_tests :
325- name : Nightly ONNXRuntime CUDA tests on Ubuntu
326- runs-on :
327- group : aws-g4dn-2xlarge
328- container :
329- image : diffusers/diffusers-onnxruntime-cuda
330- options : --gpus 0 --shm-size "16gb" --ipc host
331-
332- steps :
333- - name : Checkout diffusers
334- uses : actions/checkout@v3
335- with :
336- fetch-depth : 2
337-
338- - name : NVIDIA-SMI
339- run : nvidia-smi
340-
341- - name : Install dependencies
342- run : |
343- python -m venv /opt/venv && export PATH="/opt/venv/bin:$PATH"
344- python -m uv pip install -e [quality,test]
345- pip uninstall accelerate -y && python -m uv pip install -U accelerate@git+https://github.com/huggingface/accelerate.git
346- python -m uv pip install pytest-reportlog
347- - name : Environment
348- run : python utils/print_env.py
349-
350- - name : Run Nightly ONNXRuntime CUDA tests
351- env :
352- HF_TOKEN : ${{ secrets.DIFFUSERS_HF_HUB_READ_TOKEN }}
353- run : |
354- python -m pytest -n 1 --max-worker-restart=0 --dist=loadfile \
355- -s -v -k "Onnx" \
356- --make-reports=tests_onnx_cuda \
357- --report-log=tests_onnx_cuda.log \
358- tests/
359-
360- - name : Failure short reports
361- if : ${{ failure() }}
362- run : |
363- cat reports/tests_onnx_cuda_stats.txt
364- cat reports/tests_onnx_cuda_failures_short.txt
365-
366- - name : Test suite reports artifacts
367- if : ${{ always() }}
368- uses : actions/upload-artifact@v4
369- with :
370- name : tests_onnx_cuda_reports
371- path : reports
372-
373324 run_nightly_quantization_tests :
374325 name : Torch quantization nightly tests
375326 strategy :
@@ -382,18 +333,21 @@ jobs:
382333 additional_deps : ["peft"]
383334 - backend : " gguf"
384335 test_location : " gguf"
385- additional_deps : ["peft"]
336+ additional_deps : ["peft", "kernels" ]
386337 - backend : " torchao"
387338 test_location : " torchao"
388339 additional_deps : []
389340 - backend : " optimum_quanto"
390341 test_location : " quanto"
391342 additional_deps : []
343+ - backend : " nvidia_modelopt"
344+ test_location : " modelopt"
345+ additional_deps : []
392346 runs-on :
393347 group : aws-g6e-xlarge-plus
394348 container :
395349 image : diffusers/diffusers-pytorch-cuda
396- options : --shm-size "20gb" --ipc host --gpus 0
350+ options : --shm-size "20gb" --ipc host --gpus all
397351 steps :
398352 - name : Checkout diffusers
399353 uses : actions/checkout@v3
@@ -445,7 +399,7 @@ jobs:
445399 group : aws-g6e-xlarge-plus
446400 container :
447401 image : diffusers/diffusers-pytorch-cuda
448- options : --shm-size "20gb" --ipc host --gpus 0
402+ options : --shm-size "20gb" --ipc host --gpus all
449403 steps :
450404 - name : Checkout diffusers
451405 uses : actions/checkout@v3
@@ -485,57 +439,6 @@ jobs:
485439 name : torch_cuda_pipeline_level_quant_reports
486440 path : reports
487441
488- run_flax_tpu_tests :
489- name : Nightly Flax TPU Tests
490- runs-on :
491- group : gcp-ct5lp-hightpu-8t
492- if : github.event_name == 'schedule'
493-
494- container :
495- image : diffusers/diffusers-flax-tpu
496- options : --shm-size "16gb" --ipc host --privileged ${{ vars.V5_LITEPOD_8_ENV}} -v /mnt/hf_cache:/mnt/hf_cache
497- defaults :
498- run :
499- shell : bash
500- steps :
501- - name : Checkout diffusers
502- uses : actions/checkout@v3
503- with :
504- fetch-depth : 2
505-
506- - name : Install dependencies
507- run : |
508- python -m venv /opt/venv && export PATH="/opt/venv/bin:$PATH"
509- python -m uv pip install -e [quality,test]
510- pip uninstall accelerate -y && python -m uv pip install -U accelerate@git+https://github.com/huggingface/accelerate.git
511- python -m uv pip install pytest-reportlog
512-
513- - name : Environment
514- run : python utils/print_env.py
515-
516- - name : Run nightly Flax TPU tests
517- env :
518- HF_TOKEN : ${{ secrets.DIFFUSERS_HF_HUB_READ_TOKEN }}
519- run : |
520- python -m pytest -n 0 \
521- -s -v -k "Flax" \
522- --make-reports=tests_flax_tpu \
523- --report-log=tests_flax_tpu.log \
524- tests/
525-
526- - name : Failure short reports
527- if : ${{ failure() }}
528- run : |
529- cat reports/tests_flax_tpu_stats.txt
530- cat reports/tests_flax_tpu_failures_short.txt
531-
532- - name : Test suite reports artifacts
533- if : ${{ always() }}
534- uses : actions/upload-artifact@v4
535- with :
536- name : flax_tpu_test_reports
537- path : reports
538-
539442 generate_consolidated_report :
540443 name : Generate Consolidated Test Report
541444 needs : [
@@ -545,9 +448,9 @@ jobs:
545448 run_big_gpu_torch_tests,
546449 run_nightly_quantization_tests,
547450 run_nightly_pipeline_level_quantization_tests,
548- run_nightly_onnx_tests,
451+ # run_nightly_onnx_tests,
549452 torch_minimum_version_cuda_tests,
550- run_flax_tpu_tests
453+ # run_flax_tpu_tests
551454 ]
552455 if : always()
553456 runs-on :
0 commit comments