|
name: Slow Tests on main |
|
|
|
on: |
|
push: |
|
branches: |
|
- main |
|
paths: |
|
- "src/diffusers/**.py" |
|
- "examples/**.py" |
|
- "tests/**.py" |
|
|
|
env: |
|
DIFFUSERS_IS_CI: yes |
|
HF_HOME: /mnt/cache |
|
OMP_NUM_THREADS: 8 |
|
MKL_NUM_THREADS: 8 |
|
PYTEST_TIMEOUT: 600 |
|
RUN_SLOW: yes |
|
PIPELINE_USAGE_CUTOFF: 50000 |
|
|
|
jobs: |
|
setup_torch_cuda_pipeline_matrix: |
|
name: Setup Torch Pipelines CUDA Slow Tests Matrix |
|
runs-on: [ self-hosted, intel-cpu, 8-cpu, ci ] |
|
container: |
|
image: diffusers/diffusers-pytorch-cpu |
|
outputs: |
|
pipeline_test_matrix: ${{ steps.fetch_pipeline_matrix.outputs.pipeline_test_matrix }} |
|
steps: |
|
- name: Checkout diffusers |
|
uses: actions/checkout@v3 |
|
with: |
|
fetch-depth: 2 |
|
- name: Install dependencies |
|
run: | |
|
python -m venv /opt/venv && export PATH="/opt/venv/bin:$PATH" |
|
python -m uv pip install -e [quality,test] |
|
- name: Environment |
|
run: | |
|
python utils/print_env.py |
|
- name: Fetch Pipeline Matrix |
|
id: fetch_pipeline_matrix |
|
run: | |
|
matrix=$(python utils/fetch_torch_cuda_pipeline_test_matrix.py) |
|
echo $matrix |
|
echo "pipeline_test_matrix=$matrix" >> $GITHUB_OUTPUT |
|
- name: Pipeline Tests Artifacts |
|
if: ${{ always() }} |
|
uses: actions/upload-artifact@v2 |
|
with: |
|
name: test-pipelines.json |
|
path: reports |
|
|
|
torch_pipelines_cuda_tests: |
|
name: Torch Pipelines CUDA Slow Tests |
|
needs: setup_torch_cuda_pipeline_matrix |
|
strategy: |
|
fail-fast: false |
|
max-parallel: 8 |
|
matrix: |
|
module: ${{ fromJson(needs.setup_torch_cuda_pipeline_matrix.outputs.pipeline_test_matrix) }} |
|
runs-on: [single-gpu, nvidia-gpu, t4, ci] |
|
container: |
|
image: diffusers/diffusers-pytorch-cuda |
|
options: --shm-size "16gb" --ipc host -v /mnt/cache/.cache/huggingface/diffusers:/mnt/cache/ --gpus 0 |
|
steps: |
|
- name: Checkout diffusers |
|
uses: actions/checkout@v3 |
|
with: |
|
fetch-depth: 2 |
|
- name: NVIDIA-SMI |
|
run: | |
|
nvidia-smi |
|
- name: Install dependencies |
|
run: | |
|
python -m venv /opt/venv && export PATH="/opt/venv/bin:$PATH" |
|
python -m uv pip install -e [quality,test] |
|
python -m uv pip install accelerate@git+https://github.com/huggingface/accelerate.git |
|
- name: Environment |
|
run: | |
|
python utils/print_env.py |
|
- name: Slow PyTorch CUDA checkpoint tests on Ubuntu |
|
env: |
|
HF_TOKEN: ${{ secrets.HF_TOKEN }} |
|
|
|
CUBLAS_WORKSPACE_CONFIG: :16:8 |
|
run: | |
|
python -m pytest -n 1 --max-worker-restart=0 --dist=loadfile \ |
|
-s -v -k "not Flax and not Onnx" \ |
|
--make-reports=tests_pipeline_${{ matrix.module }}_cuda \ |
|
tests/pipelines/${{ matrix.module }} |
|
- name: Failure short reports |
|
if: ${{ failure() }} |
|
run: | |
|
cat reports/tests_pipeline_${{ matrix.module }}_cuda_stats.txt |
|
cat reports/tests_pipeline_${{ matrix.module }}_cuda_failures_short.txt |
|
- name: Test suite reports artifacts |
|
if: ${{ always() }} |
|
uses: actions/upload-artifact@v2 |
|
with: |
|
name: pipeline_${{ matrix.module }}_test_reports |
|
path: reports |
|
|
|
torch_cuda_tests: |
|
name: Torch CUDA Tests |
|
runs-on: [single-gpu, nvidia-gpu, t4, ci] |
|
container: |
|
image: diffusers/diffusers-pytorch-cuda |
|
options: --shm-size "16gb" --ipc host -v /mnt/cache/.cache/huggingface/diffusers:/mnt/cache/ --gpus 0 |
|
defaults: |
|
run: |
|
shell: bash |
|
strategy: |
|
matrix: |
|
module: [models, schedulers, lora, others, single_file] |
|
steps: |
|
- name: Checkout diffusers |
|
uses: actions/checkout@v3 |
|
with: |
|
fetch-depth: 2 |
|
|
|
- name: Install dependencies |
|
run: | |
|
python -m venv /opt/venv && export PATH="/opt/venv/bin:$PATH" |
|
python -m uv pip install -e [quality,test] |
|
python -m uv pip install accelerate@git+https://github.com/huggingface/accelerate.git |
|
|
|
- name: Environment |
|
run: | |
|
python utils/print_env.py |
|
|
|
- name: Run slow PyTorch CUDA tests |
|
env: |
|
HF_TOKEN: ${{ secrets.HF_TOKEN }} |
|
|
|
CUBLAS_WORKSPACE_CONFIG: :16:8 |
|
run: | |
|
python -m pytest -n 1 --max-worker-restart=0 --dist=loadfile \ |
|
-s -v -k "not Flax and not Onnx" \ |
|
--make-reports=tests_torch_cuda \ |
|
tests/${{ matrix.module }} |
|
|
|
- name: Failure short reports |
|
if: ${{ failure() }} |
|
run: | |
|
cat reports/tests_torch_cuda_stats.txt |
|
cat reports/tests_torch_cuda_failures_short.txt |
|
|
|
- name: Test suite reports artifacts |
|
if: ${{ always() }} |
|
uses: actions/upload-artifact@v2 |
|
with: |
|
name: torch_cuda_test_reports |
|
path: reports |
|
|
|
peft_cuda_tests: |
|
name: PEFT CUDA Tests |
|
runs-on: [single-gpu, nvidia-gpu, t4, ci] |
|
container: |
|
image: diffusers/diffusers-pytorch-cuda |
|
options: --shm-size "16gb" --ipc host -v /mnt/cache/.cache/huggingface/diffusers:/mnt/cache/ --gpus 0 |
|
defaults: |
|
run: |
|
shell: bash |
|
steps: |
|
- name: Checkout diffusers |
|
uses: actions/checkout@v3 |
|
with: |
|
fetch-depth: 2 |
|
|
|
- name: Install dependencies |
|
run: | |
|
python -m venv /opt/venv && export PATH="/opt/venv/bin:$PATH" |
|
python -m uv pip install -e [quality,test] |
|
python -m uv pip install accelerate@git+https://github.com/huggingface/accelerate.git |
|
python -m pip install -U peft@git+https://github.com/huggingface/peft.git |
|
|
|
- name: Environment |
|
run: | |
|
python utils/print_env.py |
|
|
|
- name: Run slow PEFT CUDA tests |
|
env: |
|
HF_TOKEN: ${{ secrets.HF_TOKEN }} |
|
|
|
CUBLAS_WORKSPACE_CONFIG: :16:8 |
|
run: | |
|
python -m pytest -n 1 --max-worker-restart=0 --dist=loadfile \ |
|
-s -v -k "not Flax and not Onnx and not PEFTLoRALoading" \ |
|
--make-reports=tests_peft_cuda \ |
|
tests/lora/ |
|
python -m pytest -n 1 --max-worker-restart=0 --dist=loadfile \ |
|
-s -v -k "lora and not Flax and not Onnx and not PEFTLoRALoading" \ |
|
--make-reports=tests_peft_cuda_models_lora \ |
|
tests/models/ |
|
|
|
- name: Failure short reports |
|
if: ${{ failure() }} |
|
run: | |
|
cat reports/tests_peft_cuda_stats.txt |
|
cat reports/tests_peft_cuda_failures_short.txt |
|
cat reports/tests_peft_cuda_models_lora_failures_short.txt |
|
|
|
- name: Test suite reports artifacts |
|
if: ${{ always() }} |
|
uses: actions/upload-artifact@v2 |
|
with: |
|
name: torch_peft_test_reports |
|
path: reports |
|
|
|
flax_tpu_tests: |
|
name: Flax TPU Tests |
|
runs-on: docker-tpu |
|
container: |
|
image: diffusers/diffusers-flax-tpu |
|
options: --shm-size "16gb" --ipc host -v /mnt/cache/.cache/huggingface:/mnt/cache/ --privileged |
|
defaults: |
|
run: |
|
shell: bash |
|
steps: |
|
- name: Checkout diffusers |
|
uses: actions/checkout@v3 |
|
with: |
|
fetch-depth: 2 |
|
|
|
- name: Install dependencies |
|
run: | |
|
python -m venv /opt/venv && export PATH="/opt/venv/bin:$PATH" |
|
python -m uv pip install -e [quality,test] |
|
python -m uv pip install accelerate@git+https://github.com/huggingface/accelerate.git |
|
|
|
- name: Environment |
|
run: | |
|
python utils/print_env.py |
|
|
|
- name: Run slow Flax TPU tests |
|
env: |
|
HF_TOKEN: ${{ secrets.HF_TOKEN }} |
|
run: | |
|
python -m pytest -n 0 \ |
|
-s -v -k "Flax" \ |
|
--make-reports=tests_flax_tpu \ |
|
tests/ |
|
|
|
- name: Failure short reports |
|
if: ${{ failure() }} |
|
run: | |
|
cat reports/tests_flax_tpu_stats.txt |
|
cat reports/tests_flax_tpu_failures_short.txt |
|
|
|
- name: Test suite reports artifacts |
|
if: ${{ always() }} |
|
uses: actions/upload-artifact@v2 |
|
with: |
|
name: flax_tpu_test_reports |
|
path: reports |
|
|
|
onnx_cuda_tests: |
|
name: ONNX CUDA Tests |
|
runs-on: [single-gpu, nvidia-gpu, t4, ci] |
|
container: |
|
image: diffusers/diffusers-onnxruntime-cuda |
|
options: --shm-size "16gb" --ipc host -v /mnt/cache/.cache/huggingface:/mnt/cache/ --gpus 0 |
|
defaults: |
|
run: |
|
shell: bash |
|
steps: |
|
- name: Checkout diffusers |
|
uses: actions/checkout@v3 |
|
with: |
|
fetch-depth: 2 |
|
|
|
- name: Install dependencies |
|
run: | |
|
python -m venv /opt/venv && export PATH="/opt/venv/bin:$PATH" |
|
python -m uv pip install -e [quality,test] |
|
python -m uv pip install accelerate@git+https://github.com/huggingface/accelerate.git |
|
|
|
- name: Environment |
|
run: | |
|
python utils/print_env.py |
|
|
|
- name: Run slow ONNXRuntime CUDA tests |
|
env: |
|
HF_TOKEN: ${{ secrets.HF_TOKEN }} |
|
run: | |
|
python -m pytest -n 1 --max-worker-restart=0 --dist=loadfile \ |
|
-s -v -k "Onnx" \ |
|
--make-reports=tests_onnx_cuda \ |
|
tests/ |
|
|
|
- name: Failure short reports |
|
if: ${{ failure() }} |
|
run: | |
|
cat reports/tests_onnx_cuda_stats.txt |
|
cat reports/tests_onnx_cuda_failures_short.txt |
|
|
|
- name: Test suite reports artifacts |
|
if: ${{ always() }} |
|
uses: actions/upload-artifact@v2 |
|
with: |
|
name: onnx_cuda_test_reports |
|
path: reports |
|
|
|
run_torch_compile_tests: |
|
name: PyTorch Compile CUDA tests |
|
|
|
runs-on: [single-gpu, nvidia-gpu, t4, ci] |
|
|
|
container: |
|
image: diffusers/diffusers-pytorch-compile-cuda |
|
options: --gpus 0 --shm-size "16gb" --ipc host -v /mnt/cache/.cache/huggingface:/mnt/cache/ |
|
|
|
steps: |
|
- name: Checkout diffusers |
|
uses: actions/checkout@v3 |
|
with: |
|
fetch-depth: 2 |
|
|
|
- name: NVIDIA-SMI |
|
run: | |
|
nvidia-smi |
|
- name: Install dependencies |
|
run: | |
|
python -m venv /opt/venv && export PATH="/opt/venv/bin:$PATH" |
|
python -m uv pip install -e [quality,test,training] |
|
- name: Environment |
|
run: | |
|
python utils/print_env.py |
|
- name: Run example tests on GPU |
|
env: |
|
HF_TOKEN: ${{ secrets.HF_TOKEN }} |
|
run: | |
|
python -m pytest -n 1 --max-worker-restart=0 --dist=loadfile -s -v -k "compile" --make-reports=tests_torch_compile_cuda tests/ |
|
- name: Failure short reports |
|
if: ${{ failure() }} |
|
run: cat reports/tests_torch_compile_cuda_failures_short.txt |
|
|
|
- name: Test suite reports artifacts |
|
if: ${{ always() }} |
|
uses: actions/upload-artifact@v2 |
|
with: |
|
name: torch_compile_test_reports |
|
path: reports |
|
|
|
run_xformers_tests: |
|
name: PyTorch xformers CUDA tests |
|
|
|
runs-on: [single-gpu, nvidia-gpu, t4, ci] |
|
|
|
container: |
|
image: diffusers/diffusers-pytorch-xformers-cuda |
|
options: --gpus 0 --shm-size "16gb" --ipc host -v /mnt/cache/.cache/huggingface:/mnt/cache/ |
|
|
|
steps: |
|
- name: Checkout diffusers |
|
uses: actions/checkout@v3 |
|
with: |
|
fetch-depth: 2 |
|
|
|
- name: NVIDIA-SMI |
|
run: | |
|
nvidia-smi |
|
- name: Install dependencies |
|
run: | |
|
python -m venv /opt/venv && export PATH="/opt/venv/bin:$PATH" |
|
python -m uv pip install -e [quality,test,training] |
|
- name: Environment |
|
run: | |
|
python utils/print_env.py |
|
- name: Run example tests on GPU |
|
env: |
|
HF_TOKEN: ${{ secrets.HF_TOKEN }} |
|
run: | |
|
python -m pytest -n 1 --max-worker-restart=0 --dist=loadfile -s -v -k "xformers" --make-reports=tests_torch_xformers_cuda tests/ |
|
- name: Failure short reports |
|
if: ${{ failure() }} |
|
run: cat reports/tests_torch_xformers_cuda_failures_short.txt |
|
|
|
- name: Test suite reports artifacts |
|
if: ${{ always() }} |
|
uses: actions/upload-artifact@v2 |
|
with: |
|
name: torch_xformers_test_reports |
|
path: reports |
|
|
|
run_examples_tests: |
|
name: Examples PyTorch CUDA tests on Ubuntu |
|
|
|
runs-on: [single-gpu, nvidia-gpu, t4, ci] |
|
|
|
container: |
|
image: diffusers/diffusers-pytorch-cuda |
|
options: --gpus 0 --shm-size "16gb" --ipc host -v /mnt/cache/.cache/huggingface:/mnt/cache/ |
|
|
|
steps: |
|
- name: Checkout diffusers |
|
uses: actions/checkout@v3 |
|
with: |
|
fetch-depth: 2 |
|
|
|
- name: NVIDIA-SMI |
|
run: | |
|
nvidia-smi |
|
|
|
- name: Install dependencies |
|
run: | |
|
python -m venv /opt/venv && export PATH="/opt/venv/bin:$PATH" |
|
python -m uv pip install -e [quality,test,training] |
|
|
|
- name: Environment |
|
run: | |
|
python -m venv /opt/venv && export PATH="/opt/venv/bin:$PATH" |
|
python utils/print_env.py |
|
|
|
- name: Run example tests on GPU |
|
env: |
|
HF_TOKEN: ${{ secrets.HF_TOKEN }} |
|
run: | |
|
python -m venv /opt/venv && export PATH="/opt/venv/bin:$PATH" |
|
python -m uv pip install timm |
|
python -m pytest -n 1 --max-worker-restart=0 --dist=loadfile -s -v --make-reports=examples_torch_cuda examples/ |
|
|
|
- name: Failure short reports |
|
if: ${{ failure() }} |
|
run: | |
|
cat reports/examples_torch_cuda_stats.txt |
|
cat reports/examples_torch_cuda_failures_short.txt |
|
|
|
- name: Test suite reports artifacts |
|
if: ${{ always() }} |
|
uses: actions/upload-artifact@v2 |
|
with: |
|
name: examples_test_reports |
|
path: reports |
|
|