Nightly and release tests on main/release branch #580
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
name: Nightly and release tests on main/release branch | |
on: | |
workflow_dispatch: | |
schedule: | |
- cron: "0 0 * * *" # every day at midnight | |
env: | |
DIFFUSERS_IS_CI: yes | |
HF_HOME: /mnt/cache | |
OMP_NUM_THREADS: 8 | |
MKL_NUM_THREADS: 8 | |
PYTEST_TIMEOUT: 600 | |
RUN_SLOW: yes | |
RUN_NIGHTLY: yes | |
PIPELINE_USAGE_CUTOFF: 5000 | |
SLACK_API_TOKEN: ${{ secrets.SLACK_CIFEEDBACK_BOT_TOKEN }} | |
jobs: | |
setup_torch_cuda_pipeline_matrix: | |
name: Setup Torch Pipelines Matrix | |
runs-on: diffusers/diffusers-pytorch-cpu | |
outputs: | |
pipeline_test_matrix: ${{ steps.fetch_pipeline_matrix.outputs.pipeline_test_matrix }} | |
steps: | |
- name: Checkout diffusers | |
uses: actions/checkout@v3 | |
with: | |
fetch-depth: 2 | |
- name: Set up Python | |
uses: actions/setup-python@v4 | |
with: | |
python-version: "3.8" | |
- name: Install dependencies | |
run: | | |
pip install -e . | |
pip install huggingface_hub | |
- name: Fetch Pipeline Matrix | |
id: fetch_pipeline_matrix | |
run: | | |
matrix=$(python utils/fetch_torch_cuda_pipeline_test_matrix.py) | |
echo $matrix | |
echo "pipeline_test_matrix=$matrix" >> $GITHUB_OUTPUT | |
- name: Pipeline Tests Artifacts | |
if: ${{ always() }} | |
uses: actions/upload-artifact@v2 | |
with: | |
name: test-pipelines.json | |
path: reports | |
run_nightly_tests_for_torch_pipelines: | |
name: Torch Pipelines CUDA Nightly Tests | |
needs: setup_torch_cuda_pipeline_matrix | |
strategy: | |
fail-fast: false | |
matrix: | |
module: ${{ fromJson(needs.setup_torch_cuda_pipeline_matrix.outputs.pipeline_test_matrix) }} | |
runs-on: [single-gpu, nvidia-gpu, t4, ci] | |
container: | |
image: diffusers/diffusers-pytorch-cuda | |
options: --shm-size "16gb" --ipc host -v /mnt/cache/.cache/huggingface/diffusers:/mnt/cache/ --gpus 0 | |
steps: | |
- name: Checkout diffusers | |
uses: actions/checkout@v3 | |
with: | |
fetch-depth: 2 | |
- name: NVIDIA-SMI | |
run: nvidia-smi | |
- name: Install dependencies | |
run: | | |
python -m venv /opt/venv && export PATH="/opt/venv/bin:$PATH" | |
python -m uv pip install -e [quality,test] | |
python -m uv pip install accelerate@git+https://github.com/huggingface/accelerate.git | |
python -m uv pip install pytest-reportlog | |
- name: Environment | |
run: | | |
python utils/print_env.py | |
- name: Nightly PyTorch CUDA checkpoint (pipelines) tests | |
env: | |
HF_TOKEN: ${{ secrets.HF_TOKEN }} | |
# https://pytorch.org/docs/stable/notes/randomness.html#avoiding-nondeterministic-algorithms | |
CUBLAS_WORKSPACE_CONFIG: :16:8 | |
run: | | |
python -m pytest -n 1 --max-worker-restart=0 --dist=loadfile \ | |
-s -v -k "not Flax and not Onnx" \ | |
--make-reports=tests_pipeline_${{ matrix.module }}_cuda \ | |
--report-log=tests_pipeline_${{ matrix.module }}_cuda.log \ | |
tests/pipelines/${{ matrix.module }} | |
- name: Failure short reports | |
if: ${{ failure() }} | |
run: | | |
cat reports/tests_pipeline_${{ matrix.module }}_cuda_stats.txt | |
cat reports/tests_pipeline_${{ matrix.module }}_cuda_failures_short.txt | |
- name: Test suite reports artifacts | |
if: ${{ always() }} | |
uses: actions/upload-artifact@v2 | |
with: | |
name: pipeline_${{ matrix.module }}_test_reports | |
path: reports | |
- name: Generate Report and Notify Channel | |
if: always() | |
run: | | |
pip install slack_sdk tabulate | |
python scripts/log_reports.py >> $GITHUB_STEP_SUMMARY | |
run_nightly_tests_for_other_torch_modules: | |
name: Torch Non-Pipelines CUDA Nightly Tests | |
runs-on: [single-gpu, nvidia-gpu, t4, ci] | |
container: | |
image: diffusers/diffusers-pytorch-cuda | |
options: --shm-size "16gb" --ipc host -v /mnt/hf_cache:/mnt/cache/ --gpus 0 | |
defaults: | |
run: | |
shell: bash | |
strategy: | |
matrix: | |
module: [models, schedulers, others, examples] | |
steps: | |
- name: Checkout diffusers | |
uses: actions/checkout@v3 | |
with: | |
fetch-depth: 2 | |
- name: Install dependencies | |
run: | | |
python -m venv /opt/venv && export PATH="/opt/venv/bin:$PATH" | |
python -m uv pip install -e [quality,test] | |
python -m uv pip install accelerate@git+https://github.com/huggingface/accelerate.git | |
python -m uv pip install pytest-reportlog | |
- name: Environment | |
run: python utils/print_env.py | |
- name: Run nightly PyTorch CUDA tests for non-pipeline modules | |
if: ${{ matrix.module != 'examples'}} | |
env: | |
HF_TOKEN: ${{ secrets.HF_TOKEN }} | |
# https://pytorch.org/docs/stable/notes/randomness.html#avoiding-nondeterministic-algorithms | |
CUBLAS_WORKSPACE_CONFIG: :16:8 | |
run: | | |
python -m pytest -n 1 --max-worker-restart=0 --dist=loadfile \ | |
-s -v -k "not Flax and not Onnx" \ | |
--make-reports=tests_torch_${{ matrix.module }}_cuda \ | |
--report-log=tests_torch_${{ matrix.module }}_cuda.log \ | |
tests/${{ matrix.module }} | |
- name: Run nightly example tests with Torch | |
if: ${{ matrix.module == 'examples' }} | |
env: | |
HF_TOKEN: ${{ secrets.HF_TOKEN }} | |
# https://pytorch.org/docs/stable/notes/randomness.html#avoiding-nondeterministic-algorithms | |
CUBLAS_WORKSPACE_CONFIG: :16:8 | |
run: | | |
python -m uv pip install peft@git+https://github.com/huggingface/peft.git | |
python -m pytest -n 1 --max-worker-restart=0 --dist=loadfile \ | |
-s -v --make-reports=examples_torch_cuda \ | |
--report-log=examples_torch_cuda.log \ | |
examples/ | |
- name: Failure short reports | |
if: ${{ failure() }} | |
run: | | |
cat reports/tests_torch_${{ matrix.module }}_cuda_stats.txt | |
cat reports/tests_torch_${{ matrix.module }}_cuda_failures_short.txt | |
- name: Test suite reports artifacts | |
if: ${{ always() }} | |
uses: actions/upload-artifact@v2 | |
with: | |
name: torch_${{ matrix.module }}_cuda_test_reports | |
path: reports | |
- name: Generate Report and Notify Channel | |
if: always() | |
run: | | |
pip install slack_sdk tabulate | |
python scripts/log_reports.py >> $GITHUB_STEP_SUMMARY | |
run_lora_nightly_tests: | |
name: Nightly LoRA Tests with PEFT and TORCH | |
runs-on: [single-gpu, nvidia-gpu, t4, ci] | |
container: | |
image: diffusers/diffusers-pytorch-cuda | |
options: --shm-size "16gb" --ipc host -v /mnt/hf_cache:/mnt/cache/ --gpus 0 | |
defaults: | |
run: | |
shell: bash | |
steps: | |
- name: Checkout diffusers | |
uses: actions/checkout@v3 | |
with: | |
fetch-depth: 2 | |
- name: Install dependencies | |
run: | | |
python -m venv /opt/venv && export PATH="/opt/venv/bin:$PATH" | |
python -m uv pip install -e [quality,test] | |
python -m uv pip install accelerate@git+https://github.com/huggingface/accelerate.git | |
python -m uv pip install peft@git+https://github.com/huggingface/peft.git | |
python -m uv pip install pytest-reportlog | |
- name: Environment | |
run: python utils/print_env.py | |
- name: Run nightly LoRA tests with PEFT and Torch | |
env: | |
HF_TOKEN: ${{ secrets.HF_TOKEN }} | |
# https://pytorch.org/docs/stable/notes/randomness.html#avoiding-nondeterministic-algorithms | |
CUBLAS_WORKSPACE_CONFIG: :16:8 | |
run: | | |
python -m pytest -n 1 --max-worker-restart=0 --dist=loadfile \ | |
-s -v -k "not Flax and not Onnx" \ | |
--make-reports=tests_torch_lora_cuda \ | |
--report-log=tests_torch_lora_cuda.log \ | |
tests/lora | |
- name: Failure short reports | |
if: ${{ failure() }} | |
run: | | |
cat reports/tests_torch_lora_cuda_stats.txt | |
cat reports/tests_torch_lora_cuda_failures_short.txt | |
- name: Test suite reports artifacts | |
if: ${{ always() }} | |
uses: actions/upload-artifact@v2 | |
with: | |
name: torch_lora_cuda_test_reports | |
path: reports | |
- name: Generate Report and Notify Channel | |
if: always() | |
run: | | |
pip install slack_sdk tabulate | |
python scripts/log_reports.py >> $GITHUB_STEP_SUMMARY | |
run_flax_tpu_tests: | |
name: Nightly Flax TPU Tests | |
runs-on: docker-tpu | |
if: github.event_name == 'schedule' | |
container: | |
image: diffusers/diffusers-flax-tpu | |
options: --shm-size "16gb" --ipc host -v /mnt/hf_cache:/mnt/cache/ --privileged | |
defaults: | |
run: | |
shell: bash | |
steps: | |
- name: Checkout diffusers | |
uses: actions/checkout@v3 | |
with: | |
fetch-depth: 2 | |
- name: Install dependencies | |
run: | | |
python -m venv /opt/venv && export PATH="/opt/venv/bin:$PATH" | |
python -m uv pip install -e [quality,test] | |
python -m uv pip install accelerate@git+https://github.com/huggingface/accelerate.git | |
python -m uv pip install pytest-reportlog | |
- name: Environment | |
run: python utils/print_env.py | |
- name: Run nightly Flax TPU tests | |
env: | |
HF_TOKEN: ${{ secrets.HF_TOKEN }} | |
run: | | |
python -m pytest -n 0 \ | |
-s -v -k "Flax" \ | |
--make-reports=tests_flax_tpu \ | |
--report-log=tests_flax_tpu.log \ | |
tests/ | |
- name: Failure short reports | |
if: ${{ failure() }} | |
run: | | |
cat reports/tests_flax_tpu_stats.txt | |
cat reports/tests_flax_tpu_failures_short.txt | |
- name: Test suite reports artifacts | |
if: ${{ always() }} | |
uses: actions/upload-artifact@v2 | |
with: | |
name: flax_tpu_test_reports | |
path: reports | |
- name: Generate Report and Notify Channel | |
if: always() | |
run: | | |
pip install slack_sdk tabulate | |
python scripts/log_reports.py >> $GITHUB_STEP_SUMMARY | |
run_nightly_onnx_tests: | |
name: Nightly ONNXRuntime CUDA tests on Ubuntu | |
runs-on: [single-gpu, nvidia-gpu, t4, ci] | |
container: | |
image: diffusers/diffusers-onnxruntime-cuda | |
options: --gpus 0 --shm-size "16gb" --ipc host -v /mnt/hf_cache:/mnt/cache/ | |
steps: | |
- name: Checkout diffusers | |
uses: actions/checkout@v3 | |
with: | |
fetch-depth: 2 | |
- name: NVIDIA-SMI | |
run: nvidia-smi | |
- name: Install dependencies | |
run: | | |
python -m venv /opt/venv && export PATH="/opt/venv/bin:$PATH" | |
python -m uv pip install -e [quality,test] | |
python -m uv pip install accelerate@git+https://github.com/huggingface/accelerate.git | |
python -m uv pip install pytest-reportlog | |
- name: Environment | |
run: python utils/print_env.py | |
- name: Run nightly ONNXRuntime CUDA tests | |
env: | |
HF_TOKEN: ${{ secrets.HF_TOKEN }} | |
run: | | |
python -m pytest -n 1 --max-worker-restart=0 --dist=loadfile \ | |
-s -v -k "Onnx" \ | |
--make-reports=tests_onnx_cuda \ | |
--report-log=tests_onnx_cuda.log \ | |
tests/ | |
- name: Failure short reports | |
if: ${{ failure() }} | |
run: | | |
cat reports/tests_onnx_cuda_stats.txt | |
cat reports/tests_onnx_cuda_failures_short.txt | |
- name: Test suite reports artifacts | |
if: ${{ always() }} | |
uses: actions/upload-artifact@v2 | |
with: | |
name: ${{ matrix.config.report }}_test_reports | |
path: reports | |
- name: Generate Report and Notify Channel | |
if: always() | |
run: | | |
pip install slack_sdk tabulate | |
python scripts/log_reports.py >> $GITHUB_STEP_SUMMARY | |
run_nightly_tests_apple_m1: | |
name: Nightly PyTorch MPS tests on MacOS | |
runs-on: [ self-hosted, apple-m1 ] | |
if: github.event_name == 'schedule' | |
steps: | |
- name: Checkout diffusers | |
uses: actions/checkout@v3 | |
with: | |
fetch-depth: 2 | |
- name: Clean checkout | |
shell: arch -arch arm64 bash {0} | |
run: | | |
git clean -fxd | |
- name: Setup miniconda | |
uses: ./.github/actions/setup-miniconda | |
with: | |
python-version: 3.9 | |
- name: Install dependencies | |
shell: arch -arch arm64 bash {0} | |
run: | | |
${CONDA_RUN} python -m pip install --upgrade pip uv | |
${CONDA_RUN} python -m uv pip install -e [quality,test] | |
${CONDA_RUN} python -m uv pip install torch torchvision torchaudio --extra-index-url https://download.pytorch.org/whl/cpu | |
${CONDA_RUN} python -m uv pip install accelerate@git+https://github.com/huggingface/accelerate | |
${CONDA_RUN} python -m uv pip install pytest-reportlog | |
- name: Environment | |
shell: arch -arch arm64 bash {0} | |
run: | | |
${CONDA_RUN} python utils/print_env.py | |
- name: Run nightly PyTorch tests on M1 (MPS) | |
shell: arch -arch arm64 bash {0} | |
env: | |
HF_HOME: /System/Volumes/Data/mnt/cache | |
HF_TOKEN: ${{ secrets.HF_TOKEN }} | |
run: | | |
${CONDA_RUN} python -m pytest -n 1 -s -v --make-reports=tests_torch_mps \ | |
--report-log=tests_torch_mps.log \ | |
tests/ | |
- name: Failure short reports | |
if: ${{ failure() }} | |
run: cat reports/tests_torch_mps_failures_short.txt | |
- name: Test suite reports artifacts | |
if: ${{ always() }} | |
uses: actions/upload-artifact@v2 | |
with: | |
name: torch_mps_test_reports | |
path: reports | |
- name: Generate Report and Notify Channel | |
if: always() | |
run: | | |
pip install slack_sdk tabulate | |
python scripts/log_reports.py >> $GITHUB_STEP_SUMMARY |