mirror of
https://github.com/huggingface/diffusers.git
synced 2025-12-08 21:44:27 +08:00
Compare commits
2 Commits
prompt-iso
...
enable-all
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
96ef4bf62a | ||
|
|
866740a42b |
2
.github/workflows/benchmark.yml
vendored
2
.github/workflows/benchmark.yml
vendored
@@ -25,7 +25,7 @@ jobs:
|
|||||||
group: aws-g6e-4xlarge
|
group: aws-g6e-4xlarge
|
||||||
container:
|
container:
|
||||||
image: diffusers/diffusers-pytorch-cuda
|
image: diffusers/diffusers-pytorch-cuda
|
||||||
options: --shm-size "16gb" --ipc host --gpus 0
|
options: --shm-size "16gb" --ipc host --gpus all
|
||||||
steps:
|
steps:
|
||||||
- name: Checkout diffusers
|
- name: Checkout diffusers
|
||||||
uses: actions/checkout@v3
|
uses: actions/checkout@v3
|
||||||
|
|||||||
8
.github/workflows/nightly_tests.yml
vendored
8
.github/workflows/nightly_tests.yml
vendored
@@ -61,7 +61,7 @@ jobs:
|
|||||||
group: aws-g4dn-2xlarge
|
group: aws-g4dn-2xlarge
|
||||||
container:
|
container:
|
||||||
image: diffusers/diffusers-pytorch-cuda
|
image: diffusers/diffusers-pytorch-cuda
|
||||||
options: --shm-size "16gb" --ipc host --gpus 0
|
options: --shm-size "16gb" --ipc host --gpus all
|
||||||
steps:
|
steps:
|
||||||
- name: Checkout diffusers
|
- name: Checkout diffusers
|
||||||
uses: actions/checkout@v3
|
uses: actions/checkout@v3
|
||||||
@@ -107,7 +107,7 @@ jobs:
|
|||||||
group: aws-g4dn-2xlarge
|
group: aws-g4dn-2xlarge
|
||||||
container:
|
container:
|
||||||
image: diffusers/diffusers-pytorch-cuda
|
image: diffusers/diffusers-pytorch-cuda
|
||||||
options: --shm-size "16gb" --ipc host --gpus 0
|
options: --shm-size "16gb" --ipc host --gpus all
|
||||||
defaults:
|
defaults:
|
||||||
run:
|
run:
|
||||||
shell: bash
|
shell: bash
|
||||||
@@ -222,7 +222,7 @@ jobs:
|
|||||||
group: aws-g6e-xlarge-plus
|
group: aws-g6e-xlarge-plus
|
||||||
container:
|
container:
|
||||||
image: diffusers/diffusers-pytorch-cuda
|
image: diffusers/diffusers-pytorch-cuda
|
||||||
options: --shm-size "16gb" --ipc host --gpus 0
|
options: --shm-size "16gb" --ipc host --gpus all
|
||||||
steps:
|
steps:
|
||||||
- name: Checkout diffusers
|
- name: Checkout diffusers
|
||||||
uses: actions/checkout@v3
|
uses: actions/checkout@v3
|
||||||
@@ -270,7 +270,7 @@ jobs:
|
|||||||
group: aws-g4dn-2xlarge
|
group: aws-g4dn-2xlarge
|
||||||
container:
|
container:
|
||||||
image: diffusers/diffusers-pytorch-minimum-cuda
|
image: diffusers/diffusers-pytorch-minimum-cuda
|
||||||
options: --shm-size "16gb" --ipc host --gpus 0
|
options: --shm-size "16gb" --ipc host --gpus all
|
||||||
defaults:
|
defaults:
|
||||||
run:
|
run:
|
||||||
shell: bash
|
shell: bash
|
||||||
|
|||||||
4
.github/workflows/pr_tests_gpu.yml
vendored
4
.github/workflows/pr_tests_gpu.yml
vendored
@@ -118,7 +118,7 @@ jobs:
|
|||||||
group: aws-g4dn-2xlarge
|
group: aws-g4dn-2xlarge
|
||||||
container:
|
container:
|
||||||
image: diffusers/diffusers-pytorch-cuda
|
image: diffusers/diffusers-pytorch-cuda
|
||||||
options: --shm-size "16gb" --ipc host --gpus 0
|
options: --shm-size "16gb" --ipc host --gpus all
|
||||||
steps:
|
steps:
|
||||||
- name: Checkout diffusers
|
- name: Checkout diffusers
|
||||||
uses: actions/checkout@v3
|
uses: actions/checkout@v3
|
||||||
@@ -183,7 +183,7 @@ jobs:
|
|||||||
group: aws-g4dn-2xlarge
|
group: aws-g4dn-2xlarge
|
||||||
container:
|
container:
|
||||||
image: diffusers/diffusers-pytorch-cuda
|
image: diffusers/diffusers-pytorch-cuda
|
||||||
options: --shm-size "16gb" --ipc host --gpus 0
|
options: --shm-size "16gb" --ipc host --gpus all
|
||||||
defaults:
|
defaults:
|
||||||
run:
|
run:
|
||||||
shell: bash
|
shell: bash
|
||||||
|
|||||||
4
.github/workflows/push_tests.yml
vendored
4
.github/workflows/push_tests.yml
vendored
@@ -64,7 +64,7 @@ jobs:
|
|||||||
group: aws-g4dn-2xlarge
|
group: aws-g4dn-2xlarge
|
||||||
container:
|
container:
|
||||||
image: diffusers/diffusers-pytorch-cuda
|
image: diffusers/diffusers-pytorch-cuda
|
||||||
options: --shm-size "16gb" --ipc host --gpus 0
|
options: --shm-size "16gb" --ipc host --gpus all
|
||||||
steps:
|
steps:
|
||||||
- name: Checkout diffusers
|
- name: Checkout diffusers
|
||||||
uses: actions/checkout@v3
|
uses: actions/checkout@v3
|
||||||
@@ -109,7 +109,7 @@ jobs:
|
|||||||
group: aws-g4dn-2xlarge
|
group: aws-g4dn-2xlarge
|
||||||
container:
|
container:
|
||||||
image: diffusers/diffusers-pytorch-cuda
|
image: diffusers/diffusers-pytorch-cuda
|
||||||
options: --shm-size "16gb" --ipc host --gpus 0
|
options: --shm-size "16gb" --ipc host --gpus all
|
||||||
defaults:
|
defaults:
|
||||||
run:
|
run:
|
||||||
shell: bash
|
shell: bash
|
||||||
|
|||||||
6
.github/workflows/release_tests_fast.yml
vendored
6
.github/workflows/release_tests_fast.yml
vendored
@@ -62,7 +62,7 @@ jobs:
|
|||||||
group: aws-g4dn-2xlarge
|
group: aws-g4dn-2xlarge
|
||||||
container:
|
container:
|
||||||
image: diffusers/diffusers-pytorch-cuda
|
image: diffusers/diffusers-pytorch-cuda
|
||||||
options: --shm-size "16gb" --ipc host --gpus 0
|
options: --shm-size "16gb" --ipc host --gpus all
|
||||||
steps:
|
steps:
|
||||||
- name: Checkout diffusers
|
- name: Checkout diffusers
|
||||||
uses: actions/checkout@v3
|
uses: actions/checkout@v3
|
||||||
@@ -107,7 +107,7 @@ jobs:
|
|||||||
group: aws-g4dn-2xlarge
|
group: aws-g4dn-2xlarge
|
||||||
container:
|
container:
|
||||||
image: diffusers/diffusers-pytorch-cuda
|
image: diffusers/diffusers-pytorch-cuda
|
||||||
options: --shm-size "16gb" --ipc host --gpus 0
|
options: --shm-size "16gb" --ipc host --gpus all
|
||||||
defaults:
|
defaults:
|
||||||
run:
|
run:
|
||||||
shell: bash
|
shell: bash
|
||||||
@@ -163,7 +163,7 @@ jobs:
|
|||||||
group: aws-g4dn-2xlarge
|
group: aws-g4dn-2xlarge
|
||||||
container:
|
container:
|
||||||
image: diffusers/diffusers-pytorch-minimum-cuda
|
image: diffusers/diffusers-pytorch-minimum-cuda
|
||||||
options: --shm-size "16gb" --ipc host --gpus 0
|
options: --shm-size "16gb" --ipc host --gpus all
|
||||||
defaults:
|
defaults:
|
||||||
run:
|
run:
|
||||||
shell: bash
|
shell: bash
|
||||||
|
|||||||
Reference in New Issue
Block a user