mirror of
https://github.com/huggingface/diffusers.git
synced 2025-12-12 23:44:30 +08:00
Compare commits
2 Commits
enable-cp-
...
push-test-
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
373ea245eb | ||
|
|
377dbb302c |
@@ -417,6 +417,9 @@ class ModelTesterMixin:
|
|||||||
|
|
||||||
@require_torch_gpu
|
@require_torch_gpu
|
||||||
def test_set_attn_processor_for_determinism(self):
|
def test_set_attn_processor_for_determinism(self):
|
||||||
|
if self.uses_custom_attn_processor:
|
||||||
|
return
|
||||||
|
|
||||||
torch.use_deterministic_algorithms(False)
|
torch.use_deterministic_algorithms(False)
|
||||||
if self.forward_requires_fresh_args:
|
if self.forward_requires_fresh_args:
|
||||||
model = self.model_class(**self.init_dict)
|
model = self.model_class(**self.init_dict)
|
||||||
|
|||||||
@@ -32,6 +32,9 @@ class FluxTransformerTests(ModelTesterMixin, unittest.TestCase):
|
|||||||
# We override the items here because the transformer under consideration is small.
|
# We override the items here because the transformer under consideration is small.
|
||||||
model_split_percents = [0.7, 0.6, 0.6]
|
model_split_percents = [0.7, 0.6, 0.6]
|
||||||
|
|
||||||
|
# Skip setting testing with default: AttnProcessor
|
||||||
|
uses_custom_attn_processor = True
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def dummy_input(self):
|
def dummy_input(self):
|
||||||
batch_size = 1
|
batch_size = 1
|
||||||
|
|||||||
@@ -25,6 +25,9 @@ class FluxPipelineFastTests(unittest.TestCase, PipelineTesterMixin):
|
|||||||
params = frozenset(["prompt", "height", "width", "guidance_scale", "prompt_embeds", "pooled_prompt_embeds"])
|
params = frozenset(["prompt", "height", "width", "guidance_scale", "prompt_embeds", "pooled_prompt_embeds"])
|
||||||
batch_params = frozenset(["prompt"])
|
batch_params = frozenset(["prompt"])
|
||||||
|
|
||||||
|
# there is no xformers processor for Flux
|
||||||
|
test_xformers_attention = False
|
||||||
|
|
||||||
def get_dummy_components(self):
|
def get_dummy_components(self):
|
||||||
torch.manual_seed(0)
|
torch.manual_seed(0)
|
||||||
transformer = FluxTransformer2DModel(
|
transformer = FluxTransformer2DModel(
|
||||||
|
|||||||
@@ -37,6 +37,7 @@ class StableDiffusion3PAGPipelineFastTests(unittest.TestCase, PipelineTesterMixi
|
|||||||
]
|
]
|
||||||
)
|
)
|
||||||
batch_params = frozenset(["prompt", "negative_prompt"])
|
batch_params = frozenset(["prompt", "negative_prompt"])
|
||||||
|
test_xformers_attention = False
|
||||||
|
|
||||||
def get_dummy_components(self):
|
def get_dummy_components(self):
|
||||||
torch.manual_seed(0)
|
torch.manual_seed(0)
|
||||||
|
|||||||
@@ -68,6 +68,8 @@ class StableAudioPipelineFastTests(PipelineTesterMixin, unittest.TestCase):
|
|||||||
"callback_steps",
|
"callback_steps",
|
||||||
]
|
]
|
||||||
)
|
)
|
||||||
|
# There is not xformers version of the StableAudioPipeline custom attention processor
|
||||||
|
test_xformers_attention = False
|
||||||
|
|
||||||
def get_dummy_components(self):
|
def get_dummy_components(self):
|
||||||
torch.manual_seed(0)
|
torch.manual_seed(0)
|
||||||
|
|||||||
Reference in New Issue
Block a user