mirror of
https://github.com/huggingface/diffusers.git
synced 2026-04-14 19:57:04 +08:00
[ptxla] fix pytorch xla inference on TPUs. (#13463)
Co-authored-by: Juan Acevedo <jfacevedo@google.com>
This commit is contained in:
@@ -877,10 +877,7 @@ class FluxPipeline(
|
||||
self.scheduler.config.get("max_shift", 1.15),
|
||||
)
|
||||
|
||||
if XLA_AVAILABLE:
|
||||
timestep_device = "cpu"
|
||||
else:
|
||||
timestep_device = device
|
||||
timestep_device = device
|
||||
timesteps, num_inference_steps = retrieve_timesteps(
|
||||
self.scheduler,
|
||||
num_inference_steps,
|
||||
|
||||
Reference in New Issue
Block a user