|
17 | 17 |
|
18 | 18 | import numpy as np |
19 | 19 | import torch |
| 20 | +from PIL import Image |
20 | 21 | from transformers import AutoTokenizer, T5EncoderModel |
21 | 22 |
|
22 | 23 | from diffusers import AutoencoderKLWan, UniPCMultistepScheduler, WanImageToVideoPipeline, WanTransformer3DModel |
|
27 | 28 |
|
28 | 29 | from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS |
29 | 30 | from ..test_pipelines_common import PipelineTesterMixin |
30 | | -from PIL import Image |
31 | 31 |
|
32 | 32 |
|
33 | 33 | enable_full_determinism() |
@@ -157,7 +157,10 @@ def test_inference(self): |
157 | 157 |
|
158 | 158 | generated_slice = generated_video.flatten() |
159 | 159 | generated_slice = torch.cat([generated_slice[:8], generated_slice[-8:]]) |
160 | | - self.assertTrue(torch.allclose(generated_slice, expected_slice, atol=1e-3), f"generated_slice: {generated_slice}, expected_slice: {expected_slice}") |
| 160 | + self.assertTrue( |
| 161 | + torch.allclose(generated_slice, expected_slice, atol=1e-3), |
| 162 | + f"generated_slice: {generated_slice}, expected_slice: {expected_slice}", |
| 163 | + ) |
161 | 164 |
|
162 | 165 | @unittest.skip("Test not supported") |
163 | 166 | def test_attention_slicing_forward_pass(self): |
@@ -371,7 +374,7 @@ def test_save_load_optional_components(self, expected_max_difference=1e-4): |
371 | 374 | self.assertTrue( |
372 | 375 | getattr(pipe_loaded, component) is None, |
373 | 376 | f"`{component}` did not stay set to None after loading.", |
374 | | - ) |
| 377 | + ) |
375 | 378 |
|
376 | 379 | inputs = self.get_dummy_inputs(generator_device) |
377 | 380 | torch.manual_seed(0) |
|
0 commit comments