|
24 | 24 | QwenImagePipeline,
|
25 | 25 | QwenImageTransformer2DModel,
|
26 | 26 | )
|
27 |
| -from diffusers.utils.testing_utils import enable_full_determinism, torch_device |
| 27 | +from diffusers.utils.testing_utils import CaptureLogger, enable_full_determinism, torch_device |
28 | 28 |
|
29 | 29 | from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS
|
30 | 30 | from ..test_pipelines_common import PipelineTesterMixin, to_np
|
@@ -264,27 +264,24 @@ def test_long_prompt_no_error(self):
|
264 | 264 |
|
265 | 265 | def test_long_prompt_warning(self):
|
266 | 266 | """Test that long prompts trigger appropriate warning about training limitation"""
|
267 |
| - from diffusers.utils.testing_utils import CaptureLogger |
268 | 267 | from diffusers.utils import logging
|
269 | 268 |
|
270 |
| - device = torch_device |
271 | 269 | components = self.get_dummy_components()
|
272 | 270 | pipe = self.pipeline_class(**components)
|
273 |
| - pipe.to(device) |
| 271 | + pipe.to(torch_device) |
274 | 272 |
|
275 | 273 | # Create prompt that will exceed 512 tokens to trigger warning
|
276 |
| - # Use a longer phrase and repeat more times to ensure we exceed the 512 token limit |
277 | 274 | long_phrase = "A detailed photorealistic description of a complex scene with many elements "
|
278 | 275 | long_prompt = (long_phrase * 20)[:800] # Create a prompt that will exceed 512 tokens
|
279 | 276 |
|
280 | 277 | # Capture transformer logging
|
281 | 278 | logger = logging.get_logger("diffusers.models.transformers.transformer_qwenimage")
|
282 |
| - logger.setLevel(30) # WARNING level |
| 279 | + logger.setLevel(logging.WARNING) |
283 | 280 |
|
284 | 281 | with CaptureLogger(logger) as cap_logger:
|
285 | 282 | _ = pipe(
|
286 | 283 | prompt=long_prompt,
|
287 |
| - generator=torch.Generator(device=device).manual_seed(0), |
| 284 | + generator=torch.Generator(device=torch_device).manual_seed(0), |
288 | 285 | num_inference_steps=2,
|
289 | 286 | guidance_scale=3.0,
|
290 | 287 | true_cfg_scale=1.0,
|
|
0 commit comments