2121import torch
2222
2323from diffusers import UNet2DConditionModel , UNet2DModel
24- from diffusers .utils import floats_tensor , load_numpy , require_torch_gpu , slow , torch_all_close , torch_device
24+ from diffusers .utils import floats_tensor , load_numpy , logging , require_torch_gpu , slow , torch_all_close , torch_device
2525from parameterized import parameterized
2626
2727from ..test_modeling_common import ModelTesterMixin
2828
2929
30+ logger = logging .get_logger (__name__ )
3031torch .backends .cuda .matmul .allow_tf32 = False
3132
3233
@@ -464,7 +465,7 @@ def test_compvis_sd_v1_4(self, seed, timestep, expected_slice):
464465 output_slice = sample [- 1 , - 2 :, - 2 :, :2 ].flatten ().float ().cpu ()
465466 expected_output_slice = torch .tensor (expected_slice )
466467
467- assert torch_all_close (output_slice , expected_output_slice , atol = 1e-4 )
468+ assert torch_all_close (output_slice , expected_output_slice , atol = 1e-3 )
468469
469470 @parameterized .expand (
470471 [
@@ -490,7 +491,7 @@ def test_compvis_sd_v1_4_fp16(self, seed, timestep, expected_slice):
490491 output_slice = sample [- 1 , - 2 :, - 2 :, :2 ].flatten ().float ().cpu ()
491492 expected_output_slice = torch .tensor (expected_slice )
492493
493- assert torch_all_close (output_slice , expected_output_slice , atol = 1e-4 )
494+ assert torch_all_close (output_slice , expected_output_slice , atol = 5e-3 )
494495
495496 @parameterized .expand (
496497 [
@@ -515,7 +516,7 @@ def test_compvis_sd_v1_5(self, seed, timestep, expected_slice):
515516 output_slice = sample [- 1 , - 2 :, - 2 :, :2 ].flatten ().float ().cpu ()
516517 expected_output_slice = torch .tensor (expected_slice )
517518
518- assert torch_all_close (output_slice , expected_output_slice , atol = 1e-4 )
519+ assert torch_all_close (output_slice , expected_output_slice , atol = 1e-3 )
519520
520521 @parameterized .expand (
521522 [
@@ -541,7 +542,7 @@ def test_compvis_sd_v1_5_fp16(self, seed, timestep, expected_slice):
541542 output_slice = sample [- 1 , - 2 :, - 2 :, :2 ].flatten ().float ().cpu ()
542543 expected_output_slice = torch .tensor (expected_slice )
543544
544- assert torch_all_close (output_slice , expected_output_slice , atol = 1e-4 )
545+ assert torch_all_close (output_slice , expected_output_slice , atol = 5e-3 )
545546
546547 @parameterized .expand (
547548 [
@@ -566,7 +567,7 @@ def test_compvis_sd_inpaint(self, seed, timestep, expected_slice):
566567 output_slice = sample [- 1 , - 2 :, - 2 :, :2 ].flatten ().float ().cpu ()
567568 expected_output_slice = torch .tensor (expected_slice )
568569
569- assert torch_all_close (output_slice , expected_output_slice , atol = 1e-4 )
570+ assert torch_all_close (output_slice , expected_output_slice , atol = 1e-3 )
570571
571572 @parameterized .expand (
572573 [
@@ -592,4 +593,4 @@ def test_compvis_sd_inpaint_fp16(self, seed, timestep, expected_slice):
592593 output_slice = sample [- 1 , - 2 :, - 2 :, :2 ].flatten ().float ().cpu ()
593594 expected_output_slice = torch .tensor (expected_slice )
594595
595- assert torch_all_close (output_slice , expected_output_slice , atol = 1e-4 )
596+ assert torch_all_close (output_slice , expected_output_slice , atol = 5e-3 )
0 commit comments