We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent b3e67dd commit eabcc64Copy full SHA for eabcc64
test/3x/torch/quantization/fp8_quant/test_gptq_mixed_precision.py
@@ -63,7 +63,7 @@ def teardown_class(self):
63
Calibrates the w4a16 model on N num of samples, and compares the same samples with the hybrid model.
64
"""
65
@torch.no_grad()
66
- @pytest.mark.skip(reason="SW-223106 load model error")
+ @pytest.mark.skip(reason="This test is temporarily disabled SW-224890")
67
def test_mixed_precision_gptq_fp8_quant_only_nlp(self):
68
from neural_compressor.common import set_random_seed
69
set_random_seed(12345)
0 commit comments