Skip to content

Commit de7c531

Browse files
ulivneXuehaoSun
authored andcommitted
[FSW-12066] small fixes in xpu quantized func (#145)
1 parent 65b89d5 commit de7c531

File tree

2 files changed

+1
-1
lines changed

2 files changed

+1
-1
lines changed

neural_compressor/torch/algorithms/fp8_quant/_core/quantized_func_wrappers/xpu/xpu_quantized_func_wrapper.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -65,6 +65,7 @@ def get_default_quantized_func(self):
6565

6666
_OP_TYPE_XPU_QUANTIZED_WRAPPER_CLASSES = {
6767
OP_TYPE.LINEAR_GEMM : QuantizedXPUMatmul,
68+
OP_TYPE.MATMUL_GEMM : QuantizedXPUMatmul,
6869
OP_TYPE.CAST_TO_FP8 : QuantizedXPUCastToFP8Base,
6970
OP_TYPE.CAST_FROM_FP8 : QuantizedXPUCastFromFP8Base
7071
}

test/3x/torch/algorithms/fp8_quant_xpu/unit_tests/test_xpu_basic.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -61,7 +61,6 @@ def forward(self, input):
6161
return self.my_linear(input)
6262

6363

64-
@pytest.mark.skip(reason="FSW-13402 device hangs")
6564
def test_xpu_basic_mamtul():
6665
# test convert flow and quantized func
6766
my_model = MyModelMatmul()

0 commit comments

Comments
 (0)