From edbf94f346863890f0037dbce55429665cd44925 Mon Sep 17 00:00:00 2001 From: Kaihui-intel Date: Tue, 18 Feb 2025 16:03:27 +0800 Subject: [PATCH 1/4] adpat ipex xpu transformers version Signed-off-by: Kaihui-intel --- .../transformers/models/modeling_auto.py | 17 ++++++++++++----- 1 file changed, 12 insertions(+), 5 deletions(-) diff --git a/neural_compressor/transformers/models/modeling_auto.py b/neural_compressor/transformers/models/modeling_auto.py index 281a19dcf56..38f1cdd2355 100644 --- a/neural_compressor/transformers/models/modeling_auto.py +++ b/neural_compressor/transformers/models/modeling_auto.py @@ -758,12 +758,19 @@ class AutoModelForSeq2SeqLM(_BaseINCAutoModelClass): class Qwen2VLForConditionalGeneration(_BaseINCAutoModelClass): - ORIG_MODEL = transformers.Qwen2VLForConditionalGeneration - + if transformers.__version__ >= "4.46": + ORIG_MODEL = transformers.Qwen2VLForConditionalGeneration + else: + logger.warning("please install transfomrts>=4.46 for quantizing Qwen2VLForConditionalGeneration.") class MllamaForConditionalGeneration(_BaseINCAutoModelClass): - ORIG_MODEL = transformers.MllamaForConditionalGeneration - + if transformers.__version__ >= "4.46": + ORIG_MODEL = transformers.MllamaForConditionalGeneration + else: + logger.warning("please install transfomrts>=4.46 for quantizing MllamaForConditionalGeneration.") class LlavaForConditionalGeneration(_BaseINCAutoModelClass): - ORIG_MODEL = transformers.LlavaForConditionalGeneration + if transformers.__version__ >= "4.46": + ORIG_MODEL = transformers.LlavaForConditionalGeneration + else: + logger.warning("please install transfomrts>=4.46 for quantizing LlavaForConditionalGeneration.") From e9659482b08b20a1f721af97311f26f3e60abb77 Mon Sep 17 00:00:00 2001 From: "pre-commit-ci[bot]" <66853113+pre-commit-ci[bot]@users.noreply.github.com> Date: Fri, 28 Feb 2025 06:06:59 +0000 Subject: [PATCH 2/4] [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci --- neural_compressor/transformers/models/modeling_auto.py | 2 ++ 1 file changed, 2 insertions(+) diff --git a/neural_compressor/transformers/models/modeling_auto.py b/neural_compressor/transformers/models/modeling_auto.py index 38f1cdd2355..77e354f6710 100644 --- a/neural_compressor/transformers/models/modeling_auto.py +++ b/neural_compressor/transformers/models/modeling_auto.py @@ -763,12 +763,14 @@ class Qwen2VLForConditionalGeneration(_BaseINCAutoModelClass): else: logger.warning("please install transfomrts>=4.46 for quantizing Qwen2VLForConditionalGeneration.") + class MllamaForConditionalGeneration(_BaseINCAutoModelClass): if transformers.__version__ >= "4.46": ORIG_MODEL = transformers.MllamaForConditionalGeneration else: logger.warning("please install transfomrts>=4.46 for quantizing MllamaForConditionalGeneration.") + class LlavaForConditionalGeneration(_BaseINCAutoModelClass): if transformers.__version__ >= "4.46": ORIG_MODEL = transformers.LlavaForConditionalGeneration From 989c3c643e8b06111f690ef895677e5cb7d61d19 Mon Sep 17 00:00:00 2001 From: Kaihui-intel Date: Fri, 28 Feb 2025 14:20:22 +0800 Subject: [PATCH 3/4] fix typo Signed-off-by: Kaihui-intel --- neural_compressor/transformers/models/modeling_auto.py | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/neural_compressor/transformers/models/modeling_auto.py b/neural_compressor/transformers/models/modeling_auto.py index 38f1cdd2355..29282373844 100644 --- a/neural_compressor/transformers/models/modeling_auto.py +++ b/neural_compressor/transformers/models/modeling_auto.py @@ -761,16 +761,16 @@ class Qwen2VLForConditionalGeneration(_BaseINCAutoModelClass): if transformers.__version__ >= "4.46": ORIG_MODEL = transformers.Qwen2VLForConditionalGeneration else: - logger.warning("please install transfomrts>=4.46 for quantizing Qwen2VLForConditionalGeneration.") + logger.warning("please install transfomers>=4.46 for quantizing Qwen2VLForConditionalGeneration.") class MllamaForConditionalGeneration(_BaseINCAutoModelClass): if transformers.__version__ >= "4.46": ORIG_MODEL = transformers.MllamaForConditionalGeneration else: - logger.warning("please install transfomrts>=4.46 for quantizing MllamaForConditionalGeneration.") + logger.warning("please install transfomers>=4.46 for quantizing MllamaForConditionalGeneration.") class LlavaForConditionalGeneration(_BaseINCAutoModelClass): if transformers.__version__ >= "4.46": ORIG_MODEL = transformers.LlavaForConditionalGeneration else: - logger.warning("please install transfomrts>=4.46 for quantizing LlavaForConditionalGeneration.") + logger.warning("please install transfomers>=4.46 for quantizing LlavaForConditionalGeneration.") From a374a4996fc925a4b71a6a907016f5fd9a43f14c Mon Sep 17 00:00:00 2001 From: "pre-commit-ci[bot]" <66853113+pre-commit-ci[bot]@users.noreply.github.com> Date: Fri, 28 Feb 2025 07:17:48 +0000 Subject: [PATCH 4/4] [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci --- neural_compressor/transformers/models/modeling_auto.py | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/neural_compressor/transformers/models/modeling_auto.py b/neural_compressor/transformers/models/modeling_auto.py index bc39c0f0ab8..efd38a719dc 100644 --- a/neural_compressor/transformers/models/modeling_auto.py +++ b/neural_compressor/transformers/models/modeling_auto.py @@ -761,18 +761,18 @@ class Qwen2VLForConditionalGeneration(_BaseINCAutoModelClass): if transformers.__version__ >= "4.46": ORIG_MODEL = transformers.Qwen2VLForConditionalGeneration else: - logger.warning("please install transfomers>=4.46 for quantizing Qwen2VLForConditionalGeneration.") + logger.warning("please install transformers>=4.46 for quantizing Qwen2VLForConditionalGeneration.") class MllamaForConditionalGeneration(_BaseINCAutoModelClass): if transformers.__version__ >= "4.46": ORIG_MODEL = transformers.MllamaForConditionalGeneration else: - logger.warning("please install transfomers>=4.46 for quantizing MllamaForConditionalGeneration.") + logger.warning("please install transformers>=4.46 for quantizing MllamaForConditionalGeneration.") class LlavaForConditionalGeneration(_BaseINCAutoModelClass): if transformers.__version__ >= "4.46": ORIG_MODEL = transformers.LlavaForConditionalGeneration else: - logger.warning("please install transfomers>=4.46 for quantizing LlavaForConditionalGeneration.") + logger.warning("please install transformers>=4.46 for quantizing LlavaForConditionalGeneration.")