Skip to content

Conversation

@changwangss
Copy link
Contributor

@changwangss changwangss commented Nov 30, 2022

Signed-off-by: changwa1 [email protected]

Type of Change

API changed

Description

enable examples for recommendation, speech recognition

Expected Behavior & Potential Risk

the expected behavior that triggered by this PR

How has this PR been tested?

how to reproduce the test (including hardware information)

Dependency Change?

any library dependency introduced or removed

Signed-off-by: changwa1 <[email protected]>
@changwangss changwangss changed the title update quantization api Update PyTorch exmpale quantization API Nov 30, 2022
@changwangss changwangss changed the title Update PyTorch exmpale quantization API Update PyTorch examples quantization API Nov 30, 2022
"batch_size": 16384,
"new_benchmark": false
"batch_size": 16384
},
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

require new parameter: “main_script”: “xxx.py”,

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Because test code need to replace some config class, so need to know the name scripts.

@chensuyue
Copy link
Contributor

  1. example readme need to update.
  2. benchmark part need update into Config Class as well.
  3. run_benchmark.sh input parameter update.
  4. json param missing main_script.
    Details pls check INC newAPI examples.pptx

PenghuiCheng and others added 17 commits December 1, 2022 21:45
Signed-off-by: changwa1 <[email protected]>
* Fixed UT error for bf16 op list for QAT mode

Signed-off-by: Cheng, Penghui <[email protected]>
Signed-off-by: Wang, Chang1 <[email protected]>
Signed-off-by: changwa1 <[email protected]>
Signed-off-by: Lv, Liang1 <[email protected]>

Signed-off-by: Lv, Liang1 <[email protected]>
* Fix centernet_hg104 tuning issue

Signed-off-by: sys-lpot-val <[email protected]>

* Fix TextRNN tuning issue

Signed-off-by: sys-lpot-val <[email protected]>

Signed-off-by: sys-lpot-val <[email protected]>
Signed-off-by: Lv, Liang1 <[email protected]>
Co-authored-by: sys-lpot-val <[email protected]>
…ers (#214)

* Create intel_extension_for_transformers.yaml

* change default strategy to dynamic according to huggingface sync

* change default strategy to dynamic according to HF sync

* enable intel extension for transformers

* Create change_trainer_to_nlptrainer.py

* add use_inc for not using default optimum for HF code

* add use_inc

* update optimum quant static dynamic separation

* Update interface.py

* Update interface.py

* Update autoinc_harness.py

* Update README.md

* add change_trainer_to_nlptrainer to outside_harness

* add PythonLauncher to pass spelling check CI

Signed-off-by: Yue, Wenjiao <[email protected]>

Signed-off-by: Yue, Wenjiao <[email protected]>
Co-authored-by: Yue, Wenjiao <[email protected]>
* update launcher to fit multi-item input

* Update __main__.py
…reliminary) (#217)

* Create tf_inc_static_quant.yaml

* Create inc.py

* Delete tf_inc_static_quant.yaml

* Update interface.py

* Update inc.py
XuehaoSun and others added 24 commits December 9, 2022 22:26
Signed-off-by: yiliu30 <[email protected]>
Co-authored-by: lvliang-intel <[email protected]>
Co-authored-by: chen, suyue <[email protected]>
Co-authored-by: xinhe <[email protected]>
Co-authored-by: Ray <[email protected]>
Signed-off-by: Lv, Liang1 <[email protected]>
Signed-off-by: chensuyue <[email protected]>
Co-authored-by: Lv, Liang1 <[email protected]>
Co-authored-by: chensuyue <[email protected]>
Signed-off-by: chensuyue <[email protected]>
* enable auto bench in launcher

* debug

* debug
* add pruning v2

Signed-off-by: wenhuach21 <[email protected]>

* DATASETS->Datasets

Signed-off-by: wenhuach21 <[email protected]>

* pruner README v2

Signed-off-by: Lu, Yintong <[email protected]>

* prune README v2

Signed-off-by: Lu, Yintong <[email protected]>

* prune README v2

Signed-off-by: Lu, Yintong <[email protected]>

* prune README v2

Signed-off-by: Lu, Yintong <[email protected]>

* prune README v2

Signed-off-by: Lu, Yintong <[email protected]>

* recover code in experimental

Signed-off-by: wenhuach21 <[email protected]>

* recover config

Signed-off-by: wenhuach21 <[email protected]>

Signed-off-by: wenhuach21 <[email protected]>
Signed-off-by: Lu, Yintong <[email protected]>
Co-authored-by: Lu, Yintong <[email protected]>
…arch Toolkit (#197)

Signed-off-by: Maciej Szankin <[email protected]>
Co-authored-by: Nittur Sridhar, Sharath <[email protected]>
Co-authored-by: Xinyu Ye <[email protected]>
Signed-off-by: Lv, Liang1 <[email protected]>
Co-authored-by: chen, suyue <[email protected]>
* update qa example

Signed-off-by: Zhang, Weiwei1 <[email protected]>

* pruning doc modify

Signed-off-by: Lu, Yintong <[email protected]>

* pruning doc modify

Signed-off-by: Lu, Yintong <[email protected]>

* pruning doc modify

Signed-off-by: Lu, Yintong <[email protected]>

* pruning doc modify

Signed-off-by: Lu, Yintong <[email protected]>

* pruning doc modify

Signed-off-by: Lu, Yintong <[email protected]>

Signed-off-by: Zhang, Weiwei1 <[email protected]>
Signed-off-by: Lu, Yintong <[email protected]>
Co-authored-by: Lu, Yintong <[email protected]>
* Fixed calibration sampling size error
* Update training fit API docstring
* Fixed IPEX examples error

Signed-off-by: Cheng, Penghui <[email protected]>
Signed-off-by: zehao-intel <[email protected]>
* Add recipe for TRT EP

Signed-off-by: Mengni Wang <[email protected]>

* remove codes

Signed-off-by: Mengni Wang <[email protected]>

Signed-off-by: Mengni Wang <[email protected]>
Signed-off-by: changwa1 <[email protected]>
Signed-off-by: changwa1 <[email protected]>
VincyZhang pushed a commit that referenced this pull request Feb 12, 2023
* add subfunc for dense loading

* fuse load dense & tileprod

* load idx from mem

* use subfunc_level

* dump jit binary for each cores too

* adopt unified bsr reordering to amx kernel

* vnni kernel padding outside

* chore

* chore: make gcc happy

* fix extra inst

* smaller local dense offset &  load dense first

* use static member variable as dump index

* keep param1 untouched

* try to keep reg the same

* cpplint
@chensuyue chensuyue deleted the wangchang/example branch June 15, 2023 05:50
yiliu30 pushed a commit that referenced this pull request Apr 30, 2025
* Added PatchedLinearBase

* Fixed PatchedLinear forward_qdq

* Changed quant strategy - scale to fix ci

* Renamed QuantStrategy to QuantWrapper

* Removed instance member from QuantWrapper

* [SW-224403] Added ticket and throwing error when using row_parallel_linear_allreduce_quantization

* Changed QuantWrapper to a simple method that stores scale

* [SW-224538] Added ticket to TODO comment for init_linear

* Pushed requires_grad to the tensor creation

* Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py

* Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py

* Moved copy_scale functions inside PatchedLinearBase

* Update helper_modules.py

* Update helper_modules.py

* Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py

* Update helper_modules.py

* Update helper_modules.py

* Update helper_modules.py copy scale

* Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py
xin3he pushed a commit that referenced this pull request Jul 15, 2025
* Added PatchedLinearBase

* Fixed PatchedLinear forward_qdq

* Changed quant strategy - scale to fix ci

* Renamed QuantStrategy to QuantWrapper

* Removed instance member from QuantWrapper

* [SW-224403] Added ticket and throwing error when using row_parallel_linear_allreduce_quantization

* Changed QuantWrapper to a simple method that stores scale

* [SW-224538] Added ticket to TODO comment for init_linear

* Pushed requires_grad to the tensor creation

* Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py

* Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py

* Moved copy_scale functions inside PatchedLinearBase

* Update helper_modules.py

* Update helper_modules.py

* Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py

* Update helper_modules.py

* Update helper_modules.py

* Update helper_modules.py copy scale

* Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py
XuehaoSun pushed a commit that referenced this pull request Jul 19, 2025
* Added PatchedLinearBase

* Fixed PatchedLinear forward_qdq

* Changed quant strategy - scale to fix ci

* Renamed QuantStrategy to QuantWrapper

* Removed instance member from QuantWrapper

* [SW-224403] Added ticket and throwing error when using row_parallel_linear_allreduce_quantization

* Changed QuantWrapper to a simple method that stores scale

* [SW-224538] Added ticket to TODO comment for init_linear

* Pushed requires_grad to the tensor creation

* Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py

* Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py

* Moved copy_scale functions inside PatchedLinearBase

* Update helper_modules.py

* Update helper_modules.py

* Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py

* Update helper_modules.py

* Update helper_modules.py

* Update helper_modules.py copy scale

* Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Projects

None yet

Development

Successfully merging this pull request may close these issues.