-
Notifications
You must be signed in to change notification settings - Fork 281
Update PyTorch examples quantization API #192
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Closed
Closed
Conversation
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Signed-off-by: changwa1 <[email protected]>
PenghuiCheng
approved these changes
Nov 30, 2022
chensuyue
reviewed
Nov 30, 2022
| "batch_size": 16384, | ||
| "new_benchmark": false | ||
| "batch_size": 16384 | ||
| }, |
Contributor
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
require new parameter: “main_script”: “xxx.py”,
Contributor
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Because test code need to replace some config class, so need to know the name scripts.
xin3he
approved these changes
Nov 30, 2022
Contributor
|
) Signed-off-by: Cheng, Penghui <[email protected]>
Signed-off-by: Lv, Liang1 <[email protected]>
Signed-off-by: changwa1 <[email protected]>
* Fixed UT error for bf16 op list for QAT mode Signed-off-by: Cheng, Penghui <[email protected]>
Signed-off-by: Lv, Liang1 <[email protected]>
Signed-off-by: intel-zhangyi <[email protected]>
Signed-off-by: Wang, Chang1 <[email protected]>
Signed-off-by: chensuyue <[email protected]>
Signed-off-by: Xinyu Ye <[email protected]>
…166) Signed-off-by: Xinyu Ye <[email protected]>
Signed-off-by: changwa1 <[email protected]>
Signed-off-by: Lv, Liang1 <[email protected]> Signed-off-by: Lv, Liang1 <[email protected]>
* Fix centernet_hg104 tuning issue Signed-off-by: sys-lpot-val <[email protected]> * Fix TextRNN tuning issue Signed-off-by: sys-lpot-val <[email protected]> Signed-off-by: sys-lpot-val <[email protected]> Signed-off-by: Lv, Liang1 <[email protected]> Co-authored-by: sys-lpot-val <[email protected]>
…ers (#214) * Create intel_extension_for_transformers.yaml * change default strategy to dynamic according to huggingface sync * change default strategy to dynamic according to HF sync * enable intel extension for transformers * Create change_trainer_to_nlptrainer.py * add use_inc for not using default optimum for HF code * add use_inc * update optimum quant static dynamic separation * Update interface.py * Update interface.py * Update autoinc_harness.py * Update README.md * add change_trainer_to_nlptrainer to outside_harness * add PythonLauncher to pass spelling check CI Signed-off-by: Yue, Wenjiao <[email protected]> Signed-off-by: Yue, Wenjiao <[email protected]> Co-authored-by: Yue, Wenjiao <[email protected]>
* update launcher to fit multi-item input * Update __main__.py
…reliminary) (#217) * Create tf_inc_static_quant.yaml * Create inc.py * Delete tf_inc_static_quant.yaml * Update interface.py * Update inc.py
Signed-off-by: Sun, Xuehao <[email protected]>
Signed-off-by: zehao-intel <[email protected]>
Signed-off-by: chensuyue <[email protected]>
Signed-off-by: yiliu30 <[email protected]> Co-authored-by: lvliang-intel <[email protected]> Co-authored-by: chen, suyue <[email protected]> Co-authored-by: xinhe <[email protected]> Co-authored-by: Ray <[email protected]>
Signed-off-by: Cheng, Penghui <[email protected]>
Signed-off-by: Lv, Liang1 <[email protected]> Signed-off-by: chensuyue <[email protected]> Co-authored-by: Lv, Liang1 <[email protected]> Co-authored-by: chensuyue <[email protected]>
Signed-off-by: Xin He <[email protected]> Co-authored-by: yiliu30 <[email protected]>
Signed-off-by: changwa1 <[email protected]>
Signed-off-by: chensuyue <[email protected]>
* enable auto bench in launcher * debug * debug
* add pruning v2 Signed-off-by: wenhuach21 <[email protected]> * DATASETS->Datasets Signed-off-by: wenhuach21 <[email protected]> * pruner README v2 Signed-off-by: Lu, Yintong <[email protected]> * prune README v2 Signed-off-by: Lu, Yintong <[email protected]> * prune README v2 Signed-off-by: Lu, Yintong <[email protected]> * prune README v2 Signed-off-by: Lu, Yintong <[email protected]> * prune README v2 Signed-off-by: Lu, Yintong <[email protected]> * recover code in experimental Signed-off-by: wenhuach21 <[email protected]> * recover config Signed-off-by: wenhuach21 <[email protected]> Signed-off-by: wenhuach21 <[email protected]> Signed-off-by: Lu, Yintong <[email protected]> Co-authored-by: Lu, Yintong <[email protected]>
…arch Toolkit (#197) Signed-off-by: Maciej Szankin <[email protected]> Co-authored-by: Nittur Sridhar, Sharath <[email protected]> Co-authored-by: Xinyu Ye <[email protected]>
Signed-off-by: mengniwa <[email protected]>
Signed-off-by: Lv, Liang1 <[email protected]> Co-authored-by: chen, suyue <[email protected]>
* update qa example Signed-off-by: Zhang, Weiwei1 <[email protected]> * pruning doc modify Signed-off-by: Lu, Yintong <[email protected]> * pruning doc modify Signed-off-by: Lu, Yintong <[email protected]> * pruning doc modify Signed-off-by: Lu, Yintong <[email protected]> * pruning doc modify Signed-off-by: Lu, Yintong <[email protected]> * pruning doc modify Signed-off-by: Lu, Yintong <[email protected]> Signed-off-by: Zhang, Weiwei1 <[email protected]> Signed-off-by: Lu, Yintong <[email protected]> Co-authored-by: Lu, Yintong <[email protected]>
* Fixed calibration sampling size error * Update training fit API docstring * Fixed IPEX examples error Signed-off-by: Cheng, Penghui <[email protected]>
Signed-off-by: Cheng, Penghui <[email protected]>
Signed-off-by: zehao-intel <[email protected]>
* Add recipe for TRT EP Signed-off-by: Mengni Wang <[email protected]> * remove codes Signed-off-by: Mengni Wang <[email protected]> Signed-off-by: Mengni Wang <[email protected]>
Signed-off-by: zehao-intel <[email protected]>
Signed-off-by: changwa1 <[email protected]>
Signed-off-by: changwa1 <[email protected]>
VincyZhang
pushed a commit
that referenced
this pull request
Feb 12, 2023
* add subfunc for dense loading * fuse load dense & tileprod * load idx from mem * use subfunc_level * dump jit binary for each cores too * adopt unified bsr reordering to amx kernel * vnni kernel padding outside * chore * chore: make gcc happy * fix extra inst * smaller local dense offset & load dense first * use static member variable as dump index * keep param1 untouched * try to keep reg the same * cpplint
yiliu30
pushed a commit
that referenced
this pull request
Apr 30, 2025
* Added PatchedLinearBase * Fixed PatchedLinear forward_qdq * Changed quant strategy - scale to fix ci * Renamed QuantStrategy to QuantWrapper * Removed instance member from QuantWrapper * [SW-224403] Added ticket and throwing error when using row_parallel_linear_allreduce_quantization * Changed QuantWrapper to a simple method that stores scale * [SW-224538] Added ticket to TODO comment for init_linear * Pushed requires_grad to the tensor creation * Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py * Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py * Moved copy_scale functions inside PatchedLinearBase * Update helper_modules.py * Update helper_modules.py * Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py * Update helper_modules.py * Update helper_modules.py * Update helper_modules.py copy scale * Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py
xin3he
pushed a commit
that referenced
this pull request
Jul 15, 2025
* Added PatchedLinearBase * Fixed PatchedLinear forward_qdq * Changed quant strategy - scale to fix ci * Renamed QuantStrategy to QuantWrapper * Removed instance member from QuantWrapper * [SW-224403] Added ticket and throwing error when using row_parallel_linear_allreduce_quantization * Changed QuantWrapper to a simple method that stores scale * [SW-224538] Added ticket to TODO comment for init_linear * Pushed requires_grad to the tensor creation * Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py * Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py * Moved copy_scale functions inside PatchedLinearBase * Update helper_modules.py * Update helper_modules.py * Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py * Update helper_modules.py * Update helper_modules.py * Update helper_modules.py copy scale * Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py
XuehaoSun
pushed a commit
that referenced
this pull request
Jul 19, 2025
* Added PatchedLinearBase * Fixed PatchedLinear forward_qdq * Changed quant strategy - scale to fix ci * Renamed QuantStrategy to QuantWrapper * Removed instance member from QuantWrapper * [SW-224403] Added ticket and throwing error when using row_parallel_linear_allreduce_quantization * Changed QuantWrapper to a simple method that stores scale * [SW-224538] Added ticket to TODO comment for init_linear * Pushed requires_grad to the tensor creation * Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py * Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py * Moved copy_scale functions inside PatchedLinearBase * Update helper_modules.py * Update helper_modules.py * Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py * Update helper_modules.py * Update helper_modules.py * Update helper_modules.py copy scale * Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Signed-off-by: changwa1 [email protected]
Type of Change
API changed
Description
enable examples for recommendation, speech recognition
Expected Behavior & Potential Risk
the expected behavior that triggered by this PR
How has this PR been tested?
how to reproduce the test (including hardware information)
Dependency Change?
any library dependency introduced or removed