Skip to content

Conversation

@WenjiaoYue
Copy link
Contributor

Type of Change

Io bug fix

Description

modify file Location, delete useless files

Expected Behavior & Potential Risk

click display normal

How has this PR been tested?

click IO Link

Dependency Change?

no

Signed-off-by: Yue, Wenjiao <[email protected]>
Signed-off-by: Yue, Wenjiao <[email protected]>
Signed-off-by: Yue, Wenjiao <[email protected]>
Signed-off-by: Yue, Wenjiao <[email protected]>
Signed-off-by: Yue, Wenjiao <[email protected]>
Signed-off-by: Yue, Wenjiao <[email protected]>
@chensuyue chensuyue merged commit e02da5c into master Nov 30, 2022
@chensuyue chensuyue deleted the IO-Welcome-bug-fix branch November 30, 2022 05:39
mengniwang95 pushed a commit that referenced this pull request Dec 2, 2022
Signed-off-by: Yue, Wenjiao <[email protected]>
lvliang-intel pushed a commit that referenced this pull request Dec 5, 2022
Signed-off-by: Yue, Wenjiao <[email protected]>
Signed-off-by: Lv, Liang1 <[email protected]>
yiliu30 pushed a commit that referenced this pull request Dec 7, 2022
Signed-off-by: Yue, Wenjiao <[email protected]>
Signed-off-by: yiliu30 <[email protected]>
zehao-intel pushed a commit that referenced this pull request Dec 9, 2022
Signed-off-by: Yue, Wenjiao <[email protected]>
Signed-off-by: zehao-intel <[email protected]>
zehao-intel pushed a commit that referenced this pull request Dec 20, 2022
Signed-off-by: Yue, Wenjiao <[email protected]>
Signed-off-by: zehao-intel <[email protected]>
VincyZhang pushed a commit that referenced this pull request Feb 12, 2023
* support QAT PT to ONNX

* add UT
yiliu30 pushed a commit that referenced this pull request Apr 30, 2025
…188)

* Implementing dynamic quantization design for linear ops

* Using copy_ to store scale as a member, added qdq, removed dyn

* Added PatchedLinearBase to support all linear modules

* Testing dynamic quantization with scale compare

* CR comments - calling cguid

* Added PatchedLinearBase

* Fixed PatchedLinear forward_qdq

* Changed quant strategy - scale to fix ci

* Renamed QuantStrategy to QuantWrapper

* Removed instance member from QuantWrapper

* [SW-224403] Added ticket and throwing error when using row_parallel_linear_allreduce_quantization

* Changed QuantWrapper to a simple method that stores scale

* [SW-224538] Added ticket to TODO comment for init_linear

* Pushed requires_grad to the tensor creation

* Fixed merge

* Fixed load() flow - handling meta tensors with dummy scale

* [SW-224609] removed non tested dynamic qdq

* Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py

* Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py

* Moved copy_scale functions inside PatchedLinearBase

* Added and fixed test cases

* Increased tolerance for new test cases

* Update helper_modules.py

* Update helper_modules.py

* Some tests/ci fixes

* Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py

* Update helper_modules.py

* cr comments + cguid check change

* Update helper_modules.py

* Update helper_modules.py copy scale

* Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py

* Maxabs design and some structure changes

* Merged MaxAbsDynamicPts To base + cguid comments

* changed cguid calls to functions

* Log changes

* Update neural_compressor/torch/algorithms/fp8_quant/model_configs.py

* Update neural_compressor/torch/algorithms/fp8_quant/model_configs.py

* Re-set self.scale_input as before, value is none in dynamic

* Changing back dynamic scale_input to intermediate and not member

* Disabling test_linear_dynamic_quantization: not storing scale as member

* Reintroduce MaxAbsDynamicPts: in dynamic we don't save scale as a member

* weight to hpu comment
xin3he pushed a commit that referenced this pull request Jul 15, 2025
…188)

* Implementing dynamic quantization design for linear ops

* Using copy_ to store scale as a member, added qdq, removed dyn

* Added PatchedLinearBase to support all linear modules

* Testing dynamic quantization with scale compare

* CR comments - calling cguid

* Added PatchedLinearBase

* Fixed PatchedLinear forward_qdq

* Changed quant strategy - scale to fix ci

* Renamed QuantStrategy to QuantWrapper

* Removed instance member from QuantWrapper

* [SW-224403] Added ticket and throwing error when using row_parallel_linear_allreduce_quantization

* Changed QuantWrapper to a simple method that stores scale

* [SW-224538] Added ticket to TODO comment for init_linear

* Pushed requires_grad to the tensor creation

* Fixed merge

* Fixed load() flow - handling meta tensors with dummy scale

* [SW-224609] removed non tested dynamic qdq

* Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py

* Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py

* Moved copy_scale functions inside PatchedLinearBase

* Added and fixed test cases

* Increased tolerance for new test cases

* Update helper_modules.py

* Update helper_modules.py

* Some tests/ci fixes

* Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py

* Update helper_modules.py

* cr comments + cguid check change

* Update helper_modules.py

* Update helper_modules.py copy scale

* Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py

* Maxabs design and some structure changes

* Merged MaxAbsDynamicPts To base + cguid comments

* changed cguid calls to functions

* Log changes

* Update neural_compressor/torch/algorithms/fp8_quant/model_configs.py

* Update neural_compressor/torch/algorithms/fp8_quant/model_configs.py

* Re-set self.scale_input as before, value is none in dynamic

* Changing back dynamic scale_input to intermediate and not member

* Disabling test_linear_dynamic_quantization: not storing scale as member

* Reintroduce MaxAbsDynamicPts: in dynamic we don't save scale as a member

* weight to hpu comment
XuehaoSun pushed a commit that referenced this pull request Jul 19, 2025
…188)

* Implementing dynamic quantization design for linear ops

* Using copy_ to store scale as a member, added qdq, removed dyn

* Added PatchedLinearBase to support all linear modules

* Testing dynamic quantization with scale compare

* CR comments - calling cguid

* Added PatchedLinearBase

* Fixed PatchedLinear forward_qdq

* Changed quant strategy - scale to fix ci

* Renamed QuantStrategy to QuantWrapper

* Removed instance member from QuantWrapper

* [SW-224403] Added ticket and throwing error when using row_parallel_linear_allreduce_quantization

* Changed QuantWrapper to a simple method that stores scale

* [SW-224538] Added ticket to TODO comment for init_linear

* Pushed requires_grad to the tensor creation

* Fixed merge

* Fixed load() flow - handling meta tensors with dummy scale

* [SW-224609] removed non tested dynamic qdq

* Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py

* Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py

* Moved copy_scale functions inside PatchedLinearBase

* Added and fixed test cases

* Increased tolerance for new test cases

* Update helper_modules.py

* Update helper_modules.py

* Some tests/ci fixes

* Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py

* Update helper_modules.py

* cr comments + cguid check change

* Update helper_modules.py

* Update helper_modules.py copy scale

* Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py

* Maxabs design and some structure changes

* Merged MaxAbsDynamicPts To base + cguid comments

* changed cguid calls to functions

* Log changes

* Update neural_compressor/torch/algorithms/fp8_quant/model_configs.py

* Update neural_compressor/torch/algorithms/fp8_quant/model_configs.py

* Re-set self.scale_input as before, value is none in dynamic

* Changing back dynamic scale_input to intermediate and not member

* Disabling test_linear_dynamic_quantization: not storing scale as member

* Reintroduce MaxAbsDynamicPts: in dynamic we don't save scale as a member

* weight to hpu comment
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants