-
Notifications
You must be signed in to change notification settings - Fork 282
Io welcome bug fix #188
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Merged
Merged
Io welcome bug fix #188
Conversation
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Signed-off-by: Yue, Wenjiao <[email protected]>
Signed-off-by: Yue, Wenjiao <[email protected]>
Signed-off-by: Yue, Wenjiao <[email protected]>
Signed-off-by: Yue, Wenjiao <[email protected]>
Signed-off-by: Yue, Wenjiao <[email protected]>
Signed-off-by: Yue, Wenjiao <[email protected]>
mengniwang95
pushed a commit
that referenced
this pull request
Dec 2, 2022
Signed-off-by: Yue, Wenjiao <[email protected]>
lvliang-intel
pushed a commit
that referenced
this pull request
Dec 5, 2022
Signed-off-by: Yue, Wenjiao <[email protected]> Signed-off-by: Lv, Liang1 <[email protected]>
yiliu30
pushed a commit
that referenced
this pull request
Dec 7, 2022
Signed-off-by: Yue, Wenjiao <[email protected]> Signed-off-by: yiliu30 <[email protected]>
zehao-intel
pushed a commit
that referenced
this pull request
Dec 9, 2022
Signed-off-by: Yue, Wenjiao <[email protected]> Signed-off-by: zehao-intel <[email protected]>
zehao-intel
pushed a commit
that referenced
this pull request
Dec 20, 2022
Signed-off-by: Yue, Wenjiao <[email protected]> Signed-off-by: zehao-intel <[email protected]>
VincyZhang
pushed a commit
that referenced
this pull request
Feb 12, 2023
* support QAT PT to ONNX * add UT
yiliu30
pushed a commit
that referenced
this pull request
Apr 30, 2025
…188) * Implementing dynamic quantization design for linear ops * Using copy_ to store scale as a member, added qdq, removed dyn * Added PatchedLinearBase to support all linear modules * Testing dynamic quantization with scale compare * CR comments - calling cguid * Added PatchedLinearBase * Fixed PatchedLinear forward_qdq * Changed quant strategy - scale to fix ci * Renamed QuantStrategy to QuantWrapper * Removed instance member from QuantWrapper * [SW-224403] Added ticket and throwing error when using row_parallel_linear_allreduce_quantization * Changed QuantWrapper to a simple method that stores scale * [SW-224538] Added ticket to TODO comment for init_linear * Pushed requires_grad to the tensor creation * Fixed merge * Fixed load() flow - handling meta tensors with dummy scale * [SW-224609] removed non tested dynamic qdq * Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py * Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py * Moved copy_scale functions inside PatchedLinearBase * Added and fixed test cases * Increased tolerance for new test cases * Update helper_modules.py * Update helper_modules.py * Some tests/ci fixes * Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py * Update helper_modules.py * cr comments + cguid check change * Update helper_modules.py * Update helper_modules.py copy scale * Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py * Maxabs design and some structure changes * Merged MaxAbsDynamicPts To base + cguid comments * changed cguid calls to functions * Log changes * Update neural_compressor/torch/algorithms/fp8_quant/model_configs.py * Update neural_compressor/torch/algorithms/fp8_quant/model_configs.py * Re-set self.scale_input as before, value is none in dynamic * Changing back dynamic scale_input to intermediate and not member * Disabling test_linear_dynamic_quantization: not storing scale as member * Reintroduce MaxAbsDynamicPts: in dynamic we don't save scale as a member * weight to hpu comment
xin3he
pushed a commit
that referenced
this pull request
Jul 15, 2025
…188) * Implementing dynamic quantization design for linear ops * Using copy_ to store scale as a member, added qdq, removed dyn * Added PatchedLinearBase to support all linear modules * Testing dynamic quantization with scale compare * CR comments - calling cguid * Added PatchedLinearBase * Fixed PatchedLinear forward_qdq * Changed quant strategy - scale to fix ci * Renamed QuantStrategy to QuantWrapper * Removed instance member from QuantWrapper * [SW-224403] Added ticket and throwing error when using row_parallel_linear_allreduce_quantization * Changed QuantWrapper to a simple method that stores scale * [SW-224538] Added ticket to TODO comment for init_linear * Pushed requires_grad to the tensor creation * Fixed merge * Fixed load() flow - handling meta tensors with dummy scale * [SW-224609] removed non tested dynamic qdq * Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py * Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py * Moved copy_scale functions inside PatchedLinearBase * Added and fixed test cases * Increased tolerance for new test cases * Update helper_modules.py * Update helper_modules.py * Some tests/ci fixes * Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py * Update helper_modules.py * cr comments + cguid check change * Update helper_modules.py * Update helper_modules.py copy scale * Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py * Maxabs design and some structure changes * Merged MaxAbsDynamicPts To base + cguid comments * changed cguid calls to functions * Log changes * Update neural_compressor/torch/algorithms/fp8_quant/model_configs.py * Update neural_compressor/torch/algorithms/fp8_quant/model_configs.py * Re-set self.scale_input as before, value is none in dynamic * Changing back dynamic scale_input to intermediate and not member * Disabling test_linear_dynamic_quantization: not storing scale as member * Reintroduce MaxAbsDynamicPts: in dynamic we don't save scale as a member * weight to hpu comment
XuehaoSun
pushed a commit
that referenced
this pull request
Jul 19, 2025
…188) * Implementing dynamic quantization design for linear ops * Using copy_ to store scale as a member, added qdq, removed dyn * Added PatchedLinearBase to support all linear modules * Testing dynamic quantization with scale compare * CR comments - calling cguid * Added PatchedLinearBase * Fixed PatchedLinear forward_qdq * Changed quant strategy - scale to fix ci * Renamed QuantStrategy to QuantWrapper * Removed instance member from QuantWrapper * [SW-224403] Added ticket and throwing error when using row_parallel_linear_allreduce_quantization * Changed QuantWrapper to a simple method that stores scale * [SW-224538] Added ticket to TODO comment for init_linear * Pushed requires_grad to the tensor creation * Fixed merge * Fixed load() flow - handling meta tensors with dummy scale * [SW-224609] removed non tested dynamic qdq * Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py * Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py * Moved copy_scale functions inside PatchedLinearBase * Added and fixed test cases * Increased tolerance for new test cases * Update helper_modules.py * Update helper_modules.py * Some tests/ci fixes * Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py * Update helper_modules.py * cr comments + cguid check change * Update helper_modules.py * Update helper_modules.py copy scale * Update neural_compressor/torch/algorithms/fp8_quant/_quant_common/helper_modules.py * Maxabs design and some structure changes * Merged MaxAbsDynamicPts To base + cguid comments * changed cguid calls to functions * Log changes * Update neural_compressor/torch/algorithms/fp8_quant/model_configs.py * Update neural_compressor/torch/algorithms/fp8_quant/model_configs.py * Re-set self.scale_input as before, value is none in dynamic * Changing back dynamic scale_input to intermediate and not member * Disabling test_linear_dynamic_quantization: not storing scale as member * Reintroduce MaxAbsDynamicPts: in dynamic we don't save scale as a member * weight to hpu comment
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Type of Change
Io bug fix
Description
modify file Location, delete useless files
Expected Behavior & Potential Risk
click display normal
How has this PR been tested?
click IO Link
Dependency Change?
no