Skip to content

Conversation

@xin3he
Copy link
Contributor

@xin3he xin3he commented Nov 30, 2023

Type of Change

bug fix

Description

Optimum sets bias=True for QuantLinear when packing model. Here we follow this design of huggingface format for compatibility and set use_hf_format=True as default.
change argument name from hf to optimum.

Expected Behavior & Potential Risk

UT pass

@chensuyue chensuyue added this to the v2.4 milestone Dec 1, 2023
@chensuyue chensuyue added the bug fix Something isn't working label Dec 1, 2023
@hshen14
Copy link
Contributor

hshen14 commented Dec 3, 2023

shall we specify the format, e.g., use_gptq_format? HF format sounds too general - how about AWQ and GGUF format? People also upload these formats to HF.

@xin3he
Copy link
Contributor Author

xin3he commented Dec 4, 2023

shall we specify the format, e.g., use_gptq_format? HF format sounds too general - how about AWQ and GGUF format? People also upload these formats to HF.

It's actually general, we can generate RTN, AWQ model with this format. GGUF is another format, we haven't supported it now.

Signed-off-by: Xin He <[email protected]>
Signed-off-by: Xin He <[email protected]>
Signed-off-by: Xin He <[email protected]>
@xin3he xin3he changed the title change use_hf_format=True and add bias change use_optimum_format=True and add bias Dec 5, 2023
@chensuyue
Copy link
Contributor

/azp run Code-Scan

@azure-pipelines
Copy link

Azure Pipelines successfully started running 1 pipeline(s).

@chensuyue chensuyue merged commit 0a06448 into master Dec 6, 2023
@chensuyue chensuyue deleted the xinhe/hf_format branch December 6, 2023 12:08
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

bug fix Something isn't working

Projects

None yet

Development

Successfully merging this pull request may close these issues.

5 participants