File tree Expand file tree Collapse file tree 2 files changed +2
-2
lines changed
examples/3.x_api/pytorch/nlp/huggingface_models/language-modeling/quantization/weight_only Expand file tree Collapse file tree 2 files changed +2
-2
lines changed Original file line number Diff line number Diff line change 1010 include :
1111 - neural_compressor/common
1212 - neural_compressor/torch
13- - examples/3.x_api/pytorch/nlp/huggingface_models/language-modeling/quantization/llm
13+ - examples/3.x_api/pytorch/nlp/huggingface_models/language-modeling/quantization/weight_only
1414 - setup.py
1515 - requirements_pt.txt
1616 - .azure-pipelines/scripts/models
Original file line number Diff line number Diff line change @@ -50,7 +50,7 @@ function run_tuning {
5050 extra_cmd=$extra_cmd " --double_quant_type BNB_NF4"
5151 elif [ " ${topology} " = " opt_125m_woq_gptq_int4_dq_ggml" ]; then
5252 model_name_or_path=" facebook/opt-125m"
53- extra_cmd=$extra_cmd " --woq_algo GPTQ --woq_bits 4 --woq_group_size 128 --woq_scheme asym --woq_use_mse_search --gptq_use_max_length --gptq_percdamp 0.1 --gptq_actorder"
53+ extra_cmd=$extra_cmd " --woq_algo GPTQ --woq_bits 4 --woq_group_size 128 --woq_scheme asym --woq_use_mse_search --gptq_use_max_length --gptq_percdamp 0.8 --gptq_actorder"
5454 extra_cmd=$extra_cmd " --double_quant_type GGML_TYPE_Q4_K"
5555 elif [ " ${topology} " = " llama2_7b_gptq_int4" ]; then
5656 model_name_or_path=" meta-llama/Llama-2-7b-hf"
You can’t perform that action at this time.
0 commit comments