We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent dec0bcb commit 35efed5Copy full SHA for 35efed5
examples/pytorch/nlp/huggingface_models/language-modeling/quantization/ptq_static/fx/run_clm.py
@@ -555,9 +555,10 @@ def eval_func_for_nc(model_tuned):
555
return eval_loss
556
557
if model_args.tune:
558
- from neural_compressor.config import PostTrainingQuantConfig
+ from neural_compressor.config import AccuracyCriterion, PostTrainingQuantConfig
559
from neural_compressor import quantization
560
- conf = PostTrainingQuantConfig()
+ accuracy_criterion = AccuracyCriterion(higher_is_better=False, tolerable_loss=0.5)
561
+ conf = PostTrainingQuantConfig(accuracy_criterion=accuracy_criterion)
562
q_model = quantization.fit(model,
563
conf,
564
calib_dataloader=trainer.get_eval_dataloader(),
0 commit comments