Skip to content

Commit

Permalink
Merge pull request #355 from PanQiWei/fix_pack_model_use_exllamav2
Browse files Browse the repository at this point in the history
import exllama QuantLinear instead of exllamav2's in `pack_model`
  • Loading branch information
PanQiWei committed Sep 27, 2023
2 parents 3b81fb5 + c1a3013 commit 51c043c
Showing 1 changed file with 2 additions and 2 deletions.
4 changes: 2 additions & 2 deletions auto_gptq/modeling/_utils.py
Original file line number Diff line number Diff line change
Expand Up @@ -188,15 +188,15 @@ def pack_model(
warmup_triton: bool = False,
force_layer_back_to_cpu: bool = False
):
QuantLinear = dynamically_import_QuantLinear(use_triton=use_triton, desc_act=desc_act, group_size=group_size, bits=bits)
QuantLinear = dynamically_import_QuantLinear(use_triton=use_triton, desc_act=desc_act, group_size=group_size, bits=bits, disable_exllama=False, disable_exllamav2=True)

if force_layer_back_to_cpu:
model.to(CPU)

logger.info('Packing model...')
layers = find_layers(model)
layers = {n: layers[n] for n in quantizers}
make_quant(model, quantizers, bits, group_size, use_triton=use_triton, use_cuda_fp16=use_cuda_fp16, desc_act=desc_act)
make_quant(model, quantizers, bits, group_size, use_triton=use_triton, use_cuda_fp16=use_cuda_fp16, desc_act=desc_act, disable_exllama=False, disable_exllamav2=True)
qlayers = find_layers(model, [QuantLinear])
for name in qlayers:
logger.info(name)
Expand Down

0 comments on commit 51c043c

Please sign in to comment.