-
Couldn't load subscription status.
- Fork 565
Open
Description
Running quantize.py with --mode int4-gptq does not seem to work:
- code tries to import
lm-evaluation-harnesswhich is not included/documented/used - import in
eval.pyis incorrect, should probably befrom model import Transformer as LLaMAinstead offrom model import LLaMA - after fixing two above issues, next one is a circular import
- after fixing that,
import lm_evalshould be replaced withimport lm_eval.base - there is one other circular import
- there are a few other missing imports from lm_eval
- and a few other errors
Overall here are the fixes I had to apply to make it run: lopuhin@86d990b
Based on this, could you please check if the right version of the code was included for GPTQ quantization?
Metadata
Metadata
Assignees
Labels
No labels