Skip to content

GPTQ quantization not working #12

@lopuhin

Description

@lopuhin

Running quantize.py with --mode int4-gptq does not seem to work:

  • code tries to import lm-evaluation-harness which is not included/documented/used
  • import in eval.py is incorrect, should probably be from model import Transformer as LLaMA instead of from model import LLaMA
  • after fixing two above issues, next one is a circular import
  • after fixing that, import lm_eval should be replaced with import lm_eval.base
  • there is one other circular import
  • there are a few other missing imports from lm_eval
  • and a few other errors

Overall here are the fixes I had to apply to make it run: lopuhin@86d990b

Based on this, could you please check if the right version of the code was included for GPTQ quantization?

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions