-
Notifications
You must be signed in to change notification settings - Fork 130
Labels
bugSomething isn't workingSomething isn't working
Description
An unexpected argument for snapshot_download when evaluating the model quantized with GPTQmodel.
Same issue as #1196
Traceback (most recent call last):
File "/usr/local/bin/lm_eval", line 8, in <module>
sys.exit(cli_evaluate())
^^^^^^^^^^^^^^
File "/lm_eval/__main__.py", line 459, in cli_evaluate
results = evaluator.simple_evaluate(
^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/lm_eval/utils.py", line 458, in _wrapper
return fn(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^
File "/lm_eval/evaluator.py", line 245, in simple_evaluate
lm = lm_eval.api.registry.get_model(model).create_from_arg_string(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/lm_eval/api/model.py", line 155, in create_from_arg_string
return cls(**args, **args2)
^^^^^^^^^^^^^^^^^^^^
File "/lm_eval/models/huggingface.py", line 219, in __init__
self._create_model(
File "/lm_eval/models/huggingface.py", line 675, in _create_model
self._model = GPTQModel.from_quantized(
^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/dist-packages/gptqmodel/models/auto.py", line 356, in from_quantized
return MODEL_MAP[model_type].from_quantized(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/dist-packages/gptqmodel/models/loader.py", line 316, in from_quantized
model_local_path = get_model_local_path(model_id_or_path, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/dist-packages/gptqmodel/models/loader.py", line 114, in get_model_local_path
return snapshot_download(pretrained_model_id_or_path, **download_kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/dist-packages/huggingface_hub/utils/_validators.py", line 114, in _inner_fn
return fn(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^
TypeError: snapshot_download() got an unexpected keyword argument 'max_memory'
Name: transformers
Version: 4.57.1
Summary: State-of-the-art Machine Learning for JAX, PyTorch and TensorFlow
Home-page: https://github.com/huggingface/transformers
Author: The Hugging Face team (past and future) with the help of all our contributors (https://github.com/huggingface/transformers/graphs/contributors)
Author-email: [transformers@huggingface.co](mailto:transformers@huggingface.co)
License: Apache 2.0 License
Location: /usr/local/lib/python3.12/dist-packages
Requires: filelock, huggingface-hub, numpy, packaging, pyyaml, regex, requests, safetensors, tokenizers, tqdm
Required-by: GPTQModel, lm_eval, peft, sentence-transformers, tokenicer
---
Name: GPTQModel
Version: 5.0.0
Summary: Production ready LLM model compression/quantization toolkit with hw accelerated inference support for both cpu/gpu via HF, vLLM, and SGLang.
Home-page: https://github.com/ModelCloud/GPTQModel
Author:
Author-email: ModelCloud <[qubitium@modelcloud.ai](mailto:qubitium@modelcloud.ai)>
License: Apache License
Metadata
Metadata
Assignees
Labels
bugSomething isn't workingSomething isn't working