Skip to content
This repository was archived by the owner on Oct 9, 2024. It is now read-only.
This repository was archived by the owner on Oct 9, 2024. It is now read-only.

transformers 4.21.3 doesn't support load_in_8bit #3

@SadeghRahnamoon

Description

@SadeghRahnamoon

Using transformers version 4.21.3, I see this error:

Using `load_in_8bit=True` to use quanitized model
Traceback (most recent call last):
 File "bloom-accelerate-inference.py", line 115, in <module>
   model = AutoModelForCausalLM.from_pretrained(model_name, **kwargs)
 File "/disk1/srahnamoon/llm/.lvenv/lib/python3.8/site-packages/transformers/models/auto/auto_factory.py", line 446, in from_pretrained
   return model_class.from_pretrained(pretrained_model_name_or_path, *model_args, config=config, **kwargs)
 File "/disk1/srahnamoon/llm/.lvenv/lib/python3.8/site-packages/transformers/modeling_utils.py", line 2113, in from_pretrained
   model = cls(config, *model_args, **model_kwargs)
TypeError: __init__() got an unexpected keyword argument 'load_in_8bit'

upgrading to 4.22.0 resolved the issue. Maybe the documentation for HF Accelerate solution needs to be corrected?

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions