diff --git a/src/transformers/modeling_utils.py b/src/transformers/modeling_utils.py index a1a0ad7d36fd..898fd0508997 100644 --- a/src/transformers/modeling_utils.py +++ b/src/transformers/modeling_utils.py @@ -1487,7 +1487,7 @@ def from_pretrained(cls, pretrained_model_name_or_path: Optional[Union[str, os.P Please refer to the mirror site for more information. _fast_init(`bool`, *optional*, defaults to `True`): Whether or not to disable fast initialization. - low_cpu_mem_usage(`bool``, *optional*, defaults to `False`): + low_cpu_mem_usage(`bool`, *optional*, defaults to `False`): Tries to not use more than 1x model size in CPU memory (including peak memory) while loading the model. This is an experimental feature and a subject to change at any moment. torch_dtype (`str` or `torch.dtype`, *optional*):