User Access Tokens are the preferred way to authenticate an application or notebook to Hugging Face services. You can manage your access tokens in your settings.
Access tokens allow applications and notebooks to perform specific actions specified by the scope of the roles shown in the following:
-
fine-grained
: tokens with this role can be used to provide fine-grained access to specific resources, such as a specific model or models in a specific organization. This type of token is useful in production environments, as you can use your own token without sharing access to all your resources. -
read
: tokens with this role can only be used to provide read access to repositories you could read. That includes public and private repositories that you, or an organization you're a member of, own. Use this role if you only need to read content from the Hugging Face Hub (e.g. when downloading private models or doing inference). -
write
: tokens with this role additionally grant write access to the repositories you have write access to. Use this token if you need to create or push content to a repository (e.g., when training a model or modifying a model card).
Note that Organization API Tokens have been deprecated:
If you are a member of an organization with read/write/admin role, then your User Access Tokens will be able to read/write the resources according to the token permission (read/write) and organization membership (read/write/admin).
To create an access token, go to your settings, then click on the Access Tokens tab. Click on the New token button to create a new User Access Token.
Select a role and a name for your token and voilà - you're ready to go!
You can delete and refresh User Access Tokens by clicking on the Manage button.
There are plenty of ways to use a User Access Token to access the Hugging Face Hub, granting you the flexibility you need to build awesome apps on top of it.
User Access Tokens can be:
- used in place of a password to access the Hugging Face Hub with git or with basic authentication.
- passed as a bearer token when calling the Inference API.
- used in the Hugging Face Python libraries, such as
transformers
ordatasets
:
from transformers import AutoModel
access_token = "hf_..."
model = AutoModel.from_pretrained("private/model", token=access_token)
We recommend you create one access token per app or usage. For instance, you could have a separate token for:
- A local machine.
- A Colab notebook.
- An awesome custom inference server.
This way, you can invalidate one token without impacting your other usages.
We also recommend only using fine-grained tokens for production usage. The impact, if leaked, will be reduced, and they can be shared among your organization without impacting your account.
For example, if your production application needs read access to a gated model, a member of your organization can request access to the model and then create a fine-grained token with read access to that model. This token can then be used in your production application without giving it access to all your private models.