Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Unrecognized configuration class <class 'transformers.models.distilbert.configuration_distilbert.DistilBertConfig'> for this kind of AutoModel: AutoModelForSeq2SeqLM. #13590

Closed
4 tasks
Azitt opened this issue Sep 16, 2021 · 1 comment

Comments

@Azitt
Copy link

Azitt commented Sep 16, 2021

I'm trying to use translate notebook with model_checkpoint = "distilbert-base-uncased" but it gives me below error:

image

any idea what is wrong?

Environment info

  • transformers version: distilbert
  • Platform:
  • Python version:
  • PyTorch version (GPU?):
  • Tensorflow version (GPU?):
  • Using GPU in script?:
  • Using distributed or parallel set-up in script?:

Who can help

Information

Model I am using (Bert, XLNet ...):

The problem arises when using:

  • the official example scripts: (give details below)
  • my own modified scripts: (give details below)

The tasks I am working on is:

  • an official GLUE/SQUaD task: (give the name)
  • my own task or dataset: (give details below)

To reproduce

Steps to reproduce the behavior:

Expected behavior

@LysandreJik
Copy link
Member

Hello! DistilBERT is not a seq2seq model, hence why there's no seq2seq architecture for that model!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants