Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We鈥檒l occasionally send you account related emails.

Already on GitHub? Sign in to your account

Run_squad script does not parse adapter args correctly and does not save adapters #49

Closed
2 of 4 tasks
xplip opened this issue Aug 22, 2020 · 0 comments 路 Fixed by #50
Closed
2 of 4 tasks

Run_squad script does not parse adapter args correctly and does not save adapters #49

xplip opened this issue Aug 22, 2020 · 0 comments 路 Fixed by #50
Labels
bug Something isn't working

Comments

@xplip
Copy link
Collaborator

xplip commented Aug 22, 2020

馃悰 Bug

Information

Model I am using (Bert, XLNet ...):
mBERT

Language I am using the model on (English, Chinese ...):
Arabic, but the issue is language/dataset independent

Adapter setup I am using (if any):
Arabic lang adapter from adapterhub, new squad task adapter

The problem arises when using:

  • the official example scripts: (give details below)
  • my own modified scripts: (give details below)

The tasks I am working on is:

  • an official GLUE/SQUaD task:
  • my own task or dataset: (give details below)
    Arabic Reading Comprehension Dataset (ARCD) but should be the same for any dataset

To reproduce

Steps to reproduce the behavior:

Issue of args not being parsed correctly:

  1. Try to define language adapter by providing --load_lang_adapter [adapter name]
  2. Script fails as argument parser expects it to be --load_language_adapter
  3. Try to define language adapter by providing --load_language_adapter instead
  4. Script fails as setup_task_adapter_training function of adapter_training.py expects it to be --load_lang_adapter

Issue of adapters not being saved:

  1. Run script to finetune adapters
  2. Script stores full model for each checkpoint and at the end of training

Expected behavior

I should be able to define the language adapter with the --load_lang_adapter flag and its config with the --lang_adapter_config flag. When using adapters to finetune my model, I would usually like to store the adapters, not the full model.

Environment info

  • transformers version: 2.11.0
  • Platform: macOS-10.15.6-x86_64-i386-64bit
  • Python version: 3.8.5
  • PyTorch version (GPU?): 1.5.1 (False)
  • Tensorflow version (GPU?): 2.3.0 (False)
  • Using GPU in script?: False
  • Using distributed or parallel set-up in script?: False
@xplip xplip added the bug Something isn't working label Aug 22, 2020
@calpt calpt closed this as completed in #50 Aug 25, 2020
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

Successfully merging a pull request may close this issue.

1 participant