Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Finetuning on HF using Seq2SeqTrainer. #2

Closed
varadhbhatnagar opened this issue Jun 10, 2022 · 1 comment
Closed

Finetuning on HF using Seq2SeqTrainer. #2

varadhbhatnagar opened this issue Jun 10, 2022 · 1 comment

Comments

@varadhbhatnagar
Copy link

I need to finetune csebuetnlp/mT5_m2m_crossSum using the Seq2SeqTrainer on HuggingFace. What special symbols need to be inserted during tokenization of input-output pairs?

@abhik1505040
Copy link
Collaborator

Hi, We use language-specific bos tokens only for the decoder. The language-to-bos-token mapping is present in the model config. Please have a look at the code snippet given on the model page to see how you can access it. If you want to readily use our scripts to further fine-tune this model on your own data, just replace google/mt5-base with csebuetnlp/mT5_m2m_crossSum on this line

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants