New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
bash script of fine-tuning on multinews dataset on multiple gpus using ddp #15
Comments
Hi there, The error seems to be an out-of-index error, you can double check the length limit for the tokenizer and the model, and if you truncate the inputs to the given length limit. Sorry that I do not have the bash file for fine-tuning now, but you can refer to the code for few-shot finetuning in |
hello, i meet the same issue with total same error information ,have you sloved this? |
@Caesar-666 I solved the issue by changing Line 59 in daf9f42
It seems that the problem comes from the different versions of datasets . I am using datasets==2.3.2 where the last document should not be removed by [:-1] . Otherwise there would be only one document in some cases and the sequence length will become 4097>4096 , which causes the indexing error.
|
thank you for your effort |
Hi,
I wonder if there is a script to fine-tune the pre-trained PRIMERA model on multiple GPUs using distributed data parallel (From the
run_bash
I can only find test scripts). I tried using the following command:but it prompts out errors of ddp as follows:
Are there any insights on this error? And also could you provide your bash scripts for fine-tuning the model on multi-news? Thanks much!
The text was updated successfully, but these errors were encountered: