You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I get the following error in main.py on a four GPU setup when attempting to fine-tune a BERT model:
With batch size 1: RuntimeError: split_with_sizes expects split_sizes to sum exactly to 800 (input tensor's size at dimension 0), but got split_sizes=[16]
With batch size 4: RuntimeError: split_with_sizes expects split_sizes to sum exactly to 2750 (input tensor's size at dimension 0), but got split_sizes=[25, 6, 8, 16]
I would expect number_of_chunks to be of variable size for each record in the batch, but no matter my batch size, I seem to get an error at preds_split = preds.split(number_of_chunks) in main.py.
Any idea what I might be missing?
The text was updated successfully, but these errors were encountered:
@MichalBrzozowski91, thanks for this project! Really great stuff.
I get the following error in
main.py
on a four GPU setup when attempting to fine-tune a BERT model:With batch size 1:
RuntimeError: split_with_sizes expects split_sizes to sum exactly to 800 (input tensor's size at dimension 0), but got split_sizes=[16]
With batch size 4:
RuntimeError: split_with_sizes expects split_sizes to sum exactly to 2750 (input tensor's size at dimension 0), but got split_sizes=[25, 6, 8, 16]
I would expect
number_of_chunks
to be of variable size for each record in the batch, but no matter my batch size, I seem to get an error atpreds_split = preds.split(number_of_chunks)
inmain.py
.Any idea what I might be missing?
The text was updated successfully, but these errors were encountered: