You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
If you have a large dataset and are using BERT fine-tuning, the performance difference will probably be marginal. If your data size is small, however, it could be worth a try. I have not done a formal BERT + EDA experiment.
I read the paper. EDA written by Jason Wei.
In this paper, it is said that the results in Bert won't go up much.
Will applying EDA to BERT produce better results in natural language processing?
Will Random Swap (RS) and Random Deletion (RD) work well for BERT, as BERT is besed on contextual pre-training.
Thank you very much. @jasonwei20
The text was updated successfully, but these errors were encountered: