You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I run the code using bert-base on the dataset Conll04, and got F1-scores approximately 66. I find the f1 is much lower than using albert-large. I wonder whether the comparison between this model using albert-large and the previous work using bert-base is really reasonable?
The text was updated successfully, but these errors were encountered:
Table-sequence uses Albert-xxlarge, and we want to make the experiment setting the same as the previous sota.
Also, as I mention, model is delicate in this dataset, you need to carefully tune the hyper parameters even if the only change you make is the embedding.
3-4 point difference between Bert and Albert should be reasonable, I suggest that you tune the lr, batch size and clip
I run the code using bert-base on the dataset Conll04, and got F1-scores approximately 66. I find the f1 is much lower than using albert-large. I wonder whether the comparison between this model using albert-large and the previous work using bert-base is really reasonable?
The text was updated successfully, but these errors were encountered: