You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Thanks for the excellent paper. But I have a question for the experiment setting:
As a sequence labeling task, we always use transformer encoder(like BERT) to solve the NER problem, for example, the author in Doc2EDAG use BERT as the first step backbone. However, in the paper, it is said that the vanilla transformer(encoder-decoder structure) is used in NER module, which confuse me a lot. I am wondering wher the decoder part of transformer is used for? Thanks.
The text was updated successfully, but these errors were encountered:
As @Spico197 mentioned, the decoding module refers to the module which we use to decode the event records, not Transformer decoder actually.
Besides, Doc2EDAG use vanilla Transformer instead of BERT to derive the results they report in their paper, although the BERT part is also provided in their code.
Thanks for the excellent paper. But I have a question for the experiment setting:
As a sequence labeling task, we always use transformer encoder(like BERT) to solve the NER problem, for example, the author in Doc2EDAG use BERT as the first step backbone. However, in the paper, it is said that the vanilla transformer(encoder-decoder structure) is used in NER module, which confuse me a lot. I am wondering wher the decoder part of transformer is used for? Thanks.
The text was updated successfully, but these errors were encountered: