Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Why does NER module in GIT use the encoder-decoder transformer rather than the encoder part only(like BERT used in Doc2EDAG) #3

Closed
YuanEric88 opened this issue Aug 11, 2021 · 2 comments

Comments

@YuanEric88
Copy link

YuanEric88 commented Aug 11, 2021

Thanks for the excellent paper. But I have a question for the experiment setting:
As a sequence labeling task, we always use transformer encoder(like BERT) to solve the NER problem, for example, the author in Doc2EDAG use BERT as the first step backbone. However, in the paper, it is said that the vanilla transformer(encoder-decoder structure) is used in NER module, which confuse me a lot. I am wondering wher the decoder part of transformer is used for? Thanks.
Screen Shot 2021-08-11 at 5 52 07 PM

@Spico197
Copy link

Here, the decoding means the EDAG generating, not for the Transformer decoder.
Here's the code for NER encoding.

@RunxinXu
Copy link
Owner

As @Spico197 mentioned, the decoding module refers to the module which we use to decode the event records, not Transformer decoder actually.
Besides, Doc2EDAG use vanilla Transformer instead of BERT to derive the results they report in their paper, although the BERT part is also provided in their code.

Thank you @Spico197.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants