Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

doubt regarding inputs to preprocess.py #4

Closed
ramgj28 opened this issue Jun 20, 2021 · 3 comments
Closed

doubt regarding inputs to preprocess.py #4

ramgj28 opened this issue Jun 20, 2021 · 3 comments

Comments

@ramgj28
Copy link

ramgj28 commented Jun 20, 2021

Hello there, first of all thank you so much for giving your code as open source so others like me can learn from it. I saw that the preprocess.py script requires many file inputs including the candidate summary. But those are generated by the model right? I couldn't find those in the data. Also in the example json file, I noticed that article untokenized and tokenized both seem to be sentence tokenized. So what is the difference?

@yixinL7
Copy link
Owner

yixinL7 commented Jun 21, 2021

  1. Candidate summaries are generated by a pre-trained abstractive model (in our work we use BART on CNNDM). Our code is for training the evaluation model in our paper. We've provided the preprocessed date along with the generated candidate summaries.
  2. Untokenized text is for model input, following the requirement of RoBERTa. Tokenized data is for evaluation (computing ROUGE), following the previous work.

@ramgj28
Copy link
Author

ramgj28 commented Jun 22, 2021

Oh, now I get it. Thank you so much for taking your time explaining this. Really appreciate it.

@ramgj28 ramgj28 closed this as completed Jun 22, 2021
@ramgj28
Copy link
Author

ramgj28 commented Jun 22, 2021

doubt cleared.Thanks;)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants