Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Questions on scorings in human_annotations.json #5

Open
sonsus opened this issue Sep 28, 2021 · 0 comments
Open

Questions on scorings in human_annotations.json #5

sonsus opened this issue Sep 28, 2021 · 0 comments

Comments

@sonsus
Copy link

sonsus commented Sep 28, 2021

Hi, I appreciate that I found a pretty good benchmark for the summarization metrics.
I have several things to ask after going through the codes and paper.

As far as I understand, human_annotations.json contains scores that is summary of human_annotations_sentence.json.

  1. (For sanity check of my understanding) For each sentence, major error type is considered as final label of that sentence that affects the score (such as NoE for factuality, LinkE for LinkE & Discourse Errors). So if I see one NoE sentence and 2 LinkE sentences for an article, that should be scored as {Factuality: 0.333, Discourse_Error: 0.333} while Semantic_Frame_Errors and Content Verifiability Errors being 1.0 which means the summary sentences are free from those type of errors but has Discourse Error which harms Factuality to 1/3.

  2. What I don't understand is the way of determining 'Flip' scores. First I thought it should be 1-ErrorType (e.g.Discourse Error for Flip Discourse Error). I still couldn't find any explainable way of making those scores from the labels. I tried to find some piece of code that generates human_annotations.json, but nothing indicates how to make Flip scores from the original ones explicitly. I think I got the motivation of applying flip scores for ablation study but not quite sure about how are they being generated.

Thanks again for the great piece of work. If you kindly explain this to me, it would be of great help. =]

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant