Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Computing Perplexity #6

Open
inimah opened this issue Oct 10, 2022 · 0 comments
Open

Computing Perplexity #6

inimah opened this issue Oct 10, 2022 · 0 comments

Comments

@inimah
Copy link

inimah commented Oct 10, 2022

Hi, @JianGuanTHU Thanks for making the data publicly available.

Could you please elaborate more on how the current work computes "Perplexity" metric?
Is it sentence-perplexity or perplexity of predicting a token?

The paper mentions in a footnote We take the minus of perplexity for all the following...

But I do not think the metric outputs in ~/Data/../metric_output/ppl.txt are reasonably fit with the text inputs.
What is "minus of perplexity" in this context?

For example, score on sample-ID 151 from ant_data_all
((I am using HuggingFace -- evaluate perplexity metric))

Prediction text: ["we were looking for something fun to do on a female night . Female wife and i were so excited . we went to the mall . we had a great time . we had a great time ."]

results: {'perplexities': [55.47270202636719], 'mean_perplexity': 55.47270202636719}

While, in ppl.txt the score is 2.5693

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant