-
Notifications
You must be signed in to change notification settings - Fork 25.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Guide to fixed-length model perplexity evaluation #5449
Conversation
Codecov Report
@@ Coverage Diff @@
## master #5449 +/- ##
==========================================
+ Coverage 77.82% 78.68% +0.86%
==========================================
Files 141 141
Lines 24608 24608
==========================================
+ Hits 19150 19364 +214
+ Misses 5458 5244 -214
Continue to review full report at Codecov.
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Left some comments but I really like this, and the Research section is exactly right for that kind of docs!
This post / guide is inspired by this recent Twitter discussion and this gist on the different ways that perplexity can be evaluated and the optimal strategy of a strided "sliding window".
Interested in feedback both on the guide/writing component as well as the theoretical discussion on PPL. Right now my understanding is that our language modeling script uses non-overlapping segments rather than the sliding window.
Relevant to #4415, #4219.