Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Perplexity Script and Notebook #55

Merged
merged 21 commits into from
May 17, 2023
Merged

Conversation

Stillerman
Copy link
Collaborator

#53

Here is a notebook and script that can be used to calculate perplexity. The notebook runs on colab, but I cannot test the script locally so that will need to be tested. Should output perplexity results toperplexity-results.jsonl.

Should be able to run

python3 calculate_perplexity.py --model Multi-Domain-Expert-Layers/expert-arxiv --dataset Multi-Domain-Expert-Layers/arxiv --split validation_domain

and test any model/dataset/split we want.

@Stillerman
Copy link
Collaborator Author

Stillerman commented May 15, 2023

Perplexity script produced the following results

Model arxiv freelaw github
expert-arxiv 6.588 6.077 5.675
expert-freelaw 6.744 6.038 5.673
expert-github 6.827 6.145 5.680
pythia-1b-deduped 6.560 6.013 5.604

All makes sense except github expert does worse on github dataset. Is this possible @mrcabbage972

@Stillerman
Copy link
Collaborator Author

We are still getting unexpected results but we believe it is not due to a bug in the perplexity script so undrafted. Details of weird perplexities in the issue #53

@Stillerman Stillerman changed the title Draft: Perplexity Script and Notebook Perplexity Script and Notebook May 17, 2023
@mrcabbage972 mrcabbage972 merged commit effe662 into huu4ontocord:main May 17, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

2 participants