-
-
Notifications
You must be signed in to change notification settings - Fork 179
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
🔬🔁 Evaluation loop #768
🔬🔁 Evaluation loop #768
Conversation
trigger ci
trigger ci
trigger ci
trigger ci
trigger ci
trigger ci
trigger ci
target-specific evaluation datasets have been moved to a new branch, cf. |
@mberr lgtm but would be nice to get a second reviewer |
Tried the branch on the ILPC codebase - it works, can reproduce the numbers 👍 |
Did you know about the from pykeen.pipeline import pipeline
result = pipeline(
dataset="nations",
model="mure",
result_tracker="console",
result_tracker_kwargs=dict(metric_filter=r"both.realistic.(hits_at_10|inverse_harmonic_mean_rank)"),
) will only print realistic H@10 and MRR averaged over head & tail. |
This PR adds an evaluation loop based upon torch's data loaders, and delegates the automatic batch size optimization to
torch-max-mem
. It also brings support for relation prediction evaluation.