Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

What Metrics do we want to have for training/testing here? #4

Open
1 of 5 tasks
karalets opened this issue Mar 27, 2020 · 5 comments
Open
1 of 5 tasks

What Metrics do we want to have for training/testing here? #4

karalets opened this issue Mar 27, 2020 · 5 comments
Assignees

Comments

@karalets
Copy link
Collaborator

karalets commented Mar 27, 2020

Initially, we can have things like log-likelihood in order to just be able to get some reasonable quantitative thing.

Over time, however, we may want to have more informative metrics for performance of the deep net on the task at hand, for instance downstream metrics for a chemistry application or so.

While this is not pressing to do at first, I am opening this issue so we can collect ideas for:

  1. metrics that make sense to collect for training and testing
  2. plots we may want to see down the line that would be reasonable to have

Both of those can and should also take into account the evaluation chosen in https://pubs.rsc.org/en/content/articlepdf/2019/sc/c9sc00616h as ultimately we will need to compare to it.

My first pitch is as stated:

  • compute training likelihood and test likelihood on a hold-out dataset
  • visualize this as a function of number of training datapoints
  • visualize this as a function of number of background (for semis-supervised) datapoints
  • compare different strategies for building background dataset with respect to likelihood, i.e. more diverse data leading to better representations etc.
  • ... more ways?

The nice thing about this is we can rerun the same evaluation protocols with any metric, not just LLK.

@maxentile
Copy link
Member

This sounds good to me. I guess one other basic question is whether we want to have metrics for how "well-calibrated" the predictive uncertainties are, and if so, what those should look like. If this is in-scope, perhaps @karalets can provide some references / pointers?

@karalets
Copy link
Collaborator Author

This sounds good to me. I guess one other basic question is whether we want to have metrics for how "well-calibrated" the predictive uncertainties are, and if so, what those should look like. If this is in-scope, perhaps @karalets can provide some references / pointers?

Great point, I am happy to take point on that with some references once we start having results to discuss how to evaluate such things.

The general idea is the following:
We need in distribution and out-of-distribution data sets for training in order to evaluate things.

So once we have some experiments lined up where molecules for both categories are chosen well and start plotting stuff we can discuss that subtlety.

@karalets
Copy link
Collaborator Author

But I would still like it if the chemists here could add some more informative and concrete metrics with relation to usefulness here. Examples are the things the Cambridge-group paper evaluates.

@yuanqing-wang
Copy link
Member

Personally I'd imagine real chemists care about false positive rates with a cutoff and so on. But at this stage let's just suppose they're no different.

@karalets
Copy link
Collaborator Author

I thought more about downstream quantities of interest as metrics. No clue what people care about.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

4 participants