-
Notifications
You must be signed in to change notification settings - Fork 1.3k
Calculate multiple metrics #985
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
…uto-sklearn into calculate_multiple_score
|
This implements #981 I guess? |
Development
mfeurer
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks a lot for the PR! I left some initial comments you could address, and then it would be good to incrementally work on the following things:
- rebase on the development branch to fix the current issues CI
- add an example about tracking auxiliary metrics
- think about additional unit tests to test this new functionality
Looking forward to the changes
|
This looks great, looking forward to the example and unit tests. |
mfeurer
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
A few updates from my side:
- I cannot reproduce the failure for the pandas example locally, so I think everything is fine here. Nevertheless, I restarted the tests on travis-ci.
- The code looks good and I would merge it as is, therefore, you could now move to adding unit tests (I've been thinking about one for the score calculation function, and making the tests in
test_train_evaluatorstricter by checking the entries in the additional run info, please let me know if you have any questions regarding that) and an example on how to use this new feature.
I'm really looking forward to this new feature as I actually also just found a good use case within our research :)
Codecov Report
@@ Coverage Diff @@
## development #985 +/- ##
===============================================
+ Coverage 85.34% 85.37% +0.03%
===============================================
Files 125 125
Lines 9858 9881 +23
===============================================
+ Hits 8413 8436 +23
Misses 1445 1445
Continue to review full report at Codecov.
|
|
@mfeurer Test cases and example will take some time. Currently I am on vacation for 2 weeks. I will start working on test cases and example once I am back.
If possible, can you highlight what is the use case you are looking forward to? |
|
Thanks for the heads up @rabsr enjoy your vacation |
mfeurer
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks a lot for the tests! I just made a PR which makes the examples work. How about I merge this one and you start working on a new PR updating the example for using different metrics?
If possible, can you highlight what is the use case you are looking forward to?
I'm planning to use it to generate meta-data for the Auto-sklearn 2.0 project.
Co-authored-by: Matthias Feurer <lists@matthiasfeurer.de>
…uto-sklearn into calculate_multiple_score
@mfeurer That works for me. You can merge this PR. I will open a new PR for example in couple of days. |
This change will provide users an option scoring_functions to calculate list of metrics for each pipeline. The calculated metrics will be available from attribute cv_results_.
This will provide users with more insights for each pipeline executed.
Usage: