Enable Thresholding for Time Series Binary#3140
Conversation
Codecov Report
@@ Coverage Diff @@
## main #3140 +/- ##
=======================================
- Coverage 99.7% 99.7% -0.0%
=======================================
Files 318 318
Lines 30908 30948 +40
=======================================
+ Hits 30804 30843 +39
- Misses 104 105 +1
Continue to review full report at Codecov.
|
# Conflicts: # evalml/pipelines/time_series_pipeline_base.py # evalml/tests/automl_tests/test_automl_utils.py # evalml/tests/pipeline_tests/test_time_series_pipeline.py
freddyaboulton
left a comment
There was a problem hiding this comment.
@ParthivNaresh Looks good to me! Thank you for making this change. The one thing I want to square away before merge is whether we should be tuning on forecast_horizon number of obs or use predict_proba_in_sample.
| automl_config.optimize_thresholds | ||
| and pipeline.can_tune_threshold_with_objective(threshold_tuning_objective) | ||
| ): | ||
| test_size_ = ( |
There was a problem hiding this comment.
I'm wondering if we should use predict_proba_in_sample rather than predict_proba.
My thoughts are:
- The target is known during search so we don't have to worry about the forecast horizon
- forecast horizon is probably less than 20% of the data and usually will be small I think. I wonder if that's enough data to find a good threshold.
|
|
||
| def __init__(self, parameters, random_seed=0): | ||
| def __init__( | ||
| self, parameters, custom_name=None, component_graph=None, random_seed=0 |
There was a problem hiding this comment.
Just a cosmetic change right?
…teryx/evalml into EnableThresholdingForBinaryTS
Fixes #3095