New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Automatically force_all_finite=False in non-initial warm_start runs #10600
Comments
Can I take this? |
I'd be interested in seeing if other core devs think we should do this;
there may not be consensus. But with that caveat you could certainly
propose a PR. you could test it by sneaking a np.inf into the data before a
second fit.
|
we've still not heard if other core devs have an opinion. so you can offer
a PR, but with no assurance it will ever get merged
|
…imator and GradientBoosting scikit-learn#10600
Hello @jnothman I worked on this issue and i submitted a PR for this. |
I think warm_start has inconsistent semantics across modules. In linear models, you can warm-start with different datasets and potentially get a speed benefit but essentially an equivalent solution, while for trees, you have to warm-start on the same dataset. So I don't think we can actually do this consistently and probably first maybe clarify our warm-start semantics? |
For estimators that have a
warm_start
option, we generally expect that the same data is passed in for each call tofit
as was done in the first. I think we can save a little on runtime by usingcheck_array
withforce_all_finite=False
in the case thatwarm_start=True
and the model has already been fitted.The text was updated successfully, but these errors were encountered: