-
Notifications
You must be signed in to change notification settings - Fork 38
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Isotree Isolation forest generating large model pickle file #48
Comments
Thanks for the bug report. A couple questions:
|
Below are my answers in bold
|
Thanks for the information. So it seems there's no bug in here: if you call Additionally, if using it as a missing value imputer (as it does when using As for what you could do: if the amount of rows is very large, you should call |
I am building anomaly detection model using isotree and the model pickle file if I dump via joblib without any compression, generates file of size 65GB. To load this model file for any realtime scoring requires around 256GB RAM for loading it into a python object and then scoring the new data. Is there any better way to do this or any tips on reducing the model size without impacting the accuracy of the model.
The text was updated successfully, but these errors were encountered: