-
Notifications
You must be signed in to change notification settings - Fork 417
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
spread causes system to run out of memory #13
Comments
I have not. It might be possible to replace the vectorised R code with optimised C++ code that would need less memory. |
Just a quick note that I'm having memory issues with My input dataset is 0.4MB, has 6000 rows, and 11 variables. This is the result of a filter on a dataset which is of size 200Mb. When running with Unfortunately I can't provide the exact dataset, but if there is anything I can provide to help, I'll be happy to do so. |
How many unique values are there in the variables that you are spreading? It is easy to create very very large data frames with spread. |
There are some numeric variables with a few thousand unique values, but isn't |
How long/wide can a data frame be -- going from gathered to spread form?
I have a 200,000 row data frame I'm spreading to create 200,000 columns and I'm running out of memory.
Have you tested to see limits on the operations with various machines?
The text was updated successfully, but these errors were encountered: