Replies: 1 comment
-
If the full dataset consists of just a single long time series with one ID, the features will be different. But please note that in this case also using dask will nor help (it can only parallelize over different IDs). |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I have a dataset consisting of 4 million records. I want to carry out extract_features function method on the entire dataset with Dask, but I am facing memory issues. I now want to ask
Case 1 : Carrying out feature extraction for the entire 4 million records
Case 2 : Making 4 chunks of the data (of 1 million records each) and carrying out feature extraction for individual chunks.
Would the values of feature extraction be varied in both cases?
Beta Was this translation helpful? Give feedback.
All reactions