-
-
Notifications
You must be signed in to change notification settings - Fork 97
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
append error #31
Comments
It seems one has to retrieve npartitions from original dask dataframe, and pass it to append. collection.append(symbol, df_diff, npartitions=item.data.npartitions) Will it work everytime? |
thank you @flamby this fix is working and thanks for sharing and saving time for others. the pystore notebook demo too works only with this fix, else throws an error: ValueError: Exactly one of npartitions and chunksize must be specified. great thanks to @ranaroussi for this wonderful library |
Thank you to @ranaroussi for this nices libraries and thank you to @flamby who fix this nasty bug in the Windows 10 environment ! I had exactly the same message ("Exactly one of npartitions and chunksize must be specified") and the append was impossible. Now, it's work. Thank you again. |
Hello, same here (Win10 environment)! |
The problem is that dd.from_pandas() checks: So when the append function calls dd.from_pandas(df, npartitions = None) it raises the error but if you call dd.from_pandas(df, npartitions = None, chunksize=100000) it works. Presumably dask is using npartitions = 1 as its default even though the api says npartitions is optional and doesn't list a default. The code below is what needs to be tweaked. The new variable could be set to use npartitions = 1 (new = dd.from_pandas(data, npartitions=1), since this will be superseded by the passed value after the dataframes are combined. I'm willing to bet Ran comes up with a more elegant solution though. Lines 180 to 190 in 40de1d5
|
Hello,
Have you any recommendations regarding importing data from arctic?
I'm currently using cryptostore with arctic as a backend.
Cryptostore is by the very same author of arctic, but loading trades as a dataframe takes too much time with it.
For now, this is what I did :
But I'm facing errors similar to #16 - even if rollbacking dask and fastparquet to previous releases - when append is happening.
my setup :
Thanks, and keep the good work!
The text was updated successfully, but these errors were encountered: