New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Can't upload CSV files > 1MB to Jupyterlite file system by drag and drop #741
Comments
It would also be interesting to have some browser logs (e.g. revealed by Recall: everything in this free software, hosted on free servers, is still very much in beta, and the in-browser storage stuff is all best-effort, especially with respect to the less-than-month-old "magic" mirroring of files between kernels and contents. To fix this for real, it would need to be under test in a real browser, and I haven't had to time to start building up a suite. Last a checked, galata took some shortcuts...
Right: in general, we're trying to enable flexibility for deployers without taking away power/debuggability for users. The lightweight archival data analysis environment about datasets and packages known well in advance use case is much easier to reason about than i might upload anything and try and use it with anything, but it would indeed be lovely for it to Just Work. For heavy duty files, some other things to consider:
|
Thanks for the quick response.
|
Run the following script
Download this file, then drag and drop it into Jupyterlite's file system UI to (re)upload it The (re)uploaded file looks like this
The file is now 0.1 MB instead of 1.1 MB. Probably, but when a file larger than 1 MB is uploaded, it is divided into parts of 1 MB each and uploaded.
The size of an n-byte file, after being uploaded, will be #888 Related Issue |
Thanks all for the examples. Yes I also encountered this issue recently after uploading a ~4MB file and noticed it was truncated. |
Description
I'm trying to see if Jupyterlite demo site can process some of my local CSV files with pandas. I've noticed that when dragging and dropping these files in Jupyterlite file system UI, any file that larger than ~1MB would be truncated by the system; while the smaller ones <1MB can be processed correctly.
Attached please see the screenshots for this issue:
Screenshot 1. query.csv is a 1. 9MB earthquake dataset I have, but it was NOT read correctly by file system; a lot of rows were dropped during uploading process, and thus column headers were not processed correctly.
Screenshot 2. query_2.csv is a 0.9 MB partial dataset I edited from query.csv, by deleting nearly half of the data; and Jupyterlite file system took it without any issue, with column heads processed correctly.
Screenshot 3. reading these CSV files using Pandas.
I wonder if this is a limitation of the Jupyterlite demo website, so that users are now allowed to upload large files (but I'm under the impression that these local files are uploaded to users' browser storage...); or maybe there is some site setting I need to change.
(note: I've also noticed that if I serve these csv files as static assets using my own Github Jupyterlite deployment, then file size would not matter).
Context
Screenshots
The text was updated successfully, but these errors were encountered: