-
Notifications
You must be signed in to change notification settings - Fork 41
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Issues with very wide datasets #20
Comments
Hi Ruslan, Could you please specify which version of Parso you use? Could you also tell me more about the error or ever copy its stacktrace here? It would be great to get the file where the error occurs, it is the easiest way to understand where the problem is. Thank you. |
Hi Yana, We're using 1.2.1 through this library: https://github.com/saurfang/spark-sas7bdat/ Thank you. |
Ok, let me know if it doesn't help. |
Please note that this old version doesn't support some compression mechanisms, while the latest version does. Also some other issues were fixed, as you understand 1.2.1 is significantly outdated - there was a major release and a couple of minor ones. |
We can read very wide SAS datasets with latest version of Parso library. |
There is a bug in parso when it tries to read a really wide file. When I read a sas7bdat with only 18 columns, it works perfectly, but when I read my sas7bdat with ~13000 columns, it breaks. For some reason, any of the values that aren’t zeros are replaced with null.
The text was updated successfully, but these errors were encountered: