Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Support for massive reading lists #3

jarvisms opened this issue Jul 7, 2019 · 1 comment


None yet
1 participant
Copy link

commented Jul 7, 2019

There is a limit to the number of readings that can be downloaded in one transaction based on server speeds etc. A request for massive time periods could transparently be broken into smaller transactions behind the scenes and presented as one big dataset.

@jarvisms jarvisms added the enhancement label Jul 7, 2019

@jarvisms jarvisms self-assigned this Jul 7, 2019

jarvisms added a commit that referenced this issue Jul 10, 2019

New get_mega_readings method
Enhancement proposed in issue #3 now implemented in the more advanced features module.

This comment has been minimized.

Copy link
Owner Author

commented Jul 10, 2019

Implemented in commit 189ce59. New method get_mega_readings from pythondcspro can essentially be used in place of get_readings from standard pythondcs but should avoid server timeouts and long delays if excessively large time periods are required.

@jarvisms jarvisms closed this Jul 10, 2019

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
You can’t perform that action at this time.