Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

User Story 1: Logger/streaming data #30

Closed
2 of 7 tasks
DavidBlaa opened this issue May 22, 2018 · 2 comments
Closed
2 of 7 tasks

User Story 1: Logger/streaming data #30

DavidBlaa opened this issue May 22, 2018 · 2 comments

Comments

@DavidBlaa
Copy link
Member

DavidBlaa commented May 22, 2018

capabilities to handle gradually growing tabular data up to 10Mio records/tuples.

  1. It should be possible to have multiple uploads resulting in 10 M record
  2. It should be possible to upload a 10M records.
  3. A potential performance improvement would be to reduce the validation and primary checks (David, needs to be done in any case)
  • No PK at all (2 day; prototype needed to test and measure the effect)
  • Append only file-based edits
  • Online editing for cells, without checking the PK
  • static PK at data structure level
  1. The priority is to reduce end-user’s response time versus reducing the task time. (@javadch )
  • A possible solution would be to manage long-running processes asynchronously.
  • This needs queuing (multiple priority queues), retry, and notification (offline, online), workflow resuming mechanisms.
  • Also, it may need a detailed activity logging.
@DavidBlaa
Copy link
Member Author

DavidBlaa commented Aug 7, 2018

only test the first 2 items

@DavidBlaa
Copy link
Member Author

works in v2.12 rc4

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

3 participants