Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

An exciting and daring idea about trade off between query latency and accuracy. #1465

Closed
xuguruogu opened this issue Dec 17, 2019 · 0 comments
Assignees
Labels
type/question Type: question about the product wontfix Solution: this will not be worked on recently

Comments

@xuguruogu
Copy link
Collaborator

xuguruogu commented Dec 17, 2019

Our demand to query latency is strict, at the same time, we can afford less accuracy by missing some records. Mostly, we would discard parts of the result records if more than expected.

To avoid uncontrollable massive Intermediate output, why not assign a time limit in a step in job DAG / non parallelizable execution plan. If max execution time is reached, simply go on to the next step, and drop the rest reply from peers.

As TCP is natively non parallelizable, making full use of UDP may be a better solution.

I wander if the problem be solved by provide max_exec_time/discard_rest_after_sometime semantics in the DSL.

@jude-zhu jude-zhu added the v2.0 label May 26, 2020
@CPWstatic CPWstatic added the type/question Type: question about the product label Aug 28, 2021
@Sophie-Xie Sophie-Xie added the wontfix Solution: this will not be worked on recently label Aug 31, 2021
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
type/question Type: question about the product wontfix Solution: this will not be worked on recently
Projects
None yet
Development

No branches or pull requests

5 participants