Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Batching of metrics #11

Open
akakitani opened this issue Nov 16, 2016 · 1 comment
Open

Batching of metrics #11

akakitani opened this issue Nov 16, 2016 · 1 comment

Comments

@akakitani
Copy link
Contributor

Implement batching of metrics, so that when we want to collect metrics at the record level, the driver doesn't slow down with too many metrics being sent to it.

My initial idea for this is to have some configurable batching interval of x seconds. This interval can be configured by the user to match how often some external monitoring service checks these Spark metrics, so that even time-sensitive ones can remain relatively accurate.

With Spark 2.0, we should also explore using the newer accumulator APIs to see if we can accomplish this.

@BDeus
Copy link

BDeus commented May 17, 2017

Do you have an implementation or a begin of implementation of batching ?
Use spark-metrics at record level has a too high footprint on performance.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants