Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Rdkafka consumer settings for high latency environments #696

Open
sverch opened this issue Jul 9, 2024 · 0 comments
Open

Rdkafka consumer settings for high latency environments #696

sverch opened this issue Jul 9, 2024 · 0 comments

Comments

@sverch
Copy link

sverch commented Jul 9, 2024

Our use case for kafka involves high latency (on the order of a second round trip time) and sudden bursts of messages on the order of 300k. We are seeing slowness in that situation, and I was looking into different kafka settings that might help mitigate this.

I saw that max.poll.records is set to 500 by default. I thought this might solve our problem (or would at least be worth testing), but I saw that it is not supported in rdkafka and found these two issues:

It seems like those use cases were slightly different, in that they wanted to limit the batch size due to memory concerns. In our case, processing 300k messages at a batch size of 500 would spend about ten minutes waiting for network round trips.

I also found this documentation that is related, but not quite for this case:

Specifically, I'm wondering:

  • Is it possible for this setup to be slow, or are the polls pipelined behind the scenes somehow?
  • Are there specific settings that would help with a high latency consumer and large bursts of messages? (Note, the processing itself is fast).
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant