You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Found an issue with kafka nodes are coming up again after having failed. If they wake up to a corrupted index, it will attempt to fix itself. This seems to have two major implications:
Memory consumptions goes out the roof, resulting in pod getting killed due to OOM (due to limits, of course)
Readiness probe fails, and will kill the pod if it hasn't already OOM'ed
Any thoughts on how to remedy this?
The text was updated successfully, but these errors were encountered:
stigok
changed the title
OOM when rebuilding corrupted index
Issues when forced to rebuild corrupted index
Sep 28, 2018
We didn't get around to implementing it yet, but the idea is to fix the default image to support the ./kafka-server-stop.sh command (solsson/dockerfiles@4fb7b5d) and to use a preStoppod lifecycle hook to invoke it.
Found an issue with kafka nodes are coming up again after having failed. If they wake up to a corrupted index, it will attempt to fix itself. This seems to have two major implications:
Any thoughts on how to remedy this?
The text was updated successfully, but these errors were encountered: