Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Connectivity issues seems to cause memory spike in Prometheus #13489

Open
wiardvanrij opened this issue Jan 29, 2024 · 1 comment
Open

Connectivity issues seems to cause memory spike in Prometheus #13489

wiardvanrij opened this issue Jan 29, 2024 · 1 comment

Comments

@wiardvanrij
Copy link

What did you do?

I'm running Prometheus with istio as a sidecar for proxy/mesh functionality. We potentially had a memory leak in istio, which is somewhat unrelated to Prometheus, but it did uncover some cascading issue when the sidecar goes towards 100% memory / OOM.

This causes connectivity issues for 'anything'. Meaning that scrapes will not resolve, the thanos-sidecar is unable to talk to Prometheus, queries do not hit Prometheus, etc. etc. Basically, both egress as ingress is gone. That is expected and "fine", however it directly shows an almost x2 increase in Prometheus memory usage, as seen in the following screenshot:

image

Now, I'm sorry I do not have profiles or any extra information, and I also don't find it such exiting Prometheus 'bug' (if it is anything), but I did figure to report it as it's somewhat unexpected to see.

What did you expect to see?

Prometheus not being able to scrape targets and having 'issues' in general to scrape stuff, but maintain memory usage on 'normal' levels.

What did you see instead? Under which circumstances?

Close to x2 increase in memory when there is connectivity issues due to sidecar proxy 'outage'.

I also want to mention that correlation is not causation, however it happened more than once (multi clusters and such) and that I did not dig really deep into this due to other priorities and the fact that we just need to fix the sidecar :)

System information

No response

Prometheus version

v2.48.1

Prometheus configuration file

No response

Alertmanager version

No response

Alertmanager configuration file

No response

Logs

Not much special to report, other than quite some
="http: superfluous response.WriteHeader call from github.com/opentracing-contrib/go-stdlib/nethttp.(*statusCodeTracker).WriteHeader (status-code-tracker.go:19" msg=) type of logs.
@roidelapluie
Copy link
Member

Thanks! Are you able to provide a memory trace (proof) when that happens?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants