Join GitHub today
GitHub is home to over 31 million developers working together to host and review code, manage projects, and build software together.
Sign upPrometheus OOM #5266
Comments
This comment has been minimized.
This comment has been minimized.
|
You need to use a system with more RAM and/or reduce the memory usage of Prometheus. It looks like you're using remote write and Prometheus can't send samples fast enough to the remote backend so Prometheus buffers samples in memory. Have a look at the remote_write settings to tune this part. You may also want to upgrade to the latest release of Prometheus as it may improve memory usage. I'm closing it for now. If you have further questions, please use our user mailing list, which you can also search. |
simonpasquier
closed this
Feb 25, 2019
This comment has been minimized.
This comment has been minimized.
|
Thanks for your reply ! @simonpasquier prometheus remote_write as below:
Could you give some advice about how to optimize it, and how to confirm ? |
This comment has been minimized.
This comment has been minimized.
|
I'm no remote write expert but you might want to try the defaults proposed in #5267:
Also |
This comment has been minimized.
This comment has been minimized.
|
@XSHui as Simon mentioned you either need a system with more RAM (based on your current config) or to change the config. The capacity is the number of samples each remote write shard can have buffered before it will (force) send to the endpoint. If you look at your metrics, how many shards do you have? |
XSHui commentedFeb 25, 2019
•
edited
Bug Report
What did you do?
We use prometheus to monitor tidb cluster
What did you expect to see?
Prometheus work well
What did you see instead? Under which circumstances?
Prometheus sometimes oom(only one cluster, other cluster is ok)
Environment
System information:
Linux 4.14.0-1.el7.*****.x86_64 x86_64
Prometheus version:
prometheus:v2.2.1
Logs: