Join GitHub today
GitHub is home to over 31 million developers working together to host and review code, manage projects, and build software together.
Sign upHigh CPU usage running prometheus #1301
Comments
This comment has been minimized.
This comment has been minimized.
|
Any chance your setup just hit the retention period for the first time? |
This comment has been minimized.
This comment has been minimized.
|
Not to my knowledge, it's been running for 2-3 months now. |
This comment has been minimized.
This comment has been minimized.
|
It's a metric that tells you the current state of your server, not a configuration option. |
This comment has been minimized.
This comment has been minimized.
|
Thanks: Container is running unrestricted and on a VM with 8 cores, 16GB of memory. It is however shared with some other containers. Need more detailed info? |
brian-brazil
added
the
question
label
Jan 11, 2016
This comment has been minimized.
This comment has been minimized.
|
That's plenty of resources for the size of the server. Next step would be profiling your server. If you have the go toolchain installed, run:
If you can share the resulting SVG, we might find out more. |
This comment has been minimized.
This comment has been minimized.
|
@Rucknar Just checking, did this happen without any Prometheus version upgrade? And what's the version you're running? |
This comment has been minimized.
This comment has been minimized.
|
@juliusv It's entirely possible,i pull the container image from :latest and it's been re-pulled a number of times.
|
This comment has been minimized.
This comment has been minimized.
|
@fabxc The container doesn't have the toolchain installed, i'll look at getting that on there. |
This comment has been minimized.
This comment has been minimized.
|
Since you run this command against an HTTP endpoint of this server, you can execute it from any machine that can reach the container. |
This comment has been minimized.
This comment has been minimized.
|
prof.svg - From when it's behaving: |
This comment has been minimized.
This comment has been minimized.
|
From a first look, those look basically identical. In the second one it seems to be a bit more busy – not extremely though. It's mostly spending time on query evaluation – A graph of CPU load over time would probably help to understand, too. |
This comment has been minimized.
This comment has been minimized.
|
In case the rate of incoming queries is the/a culprit, check how it changed like this: |
This comment has been minimized.
This comment has been minimized.
|
Assuming this is resolved. Please re-open if it's still an issue and you want to provide more information. |
beorn7
closed this
Feb 2, 2016
This comment has been minimized.
This comment has been minimized.
ryan5rdx
commented
Sep 21, 2016
•
|
@fabxc @beorn7 |
This comment has been minimized.
This comment has been minimized.
|
FWIW, I seem to also be seeing this issue, with prometheus 1.5.2... Has anyone had luck troubleshooting this? |
This comment has been minimized.
This comment has been minimized.
lunemec
commented
Jul 12, 2017
•
|
I have the same issue, is there any way to display queries which consume most CPU time? EDIT: I have tried running |
This comment has been minimized.
This comment has been minimized.
gauravarora
commented
Aug 25, 2017
|
I have the exact same problem as @lunemec above and the metric count is 0. |
This comment has been minimized.
This comment has been minimized.
j0nimost
commented
Aug 25, 2017
|
increase the rate time preferably 15m probably the scrape period is too long |
This comment has been minimized.
This comment has been minimized.
lock
bot
commented
Mar 22, 2019
|
This thread has been automatically locked since there has not been any recent activity after it was closed. Please open a new issue for related bugs. |

Rucknar commentedJan 11, 2016
I'm seeing an issue when running prometheus, it's currently using more CPU than it has in the past and more than i would expect it to.
For example, here is a screenshot of the prometheus container we have running:

Here is the config we are using to run prometheus:
in the STDOUT logs, we see a few like the one below but nothing strange aside from that:
INFO[0347] Done checkpointing in-memory metrics and chunks in 35.485310843s. source=persistence.go:563Tried to troubleshoot this without raising a ticket but no joy.