Join GitHub today
GitHub is home to over 31 million developers working together to host and review code, manage projects, and build software together.
Sign upincrease(...) appears to treat discontinuity as 0-value (leading to larger than expected increase) #3852
Comments
This comment has been minimized.
This comment has been minimized.
|
That looks like the effect of a counter going backwards. Can you share the raw data of one of those time series around that time? |
This comment has been minimized.
This comment has been minimized.
|
Hi - Taking one of the metrics:
http://prometheus.dev.ahl:9090/api/v1/query_range?query=message_count%7Binstance%3D~%221.*%22%2C%20env%3D%22prod%22%2C%20service%3D%22RSF%22%2C%20partition%3D%2214%22%7D&start=1518020340&end=1518034740&step=57&_=1518780931424 Increase
There's a jump between these two timesteps in
But there is no similar jump in the data. Even if I query the data at 1s precision, I don't see anything untoward in the data: other than a discontinuity -
|
This comment has been minimized.
This comment has been minimized.
This comment has been minimized.
This comment has been minimized.
Can you give me the raw data around those times? |
This comment has been minimized.
This comment has been minimized.
|
The raw data was in the and at 1s resolution Let me know if there's any other data you would like (or some other way of getting the data). |
This comment has been minimized.
This comment has been minimized.
|
The data in query1s.txt does not cover the time period of the issue in increasedata.txt. |
This comment has been minimized.
This comment has been minimized.
|
Ah, sorry about this - it does look like something strange has happened - the counter has briefly gone backwards by '163'... query1s up to 1518034134:
increase1s up to 1518034134:
|
This comment has been minimized.
This comment has been minimized.
|
We're digging deeper into this - we've recently migrated from Prometheus 1.x to 2.x and it may be that there's some funny interaction between the old and the new with remote read. |
This comment has been minimized.
This comment has been minimized.
|
So we think this is caused by a discontinuity caused by the migration from Prometheus 1.x to 2.x and remote read. Sorry for the noise! |
jamesblackburn
closed this
Feb 16, 2018
This comment has been minimized.
This comment has been minimized.
|
This is the first report of a discontinuity from such a setup. This implies that users should be careful not to have scraping working for at least a scrape interval to mitigate this. |
This comment has been minimized.
This comment has been minimized.
lock
bot
commented
Mar 22, 2019
|
This thread has been automatically locked since there has not been any recent activity after it was closed. Please open a new issue for related bugs. |




jamesblackburn commentedFeb 16, 2018
What did you do?
Attempted to use increase to see change in a value over a 24h period.
What did you expect to see?
In the data we see a large jump in increase which seems to correlate with a missed scrape for the service endpoint.
Observed data:

What did you see instead? Under which circumstances?
Large increase observed around the discontinuity.
Environment
System information:
insert output of
uname -srmhere