Join GitHub today
GitHub is home to over 31 million developers working together to host and review code, manage projects, and build software together.
Sign upOpening storage failed" err="invalid block sequence" #3714
Comments
This comment has been minimized.
This comment has been minimized.
|
I think it may like Prometheus 2.0 fails to start up after couple of restarts #3191. tell me if you need more details, Thanks. |
This comment has been minimized.
This comment has been minimized.
|
logs on k8s node here:
|
This comment has been minimized.
This comment has been minimized.
|
This comment has been minimized.
This comment has been minimized.
|
probably like Prom2: crash on opening WAL block #2795 |
This comment has been minimized.
This comment has been minimized.
|
Any updates here? |
This comment has been minimized.
This comment has been minimized.
kinghrothgar
commented
Mar 9, 2018
•
|
I am also seeing this in Prometheus v2.2.0 Edit: I will add more info as I uncover it. |
This comment has been minimized.
This comment has been minimized.
phreaker0
commented
Mar 12, 2018
|
I just hit this same error with prometheus v2.2.0 (i have installed this version fresh some days ago). details:
|
This comment has been minimized.
This comment has been minimized.
phreaker0
commented
Mar 12, 2018
|
i just checked the logs further and the issue already appeared without crashing at runtime before:
|
This comment has been minimized.
This comment has been minimized.
n0guest
commented
Mar 12, 2018
We also have 2.2.0 and this issue has few additional symptoms:
I hope it could help diagnose a problem. |
This comment has been minimized.
This comment has been minimized.
hectorag
commented
Mar 13, 2018
|
Also presenting this problem in my setup, running prometheus v2.2.0 with a empty DB. After some hours 2~3 prometheus start generating this error.
Then was not able to recover and start to fail again and again, showing the error below:
|
This comment has been minimized.
This comment has been minimized.
|
Sorry about that, this is a bug, fix is here: prometheus/tsdb#299 A new bug fix release will be out soon. |
This comment has been minimized.
This comment has been minimized.
phreaker0
commented
Mar 14, 2018
|
@gouthamve i hit it again, but this time rolling back the data to some time in the past (zfs snapshots) wouldn't work as prometheus started compacting block after startup and hit the issue after a couple of seconds. So i grabbed the linked patch, compiled prometheus and running the master + the patch version and it's fine so far, thx. |
This comment has been minimized.
This comment has been minimized.
shenshouer
commented
Mar 21, 2018
|
the same issue met at prometheus v2.2.0
|
This comment has been minimized.
This comment has been minimized.
|
Please try 2.2.1. |
This comment has been minimized.
This comment has been minimized.
shenshouer
commented
Mar 23, 2018
|
@brian-brazil It worked fine after I deleted all old data when update the prometheus v2.2.0 to v2.2.1 . |
This comment has been minimized.
This comment has been minimized.
|
Dupe of #3943. |
brian-brazil
closed this
Mar 23, 2018
This comment has been minimized.
This comment has been minimized.
bamb00
commented
Mar 30, 2018
•
|
@brian-brazil Hi, I'm hitting this issue with v2.2.1. Does this issue need to be re-open?
Thanks. |
This comment has been minimized.
This comment has been minimized.
Sriharivignesh
commented
Apr 4, 2018
|
Is there a way to recover from this error without flushing data out? I don't want to lose a chunk of my metrics data because of this :| |
This comment has been minimized.
This comment has been minimized.
zhanglijingisme
commented
Apr 18, 2018
|
@bamb00 Any update about this? |
This comment has been minimized.
This comment has been minimized.
bamb00
commented
Apr 20, 2018
|
@zhanglijingisme I have not heard back from the prometheus team. |
zhouyuanchao
referenced this issue
May 17, 2018
Closed
compact error leading to disk entire in a hour in 2.2.1 #4168
This comment has been minimized.
This comment has been minimized.
candlerb
commented
Jun 14, 2018
|
After upgrading from v2.2.1 to v2.3.0 I got this error:
I have kept the old data via Note: the thing which prompted the upgrade was that prometheus had starting doing much more disk I/O than expected, and was saturating the underlying hard drives. It's a relatively small set of time series which are being monitored - |
This comment has been minimized.
This comment has been minimized.
mysteryegg
commented
Jun 20, 2018
|
I have duplicated the behavior reported by @candlerb when upgrading from 2.2.1 to 2.3.1. |
This comment has been minimized.
This comment has been minimized.
uncleNight
commented
Jul 9, 2018
•
|
Here's how it went for me (running docker container with
Note the last two directories, they're the heaviest. If you check
Started prometheus again, voila, works again, and the data is there and accessible (I can see it by running queries from the very beginning of monitoring history). Hope it'd help somebody. |
This comment has been minimized.
This comment has been minimized.
lucasgameiro
commented
Nov 1, 2018
|
I had the same issue in windows environment with 2.3.*. I updated to 2.4.3 version and still didn't work. |
This comment has been minimized.
This comment has been minimized.
sumeshkanayi
commented
Dec 12, 2018
|
We also faced similar issue with 2.3.2 .We had to move the data from existing path mentioned under storage.tsdb.path to a new location and restart prometheus |
This comment has been minimized.
This comment has been minimized.
sanath230
commented
Dec 12, 2018
|
I faced the same issue in the version 2.3.2. I tried deleting duplicated chunks and restarting, it didn't work. Finally I had to move the whole data block to a different folder, create one more empty data folder and restart the prometheus service to make it work. |
This comment has been minimized.
This comment has been minimized.
Teriand
commented
Jan 31, 2019
|
same in 2.7.0 Help fix from @uncleNight with move bad files. |
cauwulixuan commentedJan 20, 2018
•
edited
What did you do?
I ran prometheus2.0.0 on kubernetesv1.8.5
What did you expect to see?
Everything went well.
What did you see instead? Under which circumstances?
Everything went well at beginning. But several hours later, pods' statuses turned to "CrashLoopBackOff", all prometheus turned unavaliable. After create pods, I didnt do anything.
Environment
System information:
Prometheus version:
v2.0.0
Prometheus configuration file:
Any suggestions?