Join GitHub today
GitHub is home to over 31 million developers working together to host and review code, manage projects, and build software together.
Sign uplogs filling up dockerd memory #1046
Comments
This comment has been minimized.
This comment has been minimized.
|
This is worthy of outputting, as it indicates a misconfiguration. Are you federating a timeseries that already exists in your prometheus? |
This comment has been minimized.
This comment has been minimized.
|
Any chance this is the same as #1042? |
This comment has been minimized.
This comment has been minimized.
|
@brian-brazil we don't have 100% control and what rules are being pushed to our prometheus. So yes this may indicate a problem, but I rather have some timeseries not being correct then the whole container dying. So a flag to suppress this would be welcome. :) |
This comment has been minimized.
This comment has been minimized.
|
@brian-brazil I don't see why this necessarily is a misconfiguration. |
This comment has been minimized.
This comment has been minimized.
|
@fabxc yep, def. seems related. |
This comment has been minimized.
This comment has been minimized.
|
Should be fixed by #973. |
This comment has been minimized.
This comment has been minimized.
A prometheus servers that's getting the same timeseries from two different places indicates that something is either up with your rules (rare), or you're missing distinguishing labels when federating from multiple sources (common). |
This comment has been minimized.
This comment has been minimized.
|
If that's the case, sure – but it's not the only possible or most likely cause. |
This comment has been minimized.
This comment has been minimized.
|
@miekg What version of docker are you running? Supposedly this is fixed in docker >= 1.7. See moby/moby#9139. |
This comment has been minimized.
This comment has been minimized.
|
In either case, one misconfiguration must not cause 32GiB of warnings. This will also drown out other problems, even if you're able to rotate it away quick enough. Can we rate limit this to be once per timeseries or less? |
This comment has been minimized.
This comment has been minimized.
|
Wow I didn't see 32GB - must pay closer attention ;) That's going to fill up disk too pretty quick. |
This comment has been minimized.
This comment has been minimized.
|
@matthiasr The fix will no longer show the warning for these cases at all (correctly, as there are too many false positives). |
This comment has been minimized.
This comment has been minimized.
What other causes have you seen?
I've never seen a case where this message was a false positive, in the tens of times I've investigated this over the years it was always a misconfiguration or misuse. Detecting this statically will be difficult, as you'd need to look at the label model of the entire set of federating Prometheus servers. Incorrect rules tend to be a bit easier to detect in the usual case.
When this happens the results can be surprising, this warning tells the user about a very fundamental error in their setup. |
This comment has been minimized.
This comment has been minimized.
|
Please read my reply above and the PR I linked where this log was fixed (which you |
This comment has been minimized.
This comment has been minimized.
|
|
This comment has been minimized.
This comment has been minimized.
|
I'm always thinking of HEAD :) |
This comment has been minimized.
This comment has been minimized.
This comment has been minimized.
This comment has been minimized.
|
We want to move towards an RC next week. The unclear socket leak is On Tue, Sep 1, 2015 at 6:45 PM Miek Gieben notifications@github.com wrote:
|
This comment has been minimized.
This comment has been minimized.
|
Closing this as the issue is fixed in HEAD and will be in the next release. |
juliusv
closed this
Sep 14, 2015
This comment has been minimized.
This comment has been minimized.
lock
bot
commented
Mar 24, 2019
|
This thread has been automatically locked since there has not been any recent activity after it was closed. Please open a new issue for related bugs. |
miekg commentedSep 1, 2015
In our prometheus setup we see a lot of
And with "a lot" I mean that
dockerdfills to 32GB and then gets OOM-killed. This is partly due to docker being stupid and actually caching this output or maybe this isn't worthy of outputting?