New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Restarting the exporter pod causes a crash due to port already in use #11831
Comments
|
In case anyone else is hitting this, mitigation is to scale the affected deployment to This is not debilitating but it is quite annoying =] |
|
I think I see this too. |
|
@avanthakkar Are you looking into this? |
|
The exporter is being disabled with #12077, so this issue will have more time for investigation... |
|
Should not have been closed |
|
is it even creating them now? I haven't seen the issue, but I also only have an exporter deployment for 3 of my 6 nodes |
|
I am not able to repro this again by enabling the exporter with a local code change to allow v17.2.6. The exporter log does sometimes show the following error, but then is able to retry and then successfully start. Some related issue must have fixed this in 17.2.6. About every other restart I am seeing this log error, and other times it is successful on the first try. @avanthakkar Note that I also experimented with setting |
Is this a bug report or feature request?
Deviation from expected behavior:
Restarting the exporter pod causes it to fail since the previous exporter has not yet freed the resources while its pod is still terminating.
After restarting the exporter pod again, I see this error in the log, but then the pod was able to retry and continue automatically after that.
Expected behavior:
Restarting the pod should succeed. If the previous exporter pod is not terminated yet, the new pod should wait for it to exit, or perhaps just retry until the port is available. But if it takes too long (perhaps 60s), it should give up and crash.
How to reproduce it (minimal and precise):
kubectl -n rook-ceph delete pod <exporter>File(s) to submit:
cluster.yaml, if necessaryLogs to submit:
Operator's logs, if necessary
Crashing pod(s) logs, if necessary
To get logs, use
kubectl -n <namespace> logs <pod name>When pasting logs, always surround them with backticks or use the
insert codebutton from the Github UI.Read GitHub documentation if you need help.
Cluster Status to submit:
Output of krew commands, if necessary
To get the health of the cluster, use
kubectl rook-ceph healthTo get the status of the cluster, use
kubectl rook-ceph ceph statusFor more details, see the Rook Krew Plugin
Environment:
uname -a):rook versioninside of a Rook Pod):ceph -v):kubectl version):ceph healthin the Rook Ceph toolbox):The text was updated successfully, but these errors were encountered: