New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Failed to update lock: Operation cannot be fulfilled on leases.coordination.k8s.io #16402
Comments
It took me 1 hour debugging this issue until I realized that this is open because of a PR still marked as draft #16395. This is not an issue as it is part of the k8s library and the process is handled automatically by the library which Cilium can't control. |
Are you saying the error is expected? Should it then be at least diminished to a warning? |
Unfortunately we don't control the level of klog logs so we can't set it as warning. |
Another variation seems to be:
from https://jenkins.cilium.io/job/Cilium-PR-K8s-1.16-net-next/796/testReport/junit/Suite-k8s-1/16/K8sKafkaPolicyTest_Kafka_Policy_Tests_KafkaPolicies/ I understand we can't change these messages or their level, but do we know why they happen? |
cilium#16477 was merged and a new error, cilium#16402 (comment) was discovered since the PR disallowing level=error in CI was merged. Signed-off-by: Paul Chaignon <paul@cilium.io>
#16477 was merged and a new error, #16402 (comment) was discovered since the PR disallowing level=error in CI was merged. Signed-off-by: Paul Chaignon <paul@cilium.io>
Yet another variation:
|
This error message happened in CI and seems to be a less frequent variation of known klog error messages [1]. 1 - cilium#16402 (comment) Signed-off-by: Paul Chaignon <paul@cilium.io>
This error message happened in CI and seems to be a less frequent variation of known klog error messages [1]. 1 - #16402 (comment) Signed-off-by: Paul Chaignon <paul@cilium.io>
Ran into another variation for this weeks backport:
The "Failed to release lock: resource name may not be empty" is a new one. Perhaps a worthy candidate for adding to the exceptions list? |
Occasionally the cilium-operator will run into a transient issue where it cannot get/update/release the leaselock with K8s that it uses to adjudicate its leader election. This error message is part and parcel of this failure and can be ignored. cf. #16402 Signed-off-by: Nate Sweet <nathanjsweet@pm.me>
@nathanjsweet did you check etcd logs? I think this is a similar failure as this one #17981 (comment) |
I just looked again, just some warnings about "apply request took too long". |
Occasionally the cilium-operator will run into a transient issue where it cannot get/update/release the leaselock with K8s that it uses to adjudicate its leader election. This error message is part and parcel of this failure and can be ignored. cf. #16402 Signed-off-by: Nate Sweet <nathanjsweet@pm.me>
Occasionally the cilium-operator will run into a transient issue where it cannot get/update/release the leaselock with K8s that it uses to adjudicate its leader election. This error message is part and parcel of this failure and can be ignored. cf. #16402 Signed-off-by: Nate Sweet <nathanjsweet@pm.me>
[ upstream commit 82d4422 ] Occasionally the cilium-operator will run into a transient issue where it cannot get/update/release the leaselock with K8s that it uses to adjudicate its leader election. This error message is part and parcel of this failure and can be ignored. cf. cilium#16402 Signed-off-by: Nate Sweet <nathanjsweet@pm.me> Signed-off-by: Nicolas Busseneau <nicolas@isovalent.com>
[ upstream commit 82d4422 ] [ Backport notes: had to resolve conflicts manually due to cilium#16395 previously introducing exceptions not having been backported to v1.10. The changes in this PR completely supersede cilium#16395 so there should be no need to backport it first. ] Occasionally the cilium-operator will run into a transient issue where it cannot get/update/release the leaselock with K8s that it uses to adjudicate its leader election. This error message is part and parcel of this failure and can be ignored. cf. cilium#16402 Signed-off-by: Nate Sweet <nathanjsweet@pm.me> Signed-off-by: Nicolas Busseneau <nicolas@isovalent.com>
[ upstream commit 82d4422 ] [ Backport notes: had to resolve conflicts manually due to #16395 previously introducing exceptions not having been backported to v1.10. The changes in this PR completely supersede #16395 so there should be no need to backport it first. ] Occasionally the cilium-operator will run into a transient issue where it cannot get/update/release the leaselock with K8s that it uses to adjudicate its leader election. This error message is part and parcel of this failure and can be ignored. cf. #16402 Signed-off-by: Nate Sweet <nathanjsweet@pm.me> Signed-off-by: Nicolas Busseneau <nicolas@isovalent.com>
[ upstream commit 82d4422 ] Occasionally the cilium-operator will run into a transient issue where it cannot get/update/release the leaselock with K8s that it uses to adjudicate its leader election. This error message is part and parcel of this failure and can be ignored. cf. #16402 Signed-off-by: Nate Sweet <nathanjsweet@pm.me> Signed-off-by: Nicolas Busseneau <nicolas@isovalent.com>
This comment was marked as outdated.
This comment was marked as outdated.
FWIW, we are still hitting variations of this in the CI:
Excerpt from one of the
|
Found in CI, in cilium-operator logs:
From:
https://jenkins.cilium.io/job/Cilium-PR-K8s-GKE/5586/testReport/junit/Suite-k8s-1/18/K8sCLI_CLI_Identity_CLI_testing_Test_labelsSHA256/
K8sCLI_CLI_Identity_CLI_testing_Test_labelsSHA256.zip
The text was updated successfully, but these errors were encountered: