New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add status conditions and profile applied to Profile(s) #188
Add status conditions and profile applied to Profile(s) #188
Conversation
/cc @marcel-apf |
@MarSik FYI |
// This is only to be consumed by humans. | ||
// +optional | ||
Message string `json:"message,omitempty"` | ||
} |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
@jmencak looks good, I think we better merge it and consume it with PAO, and then we can go from there. We will have a better visibility and we can come up with improvement ideas, if needed.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks, @marcel-apf . I'll aim to merge this immediately once 4.8 opens. @sjug/@dagrayvid , could you please provide a code review? Thank you!
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Except for the previous comment, the code changes look good to me.
/retest |
3 similar comments
/retest |
/retest |
/retest |
@jmencak: The following test failed, say
Full PR test history. Your PR dashboard. Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository. I understand the commands that are listed here. |
LGTM |
/retest |
/retest Please review the full test history for this PR and help us cut down flakes. |
3 similar comments
/retest Please review the full test history for this PR and help us cut down flakes. |
/retest Please review the full test history for this PR and help us cut down flakes. |
/retest Please review the full test history for this PR and help us cut down flakes. |
/hold |
1520e93
to
e3359d9
Compare
e3359d9
to
ca481f1
Compare
/hold cancel |
Type: tunedv1.TunedDegraded, | ||
} | ||
|
||
if (status & scApplied) != 0 { |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
where are these coming from ? (status, scApplied, scError, scWarn
)
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
As pointed out by Sebastian above, I "backported" part of this PR to fix rhbz1919970.
Degrading the cluster operator (NTO operator) if we get even 1 degraded profile status seems a bit harsh. Looking from Performance addons operator perspective - I think we should consume the tuned object status anyway if we want a 1:1 report for what might be wrong with performance profiles applied. |
I completely agree. If you believe that reporting the "error" status per Profile is enough, I'm more than happy not to touch the |
@jmencak ack, I think we can remove degration of the cluster operator then. We can keep status reporting maybe under the operator itself without degrading it. |
Changes: - report Tuned profile currently applied for each of the containerized Tuned daemon managed by NTO - report two Profile status conditions "Applied" and "Degraded" in every Profile indicating whether the Tuned profile was applied and whether there were issues during the profile application - cleanup of the ClusterOperator settings code; ClusterOperator now also reports Reason == ProfileDegraded for the Available condition if any of the Tuned Profiles failed to be applied cleanly for any of the containerized Tuned daemons managed by NTO - e2e test added to check the status reporting functionality - e2e basic/available test enhanced to check for not Degraded condition - using "podman build --no-cache" now. This works around issues such as: containers/buildah#2837
ca481f1
to
9a778dd
Compare
Done. |
/test e2e-aws |
/retest |
1 similar comment
/retest |
/lgtm |
[APPROVALNOTIFIER] This PR is APPROVED This pull-request has been approved by: jmencak, yanirq The full list of commands accepted by this bot can be found here. The pull request process is described here
Needs approval from an approver in each of these files:
Approvers can indicate their approval by writing |
@jmencak only thing missing here is the requested bug label |
The bug label would be needed if we wanted this in 4.7. For 4.8, the bug label is not necessary. As soon as 4.8 opens, this will merge. And it was exactly the plan to have this ready very early on for 4.8. |
Let's merge this: |
/retest Please review the full test history for this PR and help us cut down flakes. |
1 similar comment
/retest Please review the full test history for this PR and help us cut down flakes. |
Changes:
Tuned daemon managed by NTO
Applied
andDegraded
in every Profile indicating whether the Tuned profile was applied and
whether there were issues during the profile application
reports Reason == ProfileDegraded for the Available condition if any of
the Tuned Profiles failed to be applied cleanly for any of the containerized
Tuned daemons managed by NTO
podman build --no-cache
now. This works around issues such as:Podman build wrongly uses stale cache layer although build-arg changed and, thus, produces incorrect image containers/buildah#2837