Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

In autoscaling tests, add PDBs for more kube-system pods #52796

Conversation

aleksandra-malinowska
Copy link
Contributor

This adds PDBs for more kube-system pods in scale down tests. It should reduce flakes caused by evenly distributed system components blocking scale down of all nodes.

@aleksandra-malinowska aleksandra-malinowska added area/test kind/flake Categorizes issue or PR as related to a flaky test. release-note-none Denotes a PR that doesn't merit a release note. retest-not-required sig/autoscaling Categorizes an issue or PR as relevant to SIG Autoscaling. labels Sep 20, 2017
@aleksandra-malinowska aleksandra-malinowska added this to the v1.8 milestone Sep 20, 2017
@aleksandra-malinowska aleksandra-malinowska self-assigned this Sep 20, 2017
@k8s-ci-robot k8s-ci-robot added size/XS Denotes a PR that changes 0-9 lines, ignoring generated files. cncf-cla: yes Indicates the PR's author has signed the CNCF CLA. labels Sep 20, 2017
{label: "kubernetes-dashboard", min_available: 0},
{label: "l7-default-backend", min_available: 0},
{label: "heapster", min_available: 0},
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Are you sure all of those can be safely restarted? I am rather worried about restarting heapster (or any other critical system pod) in e2e. We had enough pain with rescheduler tainting our nodes already.

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

It's hard to be sure, but:

  1. in tests with broken nodes, we don't select a node based on any of those pods, so we sometimes accidentally cause them to become unavailable and rescheduled anyway,
  2. we have timeouts and retry logic which should be enough to cover these scenarios if everything works as expected,
  3. if it doesn't work as expected, I think we should fail, even if it's not CA that is responsible.

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I don't necessarily agree with point 3 above. However, we discussed offline with @aleksandra-malinowska and it looks like restarting heapster should no longer break tests.

@MaciekPytel
Copy link
Contributor

/lgtm

@k8s-ci-robot k8s-ci-robot added the lgtm "Looks good to me", indicates that a PR is ready to be merged. label Sep 20, 2017
@MaciekPytel
Copy link
Contributor

/approve no-issue

1 similar comment
@mwielgus
Copy link
Contributor

/approve no-issue

@mwielgus mwielgus added the approved Indicates a PR has been approved by an approver from all required OWNERS files. label Sep 20, 2017
@k8s-github-robot
Copy link

[APPROVALNOTIFIER] This PR is APPROVED

This pull-request has been approved by: MaciekPytel, aleksandra-malinowska, mwielgus

Associated issue requirement bypassed by: mwielgus

The full list of commands accepted by this bot can be found here.

Needs approval from an approver in each of these OWNERS Files:

You can indicate your approval by writing /approve in a comment
You can cancel your approval by writing /approve cancel in a comment

@k8s-github-robot
Copy link

Automatic merge from submit-queue. If you want to cherry-pick this change to another branch, please follow the instructions here..

@k8s-github-robot k8s-github-robot merged commit 3f447dd into kubernetes:master Sep 20, 2017
cblecker pushed a commit to cblecker/kubernetes that referenced this pull request Sep 21, 2017
…aling-test-fix-4

Automatic merge from submit-queue. If you want to cherry-pick this change to another branch, please follow the instructions <a href="https://github.com/kubernetes/community/blob/master/contributors/devel/cherry-picks.md">here</a>..

Improve cluster autoscaling tests logging and error checking during cleanup

This adds extra logs and error checks to autoscaling tests during PodDisruptionBudgets cleanup. It should help with identifying flake causes. Follow up to kubernetes#52796
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
approved Indicates a PR has been approved by an approver from all required OWNERS files. area/test cncf-cla: yes Indicates the PR's author has signed the CNCF CLA. kind/flake Categorizes issue or PR as related to a flaky test. lgtm "Looks good to me", indicates that a PR is ready to be merged. release-note-none Denotes a PR that doesn't merit a release note. sig/autoscaling Categorizes an issue or PR as relevant to SIG Autoscaling. size/XS Denotes a PR that changes 0-9 lines, ignoring generated files.
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

5 participants