Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Adjust V level for scheduler messages #17438

Merged
merged 1 commit into from
Nov 25, 2015

Conversation

spiffxp
Copy link
Member

@spiffxp spiffxp commented Nov 18, 2015

The "Combined requested resources" message becomes excessive as
the cluster fills up.

Put an explicit V(2) on the only other scheduler Infof call that didn't
have V specified already.

@k8s-bot
Copy link

k8s-bot commented Nov 18, 2015

Can one of the admins verify that this patch is reasonable to test? (reply "ok to test", or if you trust the user, reply "add to whitelist")

If this message is too spammy, please complain to ixdy.

1 similar comment
@k8s-bot
Copy link

k8s-bot commented Nov 18, 2015

Can one of the admins verify that this patch is reasonable to test? (reply "ok to test", or if you trust the user, reply "add to whitelist")

If this message is too spammy, please complain to ixdy.

@spiffxp
Copy link
Member Author

spiffxp commented Nov 18, 2015

/cc @kubernetes/sig-scalability

Started running into #14216 behavior again. I'm pretty confident something is wrong with our journald setup, but in the meantime, this seems to better follow the conventions from https://github.com/kubernetes/kubernetes/blob/master/docs/devel/logging.md

@k8s-github-robot
Copy link

Labelling this PR as size/XS

@k8s-github-robot k8s-github-robot added the size/XS Denotes a PR that changes 0-9 lines, ignoring generated files. label Nov 18, 2015
@@ -34,7 +34,7 @@ func calculateScore(requested int64, capacity int64, node string) int {
return 0
}
if requested > capacity {
glog.Infof("Combined requested resources %d from existing pods exceeds capacity %d on node %s",
glog.V(4).Infof("Combined requested resources %d from existing pods exceeds capacity %d on node %s",
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I'm torn between if this should be V(3) (Extended information about changes) or V(4) (Debug level verbosity). I veer towards V(3) as it's extended information even though it's about why the pod didn't land on a node.

Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Works for me. Could probably stand to have someone check out V level consistency across the codebase rather than case-by-case. FWIW I believe you get all of the reasons a pod failed to fit on a given node in the other V(2) statement in this PR.

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Sounds like a good idea. I'll check to see if there is an issue already created for that.

@ashcrow
Copy link
Contributor

ashcrow commented Nov 18, 2015

Looks good to me.

@davidopp

@spiffxp
Copy link
Member Author

spiffxp commented Nov 19, 2015

/cc @wojtek-t

@timothysc
Copy link
Member

lgtm.

@@ -34,7 +34,7 @@ func calculateScore(requested int64, capacity int64, node string) int {
return 0
}
if requested > capacity {
glog.Infof("Combined requested resources %d from existing pods exceeds capacity %d on node %s",
glog.V(3).Infof("Combined requested resources %d from existing pods exceeds capacity %d on node %s",
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Can you change it to V(2)?

@wojtek-t
Copy link
Member

ok to test

@wojtek-t
Copy link
Member

@spiffxp High level question: do you observe a lot of those logs in your experiments? Those shouldn't be that "popular".

@spiffxp
Copy link
Member Author

spiffxp commented Nov 19, 2015

@wojtek-t Yes, I start seeing more and more of these as the cluster fills up. In my case, they start becoming excessive to the point of blocking after ~1200 / 3000 pods have made it to Running on a 100 m3.medium cluster in AWS. Has the feeling of the algorithm rejecting more and more nodes with each pod scheduled, then forgetting that info when it comes time to schedule the next pod.

The "Combined requested resources" message becomes excessive as
the cluster fills up, drop it down to V(2)

Put an explicit V(2) on the only other scheduler Infof call that didn't
have V specified already.
@k8s-bot
Copy link

k8s-bot commented Nov 19, 2015

GCE e2e test build/test passed for commit a7178574d29e5a624ca5cbd95268133a55132e0c.

@wojtek-t wojtek-t assigned wojtek-t and unassigned davidopp Nov 19, 2015
@wojtek-t
Copy link
Member

LGTM

@wojtek-t wojtek-t added the lgtm "Looks good to me", indicates that a PR is ready to be merged. label Nov 19, 2015
@k8s-bot
Copy link

k8s-bot commented Nov 19, 2015

GCE e2e test build/test passed for commit fad1968.

@k8s-github-robot
Copy link

The author of this PR is not in the whitelist for merge, can one of the admins add the 'ok-to-merge' label?

@davidopp
Copy link
Member

LGTM

@k8s-github-robot
Copy link

@k8s-bot test this [submit-queue is verifying that this PR is safe to merge]

@k8s-bot
Copy link

k8s-bot commented Nov 25, 2015

GCE e2e test build/test passed for commit fad1968.

@k8s-github-robot
Copy link

Automatic merge from submit-queue

k8s-github-robot pushed a commit that referenced this pull request Nov 25, 2015
@k8s-github-robot k8s-github-robot merged commit a26a1ae into kubernetes:master Nov 25, 2015
@spiffxp spiffxp deleted the sched-v-levels branch November 30, 2015 19:39
k8s-github-robot referenced this pull request Dec 1, 2015
…8-upstream-release-1.1

Auto commit by PR queue bot
shyamjvs referenced this pull request in shyamjvs/kubernetes Dec 1, 2016
…k-of-#17438-upstream-release-1.1

Auto commit by PR queue bot
shouhong referenced this pull request in shouhong/kubernetes Feb 14, 2017
…k-of-#17438-upstream-release-1.1

Auto commit by PR queue bot
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
lgtm "Looks good to me", indicates that a PR is ready to be merged. size/XS Denotes a PR that changes 0-9 lines, ignoring generated files.
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

9 participants