-
Notifications
You must be signed in to change notification settings - Fork 1.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
СRI image pulling with progress notification #3542
Comments
/sig node |
@byako: The label(s) In response to this:
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository. |
/sig node |
/milestone v1.26 |
/stage alpha |
/remove-label lead-opted-in This design needs more time in SIG Node and would be reviewed during 1.26 for evaluation in 1.27. |
/label tracked/no |
The design is proposed now in KEP PR. |
Just want to clarify - in #3547, it looks like the milestones are targeting v1.26. Based on a prior comment from @derekwaynecarr, we have this enhancement as |
Yes, @rhockenbury, that is the intention, if the proposal is approved or agreed for its details to be ironed out during v1.26 cycle - we'd like to opt this into v1.26 cycle. |
I can opt it in but it's going to be pretty tight to meet the requirements for enhancements freeze since freeze will happen in a few hours. /label lead-opted-in
|
Hello 👋, 1.26 Enhancements Lead here. Unfortunately, this enhancement did not meet requirements for enhancements freeze. If you still wish to progress this enhancement in v1.26, please file an exception request. Thanks! /milestone clear |
@byako exactly. Let's also check with @kubernetes/sig-node-feature-requests that it's tracked for 1.29. 👍 |
At least it was present in tracking project https://github.com/orgs/kubernetes/projects/161 |
Yes, but SIG Node leads have to opt-in as well. |
/label lead-opted-in @byako @saschagrunert who should be marked as a primary contact for this KEP? |
@SergeyKanzhelev I'm fine with being a primary contact, I've some bandwidth for this this quarter. |
Hello @byako 👋, Enhancements team here. Just checking in as we approach enhancements freeze on Friday, 6th October 2023. This enhancement is targeting for stage Here's where this enhancement currently stands:
For this KEP, we would just need to update the following:
The status of this enhancement is marked as |
Changed status to implementable, waiting for sig-node approvers to give an approval for KEP PR. |
That's right, I've tried to address everything there was to address. Let's see if I can get anyone to approve anything. |
Hello 👋, 1.29 Enhancements Lead here. /milestone clear |
@byako I'm not sure whether I can help, but is there a particular person or group that you're waiting for review or approval from? We're running Kubernetes in edge environments with slow downlinks, sometimes as low as 10 Mbps, so this would be really valuable feature and much more friendly than my current hack ( |
@drigz, I'm happy to see that someone else is interested in this feature! I'll try to make this happen in 1.30, which starts about now. The KEP itself consists of two parts: the CRI protocol part and Kubelet implementation part. What needs to be done still is Kubelet implementation part of this KEP, it has to be re-designed. SIG-scalability had concerns about impact of this feature at scale, and there were alternative implementation suggestions. The CRI protocol change appears to be fine, but I failed to secure the approval label, that seems to be just a formality. In last cycle we agreed to remove the Kubelet implementation part from the KEP because there was too little time left to iron it out, but that didn't help either. Now that the new cycle is upon us, I think we can try to come up with new Kubelet implementation suggestion and hopefully both of them will be approved. If not - we can then push harder this time to get at least CRI change in 1.30. I'm not sure how much time I will have in 1.30 cycle for this KEP, but there will be some. We'll see at next SIG-node meeting if I'm still driving this KEP or if someone else has more time for it than I do. |
Thank you for the explanation! The scalability question does sound tricky, as a cluster administrator I can understand Wojciech's point that it's hard to stay abreast of new features and evaluate in advance whether they'll cause problems. I can also imagine a worst case where some highly-replicated pods are updated to new containers, but the registry is overloaded and, say, trickles 1 byte / second to every node, which could generate Did you look into whether there is any backpressure mechanism that could protect the apiserver in a case like this? Our clusters are small so I'm not familiar with these techniques, but API Priority & Fairness seems like it could enforce a cluster-level limit on the rate of progress events. Maybe that would address the question about a protection mechanism. However, I don't know where this traffic would belong - the default levels seem focused on requests that are important for normal cluster operation, rather than informative log events like this, which are primarily of interest to humans observing the cluster. Maybe we need a new Please ignore this question if it's an unhelpful tangent! |
I have not checked any protection measures available yet, but I'll have a look at Priority and Fairness doc, thank you. |
/remove-label lead-opted-in |
The Kubernetes project currently lacks enough contributors to adequately respond to all issues. This bot triages un-triaged issues according to the following rules:
You can:
Please send feedback to sig-contributor-experience at kubernetes/community. /lifecycle stale |
The Kubernetes project currently lacks enough active contributors to adequately respond to all issues. This bot triages un-triaged issues according to the following rules:
You can:
Please send feedback to sig-contributor-experience at kubernetes/community. /lifecycle rotten |
+1 |
/remove-lifecycle rotten |
I for one really feel blind not knowing the progress of an image pull. It really would be useful to have some insights into the speed being pulled (so I can also troubleshoot network related problems), some form of percentage progress, and maybe an estimate of completion. |
Enhancement Description
k/enhancements
) update PR(s): KEP-3542: СRI image pulling with progress notificationk/k
) update PR(s): RFC: CRI PullImageWithProgress implementation PoC kubernetes#118326k/website
) update PR(s):Please keep this description up to date. This will help the Enhancement Team to track the evolution of the enhancement efficiently.
The text was updated successfully, but these errors were encountered: