New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Remove the /cluster directory #78995
Comments
/area code-organization |
/cc @andrewsykim |
or potentially break it down and move certain sub-folders out of tree.
|
looking at the references to it... /sig testing /sig scalability |
Is this also relevant to SIG Docs? |
xref -#78543 |
Yes, I think this is a good example of where |
For v1.16: investigate if cluster API is a potential replacement and meets the same level of coverage as /assign @alejandrox1 |
Step 1 we talked about was ... look at all the CI jobs that use the kube-up in the cluster directory and inventory the knobs/settings/configurations they setup or use and cross check if cluster-api/kubeadm allows us to do the same. Both steps can be done in parallel and will need help/effort/coordination on the wg-k8s-infra and sig-testing folks |
I'll start with step 1 right away. |
@alejandrox1 sounds good. please start a google doc or something that we can use to compile notes on the various flags/options |
On the google doc: https://docs.google.com/document/d/1p3c_sOALbEzg2VH2OPz3w9yKwlwaU4jATyS0lqxFzt4/edit?usp=sharing Still got a lot to do for step 1 but will ping when this is ready |
Could I start work on step 2 (i.e. investigating/starting the refactor of I'll assign myself, but please feel free to unassign me if someone else is already working on it! /assign |
/sig cloud-provider |
/assign |
The etcd images are not kubernetes specific. The main thing they do is automatically upgrade etcd one minor at a time, per etcd administration guidelines, when the cluster administrator upgrades etcd to a new version. E.g. if the cluster administrator upgrades to etcd 3.4 and the cluster is currently on 3.1, it upgrades first to 3.2 and then to 3.3 before upgrading to 3.4. So I'm tempted to ask if the etcd community would be willing to own this. If the etcd community was okay with this, the main issue to solve is that the images are published to the k8s.io container repo. cc @gyuho, @ptabor, @wenjiaswe |
Yeah, I think we can implement some mechanism to update docs and container images in the registry for Kubernetes as part of etcd release process. @ptabor @wenjiaswe Any thoughts? |
We could also do this in the etcdadm project, as that is a kubernetes-sigs project and is thus set up to push to k8s repos / follow k8s governance etc. |
The northstar IMHO is that we should get rid of process of 'updating' etcd by running consecutive minor versions of etcd There are multiple benefits of this:
Maybe its naive, but in scope of etcd-v3 minor versions, during work on etcd storage-format documentation I haven't spotted any significant differences in the format. |
I'm a big fan of this approach. If we can get support for this one the etcd side then the solution w/r/t the Curious what others think. |
Yeah always wonder why we even need two separate etcd registries. I am open to dropping the whole container support in our release process, and let downstream project build their own. Or merge onto Kubernetes-community managed registry. |
I think that what has to be done is to have an alternative as reliable as cluster or better, I'm not saying that cluster is great, but I didn't see anything better so far ... and this is very easy to measure with testgrid |
This issue has not been updated in over 1 year, and should be re-triaged. You can:
For more details on the triage process, see https://www.kubernetes.dev/docs/guide/issue-triage/ /remove-triage accepted |
This issue is currently awaiting triage. If a SIG or subproject determines this is a relevant issue, they will accept it by applying the The Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository. |
Sketching up requirements: Here's a snapshot as of Feb27 just to be safe, please see all the comments on the hackmd.
|
I'm not sure this should be a hard requirement, as much appreciation as I have for kubeadm, it does a relatively small portion of cluster bootstrapping and we often need to test things it does not do / reconfigure things it does not directly support configuring anyhow. I'd suggest using kubeadm if we staffed writing something new from scratch, but I don't think we should preclude pre-existing options based on it.
I'm also not sure the statement about kOps in particular is accurate, we actually used to run kOps on Kubernetes PRs as recently as 2019 and it worked fine. We lost this due to the AWS bill going unpaid somewhere between Amazon and the CNCF leading to the account being terminated and no ability to run the job. #73444 (comment) We didn't spin up kops-GCE instead because we already had cluster-up and there was a strong push for Cluster-API, but kops-GCE works and passes tests. It's also relatively mature. kubespray does not support kubernetes @ HEAD and CAPA is not passing tests / adds overhead like the local cluster, AIUI. kops is already supported by kubetest + kubetest2, boskos, etc. and is passing tests on GCE + AWS. In theory it supports other providers but I don't think we have any CI there yet.
This seems impractical. Most of the tricky part here is jobs setting kube-up environment variables that reconfigure the cluster. A human will have to dive into the scripts and see what each env ultimately does and then remap it to any other tool used. Unless that tool is also written in bash, we won't have drop in identical behavior, a lot of it is expanding variables inside bash-generated configuration files. We could target the most important jobs and work from there. |
@upodroid brought a related topic today, he is working on adding ARM64 support to the CI and modifyng the cluster scripts to do that #120144 I can see kops already is running arm64 (@justinsb) https://testgrid.k8s.io/google-aws#kops-aws-arm64-ci |
I'll create the arm64 equivalent of https://testgrid.k8s.io/sig-release-master-blocking#gce-ubuntu-master-containerd on GCE using kops and then have a conversation about migrating the amd64 one to kops. |
for k0ps related questions we have #kops-dev on k8s slack. |
@upodroid how close are we to doing this? 1.31 perhaps? |
For years we have publicly stated that the /cluster directory is deprecated and not maintained. However every cycle it is updated and there are bugs found+fixed by sig-cluster-lifecycle.
I'd like to enumerate what needs to get done in order for us to wholesale remove the /cluster directory.
/assign @dims @spiffxp @justinsb @timothysc
/cc @liggitt @neolit123
The text was updated successfully, but these errors were encountered: