Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

CatalogSource pods not managed via Kubernetes controller #1514

Open
pamelachristie opened this issue May 12, 2020 · 14 comments
Open

CatalogSource pods not managed via Kubernetes controller #1514

pamelachristie opened this issue May 12, 2020 · 14 comments
Assignees
Labels
lifecycle/frozen Indicates that an issue or PR should not be auto-closed due to staleness. triage/unresolved Indicates an issue that can not or will not be resolved.

Comments

@pamelachristie
Copy link

Bug Report

The CatalogSource pod blocks kubectl drain commands as it is not managed by a Kubernetes controller.

What did you do?
Ran kubectl drain <NODE> on a cluster with a CatalogSource pod present.

What did you expect to see?
The kubectl drain should have proceeded without the pod causing a failure

What did you see instead? Under which circumstances?

cannot delete Pods not managed by ReplicationController, ReplicaSet, Job, DaemonSet or StatefulSet (use --force to override): ibm-system/addon-catalog-source-f2wnd

This result is due to the pod being not being managed by a Kubernetes controller.

Environment

  • operator-lifecycle-manager version:
    0.14.1

  • Kubernetes version information:
    1.16

  • Kubernetes cluster kind:
    IBM Cloud

Possible Solution
Create a deployment or some other kind of Kubernetes controller to manage the pod rather than just the CatalogSource custom resource.

Additional context

@kramvan1
Copy link
Contributor

kramvan1 commented Jun 8, 2020

Related to #1421

@kfox1111
Copy link

We've run into this too.

@kramvan1
Copy link
Contributor

@ecordell I think this needs a higher priority as it's a valid scenario for both upstream and RH.

@stale
Copy link

stale bot commented Aug 18, 2020

This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions.

@stale stale bot added the wontfix label Aug 18, 2020
@kramvan1
Copy link
Contributor

This is still valid bug

@openshift-ci-robot openshift-ci-robot added triage/unresolved Indicates an issue that can not or will not be resolved. and removed wontfix labels Aug 19, 2020
@mhaideibm
Copy link

mhaideibm commented Sep 1, 2020

@kramvan1 we run into the same problem. As we need to replace Ubuntu 18 nodes with Ubuntu 16. Is there any outlook for this problem as this blocks us from going forward. Don't won't to drop that node when it is not backed up by ReplicaSet!

@kramvan1
Copy link
Contributor

@mhaideibm I think we need to get this on the roadmap for OLM. https://docs.google.com/document/d/1Zuv-BoNFSwj10_zXPfaS9LWUQUCak2c8l48d0-AhpBw/edit#heading=h.8ngolbigvi7q

@pamelachristie
Copy link
Author

Bump

@stale
Copy link

stale bot commented Dec 11, 2020

This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions.

@stale stale bot added the wontfix label Dec 11, 2020
@pamelachristie
Copy link
Author

Bump

@stale stale bot removed the wontfix label Dec 11, 2020
@stale
Copy link

stale bot commented Mar 12, 2021

This issue has been automatically marked as stale because it has not had any recent activity. It will be closed in 7 days if no further activity occurs. Thank you for your contribution.
For more help on your issue, check out the olm-dev channel on the kubernetes slack [1] and the OLM Dev Working Group [2] [1] https://kubernetes.slack.com/archives/C0181L6JYQ2 [2] https://github.com/operator-framework/community#operator-lifecycle-manager-wg

@stale stale bot added the lifecycle/stale Denotes an issue or PR has remained open with no activity and has become stale. label Mar 12, 2021
@ecordell
Copy link
Member

There is a tracking kube bug here: kubernetes/kubernetes#57049

Does drain with --force not solve the problem until this is addressed upstream?

@William-Newshutz
Copy link

A drain with --force is not a viable solution in this case. The pod had to be deleted with --force (which does not truly delete) before the drain could continue. This meant manual intervention because it blocked drain.

@exdx exdx added lifecycle/frozen Indicates that an issue or PR should not be auto-closed due to staleness. and removed lifecycle/stale Denotes an issue or PR has remained open with no activity and has become stale. labels May 5, 2021
@philomory
Copy link

I noticed the PR for this was closed, is there any outlook on a replacement PR to actually get this addressed?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
lifecycle/frozen Indicates that an issue or PR should not be auto-closed due to staleness. triage/unresolved Indicates an issue that can not or will not be resolved.
Projects
None yet
Development

No branches or pull requests

9 participants