Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Promote Usage API to Beta #4622

Open
2 of 3 tasks
turkenh opened this issue Sep 14, 2023 · 4 comments
Open
2 of 3 tasks

Promote Usage API to Beta #4622

turkenh opened this issue Sep 14, 2023 · 4 comments
Labels
enhancement New feature or request feature-lifecycle roadmap Issues that have priority and are included in the roadmap, or are candidates to add to the roadmap
Milestone

Comments

@turkenh
Copy link
Member

turkenh commented Sep 14, 2023

What problem are you facing?

This is the tracking issue for promoting the Usage API (a.k.a. deletion ordering) to Beta.
With the feature recently merged into the master, we don't have any issues yet, but we will list them here as they appear.

How could Crossplane help solve your problem?

Promote Usage to Beta once we feel confident about it.

Tasks

  1. enhancement stale usage
  2. bug performance roadmap usage
    turkenh
@turkenh turkenh added the enhancement New feature or request label Sep 14, 2023
@jbw976 jbw976 added the roadmap Issues that have priority and are included in the roadmap, or are candidates to add to the roadmap label Nov 2, 2023
@jbw976 jbw976 added this to the v1.16 milestone Nov 2, 2023
@lsviben
Copy link
Contributor

lsviben commented Nov 20, 2023

Added an issue based on Community Slack feedback:
#5042

@stevendborrelli
Copy link
Contributor

I need to be able to duplicate the issue, but I have seen instances where due to long running deletions like for DBs or K8s Clusters when having usage dependences via nested XRs, there can be a 15-minute pause between attempted deletes. We should address it before beta.

@bobh66
Copy link
Contributor

bobh66 commented Jan 3, 2024

In my local implementation of this feature using a kopf operator we see this same problem - it's caused by the garbage collector doing an exponential backoff on the delete retries that maxes out at 1000 seconds (16+ minutes). So when you have something that takes a long time to delete the GC backoff will make it even longer.

I worked around this by explicitly retrying the delete on the "protected" resource when the dependency/usage itself is successfully deleted (the webhook determines that the Usage is no longer required). That seems to speed things up significantly.

I think we discussed this in the original implementation and decided to defer any solution until we had some data on how it was actually working. Explicitly retrying the delete feels risky, but if the resource is already marked for deletion (which it should be) then forcing GC to retry again sooner than it's backoff period should not cause any issues.

@stevendborrelli
Copy link
Contributor

Created issue #5193, this has impacted several users and should be addressed prior to beta graduation.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request feature-lifecycle roadmap Issues that have priority and are included in the roadmap, or are candidates to add to the roadmap
Projects
Status: Backlog
Status: Alpha to Beta
Development

No branches or pull requests

6 participants