New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[release-0.58] Use exponential backoff for failing migrations #8784
[release-0.58] Use exponential backoff for failing migrations #8784
Conversation
With this patch when migrating a VM fails an increasingly exponential backoff will be applied before retrying. Signed-off-by: Antonio Cardace <acardace@redhat.com>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
/approve
[APPROVALNOTIFIER] This PR is APPROVED This pull-request has been approved by: enp0s3 The full list of commands accepted by this bot can be found here. The pull request process is described here
Needs approval from an approver in each of these files:
Approvers can indicate their approval by writing |
@kubevirt-bot: The following tests failed, say
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository. I understand the commands that are listed here. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I am in general not too concerned about this change, but it looks like this would in general add a slightly weird cross-object relationship which is hard to understand from outside and very uncommon from a general usage perspective. The eviction problem is however real and definitely deserves a solution.
I discussed a few options with @xpivarc and I think, in order to backport this it would make sense to limit this for now the migrations which are solely created by the eviction webhook (e.g. adding a label and filtering by it) and leaving the general migration behaviour unchanged.
On main we could then decide if we want to keep it this way, or try moving the whole backoff logic to the webhook.
What do you think?
/hold |
I see your point, it does make sense.
Yes, I think this could work.
So I'd first introduce this change here in release-0.58 and then implement it in main? It would be a first. |
ah no, lets' follow the usual flow: main and back-port. |
/retest-required |
/unhold @acardace I Don't know what is the desired strategy in this case - to have a manual backport that contains the 2 PRs, or to have 2 automated backports. IMO we don't take a high risk by separating it into 2 separate automated backports. |
/retest-required |
+1 two should be fine. |
This is an automated cherry-pick of #8530
/assign acardace