-
Notifications
You must be signed in to change notification settings - Fork 1.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We鈥檒l occasionally send you account related emails.
Already on GitHub? Sign in to your account
"'subnetworks/default' is not ready" error thrown sporadically due to google_container_cluster adding ranges? #10972
Comments
Hi @grantyoung. We should already retry when APIs return this error. Would you mind sharing your debug log? |
We run into this in our nightly tests a lot. It's definitely not service-specific, I'm gonna reclassify as provider-wide. |
Discussion from triage: a possible way to fix this issue is to implement a retry in the provider |
We already have a retry in the provider for exactly this case:
|
I noticed in TestAccComputeInstance_resourcePolicyUpdate (in this execution) we're hitting a context deadline really early. 2m15s instead of 20m or so. Maybe we're attaching a short one to the retry transport? |
[upstream:ade0a1ec36b97ef2853044110fea0cdd4bec6383] Signed-off-by: Modular Magician <magic-modules@google.com>
[upstream:ade0a1ec36b97ef2853044110fea0cdd4bec6383] Signed-off-by: Modular Magician <magic-modules@google.com>
Community Note
modular-magician
user, it is either in the process of being autogenerated, or is planned to be autogenerated soon. If an issue is assigned to a user, that user is claiming responsibility for the issue. If an issue is assigned tohashibot
, a community member has claimed the issue already.Noticed recently we were getting some "'subnetworks/default' is not ready" errors in our Terraform runs on new environments:
We haven't seen this before but the reason I think this is happening is we have both VMs as well as a GKE cluster being created here at the same time.
When the GKE Cluster starts to be created it's coming in and adding it's additional ranges to the target subnet and that in turn is then "locking" the subnet for sometime - preventing other resources from being created against it.
Rerunning the apply works fine so it feels like this could be handle more gracefully by either doing some retries or holding the cluster build until other resources are done? We could workaround this on our end but at the same time it doesn't seem unreasonable to deploy both a Cluster and other resources to the same subnet so thought it was worth raising.
Terraform Version
1.1.4
Affected Resource(s)
Terraform Configuration Files
Expected Behavior
The provider should gracefully handle any timing clashes caused by the Cluster when on the same subnet
Actual Behavior
The provider creates the Cluster at the same time as VMs. The VMs as a result get a 400 error from the API as the Cluster edits the subnet to add in more ranges.
Steps to Reproduce
Important Factoids
References
b/300616739
The text was updated successfully, but these errors were encountered: