-
-
Notifications
You must be signed in to change notification settings - Fork 4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Unable to update Node Groups in place with cluster placement group strategy without EFA #3044
Comments
cc @james-masson ref #2959 |
Consider an EKS cluster deployed across 3 availability zones. A nodegroup that makes use of a "cluster" placement group will only be able to deploy into a single one of these AZs. That's the point of the "cluster" placement group, to put the instances into the same physical rack. It turns out that on initial deployment, this configuration is not a problem. The nodegroup with cluster placement successfully gets deployed into a single AZ of the 3 subnets configured with no errors. Arguably this itself is a bug. However, when it comes time to replace or upgrade the nodegroup, you get the error listed by my colleague @Josephuss This PR tries to fix this issue by adding the concept of a AZ filter into the nodegroups, to allow nodegroups to be deployed into a subset of the subnets that the rest of the cluster is configured with. We're aware that this can be currently worked around by overriding subnet ids. https://github.com/terraform-aws-modules/terraform-aws-eks/blob/master/node_groups.tf#L308 eg.
But this is quite fragile, because:
Hence seeing the general value in an interface like this - it's more practical, understandable and portable, and it has value outside of placement groups too.
|
Description
When using a node group without EFA enabled and a placement group with cluster strategy, updates of the node group fail because the auto scaling group does not restrict the list of availability zones.
The node group with cluster placement successfully gets deployed into a single AZ of the 3 subnets configured with no errors, However, replacing or upgrading the node group fails as listed below due to the availability zone is not filtered and the update does not take place unless the subnet id is overridden in the configs.
Versions
Module version [Required]: 20.8.5
Terraform version: 1.5.7
Provider version(s): 5.40.0
Reproduction Code
This is a copy of the managed node group example with a placement group created. The default placement group strategy is cluster.
Steps to reproduce the behavior:
Expected behavior
Node group is increased as requested.
Actual behavior
Node group update may fail with:
Additional context
The text was updated successfully, but these errors were encountered: