Skip to content

Downtime during migration to V2 #1650

@tibin-mfl

Description

@tibin-mfl

I am trying to upgrade alb ingress controller v1.1.8 to new LB controller v2.0.0, as per this doc -> https://kubernetes-sigs.github.io/aws-load-balancer-controller/guide/upgrade/migrate_v1_v2/

I tried upgrading in my dev cluster and we saw that we are facing downtime during the upgrade process.

Result of the test during the upgrade process 👇

Test-1:

fortio load -a -c 50 -qps 500 -t 300s http://xxx-2048game-2048ingr-6fa0-xxx.ap-south-1.elb.amazonaws.com 
Code 200 : 148983 (99.3 %)
Code 503 : 1017 (0.7 %)

Test-2:

fortio load -a -c 50 -qps 500 -t 200s http://xxx-2048game-2048ingr-6fa0-xxx.ap-south-1.elb.amazonaws.com 
Code 200 : 87031 (87.0 %)
Code 503 : 12969 (13.0 %)

Currently, during the upgrade process before the targets are registered to the new target group, the target group is updated in ALB listeners, so where there are 0 targets in the new target group the ALB listener will be pointing to that target group, that is causing the downtime.
This shouldn’t be the case ALB listener should be updated only once the targets are registered to the new target group.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions