Skip to content
Closed
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 2 additions & 0 deletions go.mod
Original file line number Diff line number Diff line change
Expand Up @@ -91,3 +91,5 @@ require (
sigs.k8s.io/structured-merge-diff/v4 v4.4.2 // indirect
sigs.k8s.io/yaml v1.4.0 // indirect
)

replace github.com/aws-controllers-k8s/runtime => github.com/gustavodiaz7722/ack-runtime v0.51.0
4 changes: 2 additions & 2 deletions go.sum
Original file line number Diff line number Diff line change
Expand Up @@ -2,8 +2,6 @@ github.com/aws-controllers-k8s/ec2-controller v1.0.7 h1:7MDu2bq8NFKbgzzgHYPFRT7b
github.com/aws-controllers-k8s/ec2-controller v1.0.7/go.mod h1:PvsQehgncHgcu9FiY13M45+GkVsKI98g7G83SrgH7vY=
github.com/aws-controllers-k8s/kms-controller v1.0.2 h1:v8nh/oaX/U6spCwBDaWyem7XXpzoP/MnkJyEjNOZN9s=
github.com/aws-controllers-k8s/kms-controller v1.0.2/go.mod h1:BeoijsyGjJ9G5VcDjpFdxBW0IxaeKXYX497XmUJiPSQ=
github.com/aws-controllers-k8s/runtime v0.52.0 h1:Q5UIAn6SSBr60t/DiU/zr6NLBlUuK2AG3yy2ma/9gDU=
github.com/aws-controllers-k8s/runtime v0.52.0/go.mod h1:OkUJN+Ds799JLYZsMJrO2vDJ4snxUeHK2MgrQHbU+Qc=
github.com/aws-controllers-k8s/sns-controller v1.0.11 h1:nnkywTHzO64y7RrrfoPNyYf1TOkkQHtlg+S0jEPKUZ8=
github.com/aws-controllers-k8s/sns-controller v1.0.11/go.mod h1:ODQIDZR3hHQqcyif4UXVFQfEzTaWU1jqFtVr83K2p9M=
github.com/aws/aws-sdk-go v1.49.0 h1:g9BkW1fo9GqKfwg2+zCD+TW/D36Ux+vtfJ8guF4AYmY=
Expand Down Expand Up @@ -90,6 +88,8 @@ github.com/google/pprof v0.0.0-20241029153458-d1b30febd7db h1:097atOisP2aRj7vFgY
github.com/google/pprof v0.0.0-20241029153458-d1b30febd7db/go.mod h1:vavhavw2zAxS5dIdcRluK6cSGGPlZynqzFM8NdvU144=
github.com/google/uuid v1.6.0 h1:NIvaJDMOsjHA8n1jAhLSgzrAzy1Hgr+hNrb57e+94F0=
github.com/google/uuid v1.6.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo=
github.com/gustavodiaz7722/ack-runtime v0.51.0 h1:dgf25lfg5r1kjJtHzdbNrMtQVloYn77WLYi1X41NXhw=
github.com/gustavodiaz7722/ack-runtime v0.51.0/go.mod h1:OkUJN+Ds799JLYZsMJrO2vDJ4snxUeHK2MgrQHbU+Qc=
github.com/itchyny/gojq v0.12.6 h1:VjaFn59Em2wTxDNGcrRkDK9ZHMNa8IksOgL13sLL4d0=
github.com/itchyny/gojq v0.12.6/go.mod h1:ZHrkfu7A+RbZLy5J1/JKpS4poEqrzItSTGDItqsfP0A=
github.com/itchyny/timefmt-go v0.1.3 h1:7M3LGVDsqcd0VZH2U+x393obrzZisp7C0uEe921iRkU=
Expand Down
2 changes: 1 addition & 1 deletion test/e2e/requirements.txt
Original file line number Diff line number Diff line change
@@ -1 +1 @@
acktest @ git+https://github.com/aws-controllers-k8s/test-infra.git@5a09bbdb961ea14a65b15b63769134125023ac61
acktest @ git+https://github.com/gustavodiaz7722/ack-test-infra.git@075991b980ffbc9177f0427270be707359240f89
4 changes: 2 additions & 2 deletions test/e2e/tests/test_cache_cluster.py
Original file line number Diff line number Diff line change
Expand Up @@ -70,9 +70,9 @@ def get_and_assert_status(ref: k8s.CustomResourceReference, expected_status: str
assert cr['status']['cacheClusterStatus'] == expected_status

if expected_synced:
condition.assert_synced(ref)
condition.assert_ready(ref)
else:
condition.assert_not_synced(ref)
condition.assert_not_ready(ref)


@pytest.fixture(scope="module")
Expand Down
16 changes: 8 additions & 8 deletions test/e2e/tests/test_replicationgroup.py
Original file line number Diff line number Diff line change
Expand Up @@ -171,7 +171,7 @@ class TestReplicationGroup:
def test_rg_cmd_fromsnapshot(self, rg_cmd_fromsnapshot):
(reference, _) = rg_cmd_fromsnapshot
assert k8s.wait_on_condition(
reference, "ACK.ResourceSynced", "True", wait_periods=90)
reference, "Ready", "True", wait_periods=90)

def test_rg_invalid_primary(self, make_rg_name, make_replication_group, rg_deletion_waiter):
input_dict = {
Expand All @@ -196,7 +196,7 @@ def test_rg_invalid_primary(self, make_rg_name, make_replication_group, rg_delet
def test_rg_update(self, rg_update_input, rg_update):
(reference, _) = rg_update
assert k8s.wait_on_condition(
reference, "ACK.ResourceSynced", "True", wait_periods=90)
reference, "Ready", "True", wait_periods=90)

# desired initial state
cr = k8s.get_resource(reference)
Expand Down Expand Up @@ -246,7 +246,7 @@ def test_rg_update(self, rg_update_input, rg_update):
_ = k8s.patch_custom_resource(reference, patch)
sleep(DEFAULT_WAIT_SECS)
assert k8s.wait_on_condition(
reference, "ACK.ResourceSynced", "True", wait_periods=90)
reference, "Ready", "True", wait_periods=90)

# Assert new state
resource = k8s.get_resource(reference)
Expand All @@ -266,7 +266,7 @@ def test_rg_update(self, rg_update_input, rg_update):
LONG_WAIT_SECS = 180
sleep(LONG_WAIT_SECS)
assert k8s.wait_on_condition(
reference, "ACK.ResourceSynced", "True", wait_periods=90)
reference, "Ready", "True", wait_periods=90)

# assert new tags
assert_spec_tags(rg_id, new_tags)
Expand All @@ -275,7 +275,7 @@ def test_rg_update(self, rg_update_input, rg_update):
def test_rg_fault_tolerance(self, rg_fault_tolerance):
(reference, _) = rg_fault_tolerance
assert k8s.wait_on_condition(
reference, "ACK.ResourceSynced", "True", wait_periods=90)
reference, "Ready", "True", wait_periods=90)

# assert initial state
resource = k8s.get_resource(reference)
Expand All @@ -300,7 +300,7 @@ def test_rg_fault_tolerance(self, rg_fault_tolerance):
_ = k8s.patch_custom_resource(reference, patch)
sleep(DEFAULT_WAIT_SECS)
assert k8s.wait_on_condition(
reference, "ACK.ResourceSynced", "True", wait_periods=90)
reference, "Ready", "True", wait_periods=90)

# assert new state
resource = k8s.get_resource(reference)
Expand All @@ -313,7 +313,7 @@ def test_rg_fault_tolerance(self, rg_fault_tolerance):
_ = k8s.patch_custom_resource(reference, patch)
sleep(DEFAULT_WAIT_SECS)
assert k8s.wait_on_condition(
reference, "ACK.ResourceSynced", "True", wait_periods=90)
reference, "Ready", "True", wait_periods=90)

# assert roles
resource = k8s.get_resource(reference)
Expand Down Expand Up @@ -343,7 +343,7 @@ def test_rg_creation_deletion(self, make_rg_name, make_replication_group, rg_del
"replicationgroup_create_delete", input_dict, input_dict["RG_ID"])

assert k8s.wait_on_condition(
reference, "ACK.ResourceSynced", "True", wait_periods=90)
reference, "Ready", "True", wait_periods=90)

# assertions after initial creation
resource = k8s.get_resource(reference)
Expand Down
12 changes: 6 additions & 6 deletions test/e2e/tests/test_serverless_cache.py
Original file line number Diff line number Diff line change
Expand Up @@ -58,9 +58,9 @@ def get_and_assert_status(ref: k8s.CustomResourceReference, expected_status: str
assert cr['status']['status'] == expected_status

if expected_synced:
condition.assert_synced(ref)
condition.assert_ready(ref)
else:
condition.assert_not_synced(ref)
condition.assert_not_ready(ref)


@pytest.fixture(scope="module")
Expand Down Expand Up @@ -129,7 +129,7 @@ def test_create_update_delete_serverless_cache(self, simple_serverless_cache, el
(ref, _) = simple_serverless_cache

assert k8s.wait_on_condition(
ref, "ACK.ResourceSynced", "True", wait_periods=90
ref, "Ready", "True", wait_periods=90
)
get_and_assert_status(ref, "available", True)

Expand Down Expand Up @@ -162,7 +162,7 @@ def test_create_update_delete_serverless_cache(self, simple_serverless_cache, el

# Wait for update to be synced
assert k8s.wait_on_condition(
ref, "ACK.ResourceSynced", "True", wait_periods=90
ref, "Ready", "True", wait_periods=90
)

# Verify the update was applied
Expand All @@ -178,7 +178,7 @@ def test_upgrade_redis_to_valkey(self, upgrade_serverless_cache, elasticache_cli

# Wait for the serverless cache to be created and become available
assert k8s.wait_on_condition(
ref, "ACK.ResourceSynced", "True", wait_periods=90
ref, "Ready", "True", wait_periods=90
)
get_and_assert_status(ref, "available", True)

Expand Down Expand Up @@ -206,7 +206,7 @@ def test_upgrade_redis_to_valkey(self, upgrade_serverless_cache, elasticache_cli

# Wait for upgrade to be synced
assert k8s.wait_on_condition(
ref, "ACK.ResourceSynced", "True", wait_periods=90
ref, "Ready", "True", wait_periods=90
)

# Wait for it to be available again after upgrade
Expand Down
4 changes: 2 additions & 2 deletions test/e2e/tests/test_serverless_cache_snapshot.py
Original file line number Diff line number Diff line change
Expand Up @@ -64,7 +64,7 @@ def serverless_cache_for_snapshot(elasticache_client):

# Wait for serverless cache to be available
assert k8s.wait_on_condition(
ref, "ACK.ResourceSynced", "True", wait_periods=90
ref, "Ready", "True", wait_periods=90
)

yield ref, cr
Expand Down Expand Up @@ -112,7 +112,7 @@ def test_create_delete_serverless_cache_snapshot(self, simple_serverless_cache_s
(ref, _) = simple_serverless_cache_snapshot

assert k8s.wait_on_condition(
ref, "ACK.ResourceSynced", "True", wait_periods=120
ref, "Ready", "True", wait_periods=120
)

tag_updates = {
Expand Down
2 changes: 1 addition & 1 deletion test/e2e/tests/test_snapshot.py
Original file line number Diff line number Diff line change
Expand Up @@ -82,4 +82,4 @@ class TestSnapshot:
# test create of snapshot while providing KMS key
def test_snapshot_kms(self, snapshot_kms):
(reference, _) = snapshot_kms
assert k8s.wait_on_condition(reference, "ACK.ResourceSynced", "True", wait_periods=15)
assert k8s.wait_on_condition(reference, "Ready", "True", wait_periods=15)
6 changes: 3 additions & 3 deletions test/e2e/tests/test_user.py
Original file line number Diff line number Diff line change
Expand Up @@ -115,7 +115,7 @@ def test_user_nopass(self, user_nopass, user_nopass_input):
(reference, resource) = user_nopass
assert k8s.get_resource_exists(reference)

assert k8s.wait_on_condition(reference, "ACK.ResourceSynced", "True", wait_periods=5)
assert k8s.wait_on_condition(reference, "Ready", "True", wait_periods=5)
resource = k8s.get_resource(reference)
assert resource["status"]["lastRequestedAccessString"] == user_nopass_input["ACCESS_STRING"]

Expand All @@ -124,7 +124,7 @@ def test_user_nopass(self, user_nopass, user_nopass_input):
_ = k8s.patch_custom_resource(reference, user_patch)
sleep(DEFAULT_WAIT_SECS)

assert k8s.wait_on_condition(reference, "ACK.ResourceSynced", "True", wait_periods=5)
assert k8s.wait_on_condition(reference, "Ready", "True", wait_periods=5)
resource = k8s.get_resource(reference)
assert resource["status"]["lastRequestedAccessString"] == new_access_string

Expand All @@ -133,7 +133,7 @@ def test_user_password(self, user_password, user_password_input):
(reference, resource) = user_password
assert k8s.get_resource_exists(reference)

assert k8s.wait_on_condition(reference, "ACK.ResourceSynced", "True", wait_periods=5)
assert k8s.wait_on_condition(reference, "Ready", "True", wait_periods=5)
resource = k8s.get_resource(reference)
assert resource["status"]["authentication"] is not None
assert resource["status"]["authentication"]["type_"] == "password"
Expand Down
4 changes: 2 additions & 2 deletions test/e2e/tests/test_usergroup.py
Original file line number Diff line number Diff line change
Expand Up @@ -66,15 +66,15 @@ def user_group_create(get_user_group_yaml):
class TestUserGroup:
def test_user_group_create_update(self, user_group_create, get_user_group_yaml, bootstrap_resources):
(reference, resource) = user_group_create
assert k8s.wait_on_condition(reference, "ACK.ResourceSynced", "True", wait_periods=15)
assert k8s.wait_on_condition(reference, "Ready", "True", wait_periods=15)

# Update the usergroup to include one more user
updated_user_group = get_user_group_yaml(reference.name)
updated_user_group["spec"]["userIDs"].append(bootstrap_resources.NonDefaultUser)

k8s.patch_custom_resource(reference, updated_user_group)

assert k8s.wait_on_condition(reference, "ACK.ResourceSynced", "True", wait_periods=15)
assert k8s.wait_on_condition(reference, "Ready", "True", wait_periods=15)
resource = k8s.get_resource(reference)
assert len(resource["spec"]["userIDs"]) == 2
assert resource["status"]["status"] == "active"