You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Deviation from expected behavior:
The daily octopus upgrade test is failing. See this test for example.
The test log shows that the rgw daemon is not updating, then the test times out and fails:
2021-11-03 00:25:29.028371 I | testutil: waiting for 1 deployments (found 0) with label app=rook-ceph-rgw,ceph-version!=15.2.15-0 in namespace upgrade-ns
The operator log shows that there is an unexpected version in the octopus image. Since it looks like a downgrade, rgw refuses to upgrade even though all the other daemons including mon,mgr,osd choose to upgrade.
2021-11-03 22:19:10.694041 I | ceph-cluster-controller: CR has changed for "upgrade-ns". diff= v1.ClusterSpec{
CephVersion: v1.CephVersionSpec{
- Image: "quay.io/ceph/ceph:v15",
+ Image: "quay.io/ceph/daemon-base:latest-octopus-devel",
AllowUnsupported: false,
},
Storage: {UseAllNodes: true, Config: {"databaseSizeMB": "1024", "journalSizeMB": "1024"}, Selection: {UseAllDevices: &true}},
Annotations: nil,
... // 21 identical fields
}
2021-11-03 22:19:10.694939 I | ceph-cluster-controller: reconciling ceph cluster in namespace "upgrade-ns"
2021-11-03 22:19:10.718721 I | op-mon: parsing mon endpoints: a=10.101.175.6:6789
2021-11-03 22:19:10.770814 I | ceph-cluster-controller: detecting the ceph image version for image quay.io/ceph/daemon-base:latest-octopus-devel...
2021-11-03 22:19:15.011090 I | ceph-cluster-controller: detected ceph image version: "15.2.14-133 octopus"
2021-11-03 22:19:15.032465 I | ceph-cluster-controller: validating ceph version from provided image
2021-11-03 22:19:15.052961 I | op-mon: parsing mon endpoints: a=10.101.175.6:6789
2021-11-03 22:19:15.059380 I | cephclient: writing config file /var/lib/rook/upgrade-ns/upgrade-ns.config
2021-11-03 22:19:15.059739 I | cephclient: generated admin config in /var/lib/rook/upgrade-ns
2021-11-03 22:19:15.816626 E | ceph-cluster-controller: failed to determine if we should upgrade or not. image spec version 15.2.14-133 octopus is lower than the running cluster version 15.2.15-0 octopus, downgrading is not supported
2021-11-03 22:19:15.816714 I | ceph-cluster-controller: cluster "upgrade-ns": version "15.2.14-133 octopus" detected for image "quay.io/ceph/daemon-base:latest-octopus-devel"
While the downgraded version is unexpected, the operator should downgrade the daemons consistently instead of proceeding with some and skipping others.
The operator is failing to decide if it's an upgrade here, and the isUpgrade flag is not set as seen later in the method. If the images changed in any way, we should just assume it's an upgrade.
Expected behavior:
The daily tests should pass consistently.
The text was updated successfully, but these errors were encountered:
Is this a bug report or feature request?
Deviation from expected behavior:
The daily octopus upgrade test is failing. See this test for example.
The test log shows that the rgw daemon is not updating, then the test times out and fails:
The operator log shows that there is an unexpected version in the octopus image. Since it looks like a downgrade, rgw refuses to upgrade even though all the other daemons including mon,mgr,osd choose to upgrade.
While the downgraded version is unexpected, the operator should downgrade the daemons consistently instead of proceeding with some and skipping others.
The operator is failing to decide if it's an upgrade here, and the
isUpgrade
flag is not set as seen later in the method. If the images changed in any way, we should just assume it's an upgrade.Expected behavior:
The daily tests should pass consistently.
The text was updated successfully, but these errors were encountered: