Replies: 1 comment 1 reply
-
It's expected that the osd purge will also delete the osd auth. That said, I also saw recently during a test of osd purge that the osd auth did not get purged as expected and I had to delete the auth before I could create a new OSD. I believe what happened is that when the OSD was purged, the operator immediately tried to re-create the OSD before I had a chance to wipe the disk, so the same OSD tried to start up again and created the auth again. If that's the case, the operator should be stopped until the disk is wiped and ready to be re-created. |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi,
Is the
ceph auth del osd.<ID>
command necessary when replacing an OSD? I ask this because in my case, the OSD was not recreated and I got the following error in the rook-ceph-operator log:After running the
ceph auth del osd.0
command and deleting therook-ceph-operator
pod, the OSD was recreated as expected.Context
CrashLoopBackoff
. This led me to vMotion the disk to a new storage and follow https://rook.io/docs/rook/v1.11/Storage-Configuration/Advanced/ceph-osd-mgmt/#remove-an-osd to remove the OSD, wipe the disk, and add it back to the cluster.osd.0
, located at/dev/sdb
on nodeworker01.k8s.redacted
.osd.0
is down, many PGs weredegraded+undersized
.osd.0
with a job, but failed, so I had to purge it manually.ceph osd out osd.0
command, backfilling did not happen, so I had to proceed without waiting for all of the PGs to beactive+clean
. (Perhaps the PGs were undersized?)Is this only specific to my situation? Would it be helpful if I open a PR to edit the document?
Thanks.
Beta Was this translation helpful? Give feedback.
All reactions