Skip to content

Disaster recovery (1/3 nodes down; 2 OSDs also down) #14102

Answered by travisn
daaang asked this question in Q&A
Discussion options

You must be logged in to vote

Next step I'd suggest is to reduce the pool replication to 2 and see if you can get healthy PGs. You can either update the CephBlockPool CR with the new replication.size. Since the ceph status is HEALTH_WARN, rook should be able to reconcile the pool changes successfully.

Then you could purge the OSDs on the third host since you don't expect to bring that host back up.

Also, what version of Rook and Ceph? The latest (or near-latest) versions?

Replies: 3 comments 3 replies

Comment options

You must be logged in to vote
3 replies
@daaang
Comment options

@travisn
Comment options

Answer selected by daaang
@daaang
Comment options

Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
0 replies
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants