-
First of all: this isn't urgent. Any potential data loss at this point is tolerable. I have many copies of personal things (e.g. family photos). Everything that's at risk, if lost, really just means I'll have to re-rip a lot of DVDs. Not fun but not the end of the world. Ok! A couple years ago, I got a kubernetes cluster up with 3 worker nodes and a bunch of disks, and I installed rook/ceph. Good so far. Used it for a while, pretty cool stuff. Then 2 OSDs went down at the same time on 1 worker node. I ordered new disks and made sure I had backups of everything important. It's a good thing I did because, before the new disks arrived, a different worker node's power supply died. State so far: 1 worker node completely down, 1 worker node up but 3/5 OSDs up, and 1 worker node up (6/6 OSDs up). I figured I should start with the down worker node. I got a new power supply, but it still wouldn't turn on. I could go on, but it's just computer-building details. The important thing is that I kept buying different replacement parts while experiencing failure after failure, and I grew very stressed about the whole project. I abandoned it and turned everything off. Two years pass At this point, my goal is to get it working enough to make a backup, shut everything back down, then repair or replace the broken node, then start over fresh with a restore from backup. (And then I'll have backups going forward!) I created a new 2-node kubernetes cluster and followed the directions for running an existing ceph cluster in a new kubernetes cluster, and that went great! But I still have 2 OSDs down. I tried replacing one of them, but Operator logs keep repeating this message, twice every 30 seconds:
As I say at the top, it's not the end of the world if I have to just give up and start over completely. That said, in theory, I have one worker node up with 6/6 OSDs up and happy, so, again in theory, the data does exist, even if it's basically RAID-0 at this point. Any ideas for how I can get ceph to accept that the down OSDs are never coming back, but here are some new empty disks? Or if that's a lost cause, any ideas for how I can focus on the one node which should (again in theory) have 1 complete copy of everything, and if there's a way for me to access and back up what's there? Here's ceph status:
I set The following commands hang indefinitely with no output:
Is this a lost cause, or do you have some ideas? Either way, thank you so much for reading. Definitely please do point out any of the very stupid decisions I made, as I'm sure there were more than a few. |
Beta Was this translation helpful? Give feedback.
Replies: 3 comments 3 replies
-
That's quite an interesting journey! Some questions for clarity...
|
Beta Was this translation helpful? Give feedback.
-
Great to hear that helped. Someday if you're at Kubecon, do stop by the Rook booth to say hi. :)
Agreed, upgrading at the same time as disaster recovery isn't generally expect, just curious on the versions in case follow up was still needed. |
Beta Was this translation helpful? Give feedback.
-
I said I'd try and restore health to the existing cluster before starting over and restoring from backup, as an exercise. As suggested, I purged all the no-longer-in-use OSDs. Then over the next week or so, all the Anticlimactic—but in the good way! |
Beta Was this translation helpful? Give feedback.
Next step I'd suggest is to reduce the pool replication to 2 and see if you can get healthy PGs. You can either update the CephBlockPool CR with the new replication.size. Since the ceph status is HEALTH_WARN, rook should be able to reconcile the pool changes successfully.
Then you could purge the OSDs on the third host since you don't expect to bring that host back up.
Also, what version of Rook and Ceph? The latest (or near-latest) versions?