Respect for #Ceph.
Managed to kill a SATA-controller, let the cluster run with one of three legs cut off for a week (decided halfway through to mark the dead OSDs out, so I wouldn‘t have stale objects on my 2/1-pools).
Got it to recover/rebalance again, and while scrubbing two drives (15 y/o) started throwing read errors (on a previously healthy node).
Well, gotta get new drives I guess.
And never did I lose any data, or had Ceph refuse to operate with two of the three working.


