Subj : Re: CEPH To : deon From : tassiebob Date : Wed Oct 16 2024 18:44:30 de> So this weekend, I did some updates to the hosts running ceph (updating de> packages, etc), and rebooted each host after their updates (one at a de> time). de> de> While I didnt do much testing for stuff being accessible while a host de> was down, it all appeared to be ok - even though there was a delay I de> guess while ceph figured out a node was down and had to shuffle around de> who was the next "master" to handle the IO. Nice. Maybe I should add it to the list of things to look at sometime (Christmas maybe)... de> Pretty happy with this setup - I was prevously using a proprietary file de> system, which I had to nurse if I rebooted nodes - and occassionally de> drives would go offline, especially if there was busy I/O going on (all de> three nodes are VMs of the same host). I had that kind of experience with a docker swarm not that long ago. I had 3 manager nodes and 2 workers. Upgraded one of the manager nodes and the swarm fell apart. Probably something specific to the specific from/to versions, as it had always worked prior, and since (until I retired the swarm maybe a month ago). de> So the only thing I need to figure out (learn) if single node dies, de> rebuilding back the third node and hopefully not loosing data along the de> way. I'll tackle that when I get to it... ;) :-) --- Mystic BBS v1.12 A48 (Linux/64) * Origin: TassieBob BBS, Hobart, Tasmania (1337:2/106) .