You may also want to check your 'min_size'... if it's 2, then you'll be incomplete even with 1 complete copy.
ceph osd dump | grep pool You can reduce the min size with the following syntax: ceph osd pool set <poolname> min_size 1 Thanks, Michael J. Kidd Sent from my mobile device. Please excuse brevity and typographical errors. On Apr 19, 2014 12:50 PM, "Jean-Charles Lopez" <jc.lo...@inktank.com> wrote: > Hi again > > Looked at your ceph -s. > > You have only 2 OSDs, one on each node. The default replica count is 2, > the default crush map says each replica on a different host, or may be you > set it to 2 different OSDs. Anyway, when one of your OSD goes down, Ceph > can no longer find another OSDs to host the second replica it must create. > > Looking at your crushmap we would know better. > > Recommendation: for testing efficiently and most options available, > functionnally speaking, deploy a cluster with 3 nodes, 3 OSDs each is my > best practice. > > Or make 1 node with 3 OSDs modifying your crushmap to "choose type osd" in > your rulesets. > > JC > > > On Saturday, April 19, 2014, Gonzalo Aguilar Delgado < > gagui...@aguilardelgado.com> wrote: > >> Hi, >> >> I'm building a cluster where two nodes replicate objects inside. I found >> that shutting down just one of the nodes (the second one), makes everything >> "incomplete". >> >> I cannot find why, since crushmap looks good to me. >> >> after shutting down one node >> >> cluster 9028f4da-0d77-462b-be9b-dbdf7fa57771 >> health HEALTH_WARN 192 pgs incomplete; 96 pgs stuck inactive; 96 pgs >> stuck unclean; 1/2 in osds are down >> monmap e9: 1 mons at {blue-compute=172.16.0.119:6789/0}, election >> epoch 1, quorum 0 blue-compute >> osdmap e73: 2 osds: 1 up, 2 in >> pgmap v172: 192 pgs, 3 pools, 275 bytes data, 1 objects >> 7552 kB used, 919 GB / 921 GB avail >> 192 incomplete >> >> >> Both nodes has WD Caviar Black 500MB disk with btrfs filesystem on it. >> Full disk used. >> >> I cannot understand why does not replicate to both nodes. >> >> Someone can help? >> >> Best regards, >> > > > -- > Sent while moving > Pardon my French and any spelling &| grammar glitches > > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com