Hello David,
so in a 3 node Cluster how should I set min_size if I want that 2 nodes
could fail?
Regards - Willi
Am 28.06.16 um 13:07 schrieb David:
Hi,
This is probably the min_size on your cephfs data and/or metadata
pool. I believe the default is 2, if you have less than 2 replicas
available I/O will stop. See:
http://docs.ceph.com/docs/master/rados/operations/pools/#set-the-number-of-object-replicas
On Tue, Jun 28, 2016 at 10:23 AM, willi.feh...@t-online.de
<mailto:willi.feh...@t-online.de> <willi.feh...@t-online.de
<mailto:willi.feh...@t-online.de>> wrote:
Hello,
I'm still very new to Ceph. I've created a small test Cluster.
ceph-node1
osd0
osd1
osd2
ceph-node2
osd3
osd4
osd5
ceph-node3
osd6
osd7
osd8
My pool for CephFS has a replication count of 3. I've powered of 2
nodes(6 OSDs went down) and my cluster status became critical and
my ceph clients(cephfs) run into a timeout. My data(I had only one
file on my pool) was still on one of the active OSDs. Is this the
expected behaviour that the Cluster status became critical and my
Clients run into a timeout?
Many thanks for your feedback.
Regards - Willi
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com>
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com