I am kind of a newbie but I thought you needed 2 mons working at a minimum.

You should split those away onto some really budget hardware.

//Tu
Hello Tu,

yes that's correct. The mon nodes run as well on the OSD nodes. So I have

3 nodes in total. OSD, MDS and Mon on each Node.

Regards - Willi

Am 03.07.16 um 09:56 schrieb Tu Holmes:

Where are your mon nodes?

Were you mixing mon and OSD together?

Are 2 of the mon nodes down as well?
On Jul 3, 2016 12:53 AM, "Willi Fehler" <willi.feh...@t-online.de> wrote:

> Hello Sean,
>
> I've powered down 2 nodes. So 6 of 9 OSD are down. But my client can't
> write and read anymore from my Ceph mount. Also 'ceph -s' hangs.
>
> pool 1 'cephfs_data' replicated size 3 min_size 1 crush_ruleset 0
> object_hash rjenkins pg_num 300 pgp_num 300 last_change 447 flags
> hashpspool crash_replay_interval 45 stripe_width 0
> pool 2 'cephfs_metadata' replicated size 3 min_size 1 crush_ruleset 0
> object_hash rjenkins pg_num 300 pgp_num 300 last_change 445 flags
> hashpspool stripe_width 0
>
> 2016-07-03 09:49:40.695953 7f3da56f9700  0 -- 192.168.0.5:0/2773396901 >>
> 192.168.0.7:6789/0 pipe(0x7f3da0001f50 sd=3 :0 s=1 pgs=0 cs=0 l=1
> c=0x7f3da0000f20).fault
> 2016-07-03 09:49:44.195029 7f3da57fa700  0 -- 192.168.0.5:0/2773396901 >>
> 192.168.0.6:6789/0 pipe(0x7f3da0005500 sd=4 :0 s=1 pgs=0 cs=0 l=1
> c=0x7f3da00067c0).fault
> 2016-07-03 09:49:50.205788 7f3da55f8700  0 -- 192.168.0.5:0/2773396901 >>
> 192.168.0.6:6789/0 pipe(0x7f3da0005500 sd=3 :0 s=1 pgs=0 cs=0 l=1
> c=0x7f3da0004c40).fault
> 2016-07-03 09:49:52.720116 7f3da57fa700  0 -- 192.168.0.5:0/2773396901 >>
> 192.168.0.7:6789/0 pipe(0x7f3da00023f0 sd=4 :0 s=1 pgs=0 cs=0 l=1
> c=0x7f3da00036b0).fault
>
> Regards - Willi
>
> Am 03.07.16 um 09:36 schrieb Sean Redmond:
>
> It would need to be set to 1
> On 3 Jul 2016 8:17 a.m., "Willi Fehler" <willi.feh...@t-online.de> wrote:
>
>> Hello David,
>>
>> so in a 3 node Cluster how should I set min_size if I want that 2 nodes
>> could fail?
>>
>> Regards - Willi
>>
>> Am 28.06.16 um 13:07 schrieb David:
>>
>> Hi,
>>
>> This is probably the min_size on your cephfs data and/or metadata pool. I
>> believe the default is 2, if you have less than 2 replicas available I/O
>> will stop. See:
>> http://docs.ceph.com/docs/master/rados/operations/pools/#set-the-number-of-object-replicas
>>
>> On Tue, Jun 28, 2016 at 10:23 AM, willi.feh...@t-online.de <
>> willi.feh...@t-online.de> wrote:
>>
>>> Hello,
>>>
>>> I'm still very new to Ceph. I've created a small test Cluster.
>>>
>>>
>>>
>>> ceph-node1
>>>
>>> osd0
>>>
>>> osd1
>>>
>>> osd2
>>>
>>> ceph-node2
>>>
>>> osd3
>>>
>>> osd4
>>>
>>> osd5
>>>
>>> ceph-node3
>>>
>>> osd6
>>>
>>> osd7
>>>
>>> osd8
>>>
>>>
>>>
>>> My pool for CephFS has a replication count of 3. I've powered of 2
>>> nodes(6 OSDs went down) and my cluster status became critical and my ceph
>>> clients(cephfs) run into a timeout. My data(I had only one file on my pool)
>>> was still on one of the active OSDs. Is this the expected behaviour that
>>> the Cluster status became critical and my Clients run into a timeout?
>>>
>>>
>>>
>>> Many thanks for your feedback.
>>>
>>>
>>>
>>> Regards - Willi
>>>
>>>
>>> 
>>>
>>> _______________________________________________
>>> ceph-users mailing list
>>> ceph-users@lists.ceph.com
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>
>>>
>>
>>
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to