Hello Tu,

yes that's correct. The mon nodes run as well on the OSD nodes. So I have

3 nodes in total. OSD, MDS and Mon on each Node.

Regards - Willi

Am 03.07.16 um 09:56 schrieb Tu Holmes:

Where are your mon nodes?

Were you mixing mon and OSD together?

Are 2 of the mon nodes down as well?

On Jul 3, 2016 12:53 AM, "Willi Fehler" <willi.feh...@t-online.de <mailto:willi.feh...@t-online.de>> wrote:

    Hello Sean,

    I've powered down 2 nodes. So 6 of 9 OSD are down. But my client
    can't write and read anymore from my Ceph mount. Also 'ceph -s' hangs.

    pool 1 'cephfs_data' replicated size 3 min_size 1 crush_ruleset 0
    object_hash rjenkins pg_num 300 pgp_num 300 last_change 447 flags
    hashpspool crash_replay_interval 45 stripe_width 0
    pool 2 'cephfs_metadata' replicated size 3 min_size 1
    crush_ruleset 0 object_hash rjenkins pg_num 300 pgp_num 300
    last_change 445 flags hashpspool stripe_width 0

    2016-07-03 09:49:40.695953 7f3da56f9700  0 --
    192.168.0.5:0/2773396901 <http://192.168.0.5:0/2773396901> >>
    192.168.0.7:6789/0 <http://192.168.0.7:6789/0> pipe(0x7f3da0001f50
    sd=3 :0 s=1 pgs=0 cs=0 l=1 c=0x7f3da0000f20).fault
    2016-07-03 09:49:44.195029 7f3da57fa700  0 --
    192.168.0.5:0/2773396901 <http://192.168.0.5:0/2773396901> >>
    192.168.0.6:6789/0 <http://192.168.0.6:6789/0> pipe(0x7f3da0005500
    sd=4 :0 s=1 pgs=0 cs=0 l=1 c=0x7f3da00067c0).fault
    2016-07-03 09:49:50.205788 7f3da55f8700  0 --
    192.168.0.5:0/2773396901 <http://192.168.0.5:0/2773396901> >>
    192.168.0.6:6789/0 <http://192.168.0.6:6789/0> pipe(0x7f3da0005500
    sd=3 :0 s=1 pgs=0 cs=0 l=1 c=0x7f3da0004c40).fault
    2016-07-03 09:49:52.720116 7f3da57fa700  0 --
    192.168.0.5:0/2773396901 <http://192.168.0.5:0/2773396901> >>
    192.168.0.7:6789/0 <http://192.168.0.7:6789/0> pipe(0x7f3da00023f0
    sd=4 :0 s=1 pgs=0 cs=0 l=1 c=0x7f3da00036b0).fault

    Regards - Willi

    Am 03.07.16 um 09:36 schrieb Sean Redmond:

    It would need to be set to 1

    On 3 Jul 2016 8:17 a.m., "Willi Fehler" <willi.feh...@t-online.de
    <mailto:willi.feh...@t-online.de>> wrote:

        Hello David,

        so in a 3 node Cluster how should I set min_size if I want
        that 2 nodes could fail?

        Regards - Willi

        Am 28.06.16 um 13:07 schrieb David:
        Hi,

        This is probably the min_size on your cephfs data and/or
        metadata pool. I believe the default is 2, if you have less
        than 2 replicas available I/O will stop. See:
        
http://docs.ceph.com/docs/master/rados/operations/pools/#set-the-number-of-object-replicas

        On Tue, Jun 28, 2016 at 10:23 AM, willi.feh...@t-online.de
        <mailto:willi.feh...@t-online.de> <willi.feh...@t-online.de
        <mailto:willi.feh...@t-online.de>> wrote:

            Hello,

            I'm still very new to Ceph. I've created a small test
            Cluster.

            ceph-node1

            osd0

            osd1

            osd2

            ceph-node2

            osd3

            osd4

            osd5

            ceph-node3

            osd6

            osd7

            osd8

            My pool for CephFS has a replication count of 3. I've
            powered of 2 nodes(6 OSDs went down) and my cluster
            status became critical and my ceph clients(cephfs) run
            into a timeout. My data(I had only one file on my pool)
            was still on one of the active OSDs. Is this the
            expected behaviour that the Cluster status became
            critical and my Clients run into a timeout?

            Many thanks for your feedback.

            Regards - Willi

            

            _______________________________________________
            ceph-users mailing list
            ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com>
            http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




        _______________________________________________
        ceph-users mailing list
        ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com>
        http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



    _______________________________________________
    ceph-users mailing list
    ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com>
    http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to