Hi Caspar,

Yes, cluster was working fine with number of PGs per OSD warning up until
now. I am not sure how to recover from stale down/inactive PGs. If you
happen to know about this can you let me know?

Current State:

[root@fre101 ~]# ceph -s
2019-01-04 05:22:05.942349 7f314f613700 -1 asok(0x7f31480017a0)
AdminSocketConfigObs::init: failed: AdminSocket::bind_and_listen: failed to
bind the UNIX domain socket to
'/var/run/ceph-guests/ceph-client.admin.1053724.139849638091088.asok': (2)
No such file or directory
  cluster:
    id:     adb9ad8e-f458-4124-bf58-7963a8d1391f
    health: HEALTH_ERR
            3 pools have many more objects per pg than average
            505714/12392650 objects misplaced (4.081%)
            3883 PGs pending on creation
            Reduced data availability: 6519 pgs inactive, 1870 pgs down, 1
pg peering, 886 pgs stale
            Degraded data redundancy: 42987/12392650 objects degraded
(0.347%), 634 pgs degraded, 16 pgs undersized
            125827 slow requests are blocked > 32 sec
            2 stuck requests are blocked > 4096 sec
            too many PGs per OSD (2758 > max 200)

  services:
    mon: 3 daemons, quorum ceph-mon01,ceph-mon02,ceph-mon03
    mgr: ceph-mon03(active), standbys: ceph-mon01, ceph-mon02
    osd: 39 osds: 39 up, 39 in; 76 remapped pgs
    rgw: 1 daemon active

  data:
    pools:   18 pools, 54656 pgs
    objects: 6051k objects, 10944 GB
    usage:   21933 GB used, 50688 GB / 72622 GB avail
    pgs:     11.927% pgs not active
             42987/12392650 objects degraded (0.347%)
             505714/12392650 objects misplaced (4.081%)
             48080 active+clean
             3885  activating
             1111  down
             759   stale+down
             614   activating+degraded
             74    activating+remapped
             46    stale+active+clean
             35    stale+activating
             21    stale+activating+remapped
             9     stale+active+undersized
             9     stale+activating+degraded
             5     stale+activating+undersized+degraded+remapped
             3     activating+degraded+remapped
             1     stale+activating+degraded+remapped
             1     stale+active+undersized+degraded
             1     remapped+peering
             1     active+clean+remapped
             1     activating+undersized+degraded+remapped

  io:
    client:   0 B/s rd, 25397 B/s wr, 4 op/s rd, 4 op/s wr

I will update number of PGs per OSD once these inactive or stale PGs come
online. I am not able to access VMs (VMs, Images) which are using Ceph.

Thanks
Arun

On Fri, Jan 4, 2019 at 4:53 AM Caspar Smit <caspars...@supernas.eu> wrote:

> Hi Arun,
>
> How did you end up with a 'working' cluster with so many pgs per OSD?
>
> "too many PGs per OSD (2968 > max 200)"
>
> To (temporarily) allow this kind of pgs per osd you could try this:
>
> Change these values in the global section in your ceph.conf:
>
> mon max pg per osd = 200
> osd max pg per osd hard ratio = 2
>
> It allows 200*2 = 400 Pgs per OSD before disabling the creation of new
> pgs.
>
> Above are the defaults (for Luminous, maybe other versions too)
> You can check your current settings with:
>
> ceph daemon mon.ceph-mon01 config show |grep pg_per_osd
>
> Since your current pgs per osd ratio is way higher then the default you
> could set them to for instance:
>
> mon max pg per osd = 1000
> osd max pg per osd hard ratio = 5
>
> Which allow for 5000 pgs per osd before disabling creation of new pgs.
>
> You'll need to inject the setting into the mons/osds and restart mgrs to
> make them active.
>
> ceph tell mon.* injectargs ‘--mon_max_pg_per_osd 1000’
> ceph tell mon.* injectargs ‘--osd_max_pg_per_osd_hard_ratio 5’
> ceph tell osd.* injectargs ‘--mon_max_pg_per_osd 1000’
> ceph tell osd.* injectargs ‘--osd_max_pg_per_osd_hard_ratio 5’
> restart mgrs
>
> Kind regards,
> Caspar
>
>
> Op vr 4 jan. 2019 om 04:28 schreef Arun POONIA <
> arun.poo...@nuagenetworks.net>:
>
>> Hi Chris,
>>
>> Indeed that's what happened. I didn't set noout flag either and I did
>> zapped disk on new server every time. In my cluster status fre201 is only
>> new server.
>>
>> Current Status after enabling 3 OSDs on fre201 host.
>>
>> [root@fre201 ~]# ceph osd tree
>> ID  CLASS WEIGHT   TYPE NAME       STATUS REWEIGHT PRI-AFF
>>  -1       70.92137 root default
>>  -2        5.45549     host fre101
>>   0   hdd  1.81850         osd.0       up  1.00000 1.00000
>>   1   hdd  1.81850         osd.1       up  1.00000 1.00000
>>   2   hdd  1.81850         osd.2       up  1.00000 1.00000
>>  -9        5.45549     host fre103
>>   3   hdd  1.81850         osd.3       up  1.00000 1.00000
>>   4   hdd  1.81850         osd.4       up  1.00000 1.00000
>>   5   hdd  1.81850         osd.5       up  1.00000 1.00000
>>  -3        5.45549     host fre105
>>   6   hdd  1.81850         osd.6       up  1.00000 1.00000
>>   7   hdd  1.81850         osd.7       up  1.00000 1.00000
>>   8   hdd  1.81850         osd.8       up  1.00000 1.00000
>>  -4        5.45549     host fre107
>>   9   hdd  1.81850         osd.9       up  1.00000 1.00000
>>  10   hdd  1.81850         osd.10      up  1.00000 1.00000
>>  11   hdd  1.81850         osd.11      up  1.00000 1.00000
>>  -5        5.45549     host fre109
>>  12   hdd  1.81850         osd.12      up  1.00000 1.00000
>>  13   hdd  1.81850         osd.13      up  1.00000 1.00000
>>  14   hdd  1.81850         osd.14      up  1.00000 1.00000
>>  -6        5.45549     host fre111
>>  15   hdd  1.81850         osd.15      up  1.00000 1.00000
>>  16   hdd  1.81850         osd.16      up  1.00000 1.00000
>>  17   hdd  1.81850         osd.17      up  0.79999 1.00000
>>  -7        5.45549     host fre113
>>  18   hdd  1.81850         osd.18      up  1.00000 1.00000
>>  19   hdd  1.81850         osd.19      up  1.00000 1.00000
>>  20   hdd  1.81850         osd.20      up  1.00000 1.00000
>>  -8        5.45549     host fre115
>>  21   hdd  1.81850         osd.21      up  1.00000 1.00000
>>  22   hdd  1.81850         osd.22      up  1.00000 1.00000
>>  23   hdd  1.81850         osd.23      up  1.00000 1.00000
>> -10        5.45549     host fre117
>>  24   hdd  1.81850         osd.24      up  1.00000 1.00000
>>  25   hdd  1.81850         osd.25      up  1.00000 1.00000
>>  26   hdd  1.81850         osd.26      up  1.00000 1.00000
>> -11        5.45549     host fre119
>>  27   hdd  1.81850         osd.27      up  1.00000 1.00000
>>  28   hdd  1.81850         osd.28      up  1.00000 1.00000
>>  29   hdd  1.81850         osd.29      up  1.00000 1.00000
>> -12        5.45549     host fre121
>>  30   hdd  1.81850         osd.30      up  1.00000 1.00000
>>  31   hdd  1.81850         osd.31      up  1.00000 1.00000
>>  32   hdd  1.81850         osd.32      up  1.00000 1.00000
>> -13        5.45549     host fre123
>>  33   hdd  1.81850         osd.33      up  1.00000 1.00000
>>  34   hdd  1.81850         osd.34      up  1.00000 1.00000
>>  35   hdd  1.81850         osd.35      up  1.00000 1.00000
>> -27        5.45549     host fre201
>>  36   hdd  1.81850         osd.36      up  1.00000 1.00000
>>  37   hdd  1.81850         osd.37      up  1.00000 1.00000
>>  38   hdd  1.81850         osd.38      up  1.00000 1.00000
>> [root@fre201 ~]#
>> [root@fre201 ~]#
>> [root@fre201 ~]#
>> [root@fre201 ~]#
>> [root@fre201 ~]#
>> [root@fre201 ~]# ceph -s
>>   cluster:
>>     id:     adb9ad8e-f458-4124-bf58-7963a8d1391f
>>     health: HEALTH_ERR
>>             3 pools have many more objects per pg than average
>>             585791/12391450 objects misplaced (4.727%)
>>             2 scrub errors
>>             2374 PGs pending on creation
>>             Reduced data availability: 6578 pgs inactive, 2025 pgs down,
>> 74 pgs peering, 1234 pgs stale
>>             Possible data damage: 2 pgs inconsistent
>>             Degraded data redundancy: 64969/12391450 objects degraded
>> (0.524%), 616 pgs degraded, 20 pgs undersized
>>             96242 slow requests are blocked > 32 sec
>>             228 stuck requests are blocked > 4096 sec
>>             too many PGs per OSD (2768 > max 200)
>>
>>   services:
>>     mon: 3 daemons, quorum ceph-mon01,ceph-mon02,ceph-mon03
>>     mgr: ceph-mon03(active), standbys: ceph-mon01, ceph-mon02
>>     osd: 39 osds: 39 up, 39 in; 96 remapped pgs
>>     rgw: 1 daemon active
>>
>>   data:
>>     pools:   18 pools, 54656 pgs
>>     objects: 6050k objects, 10942 GB
>>     usage:   21900 GB used, 50721 GB / 72622 GB avail
>>     pgs:     0.002% pgs unknown
>>              12.050% pgs not active
>>              64969/12391450 objects degraded (0.524%)
>>              585791/12391450 objects misplaced (4.727%)
>>              47489 active+clean
>>              3670  activating
>>              1098  stale+down
>>              923   down
>>              575   activating+degraded
>>              563   stale+active+clean
>>              105   stale+activating
>>              78    activating+remapped
>>              72    peering
>>              25    stale+activating+degraded
>>              23    stale+activating+remapped
>>              9     stale+active+undersized
>>              6     stale+activating+undersized+degraded+remapped
>>              5     stale+active+undersized+degraded
>>              4     down+remapped
>>              4     activating+degraded+remapped
>>              2     active+clean+inconsistent
>>              1     stale+activating+degraded+remapped
>>              1     stale+active+clean+remapped
>>              1     stale+remapped+peering
>>              1     remapped+peering
>>              1     unknown
>>
>>   io:
>>     client:   0 B/s rd, 208 kB/s wr, 22 op/s rd, 22 op/s wr
>>
>>
>>
>> Thanks
>> Arun
>>
>>
>> On Thu, Jan 3, 2019 at 7:19 PM Chris <bitskr...@bitskrieg.net> wrote:
>>
>>> If you added OSDs and then deleted them repeatedly without waiting for
>>> replication to finish as the cluster attempted to re-balance across them,
>>> its highly likely that you are permanently missing PGs (especially if the
>>> disks were zapped each time).
>>>
>>> If those 3 down OSDs can be revived there is a (small) chance that you
>>> can right the ship, but 1400pg/OSD is pretty extreme.  I'm surprised
>>> the cluster even let you do that - this sounds like a data loss event.
>>>
>>> Bring back the 3 OSD and see what those 2 inconsistent pgs look like
>>> with ceph pg query.
>>>
>>> On January 3, 2019 21:59:38 Arun POONIA <arun.poo...@nuagenetworks.net>
>>> wrote:
>>>
>>>> Hi,
>>>>
>>>> Recently I tried adding a new node (OSD) to ceph cluster using
>>>> ceph-deploy tool. Since I was experimenting with tool and ended up deleting
>>>> OSD nodes on new server couple of times.
>>>>
>>>> Now since ceph OSDs are running on new server cluster PGs seems to be
>>>> inactive (10-15%) and they are not recovering or rebalancing. Not sure what
>>>> to do. I tried shutting down OSDs on new server.
>>>>
>>>> Status:
>>>> [root@fre105 ~]# ceph -s
>>>> 2019-01-03 18:56:42.867081 7fa0bf573700 -1 asok(0x7fa0b80017a0)
>>>> AdminSocketConfigObs::init: failed: AdminSocket::bind_and_listen: failed to
>>>> bind the UNIX domain socket to
>>>> '/var/run/ceph-guests/ceph-client.admin.4018644.140328258509136.asok': (2)
>>>> No such file or directory
>>>>   cluster:
>>>>     id:     adb9ad8e-f458-4124-bf58-7963a8d1391f
>>>>     health: HEALTH_ERR
>>>>             3 pools have many more objects per pg than average
>>>>             373907/12391198 objects misplaced (3.018%)
>>>>             2 scrub errors
>>>>             9677 PGs pending on creation
>>>>             Reduced data availability: 7145 pgs inactive, 6228 pgs
>>>> down, 1 pg peering, 2717 pgs stale
>>>>             Possible data damage: 2 pgs inconsistent
>>>>             Degraded data redundancy: 178350/12391198 objects degraded
>>>> (1.439%), 346 pgs degraded, 1297 pgs undersized
>>>>             52486 slow requests are blocked > 32 sec
>>>>             9287 stuck requests are blocked > 4096 sec
>>>>             too many PGs per OSD (2968 > max 200)
>>>>
>>>>   services:
>>>>     mon: 3 daemons, quorum ceph-mon01,ceph-mon02,ceph-mon03
>>>>     mgr: ceph-mon03(active), standbys: ceph-mon01, ceph-mon02
>>>>     osd: 39 osds: 36 up, 36 in; 51 remapped pgs
>>>>     rgw: 1 daemon active
>>>>
>>>>   data:
>>>>     pools:   18 pools, 54656 pgs
>>>>     objects: 6050k objects, 10941 GB
>>>>     usage:   21727 GB used, 45308 GB / 67035 GB avail
>>>>     pgs:     13.073% pgs not active
>>>>              178350/12391198 objects degraded (1.439%)
>>>>              373907/12391198 objects misplaced (3.018%)
>>>>              46177 active+clean
>>>>              5054  down
>>>>              1173  stale+down
>>>>              1084  stale+active+undersized
>>>>              547   activating
>>>>              201   stale+active+undersized+degraded
>>>>              158   stale+activating
>>>>              96    activating+degraded
>>>>              46    stale+active+clean
>>>>              42    activating+remapped
>>>>              34    stale+activating+degraded
>>>>              23    stale+activating+remapped
>>>>              6     stale+activating+undersized+degraded+remapped
>>>>              6     activating+undersized+degraded+remapped
>>>>              2     activating+degraded+remapped
>>>>              2     active+clean+inconsistent
>>>>              1     stale+activating+degraded+remapped
>>>>              1     stale+active+clean+remapped
>>>>              1     stale+remapped
>>>>              1     down+remapped
>>>>              1     remapped+peering
>>>>
>>>>   io:
>>>>     client:   0 B/s rd, 208 kB/s wr, 28 op/s rd, 28 op/s wr
>>>>
>>>> Thanks
>>>> --
>>>> Arun Poonia
>>>>
>>>> _______________________________________________
>>>> ceph-users mailing list
>>>> ceph-users@lists.ceph.com
>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>
>>>>
>>>
>>
>> --
>> Arun Poonia
>>
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>


-- 
Arun Poonia
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to