Just an update.

Recovery completed but the PGS are still inactive.

Still having a hard time understanding why adding OSDs caused this. I'm on
12.2.2

user@admin:~$ ceph -s
  cluster:
    id:     a3672c60-3051-440c-bd83-8aff7835ce53
    health: HEALTH_WARN
            Reduced data availability: 307 pgs inactive
            Degraded data redundancy: 307 pgs unclean

  services:
    mon: 5 daemons, quorum
stor585r2u8a,stor585r2u12a,stor585r2u16a,stor585r2u20a,stor585r2u24a
    mgr: stor585r2u8a(active)
    osd: 88 osds: 87 up, 87 in; 133 remapped pgs

  data:
    pools:   12 pools, 3016 pgs
    objects: 387k objects, 1546 GB
    usage:   3313 GB used, 186 TB / 189 TB avail
    pgs:     10.179% pgs not active
             2709 active+clean
             174  activating
             133  activating+remapped

  io:
    client:   8436 kB/s rd, 935 kB/s wr, 140 op/s rd, 64 op/s wr


On Tue, Dec 19, 2017 at 8:57 PM, Daniel K <satha...@gmail.com> wrote:

> I'm trying to understand why adding OSDs would cause pgs to go inactive.
>
> This cluster has 88 OSDs, and had 6 OSD with device class "hdd_10TB_7.2k"
>
> I added two more OSDs, set the device class to "hdd_10TB_7.2k" and 10% of
> pgs went inactive.
>
> I have an EC pool on these OSDs with the profile:
> user@admin:~$ ceph osd erasure-code-profile get ISA_10TB_7.2k_4.2
> crush-device-class=hdd_10TB_7.2k
> crush-failure-domain=host
> crush-root=default
> k=4
> m=2
> plugin=isa
> technique=reed_sol_van.
>
> some outputs of ceph health detail and ceph osd df
> user@admin:~$ ceph osd df |grep 10TB
> 76 hdd_10TB_7.2k 9.09509  1.00000 9313G   349G 8963G 3.76 2.20 488
> 20 hdd_10TB_7.2k 9.09509  1.00000 9313G   345G 8967G 3.71 2.17 489
> 28 hdd_10TB_7.2k 9.09509  1.00000 9313G   344G 8968G 3.70 2.17 484
> 36 hdd_10TB_7.2k 9.09509  1.00000 9313G   345G 8967G 3.71 2.17 484
> 87 hdd_10TB_7.2k 9.09560  1.00000 9313G  8936M 9305G 0.09 0.05 311
> 86 hdd_10TB_7.2k 9.09560  1.00000 9313G  8793M 9305G 0.09 0.05 304
>  6 hdd_10TB_7.2k 9.09509  1.00000 9313G   344G 8969G 3.70 2.16 471
> 68 hdd_10TB_7.2k 9.09509  1.00000 9313G   344G 8969G 3.70 2.17 480
> user@admin:~$ ceph health detail|grep inactive
> HEALTH_WARN 68287/1928007 objects misplaced (3.542%); Reduced data
> availability: 307 pgs inactive; Degraded data redundancy: 341 pgs unclean
> PG_AVAILABILITY Reduced data availability: 307 pgs inactive
>     pg 24.60 is stuck inactive for 1947.792377, current state
> activating+remapped, last acting [36,20,76,6,68,28]
>     pg 24.63 is stuck inactive for 1946.571425, current state
> activating+remapped, last acting [28,76,6,20,68,36]
>     pg 24.71 is stuck inactive for 1947.625988, current state
> activating+remapped, last acting [6,68,20,36,28,76]
>     pg 24.73 is stuck inactive for 1947.705250, current state
> activating+remapped, last acting [36,6,20,76,68,28]
>     pg 24.74 is stuck inactive for 1947.828063, current state
> activating+remapped, last acting [68,36,28,20,6,76]
>     pg 24.75 is stuck inactive for 1947.475644, current state
> activating+remapped, last acting [6,28,76,36,20,68]
>     pg 24.76 is stuck inactive for 1947.712046, current state
> activating+remapped, last acting [20,76,6,28,68,36]
>     pg 24.78 is stuck inactive for 1946.576304, current state
> activating+remapped, last acting [76,20,68,36,6,28]
>     pg 24.7a is stuck inactive for 1947.820932, current state
> activating+remapped, last acting [36,20,28,68,6,76]
>     pg 24.7b is stuck inactive for 1947.858305, current state
> activating+remapped, last acting [68,6,20,28,76,36]
>     pg 24.7c is stuck inactive for 1947.753917, current state
> activating+remapped, last acting [76,6,20,36,28,68]
>     pg 24.7d is stuck inactive for 1947.803229, current state
> activating+remapped, last acting [68,6,28,20,36,76]
>     pg 24.7f is stuck inactive for 1947.792506, current state
> activating+remapped, last acting [28,20,76,6,68,36]
>     pg 24.8a is stuck inactive for 1947.823189, current state
> activating+remapped, last acting [28,76,20,6,36,68]
>     pg 24.8b is stuck inactive for 1946.579755, current state
> activating+remapped, last acting [76,68,20,28,6,36]
>     pg 24.8c is stuck inactive for 1947.555872, current state
> activating+remapped, last acting [76,36,68,6,28,20]
>     pg 24.8d is stuck inactive for 1946.589814, current state
> activating+remapped, last acting [36,6,28,76,68,20]
>     pg 24.8e is stuck inactive for 1947.802894, current state
> activating+remapped, last acting [28,6,68,36,76,20]
>     pg 24.8f is stuck inactive for 1947.528603, current state
> activating+remapped, last acting [76,28,6,68,20,36]
>     pg 25.60 is stuck inactive for 1947.620823, current state activating,
> last acting [20,6,87,36,28,68]
>     pg 25.61 is stuck inactive for 1947.883517, current state activating,
> last acting [28,36,86,76,6,87]
>     pg 25.62 is stuck inactive for 1542089.552271, current state
> activating, last acting [86,6,76,20,87,68]
>     pg 25.70 is stuck inactive for 1542089.729631, current state
> activating, last acting [86,87,76,20,68,28]
>     pg 25.71 is stuck inactive for 1947.642271, current state activating,
> last acting [28,86,68,20,6,36]
>     pg 25.75 is stuck inactive for 1947.825872, current state activating,
> last acting [68,86,36,20,76,6]
>     pg 25.76 is stuck inactive for 1947.737307, current state activating,
> last acting [36,87,28,6,68,76]
>     pg 25.77 is stuck inactive for 1947.218420, current state activating,
> last acting [87,36,86,28,76,6]
>     pg 25.79 is stuck inactive for 1947.253871, current state activating,
> last acting [6,36,86,28,68,76]
>     pg 25.7a is stuck inactive for 1542089.794085, current state
> activating, last acting [86,36,68,20,76,87]
>     pg 25.7c is stuck inactive for 1947.666774, current state activating,
> last acting [20,86,36,6,76,87]
>     pg 25.8a is stuck inactive for 1542089.687299, current state
> activating, last acting [87,36,68,20,86,28]
>     pg 25.8c is stuck inactive for 1947.545965, current state activating,
> last acting [76,6,28,87,36,86]
>     pg 25.8d is stuck inactive for 1947.213627, current state activating,
> last acting [86,36,87,20,28,76]
>     pg 25.8e is stuck inactive for 1947.230754, current state activating,
> last acting [87,86,68,28,76,20]
>     pg 25.8f is stuck inactive for 1542089.800416, current state
> activating, last acting [86,76,20,68,36,28]
>     pg 34.40 is stuck inactive for 1947.641110, current state activating,
> last acting [20,36,87,6,86,28]
>     pg 34.41 is stuck inactive for 1947.759524, current state activating,
> last acting [28,86,36,68,76,87]
>     pg 34.42 is stuck inactive for 1947.656110, current state activating,
> last acting [68,36,87,28,6,86]
>     pg 34.44 is stuck inactive for 1947.659653, current state activating,
> last acting [28,68,6,36,87,20]
>     pg 34.45 is stuck inactive for 1542089.795364, current state
> activating, last acting [86,28,76,36,6,68]
>     pg 34.46 is stuck inactive for 1947.570029, current state activating,
> last acting [28,20,87,6,86,36]
>     pg 34.47 is stuck inactive for 1947.667102, current state activating,
> last acting [20,86,68,76,36,87]
>     pg 34.48 is stuck inactive for 1947.632449, current state activating,
> last acting [28,76,6,86,87,20]
>     pg 34.4b is stuck inactive for 1947.671088, current state activating,
> last acting [36,87,68,28,6,20]
>     pg 34.4c is stuck inactive for 1947.699305, current state activating,
> last acting [20,6,86,28,87,68]
>     pg 34.4d is stuck inactive for 1542089.756804, current state
> activating, last acting [87,36,20,86,6,68]
>     pg 34.58 is stuck inactive for 1947.749120, current state activating,
> last acting [28,86,87,76,6,20]
>     pg 34.59 is stuck inactive for 1947.584327, current state activating,
> last acting [28,20,87,6,86,76]
>     pg 34.5a is stuck inactive for 1947.670953, current state activating,
> last acting [6,87,36,68,86,76]
>     pg 34.5b is stuck inactive for 1947.692114, current state activating,
> last acting [68,76,86,6,20,36]
>     pg 34.5e is stuck inactive for 1542089.773455, current state
> activating, last acting [86,68,28,87,6,36]
>
>
> It looks like recovery is happening, so they will eventually be active,
> but I'm trying to figure out what I did wrong, and how I could do this in
> the future to keep from taking 10% of my pgs offline.
>
>
> Thanks!
>
>
> Dan
>
>
>
>
>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to