[Hrm, this email was in my spam folder.]

At a quick glance, you're probably running into some issues because
you've got two racks of very different weights. Things will probably
get better if you enable the optimal "crush tunables"; check out the
docs on that and see if you can switch to them.
-Greg
Software Engineer #42 @ http://inktank.com | http://ceph.com


On Fri, Dec 27, 2013 at 3:58 AM, Sreejith Vijayendran
<sreejith.vijayend...@inmobi.com> wrote:
> Hello,
>
> We have a 3 node cluster set up with OSDs created on all the 3 nodes. The
> replication was set to 2.
> 1>> We are in testing phase and tried to bring down all the OSDs in a
> particular node and was testing the migration of PGs to other OSDs.
> But the PGs are not getting replicated on other OSDs and the status of the
> replication also was not clear at all.
> Below is the ceph status at that point:
>
> ===============
> sreejith@sb1001:/var/run/ceph$ sudo ceph status
>     cluster 9b48b60c-bebe-4714-8a61-91ca5b388a17
>      health HEALTH_WARN 885 pgs degraded; 885 pgs stuck unclean; recovery
> 59/232 objects degraded (25.431%); 22/60 in osds are down
>      monmap e2: 3 mons at
> {sb1001=10.2.4.90:6789/0,sb1002=10.2.4.202:6789/0,sb1004=10.2.4.203:6789/0},
> election epoch 22, quorum 0,1,2 sb1001,sb1002,sb1004
>      osdmap e378: 68 osds: 38 up, 60 in
>       pgmap v4490: 1564 pgs, 25 pools, 1320 MB data, 116 objects
>             8415 MB used, 109 TB / 109 TB avail
>             59/232 objects degraded (25.431%)
>                  679 active+clean
>                  862 active+degraded
>                   23 active+degraded+remapped
> ===============
>
> We waited for around 4-5 hours and the status just increased marginally to
> (25.431%) from (27%) at the start.
>
> 2>> we then tweaked some OSD values to speed up the recovery
> namely(osd_recovery_threads, osd_recovery_max_active,
> osd_recovery_max_chunk, osd_max_backfills, osd_backfill_retry_interval etc)
> as we were only concerned about getting the OSDs rebalanced as of now. But
> this didnt improve at all all over the night till the morning.
>
> 3>> We then manually started all the OSDs in that specific node and the
> status came back up:
> But then we could see that there were 23PGs stuck unclean and in
> 'active+remapped state'
>
> =================
> sreejith@sb1001:~$ sudo ceph status
> [sudo] password for sreejith:
>     cluster 9b48b60c-bebe-4714-8a61-91ca5b388a17
>      health HEALTH_WARN 23 pgs stuck unclean
>      monmap e2: 3 mons at
> {sb1001=10.2.4.90:6789/0,sb1002=10.2.4.202:6789/0,sb1004=10.2.4.203:6789/0},
> election epoch 22, quorum 0,1,2 sb1001,sb1002,sb1004
>      osdmap e382: 68 osds: 61 up, 61 in
>       pgmap v4931: 1564 pgs, 25 pools, 1320 MB data, 116 objects
>             7931 MB used, 110 TB / 110 TB avail
>                 1541 active+clean
>                   23 active+remapped
> =================
>
> The 'pg dump_stuck unclean' was showing  that all the PGs were on 4 OSDs and
> no other PG were on those same OSDs.
> SO:
> 4>> we took those OSDs out of the cluster using 'ceph osd out {id}'. Then
> the unclean PG number increased to 52. Even after making the OSDs back 'IN',
> the situation didn't improve.
>
> =============
> root@sb1001:/home/sreejith# ceph health detail
> HEALTH_WARN 52 pgs stuck unclean
> pg 9.63 is stuck unclean since forever, current state active+remapped, last
> acting [47,7]
> pg 11.61 is stuck unclean since forever, current state active+remapped, last
> acting [47,7]
> pg 10.62 is stuck unclean since forever, current state active+remapped, last
> acting [47,7]
> pg 13.5f is stuck unclean since forever, current state active+remapped, last
> acting [47,7]
> pg 15.5d is stuck unclean since forever, current state active+remapped, last
> acting [47,7]
> pg 14.5e is stuck unclean since forever, current state active+remapped, last
> acting [47,7]
> pg 9.47 is stuck unclean for 530.594604, current state active+remapped, last
> acting [66,43]
> pg 7.49 is stuck unclean for 530.594593, current state active+remapped, last
> acting [66,43]
> pg 5.4b is stuck unclean for 530.594481, current state active+remapped, last
> acting [66,43]
> pg 3.4d is stuck unclean for 530.594449, current state active+remapped, last
> acting [66,43]
> pg 11.45 is stuck unclean for 530.594635, current state active+remapped,
> last acting [66,43]
> pg 13.43 is stuck unclean for 530.594654, current state active+remapped,
> last acting [66,43]
> pg 15.41 is stuck unclean for 530.594695, current state active+remapped,
> last acting [66,43]
> pg 6.4a is stuck unclean for 530.594366, current state active+remapped, last
> acting [66,43]
> pg 10.46 is stuck unclean for 530.594387, current state active+remapped,
> last acting [66,43]
> pg 14.42 is stuck unclean for 530.594422, current state active+remapped,
> last acting [66,43]
> pg 4.4c is stuck unclean for 530.594341, current state active+remapped, last
> acting [66,43]
> pg 12.44 is stuck unclean for 530.594361, current state active+remapped,
> last acting [66,43]
> pg 8.48 is stuck unclean for 530.594294, current state active+remapped, last
> acting [66,43]
> pg 8.30 is stuck unclean for 175428.682512, current state active+remapped,
> last acting [51,45]
> pg 0.38 is stuck unclean for 175428.682498, current state active+remapped,
> last acting [51,45]
> pg 7.31 is stuck unclean since forever, current state active+remapped, last
> acting [51,45]
> pg 5.33 is stuck unclean since forever, current state active+remapped, last
> acting [51,45]
> pg 3.35 is stuck unclean since forever, current state active+remapped, last
> acting [51,45]
> pg 1.37 is stuck unclean since forever, current state active+remapped, last
> acting [51,45]
> pg 6.32 is stuck unclean for 154614.490724, current state active+remapped,
> last acting [51,45]
> pg 2.36 is stuck unclean since forever, current state active+remapped, last
> acting [51,45]
> pg 4.34 is stuck unclean since forever, current state active+remapped, last
> acting [51,45]
> pg 9.24 is stuck unclean since forever, current state active+remapped, last
> acting [48,24]
> pg 8.25 is stuck unclean for 482.144359, current state active+remapped, last
> acting [48,24]
> pg 5.28 is stuck unclean since forever, current state active+remapped, last
> acting [48,24]
> pg 4.29 is stuck unclean since forever, current state active+remapped, last
> acting [48,24]
> pg 12.21 is stuck unclean since forever, current state active+remapped, last
> acting [48,24]
> pg 0.2d is stuck unclean for 482.144315, current state active+remapped, last
> acting [48,24]
> pg 1.2c is stuck unclean since forever, current state active+remapped, last
> acting [48,24]
> pg 13.20 is stuck unclean since forever, current state active+remapped, last
> acting [48,24]
> pg 3.2a is stuck unclean since forever, current state active+remapped, last
> acting [48,24]
> pg 11.22 is stuck unclean since forever, current state active+remapped, last
> acting [48,24]
> pg 10.23 is stuck unclean since forever, current state active+remapped, last
> acting [48,24]
> pg 2.2b is stuck unclean since forever, current state active+remapped, last
> acting [48,24]
> pg 9.2f is stuck unclean since forever, current state active+remapped, last
> acting [51,45]
> pg 11.2d is stuck unclean since forever, current state active+remapped, last
> acting [51,45]
> pg 13.2b is stuck unclean since forever, current state active+remapped, last
> acting [51,45]
> pg 15.29 is stuck unclean since forever, current state active+remapped, last
> acting [51,45]
> pg 10.2e is stuck unclean since forever, current state active+remapped, last
> acting [51,45]
> pg 14.2a is stuck unclean since forever, current state active+remapped, last
> acting [51,45]
> pg 7.26 is stuck unclean since forever, current state active+remapped, last
> acting [48,24]
> pg 6.27 is stuck unclean since forever, current state active+remapped, last
> acting [48,24]
> pg 12.2c is stuck unclean since forever, current state active+remapped, last
> acting [51,45]
> pg 14.1f is stuck unclean since forever, current state active+remapped, last
> acting [48,24]
> pg 15.1e is stuck unclean since forever, current state active+remapped, last
> acting [48,24]
> pg 12.60 is stuck unclean for 175428.710580, current state active+remapped,
> last acting [47,7]
> root@sb1001:/home/sreejith#
> =============
>
> Now we could see that some other OSDs also came in the list of unclean PGs.
> And the number of unclean PGs just increased to 52
>
>
> So we have two questions:
> 1> why isn't the unclean PGs getting cleared, how to debug more.
> 2> how to check the recover/backfill status of OSDs.(ceph status, -w, all
> commands show same data). It just shows this many are up, in, this many PGs
> are clean, remapped. There is no status update of the progress if any.
> I referred this bug:
> http://tracker.ceph.com/issues/6736
> where the user could see status update like(peering, recovery etc), But for
> us these are not shown.
>
>
> More details:
> we have all our OSDs of 1.7TB size.
> we have 3 nodes in cluster (all 3 having one mon instance each and OSDs on
> all)
> replication set to default 2
> ceph version 0.72.2
>
> Attaching the crush rule set, osd dump, pg dump.
>
>
> --
>
> Regards,
> Sreejith
>
> _____________________________________________________________
> The information contained in this communication is intended solely for the
> use of the individual or entity to whom it is addressed and others
> authorized to receive it. It may contain confidential or legally privileged
> information. If you are not the intended recipient you are hereby notified
> that any disclosure, copying, distribution or taking any action in reliance
> on the contents of this information is strictly prohibited and may be
> unlawful. If you have received this communication in error, please notify us
> immediately by responding to this email and then delete it from your system.
> The firm is neither liable for the proper and complete transmission of the
> information contained in this communication nor for any delay in its
> receipt.
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to