On Mon, 22 Jul 2013, Gaylord Holder wrote:
> Sage,
>
> The crush tunables did the trick.
>
> why? Could you explain what was causing the problem?
This has a good explanation, I think:
http://ceph.com/docs/master/rados/operations/crush-map/#tunables
> I've haven't installed 3.9 on my RBD servers yet. Will setting crush tunables
> back to default or legacy cause me similar problems in the future?
Yeah. For 3.6+ kernels, you can set slightly different tunables and it
will be very close to optimal...
sage
>
> Thank you again Sage!
>
> -Gaylord
>
> On 07/22/2013 02:27 PM, Sage Weil wr:
> > On Mon, 22 Jul 2013, Gaylord Holder wrote:
> > >
> > > I have a 12 OSD/3 host set up, and have be stuck with a bunch of stuck
> > > pages.
> > >
> > > I've verified the OSDs are all up and in. The crushmap looks fine.
> > > I've tried restarting all the daemons.
> > >
> > >
> > >
> > > root@never:/var/lib/ceph/mon# ceph status
> > > health HEALTH_WARN 139 pgs degraded; 461 pgs stuck unclean; recovery
> > > 216/6213 degraded (3.477%)
> > > monmap e4: 2 mons at {a=192.168.225.9:6789/0,b=192.168.225.10:6789/0},
> > > election epoch 14, quorum 0,1 a,b
> >
> > Add another monitor; right now if 1 fails the cluster is unavailable.
> >
> > > osdmap e238: 12 osds: 12 up, 12 in
> > > pgmap v7396: 2528 pgs: 2067 active+clean, 322 active+remapped, 139
> > > active+degraded; 8218 MB data, 103 GB used, 22241 GB / 22345 GB avail;
> > > 216/6213 degraded (3.477%)
> > > mdsmap e1: 0/0/1 up
> >
> > My guess crush tunables. Try
> >
> > ceph osd crush tunables optimal
> >
> > unless you are using a pre-3.8(ish) kernel or other very old (pre-bobtail)
> > clients.
> >
> > sage
> >
> >
> > >
> > >
> > > I have one non-default pool with 3x replication. Fewer than half of the
> > > pg
> > > have expanded to 3x (278/400 pgs still have acting 2x sets).
> > >
> > > Where can I go look for the trouble?
> > >
> > > Thank you for any light someone can shed on this.
> > >
> > > Cheers,
> > > -Gaylord
> > > _______________________________________________
> > > ceph-users mailing list
> > > [email protected]
> > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> > >
> > >
>
>
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com