Hi Sage,

Thanks, I noticed after re-reading the documentation.

I realized that osd.8 was not in host3. After adding osd.8 to host3, the
PGs are now in "active+remapped"

# ceph pg 3.45 query

{ "state": "active+remapped",
  "epoch": 1374,
  "up": [
        4,
        8],
  "acting": [
        4,
        8,
        6],
<snip>

Still, nothing is happening. What can be wrong?

- WP

On Tue, Jun 4, 2013 at 12:26 PM, Sage Weil <s...@inktank.com> wrote:

> On Tue, 4 Jun 2013, YIP Wai Peng wrote:
> > Sorry, to set things in context, I had some other problems last weekend.
> > Setting it to optimal tunables helped (although I am on the older
> kernel).
> > Since it worked, I was inclined to believed that the tunables do work on
> the
> > older kernel.
> > That being said, I will upgrade the kernel to see if this issue goes
> away.
>
> The kernel version is only an issue if you are using the cephfs or rbd
> *client* from the kernel (e.g., rbd map ... or mount -t ceph ...).  (Ceph
> didn't appear upstream until 2.6.35 or thereabouts, and fixes are only
> backported as far as v3.4.)
>
> sage
>
> >
> > Regards,
> > Wai Peng
> >
> >
> > On Tue, Jun 4, 2013 at 12:01 PM, YIP Wai Peng <yi...@comp.nus.edu.sg>
> wrote:
> >       Hi Sage,
> > It is on optimal tunables already. However, I'm on kernel
> > 2.6.32-358.6.2.el6.x86_64. Will the tunables take effect or do I have
> > to upgrade to something newer?
> >
> > - WP
> >
> >
> > On Tue, Jun 4, 2013 at 11:58 AM, Sage Weil <s...@inktank.com> wrote:
> >       On Tue, 4 Jun 2013, YIP Wai Peng wrote:
> >       > Hi all,
> >       > I'm running ceph on CentOS6 on 3 hosts, with 3 OSD each
> >       (total 9 OSD).
> >       > When I increased one of my pool rep size from 2 to 3,
> >       just 6 PGs will get
> >       > stuck in active+clean+degraded mode, but it doesn't
> >       create new replicas.
> >
> > My first guess is that you do not have the newer crush tunables
> > set and
> > some placements are not quite right.  If you are prepared for
> > some data
> > migration, and are not using an older kernel client, try
> >
> >  ceph osd crush tunables optimal
> >
> > sage
> >
> >
> > >
> > > One of the problematic PG has the following (snipped for
> > brevity)
> > >
> > > { "state": "active+clean+degraded",
> > >   "epoch": 1329,
> > >   "up": [
> > >         4,
> > >         6],
> > >   "acting": [
> > >         4,
> > >         6],
> > > <snip>
> > >   "recovery_state": [
> > >         { "name": "Started\/Primary\/Active",
> > >           "enter_time": "2013-06-04 01:10:30.092977",
> > >           "might_have_unfound": [
> > >                 { "osd": 3,
> > >                   "status": "already probed"},
> > >                 { "osd": 5,
> > >                   "status": "not queried"},
> > >                 { "osd": 6,
> > >                   "status": "already probed"}],
> > > <snip>
> > >
> > >
> > > I tried force_create_pg but it gets stuck in "creating". Any
> > ideas on how to
> > > "kickstart" this node to create the correct numbers of
> > replicas?
> > >
> > >
> > > PS: I have the following crush rule for the pool, which makes
> > the replicas
> > > go to different hosts.
> > > host1 has OSD 0,1,2
> > > host2 has OSD 3,4,5
> > > host3 has OSD 6,7,8
> > > Looking at it, the new replica should be going to OSD 0,1,2,
> > but ceph is not
> > > creating it?
> > >
> > > rule different_host {
> > >         ruleset 3
> > >         type replicated
> > >         min_size 1
> > >         max_size 10
> > >         step take default
> > >         step chooseleaf firstn 0 type host
> > >         step emit
> > > }
> > >
> > >
> > > Any help will be much appreciated. Cheers
> > > - Wai Peng
> > >
> > >
> >
> >
> >
> >
> >
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to