Just to add, this doesn't happen in just one pool. When I change "data" pool replicate size from 2 to 3, a few PGs (3) got stuck too.
pg 0.7c is active+clean+degraded, acting [8,2] pg 0.48 is active+clean+degraded, acting [4,8] pg 0.1f is active+clean+degraded, acting [5,7] I am already on tunables optimal. # ceph osd crush tunables optimal adjusted tunables profile to optimal What's wrong? - WP On Tue, Jun 4, 2013 at 1:23 PM, YIP Wai Peng <yi...@comp.nus.edu.sg> wrote: > Hi Sage, > > Thanks, I noticed after re-reading the documentation. > > I realized that osd.8 was not in host3. After adding osd.8 to host3, the > PGs are now in "active+remapped" > > # ceph pg 3.45 query > > { "state": "active+remapped", > "epoch": 1374, > "up": [ > 4, > 8], > "acting": [ > 4, > 8, > 6], > <snip> > > Still, nothing is happening. What can be wrong? > > - WP > > On Tue, Jun 4, 2013 at 12:26 PM, Sage Weil <s...@inktank.com> wrote: > >> On Tue, 4 Jun 2013, YIP Wai Peng wrote: >> > Sorry, to set things in context, I had some other problems last weekend. >> > Setting it to optimal tunables helped (although I am on the older >> kernel). >> > Since it worked, I was inclined to believed that the tunables do work >> on the >> > older kernel. >> > That being said, I will upgrade the kernel to see if this issue goes >> away. >> >> The kernel version is only an issue if you are using the cephfs or rbd >> *client* from the kernel (e.g., rbd map ... or mount -t ceph ...). (Ceph >> didn't appear upstream until 2.6.35 or thereabouts, and fixes are only >> backported as far as v3.4.) >> >> sage >> >> > >> > Regards, >> > Wai Peng >> > >> > >> > On Tue, Jun 4, 2013 at 12:01 PM, YIP Wai Peng <yi...@comp.nus.edu.sg> >> wrote: >> > Hi Sage, >> > It is on optimal tunables already. However, I'm on kernel >> > 2.6.32-358.6.2.el6.x86_64. Will the tunables take effect or do I have >> > to upgrade to something newer? >> > >> > - WP >> > >> > >> > On Tue, Jun 4, 2013 at 11:58 AM, Sage Weil <s...@inktank.com> wrote: >> > On Tue, 4 Jun 2013, YIP Wai Peng wrote: >> > > Hi all, >> > > I'm running ceph on CentOS6 on 3 hosts, with 3 OSD each >> > (total 9 OSD). >> > > When I increased one of my pool rep size from 2 to 3, >> > just 6 PGs will get >> > > stuck in active+clean+degraded mode, but it doesn't >> > create new replicas. >> > >> > My first guess is that you do not have the newer crush tunables >> > set and >> > some placements are not quite right. If you are prepared for >> > some data >> > migration, and are not using an older kernel client, try >> > >> > ceph osd crush tunables optimal >> > >> > sage >> > >> > >> > > >> > > One of the problematic PG has the following (snipped for >> > brevity) >> > > >> > > { "state": "active+clean+degraded", >> > > "epoch": 1329, >> > > "up": [ >> > > 4, >> > > 6], >> > > "acting": [ >> > > 4, >> > > 6], >> > > <snip> >> > > "recovery_state": [ >> > > { "name": "Started\/Primary\/Active", >> > > "enter_time": "2013-06-04 01:10:30.092977", >> > > "might_have_unfound": [ >> > > { "osd": 3, >> > > "status": "already probed"}, >> > > { "osd": 5, >> > > "status": "not queried"}, >> > > { "osd": 6, >> > > "status": "already probed"}], >> > > <snip> >> > > >> > > >> > > I tried force_create_pg but it gets stuck in "creating". Any >> > ideas on how to >> > > "kickstart" this node to create the correct numbers of >> > replicas? >> > > >> > > >> > > PS: I have the following crush rule for the pool, which makes >> > the replicas >> > > go to different hosts. >> > > host1 has OSD 0,1,2 >> > > host2 has OSD 3,4,5 >> > > host3 has OSD 6,7,8 >> > > Looking at it, the new replica should be going to OSD 0,1,2, >> > but ceph is not >> > > creating it? >> > > >> > > rule different_host { >> > > ruleset 3 >> > > type replicated >> > > min_size 1 >> > > max_size 10 >> > > step take default >> > > step chooseleaf firstn 0 type host >> > > step emit >> > > } >> > > >> > > >> > > Any help will be much appreciated. Cheers >> > > - Wai Peng >> > > >> > > >> > >> > >> > >> > >> > >> > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com