Hi,

As far as I know, this is exactly the problem why the new tunables where 
introduced, If you use 3 Replicas with only 3 hosts, crush sometimes doesn't 
find a solution to place all pgs.

If you are really stuck with bobtail turntables, I can think of 2 possible 
workarounds:

1. Add another osd Server.
2. Bad idea, but could work: build your crush rule manually, e.g.: set all 
primary pgs to host ceph1, first copy to host ceph2 and second copy to host3.

Micha Krause

Am 08.07.2016 um 05:47 schrieb Nathanial Byrnes:
Hello,
     I've got a Jewel Cluster (3 nodes, 15 OSD's) running with bobtail tunables 
(my xenserver cluster uses 3.10 as the kernel and there's no upgrading 
that....). I started the cluster out on Hammer, upgraded to Jewel, discovered 
that optimal tunables would not work, and then set the tunables back to 
bobtail. Once the re-balancing completed, I was stuck with 1 pg in 
active+remapped. Repair didn't fix the pg.  I then upped the number of pgs from 
328 to 712 (oddly I asked for 512, but ended p with 712...), now I have 5 pgs 
stuck in active+remapped. I also tried re-weighting the pgs a couple times, but 
no change.... Here is my osd tree:

ID WEIGHT   TYPE NAME      UP/DOWN REWEIGHT PRIMARY-AFFINITY
-1 15.00000 root default
-2  5.00000     host ceph1
  0  1.00000         osd.0       up  0.95001          1.00000
  1  1.00000         osd.1       up  1.00000          1.00000
  2  1.00000         osd.2       up  1.00000          1.00000
  3  1.00000         osd.3       up  0.90002          1.00000
  4  1.00000         osd.4       up  1.00000          1.00000
-3  5.00000     host ceph3
10  1.00000         osd.10      up  1.00000          1.00000
11  1.00000         osd.11      up  1.00000          1.00000
12  1.00000         osd.12      up  1.00000          1.00000
13  1.00000         osd.13      up  1.00000          1.00000
14  1.00000         osd.14      up  1.00000          1.00000
-4  5.00000     host ceph2
  5  1.00000         osd.5       up  1.00000          1.00000
  6  1.00000         osd.6       up  1.00000          1.00000
  7  1.00000         osd.7       up  1.00000          1.00000
  8  1.00000         osd.8       up  1.00000          1.00000
  9  1.00000         osd.9       up  1.00000          1.00000


     Any suggestions on how to troubleshoot or repair this?

     Thanks and Regards,
     Nate



_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to