Thank you for the explanation, Brad. I will change that setting and see how
it goes.

Subhachandra

On Thu, Jan 11, 2018 at 10:38 PM, Brad Hubbard <bhubb...@redhat.com> wrote:

> On Fri, Jan 12, 2018 at 11:27 AM, Subhachandra Chandra
> <schan...@grailbio.com> wrote:
> > Hello,
> >
> >      We are running experiments on a Ceph cluster before we move data on
> it.
> > While trying to increase the number of PGs on one of the pools it threw
> the
> > following error
> >
> > root@ctrl1:/# ceph osd pool set data pg_num 65536
> > Error E2BIG: specified pg_num 65536 is too large (creating 32768 new PGs
> on
> > ~540 OSDs exceeds per-OSD max of 32)
>
> That comes from here:
>
> https://github.com/ceph/ceph/blob/5d7813f612aea59239c8375aaa0091
> 9ae32f952f/src/mon/OSDMonitor.cc#L6027
>
> So the warning is triggered because new_pgs (65536) >
> g_conf->mon_osd_max_split_count (32) * expected_osds (540)
>
> >
> > There are 2 pools named "data" and "metadata". "data" is an erasure coded
> > pool (6,3) and "metadata" is a replicated pool with a replication factor
> of
> > 3.
> >
> > root@ctrl1:/# ceph osd lspools
> > 1 metadata,2 data,
> > root@ctrl1:/# ceph osd pool get metadata pg_num
> > pg_num: 512
> > root@ctrl1:/# ceph osd pool get data pg_num
> > pg_num: 32768
> >
> >     osd: 540 osds: 540 up, 540 in
> >          flags noout,noscrub,nodeep-scrub
> >
> >   data:
> >     pools:   2 pools, 33280 pgs
> >     objects: 7090k objects, 1662 TB
> >     usage:   2501 TB used, 1428 TB / 3929 TB avail
> >     pgs:     33280 active+clean
> >
> > The current PG/OSD ratio according to my calculation should be 549
> >>>> (32768 * 9 + 512 * 3 ) / 540.0
> > 548.9777777777778
> >
> > Increasing the number of PGs in the "data" pool should increase the
> PG/OSD
> > ratio to about 1095
> >>>> (65536 * 9 + 512 * 3 ) / 540.0
> > 1095.111111111111
> >
> > In the config, settings related to PG/OSD ratio look like
> > mon_max_pg_per_osd = 1500
> > osd_max_pg_per_osd_hard_ratio = 1.0
> >
> > Trying to increase the number of PGs to 65536 throws the previously
> > mentioned error. The new PG/OSD ratio is still under the configured
> limit.
> > Why do we see the error? Further, there seems to be a bug in the error
> > message where it says "exceeds per-OSD max of 32" in terms of where does
> > "32" comes from?
>
> Maybe the wording could be better. Perhaps "exceeds per-OSD max with
> mon_osd_max_split_count of 32". I'll submit this and see how it goes.
>
> >
> > P.S. I understand that the PG/OSD ratio configured on this cluster far
> > exceeds the recommended values. The experiment is to find scaling limits
> and
> > try out expansion scenarios.
> >
> > Thanks
> > Subhachandra
> >
> >
> >
> >
> >
> > _______________________________________________
> > ceph-users mailing list
> > ceph-users@lists.ceph.com
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >
>
>
>
> --
> Cheers,
> Brad
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to