rgw uses more than one pool. (5 or 6 IIRC)

-- 
Paul Emmerich

Looking for help with your Ceph cluster? Contact us at https://croit.io

croit GmbH
Freseniusstr. 31h
81247 München
www.croit.io
Tel: +49 89 1896585 90


On Sun, Jun 9, 2019 at 7:00 PM <dhils...@performair.com> wrote:

> Huan;
>
> I get that, but the pool already exists, why is radosgw trying to create
> one?
>
> Dominic Hilsbos
>
> Get Outlook for Android <https://aka.ms/ghei36>
>
>
>
>
> On Sat, Jun 8, 2019 at 2:55 AM -0700, "huang jun" <hjwsm1...@gmail.com>
> wrote:
>
> From the error message, i'm decline to that 'mon_max_pg_per_osd' was exceed,
>> you can check the value of it, and its default value is 250, so you
>> can at most have 1500pgs(250*6osds),
>> and for replicated pools with size=3, you can have 500pgs for all pools,
>> you already have 448pgs, so the next pool can create at most 500-448=52pgs.
>>  于2019年6月8日周六 下午2:41写道:
>> >
>> > All;
>> >
>> > I have a test and demonstration cluster running (3 hosts, MON, MGR, 2x OSD 
>> > per host), and I'm trying to add a 4th host for gateway purposes.
>> >
>> > The radosgw process keeps dying with:
>> > 2019-06-07 15:59:50.700 7fc4ef273780  0 ceph version 14.2.1 
>> > (d555a9489eb35f84f2e1ef49b77e19da9d113972) nautilus (stable), process 
>> > radosgw, pid 17588
>> > 2019-06-07 15:59:51.358 7fc4ef273780  0 rgw_init_ioctx ERROR: 
>> > librados::Rados::pool_create returned (34) Numerical result out of range 
>> > (this can be due to a pool or placement group misconfiguration, e.g. 
>> > pg_num < pgp_num or mon_max_pg_per_osd exceeded)
>> > 2019-06-07 15:59:51.396 7fc4ef273780 -1 Couldn't init storage provider 
>> > (RADOS)
>> >
>> > The .rgw.root pool already exists.
>> >
>> > ceph status returns:
>> >   cluster:
>> >     id:     1a8a1693-fa54-4cb3-89d2-7951d4cee6a3
>> >     health: HEALTH_OK
>> >
>> >   services:
>> >     mon: 3 daemons, quorum S700028,S700029,S700030 (age 30m)
>> >     mgr: S700028(active, since 47h), standbys: S700030, S700029
>> >     osd: 6 osds: 6 up (since 2d), 6 in (since 3d)
>> >
>> >   data:
>> >     pools:   5 pools, 448 pgs
>> >     objects: 12 objects, 1.2 KiB
>> >     usage:   722 GiB used, 65 TiB / 66 TiB avail
>> >     pgs:     448 active+clean
>> >
>> > and ceph osd tree returns:
>> > ID CLASS WEIGHT   TYPE NAME        STATUS REWEIGHT PRI-AFF
>> > -1       66.17697 root default
>> > -5       22.05899     host S700029
>> >  2   hdd 11.02950         osd.2        up  1.00000 1.00000
>> >  3   hdd 11.02950         osd.3        up  1.00000 1.00000
>> > -7       22.05899     host S700030
>> >  4   hdd 11.02950         osd.4        up  1.00000 1.00000
>> >  5   hdd 11.02950         osd.5        up  1.00000 1.00000
>> > -3       22.05899     host s700028
>> >  0   hdd 11.02950         osd.0        up  1.00000 1.00000
>> >  1   hdd 11.02950         osd.1        up  1.00000 1.00000
>> >
>> > Any thoughts on what I'm missing?
>> >
>> > Thank you,
>> >
>> > Dominic L. Hilsbos, MBA
>> > Director - Information Technology
>> > Perform Air International Inc.
>> > dhils...@performair.com
>> > www.PerformAir.com
>> >
>> >
>> >
>> > _______________________________________________
>> > ceph-users mailing list
>> > ceph-users@lists.ceph.com
>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>>
>>
>> --
>> Thank you!
>> HuangJun
>>
>> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to