Bryan, I just went through this myself, also on Hammer, as offline bucket index resharding was backported. I had three buckets with > 10 million objects each, one of them with 30 million. I was experiencing the typical blocked request issue during scrubs, when the placement group containing the bucket index got hit.
I solved it in two steps. First, I added an SSD only pool, and moved the bucket index to this new SSD pool. This is an on-line operation. After that was complete I scheduled some downtime (we run a highly available consumer facing website), and made a plan to reshard the bucket indexes. I did some tests with buckets containing 100,000 test objects and found the performance to be satisfactory. Once my maintenance window hit and I stopped all access to RGW, I was able to reshard all my bucket indexes in 20 minutes. I can't remember exact numbers, but I believe I did a 20+ million bucket in about 5 minutes. It was extremely fast, but again I had moved my bucket indexes to an SSD backed pool of fast enterprise SSDs (three hosts, one SSD per host, Samsung 3.84tb PM863a for what it's worth). Once I finished this, all my ceph performance issues disappeared. I'll slowly upgrade my cluster with the end goal of moving to the more efficient bluestore, but I no longer feel the rush. Last detail: I used 100 shards per bucket which seems to be a good compromise. Cullen > Date: Fri, 9 Jun 2017 14:58:41 -0700 > From: Yehuda Sadeh-Weinraub <yeh...@redhat.com> > To: Dan van der Ster <d...@vanderster.com> > Cc: "ceph-users@lists.ceph.com" <ceph-users@lists.ceph.com> > Subject: Re: [ceph-users] Living with huge bucket sizes > Message-ID: > <CADRKj5SMdrA2FZW8W5VY0tzNi_21dAo7r9gWLevV9CMygkiuTw@mail. > gmail.com> > Content-Type: text/plain; charset="UTF-8" > > On Fri, Jun 9, 2017 at 2:21 AM, Dan van der Ster <d...@vanderster.com> > wrote: > > Hi Bryan, > > > > On Fri, Jun 9, 2017 at 1:55 AM, Bryan Stillwell <bstillw...@godaddy.com> > wrote: > >> This has come up quite a few times before, but since I was only working > with > >> RBD before I didn't pay too close attention to the conversation. I'm > >> looking > >> for the best way to handle existing clusters that have buckets with a > large > >> number of objects (>20 million) in them. The cluster I'm doing test on > is > >> currently running hammer (0.94.10), so if things got better in jewel I > would > >> love to hear about it! > >> ... > >> Has anyone found a good solution for this for existing large buckets? I > >> know sharding is the solution going forward, but afaik it can't be done > >> on existing buckets yet (although the dynamic resharding work mentioned > >> on today's performance call sounds promising). > > > > I haven't tried it myself, but 0.94.10 should have the (offline) > > resharding feature. From the release notes: > > > > Right. We did add automatic dynamic resharding to Luminous, but > offline resharding should be enough. > > > >> * In RADOS Gateway, it is now possible to reshard an existing bucket's > index > >> using an off-line tool. > >> > >> Usage: > >> > >> $ radosgw-admin bucket reshard --bucket=<bucket_name> > --num_shards=<num_shards> > >> > >> This will create a new linked bucket instance that points to the newly > created > >> index objects. The old bucket instance still exists and currently it's > up to > >> the user to manually remove the old bucket index objects. (Note that > bucket > >> resharding currently requires that all IO (especially writes) to the > specific > >> bucket is quiesced.) > > Once resharding is done, use the radosgw-admin bi purge command to > remove the old bucket indexes. > > Yehuda > > > > > -- Dan > > _______________________________________________ > > ceph-users mailing list > > ceph-users@lists.ceph.com > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > ------------------------------ > > Subject: Digest Footer > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > ------------------------------ > > End of ceph-users Digest, Vol 53, Issue 9 > ***************************************** >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com