Casey,

This was the result of trying 'data sync init':

root@c2-rgw1:~# radosgw-admin data sync init
ERROR: source zone not specified
root@c2-rgw1:~# radosgw-admin data sync init --source-zone=<master zone
uuid>
WARNING: cannot find source zone id for name=<master zone uuid>
ERROR: sync.init_sync_status() returned ret=-2
root@c2-rgw1:~# radosgw-admin data sync init --source-zone=c1-zone
ERROR: sync.init() returned ret=-5
2019-03-06 10:14:59.815735 7fecb214fe40  0 data sync: ERROR: failed to
fetch datalog info
root@c2-rgw1:~#

Do you have any further advice or info?

Thanks again,

Trey


On Wed, Mar 6, 2019 at 11:47 AM Casey Bodley <cbod...@redhat.com> wrote:

> Hi Trey,
>
> I think it's more likely that these stale metadata entries are from
> deleted buckets, rather than accidental bucket reshards. When a bucket
> is deleted in a multisite configuration, we don't delete its bucket
> instance because other zones may still need to sync the object deletes -
> and they can't make progress on sync if the bucket metadata disappears.
> These leftover bucket instances look the same to the 'reshard
> stale-instances' commands, but I'd be cautious about using that to
> remove them in multisite, as it may cause more sync errors and
> potentially leak storage if they still contain objects.
>
> Regarding 'datalog trim', that alone isn't safe because it could trim
> entries that hadn't been applied on other zones yet, causing them to
> miss some updates. What you can do is run 'data sync init' on each zone,
> and restart gateways. This will restart with a data full sync (which
> will scan all buckets for changes), and skip past any datalog entries
> from before the full sync. I was concerned that the bug in error
> handling (ie "ERROR: init sync on...") would also affect full sync, but
> that doesn't appear to be the case - so I do think that's worth trying.
>
> On 3/5/19 6:24 PM, Trey Palmer wrote:
> > Casey,
> >
> > Thanks very much for the reply!
> >
> > We definitely have lots of errors on sync-disabled buckets and the
> > workaround for that is obvious (most of them are empty anyway).
> >
> > Our second form of error is stale buckets.  We had dynamic resharding
> > enabled but have now disabled it (having discovered it was on by
> > default, and not supported in multisite).
> >
> > We removed several hundred stale buckets via 'radosgw-admin sharding
> > stale-instances rm', but they are still giving us sync errors.
> >
> > I have found that these buckets do have entries in 'radosgw-admin
> > datalog list', and my guess is this could be fixed by doing a
> > 'radosgw-admin datalog trim' for each entry on the master zone.
> >
> > Does that sound right?  :-)
> >
> > Thanks again for the detailed explanation,
> >
> > Trey Palmer
> >
> > On Tue, Mar 5, 2019 at 5:55 PM Casey Bodley <cbod...@redhat.com
> > <mailto:cbod...@redhat.com>> wrote:
> >
> >     Hi Christian,
> >
> >     I think you've correctly intuited that the issues are related to
> >     the use
> >     of 'bucket sync disable'. There was a bug fix for that feature in
> >     http://tracker.ceph.com/issues/26895, and I recently found that a
> >     block
> >     of code was missing from its luminous backport. That missing code is
> >     what handled those "ERROR: init sync on <bucket instance> failed,
> >     retcode=-2" errors.
> >
> >     I included a fix for that in a later backport
> >     (https://github.com/ceph/ceph/pull/26549), which I'm still working
> to
> >     get through qa. I'm afraid I can't really recommend a workaround
> >     for the
> >     issue in the meantime.
> >
> >     Looking forward though, we do plan to support something like s3's
> >     cross
> >     region replication so you can enable replication on a specific bucket
> >     without having to enable it globally.
> >
> >     Casey
> >
> >
> >     On 3/5/19 2:32 PM, Christian Rice wrote:
> >     >
> >     > Much appreciated.  We’ll continue to poke around and certainly will
> >     > disable the dynamic resharding.
> >     >
> >     > We started with 12.2.8 in production.  We definitely did not
> >     have it
> >     > enabled in ceph.conf
> >     >
> >     > *From: *Matthew H <matthew.he...@hotmail.com
> >     <mailto:matthew.he...@hotmail.com>>
> >     > *Date: *Tuesday, March 5, 2019 at 11:22 AM
> >     > *To: *Christian Rice <cr...@pandora.com
> >     <mailto:cr...@pandora.com>>, ceph-users
> >     > <ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com>>
> >     > *Cc: *Trey Palmer <nerdmagic...@gmail.com
> >     <mailto:nerdmagic...@gmail.com>>
> >     > *Subject: *Re: radosgw sync falling behind regularly
> >     >
> >     > Hi Christian,
> >     >
> >     > To be on the safe side and future proof yourself will want to go
> >     ahead
> >     > and set the following in your ceph.conf file, and then issue a
> >     restart
> >     > to your RGW instances.
> >     >
> >     > rgw_dynamic_resharding = false
> >     >
> >     > There are a number of issues with dynamic resharding, multisite rgw
> >     > problems being just one of them. However I thought it was disabled
> >     > automatically when multisite rgw is used (but I will have to double
> >     > check the code on that). What version of Ceph did you initially
> >     > install the cluster with? Prior to v12.2.2 this feature was
> >     enabled by
> >     > default for all rgw use cases.
> >     >
> >     > Thanks,
> >     >
> >     >
> >
>  ------------------------------------------------------------------------
> >     >
> >     > *From:*Christian Rice <cr...@pandora.com <mailto:cr...@pandora.com
> >>
> >     > *Sent:* Tuesday, March 5, 2019 2:07 PM
> >     > *To:* Matthew H; ceph-users
> >     > *Subject:* Re: radosgw sync falling behind regularly
> >     >
> >     > Matthew, first of all, let me say we very much appreciate your
> help!
> >     >
> >     > So I don’t think we turned dynamic resharding on, nor did we
> >     manually
> >     > reshard buckets. Seems like it defaults to on for luminous but the
> >     > mimic docs say it’s not supported in multisite.  So do we need to
> >     > disable it manually via tell and ceph.conf?
> >     >
> >     > Also, after running the command you suggested, all the stale
> >     instances
> >     > are gone…these from my examples were in output:
> >     >
> >     >         "bucket_instance":
> >     >
> >
>  "sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18467.303",
> >     >
> >     >         "bucket_instance":
> >     >
> >
>  "sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299",
> >     >
> >     >         "bucket_instance":
> >     >
> >
>  "sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.301",
> >     >
> >     > Though we still get lots of log messages like so in rgw:
> >     >
> >     > 2019-03-05 11:01:09.526120 7f64120ae700  0 ERROR: failed to get
> >     bucket
> >     > instance info for
> >     >
> >
>  
> .bucket.meta.sysad_task:sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299
> >     >
> >     > 2019-03-05 11:01:09.528664 7f63e5016700  1 civetweb:
> >     0x55976f1c2000:
> >     > 172.17.136.17 - - [05/Mar/2019:10:54:06 -0800] "GET
> >     >
> >
>  
> /admin/metadata/bucket.instance/sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299?key=sysad_task%2Fsysad-task%3A1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299&rgwx-zonegroup=de6af748-1a2f-44a1-9d44-30799cf1313e
> >
> >     > HTTP/1.1" 404 0 - -
> >     >
> >     > 2019-03-05 11:01:09.529648 7f64130b0700  0 meta sync: ERROR: can't
> >     > remove key:
> >     >
> >
>  
> bucket.instance:sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299
> >
> >     > ret=-2
> >     >
> >     > 2019-03-05 11:01:09.530324 7f64138b1700  0 ERROR: failed to get
> >     bucket
> >     > instance info for
> >     >
> >
>  
> .bucket.meta.sysad_task:sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299
> >     >
> >     > 2019-03-05 11:01:09.530345 7f6405094700  0 data sync: ERROR:
> >     failed to
> >     > retrieve bucket info for
> >     >
> >
>  bucket=sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299
> >     >
> >     > 2019-03-05 11:01:09.531774 7f6405094700  0 data sync: WARNING:
> >     > skipping data log entry for missing bucket
> >     >
> sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299
> >     >
> >     > 2019-03-05 11:01:09.571680 7f6405094700  0 data sync: ERROR:
> >     init sync
> >     > on
> >     >
> >     sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.302
> >     > failed, retcode=-2
> >     >
> >     > 2019-03-05 11:01:09.573179 7f6405094700  0 data sync: WARNING:
> >     > skipping data log entry for missing bucket
> >     >
> sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.302
> >     >
> >     > 2019-03-05 11:01:13.504308 7f63f903e700  1 civetweb:
> >     0x55976f0f2000:
> >     > 10.105.18.20 - - [05/Mar/2019:11:00:57 -0800] "GET
> >     >
> >
>  
> /admin/metadata/bucket.instance/sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299?key=sysad_task%2Fsysad-task%3A1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299&rgwx-zonegroup=de6af748-1a2f-44a1-9d44-30799cf1313e
> >
> >     > HTTP/1.1" 404 0 - -
> >     >
> >     > *From: *Matthew H <matthew.he...@hotmail.com
> >     <mailto:matthew.he...@hotmail.com>>
> >     > *Date: *Tuesday, March 5, 2019 at 10:03 AM
> >     > *To: *Christian Rice <cr...@pandora.com
> >     <mailto:cr...@pandora.com>>, ceph-users
> >     > <ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com>>
> >     > *Subject: *Re: radosgw sync falling behind regularly
> >     >
> >     > Hi Christian,
> >     >
> >     > You have stale bucket instances that need to be clean up, which is
> >     > what 'radosgw-admin reshard stale-instances list' is showing
> >     you. Have
> >     > you or were you manually resharding your buckets? The errors you
> >     are
> >     > seeing in the logs are related to these stale instances being kept
> >     > around.
> >     >
> >     > In v12.2.11 this command along with 'radosgw-admin reshard
> >     > stale-instance rm' was introduced [1].
> >     >
> >     > Hopefully this helps.
> >     >
> >     > [1]
> >     >
> >     > https://ceph.com/releases/v12-2-11-luminous-released/
> >     >
> >     <
> https://urldefense.proofpoint.com/v2/url?u=https-3A__ceph.com_releases_v12-2D2-2D11-2Dluminous-2Dreleased_&d=DwMF-g&c=gFTBenQ7Vj71sUi1A4CkFnmPzqwDo07QsHw-JRepxyw&r=NE1NbWtVhgG-K7YvLdoLZigfFx8zGPwOGk6HWpYK04I&m=vdtYIn6lEKaWD9wW297aHjQLpmQdHZrOVpOhmCBqkqo&s=nGCpS4p5jnaSpPUFlziSi3Y3pFijhVDy6e3867jA9BE&e=
> >
> >     >
> >     > /"There have been fixes to RGW dynamic and manual resharding,
> >     which no
> >     > longer
> >     > leaves behind stale bucket instances to be removed manually. For
> >     > finding and
> >     > cleaning up older instances from a reshard a radosgw-admin
> >     command reshard
> >     > stale-instances list and reshard stale-instances rm should do the
> >     > necessary
> >     > cleanup."/
> >     >
> >     >
> >
>  ------------------------------------------------------------------------
> >     >
> >     > *From:*Christian Rice <cr...@pandora.com <mailto:cr...@pandora.com
> >>
> >     > *Sent:* Tuesday, March 5, 2019 11:34 AM
> >     > *To:* Matthew H; ceph-users
> >     > *Subject:* Re: radosgw sync falling behind regularly
> >     >
> >     > The output of “radosgw-admin reshard stale-instances list” shows
> >     242
> >     > entries, which might embed too much proprietary info for me to
> >     list,
> >     > but here’s a tiny sample:
> >     >
> >     >
> >
>  "sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18467.303",
> >     >
> >     >
> >
>  "sysad_task/sysad_task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18467.281",
> >     >
> >     >
> >
>  "sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299",
> >     >
> >     >
> >
>  "sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.301",
> >     >
> >     > Some of appear repeatedly in the radosgw error logs like so:
> >     >
> >     > 2019-03-05 08:13:08.929206 7f6405094700  0 data sync: ERROR:
> >     init sync
> >     > on
> >     >
> >     sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.302
> >     > failed, retcode=-2
> >     >
> >     > 2019-03-05 08:13:08.930581 7f6405094700  0 data sync: WARNING:
> >     > skipping data log entry for missing bucket
> >     >
> sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.302
> >     >
> >     > 2019-03-05 08:13:08.972053 7f6405094700  0 data sync: ERROR:
> >     init sync
> >     > on
> >     >
> >     sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299
> >     > failed, retcode=-2
> >     >
> >     > 2019-03-05 08:13:08.973442 7f6405094700  0 data sync: WARNING:
> >     > skipping data log entry for missing bucket
> >     >
> sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299
> >     >
> >     > 2019-03-05 08:13:19.528295 7f6406897700  0 data sync: ERROR:
> >     init sync
> >     > on
> >     >
> >     sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299
> >     > failed, retcode=-2
> >     >
> >     > Notably, “Sync is disabled for bucket sysad-task.”  We use “bucket
> >     > sync disable” A LOT.  It’s the only way we’ve been able to use
> >     > multisite and a single namespace and not replicate things that are
> >     > unneeded to every zone.  Perhaps there’s a bug in the
> >     implementation
> >     > that’s tripping us up now, with the new sync multisite sync code
> >     from
> >     > 12.2.9 onward?
> >     >
> >     > What might we do with stale bucket instances, then?
> >     >
> >     > Of note, our master zone endpoint, which was timing out health
> >     checks
> >     > most of the day after the upgrade (was running but overworked by
> >     > cluster confusion, so we couldn’t create new buckets or do user
> >     ops),
> >     > has returned to availability late last night.  There’s a lot of
> >     data
> >     > to look at, but in my estimation, due to lack of user complaints
> >     (or
> >     > their unawareness of specific issues), it seems the zones are
> >     > nominally available, even with all the errors and warnings being
> >     > logged.  We’ve tested simple zone replication by creating a few
> >     files
> >     > in one zone and seeing them show up elsewhere…
> >     >
> >     > here’s “period get” output:
> >     >
> >     > sv5-ceph-rgw1
> >     >
> >     > {
> >     >
> >     >     "id": "3d0d40ef-90de-40ea-8c44-caa20ea8dc53",
> >     >
> >     > "epoch": 16,
> >     >
> >     > "predecessor_uuid": "926c74c7-c1a7-46b1-9f25-eb5c392a7fbb",
> >     >
> >     > "sync_status": [],
> >     >
> >     > "period_map": {
> >     >
> >     > "id": "3d0d40ef-90de-40ea-8c44-caa20ea8dc53",
> >     >
> >     > "zonegroups": [
> >     >
> >     > {
> >     >
> >     > "id": "de6af748-1a2f-44a1-9d44-30799cf1313e",
> >     >
> >     > "name": "us",
> >     >
> >     > "api_name": "us",
> >     >
> >     > "is_master": "true",
> >     >
> >     > "endpoints": [
> >     >
> >     >                    "http://sv5-ceph-rgw1.savagebeast.com:8080";
> >     >
> >     > ],
> >     >
> >     > "hostnames": [],
> >     >
> >     > "hostnames_s3website": [],
> >     >
> >     > "master_zone": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04",
> >     >
> >     > "zones": [
> >     >
> >     > {
> >     >
> >     > "id": "107d29a0-b732-4bf1-a26e-1f64f820e839",
> >     >
> >     > "name": "dc11-prod",
> >     >
> >     > "endpoints": [
> >     >
> >     > "http://dc11-ceph-rgw1:8080";
> >     >
> >     >  ],
> >     >
> >     > "log_meta": "false",
> >     >
> >     > "log_data": "true",
> >     >
> >     > "bucket_index_max_shards": 0,
> >     >
> >     > "read_only": "false",
> >     >
> >     > "tier_type": "",
> >     >
> >     >   "sync_from_all": "true",
> >     >
> >     > "sync_from": []
> >     >
> >     > },
> >     >
> >     > {
> >     >
> >     > "id": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04",
> >     >
> >     > "name": "sv5-corp",
> >     >
> >     > "endpoints": [
> >     >
> >     > "http://sv5-ceph-rgw1.savagebeast.com:8080";
> >     >
> >     > ],
> >     >
> >     > "log_meta": "false",
> >     >
> >     > "log_data": "true",
> >     >
> >     > "bucket_index_max_shards": 0,
> >     >
> >     >                    "read_only": "false",
> >     >
> >     > "tier_type": "",
> >     >
> >     > "sync_from_all": "true",
> >     >
> >     > "sync_from": []
> >     >
> >     > },
> >     >
> >     > {
> >     >
> >     > "id": "331d3f1e-1b72-4c56-bb5a-d1d0fcf6d0b8",
> >     >
> >     > "name": "sv3-prod",
> >     >
> >     > "endpoints": [
> >     >
> >     > "http://sv3-ceph-rgw1:8080";
> >     >
> >     > ],
> >     >
> >     >     "log_meta": "false",
> >     >
> >     > "log_data": "true",
> >     >
> >     > "bucket_index_max_shards": 0,
> >     >
> >     > "read_only": "false",
> >     >
> >     > "tier_type": "",
> >     >
> >     > "sync_from_all": "true",
> >     >
> >     > "sync_from": []
> >     >
> >     > }
> >     >
> >     > ],
> >     >
> >     > "placement_targets": [
> >     >
> >     > {
> >     >
> >     > "name": "default-placement",
> >     >
> >     > "tags": []
> >     >
> >     > }
> >     >
> >     >                ],
> >     >
> >     > "default_placement": "default-placement",
> >     >
> >     > "realm_id": "b3e2afe7-2254-494a-9a34-ce50358779fd"
> >     >
> >     > }
> >     >
> >     >         ],
> >     >
> >     > "short_zone_ids": [
> >     >
> >     > {
> >     >
> >     > "key": "107d29a0-b732-4bf1-a26e-1f64f820e839",
> >     >
> >     > "val": 1720993486
> >     >
> >     > },
> >     >
> >     > {
> >     >
> >     > "key": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04",
> >     >
> >     > "val": 2301637458
> >     >
> >     > },
> >     >
> >     > {
> >     >
> >     > "key": "331d3f1e-1b72-4c56-bb5a-d1d0fcf6d0b8",
> >     >
> >     > "val": 1449486239
> >     >
> >     > }
> >     >
> >     >         ]
> >     >
> >     >     },
> >     >
> >     > "master_zonegroup": "de6af748-1a2f-44a1-9d44-30799cf1313e",
> >     >
> >     > "master_zone": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04",
> >     >
> >     > "period_config": {
> >     >
> >     > "bucket_quota": {
> >     >
> >     > "enabled": false,
> >     >
> >     > "check_on_raw": false,
> >     >
> >     > "max_size": -1,
> >     >
> >     > "max_size_kb": 0,
> >     >
> >     > "max_objects": -1
> >     >
> >     >         },
> >     >
> >     > "user_quota": {
> >     >
> >     > "enabled": false,
> >     >
> >     > "check_on_raw": false,
> >     >
> >     > "max_size": -1,
> >     >
> >     > "max_size_kb": 0,
> >     >
> >     > "max_objects": -1
> >     >
> >     >         }
> >     >
> >     >     },
> >     >
> >     > "realm_id": "b3e2afe7-2254-494a-9a34-ce50358779fd",
> >     >
> >     > "realm_name": "savagebucket",
> >     >
> >     > "realm_epoch": 2
> >     >
> >     > }
> >     >
> >     > sv3-ceph-rgw1
> >     >
> >     > {
> >     >
> >     >     "id": "3d0d40ef-90de-40ea-8c44-caa20ea8dc53",
> >     >
> >     > "epoch": 16,
> >     >
> >     > "predecessor_uuid": "926c74c7-c1a7-46b1-9f25-eb5c392a7fbb",
> >     >
> >     > "sync_status": [],
> >     >
> >     > "period_map": {
> >     >
> >     > "id": "3d0d40ef-90de-40ea-8c44-caa20ea8dc53",
> >     >
> >     > "zonegroups": [
> >     >
> >     > {
> >     >
> >     > "id": "de6af748-1a2f-44a1-9d44-30799cf1313e",
> >     >
> >     > "name": "us",
> >     >
> >     > "api_name": "us",
> >     >
> >     > "is_master": "true",
> >     >
> >     > "endpoints": [
> >     >
> >     >         "http://sv5-ceph-rgw1.savagebeast.com:8080";
> >     >
> >     > ],
> >     >
> >     > "hostnames": [],
> >     >
> >     > "hostnames_s3website": [],
> >     >
> >     > "master_zone": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04",
> >     >
> >     > "zones": [
> >     >
> >     >          {
> >     >
> >     > "id": "107d29a0-b732-4bf1-a26e-1f64f820e839",
> >     >
> >     > "name": "dc11-prod",
> >     >
> >     > "endpoints": [
> >     >
> >     > "http://dc11-ceph-rgw1:8080";
> >     >
> >     > ],
> >     >
> >     > "log_meta": "false",
> >     >
> >     > "log_data": "true",
> >     >
> >     > "bucket_index_max_shards": 0,
> >     >
> >     > "read_only": "false",
> >     >
> >     > "tier_type": "",
> >     >
> >     > "sync_from_all": "true",
> >     >
> >     > "sync_from": []
> >     >
> >     > },
> >     >
> >     > {
> >     >
> >     > "id": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04",
> >     >
> >     > "name": "sv5-corp",
> >     >
> >     > "endpoints": [
> >     >
> >     > "http://sv5-ceph-rgw1.savagebeast.com:8080";
> >     >
> >     > ],
> >     >
> >     > "log_meta": "false",
> >     >
> >     > "log_data": "true",
> >     >
> >     > "bucket_index_max_shards": 0,
> >     >
> >     >                "read_only": "false",
> >     >
> >     > "tier_type": "",
> >     >
> >     > "sync_from_all": "true",
> >     >
> >     > "sync_from": []
> >     >
> >     > },
> >     >
> >     > {
> >     >
> >     > "id": "331d3f1e-1b72-4c56-bb5a-d1d0fcf6d0b8",
> >     >
> >     > "name": "sv3-prod",
> >     >
> >     > "endpoints": [
> >     >
> >     > "http://sv3-ceph-rgw1:8080";
> >     >
> >     > ],
> >     >
> >     > "log_meta": "false",
> >     >
> >     > "log_data": "true",
> >     >
> >     > "bucket_index_max_shards": 0,
> >     >
> >     > "read_only": "false",
> >     >
> >     > "tier_type": "",
> >     >
> >     > "sync_from_all": "true",
> >     >
> >     >   "sync_from": []
> >     >
> >     > }
> >     >
> >     > ],
> >     >
> >     > "placement_targets": [
> >     >
> >     > {
> >     >
> >     > "name": "default-placement",
> >     >
> >     > "tags": []
> >     >
> >     > }
> >     >
> >     > ],
> >     >
> >     >          "default_placement": "default-placement",
> >     >
> >     > "realm_id": "b3e2afe7-2254-494a-9a34-ce50358779fd"
> >     >
> >     > }
> >     >
> >     >         ],
> >     >
> >     > "short_zone_ids": [
> >     >
> >     > {
> >     >
> >     > "key": "107d29a0-b732-4bf1-a26e-1f64f820e839",
> >     >
> >     >         "val": 1720993486
> >     >
> >     > },
> >     >
> >     > {
> >     >
> >     > "key": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04",
> >     >
> >     > "val": 2301637458
> >     >
> >     > },
> >     >
> >     > {
> >     >
> >     > "key": "331d3f1e-1b72-4c56-bb5a-d1d0fcf6d0b8",
> >     >
> >     > "val": 1449486239
> >     >
> >     > }
> >     >
> >     >         ]
> >     >
> >     >     },
> >     >
> >     > "master_zonegroup": "de6af748-1a2f-44a1-9d44-30799cf1313e",
> >     >
> >     > "master_zone": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04",
> >     >
> >     > "period_config": {
> >     >
> >     > "bucket_quota": {
> >     >
> >     > "enabled": false,
> >     >
> >     > "check_on_raw": false,
> >     >
> >     > "max_size": -1,
> >     >
> >     > "max_size_kb": 0,
> >     >
> >     > "max_objects": -1
> >     >
> >     >         },
> >     >
> >     > "user_quota": {
> >     >
> >     > "enabled": false,
> >     >
> >     > "check_on_raw": false,
> >     >
> >     > "max_size": -1,
> >     >
> >     > "max_size_kb": 0,
> >     >
> >     > "max_objects": -1
> >     >
> >     >         }
> >     >
> >     >     },
> >     >
> >     > "realm_id": "b3e2afe7-2254-494a-9a34-ce50358779fd",
> >     >
> >     > "realm_name": "savagebucket",
> >     >
> >     > "realm_epoch": 2
> >     >
> >     > }
> >     >
> >     > dc11-ceph-rgw1
> >     >
> >     > {
> >     >
> >     >     "id": "3d0d40ef-90de-40ea-8c44-caa20ea8dc53",
> >     >
> >     > "epoch": 16,
> >     >
> >     > "predecessor_uuid": "926c74c7-c1a7-46b1-9f25-eb5c392a7fbb",
> >     >
> >     > "sync_status": [],
> >     >
> >     > "period_map": {
> >     >
> >     > "id": "3d0d40ef-90de-40ea-8c44-caa20ea8dc53",
> >     >
> >     > "zonegroups": [
> >     >
> >     >         {
> >     >
> >     > "id": "de6af748-1a2f-44a1-9d44-30799cf1313e",
> >     >
> >     > "name": "us",
> >     >
> >     > "api_name": "us",
> >     >
> >     > "is_master": "true",
> >     >
> >     > "endpoints": [
> >     >
> >     > "http://sv5-ceph-rgw1.savagebeast.com:8080";
> >     >
> >     > ],
> >     >
> >     > "hostnames": [],
> >     >
> >     > "hostnames_s3website": [],
> >     >
> >     > "master_zone": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04",
> >     >
> >     > "zones": [
> >     >
> >     > {
> >     >
> >     > "id": "107d29a0-b732-4bf1-a26e-1f64f820e839",
> >     >
> >     > "name": "dc11-prod",
> >     >
> >     > "endpoints": [
> >     >
> >     > "http://dc11-ceph-rgw1:8080";
> >     >
> >     > ],
> >     >
> >     > "log_meta": "false",
> >     >
> >     >                       "log_data": "true",
> >     >
> >     > "bucket_index_max_shards": 0,
> >     >
> >     > "read_only": "false",
> >     >
> >     > "tier_type": "",
> >     >
> >     > "sync_from_all": "true",
> >     >
> >     > "sync_from": []
> >     >
> >     > },
> >     >
> >     > {
> >     >
> >     > "id": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04",
> >     >
> >     > "name": "sv5-corp",
> >     >
> >     > "endpoints": [
> >     >
> >     >           "http://sv5-ceph-rgw1.savagebeast.com:8080";
> >     >
> >     > ],
> >     >
> >     > "log_meta": "false",
> >     >
> >     > "log_data": "true",
> >     >
> >     > "bucket_index_max_shards": 0,
> >     >
> >     > "read_only": "false",
> >     >
> >     > "tier_type": "",
> >     >
> >     > "sync_from_all": "true",
> >     >
> >     > "sync_from": []
> >     >
> >     > },
> >     >
> >     > {
> >     >
> >     > "id": "331d3f1e-1b72-4c56-bb5a-d1d0fcf6d0b8",
> >     >
> >     > "name": "sv3-prod",
> >     >
> >     > "endpoints": [
> >     >
> >     > "http://sv3-ceph-rgw1:8080";
> >     >
> >     > ],
> >     >
> >     > "log_meta": "false",
> >     >
> >     > "log_data": "true",
> >     >
> >     > "bucket_index_max_shards": 0,
> >     >
> >     > "read_only": "false",
> >     >
> >     > "tier_type": "",
> >     >
> >     > "sync_from_all": "true",
> >     >
> >     > "sync_from": []
> >     >
> >     > }
> >     >
> >     > ],
> >     >
> >     > "placement_targets": [
> >     >
> >     > {
> >     >
> >     > "name": "default-placement",
> >     >
> >     > "tags": []
> >     >
> >     > }
> >     >
> >     > ],
> >     >
> >     > "default_placement": "default-placement",
> >     >
> >     > "realm_id": "b3e2afe7-2254-494a-9a34-ce50358779fd"
> >     >
> >     > }
> >     >
> >     >         ],
> >     >
> >     > "short_zone_ids": [
> >     >
> >     > {
> >     >
> >     > "key": "107d29a0-b732-4bf1-a26e-1f64f820e839",
> >     >
> >     > "val": 1720993486
> >     >
> >     > },
> >     >
> >     > {
> >     >
> >     > "key": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04",
> >     >
> >     > "val": 2301637458
> >     >
> >     > },
> >     >
> >     > {
> >     >
> >     > "key": "331d3f1e-1b72-4c56-bb5a-d1d0fcf6d0b8",
> >     >
> >     > "val": 1449486239
> >     >
> >     > }
> >     >
> >     >         ]
> >     >
> >     >     },
> >     >
> >     > "master_zonegroup": "de6af748-1a2f-44a1-9d44-30799cf1313e",
> >     >
> >     > "master_zone": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04",
> >     >
> >     > "period_config": {
> >     >
> >     > "bucket_quota": {
> >     >
> >     > "enabled": false,
> >     >
> >     > "check_on_raw": false,
> >     >
> >     >           "max_size": -1,
> >     >
> >     > "max_size_kb": 0,
> >     >
> >     > "max_objects": -1
> >     >
> >     >         },
> >     >
> >     > "user_quota": {
> >     >
> >     > "enabled": false,
> >     >
> >     > "check_on_raw": false,
> >     >
> >     > "max_size": -1,
> >     >
> >     > "max_size_kb": 0,
> >     >
> >     > "max_objects": -1
> >     >
> >     >         }
> >     >
> >     >     },
> >     >
> >     > "realm_id": "b3e2afe7-2254-494a-9a34-ce50358779fd",
> >     >
> >     > "realm_name": "savagebucket",
> >     >
> >     > "realm_epoch": 2
> >     >
> >     > }
> >     >
> >     > *From: *Matthew H <matthew.he...@hotmail.com
> >     <mailto:matthew.he...@hotmail.com>>
> >     > *Date: *Tuesday, March 5, 2019 at 4:31 AM
> >     > *To: *Christian Rice <cr...@pandora.com
> >     <mailto:cr...@pandora.com>>, ceph-users
> >     > <ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com>>
> >     > *Subject: *Re: radosgw sync falling behind regularly
> >     >
> >     > Hi Christian,
> >     >
> >     > You haven't resharded any of your buckets have you?  You can run
> >     the
> >     > command below in v12.2.11 to list stale bucket instances.
> >     >
> >     > radosgw-admin reshard stale-instances list
> >     >
> >     > Can you also send the output from the following command on each
> rgw?
> >     >
> >     > radosgw-admin period get
> >     >
> >     >
> >     > _______________________________________________
> >     > ceph-users mailing list
> >     > ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com>
> >     > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >     _______________________________________________
> >     ceph-users mailing list
> >     ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com>
> >     http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to