It appears we eventually got 'data sync init' working.

At least, it's worked on 5 of the 6 sync directions in our 3-node cluster.
 The sixth has not run without an error returned, although 'sync status'
does say "preparing for full sync".

Thanks,

Trey

On Wed, Mar 6, 2019 at 1:22 PM Trey Palmer <nerdmagic...@gmail.com> wrote:

> Casey,
>
> This was the result of trying 'data sync init':
>
> root@c2-rgw1:~# radosgw-admin data sync init
> ERROR: source zone not specified
> root@c2-rgw1:~# radosgw-admin data sync init --source-zone=<master zone
> uuid>
> WARNING: cannot find source zone id for name=<master zone uuid>
> ERROR: sync.init_sync_status() returned ret=-2
> root@c2-rgw1:~# radosgw-admin data sync init --source-zone=c1-zone
> ERROR: sync.init() returned ret=-5
> 2019-03-06 10:14:59.815735 7fecb214fe40  0 data sync: ERROR: failed to
> fetch datalog info
> root@c2-rgw1:~#
>
> Do you have any further advice or info?
>
> Thanks again,
>
> Trey
>
>
> On Wed, Mar 6, 2019 at 11:47 AM Casey Bodley <cbod...@redhat.com> wrote:
>
>> Hi Trey,
>>
>> I think it's more likely that these stale metadata entries are from
>> deleted buckets, rather than accidental bucket reshards. When a bucket
>> is deleted in a multisite configuration, we don't delete its bucket
>> instance because other zones may still need to sync the object deletes -
>> and they can't make progress on sync if the bucket metadata disappears.
>> These leftover bucket instances look the same to the 'reshard
>> stale-instances' commands, but I'd be cautious about using that to
>> remove them in multisite, as it may cause more sync errors and
>> potentially leak storage if they still contain objects.
>>
>> Regarding 'datalog trim', that alone isn't safe because it could trim
>> entries that hadn't been applied on other zones yet, causing them to
>> miss some updates. What you can do is run 'data sync init' on each zone,
>> and restart gateways. This will restart with a data full sync (which
>> will scan all buckets for changes), and skip past any datalog entries
>> from before the full sync. I was concerned that the bug in error
>> handling (ie "ERROR: init sync on...") would also affect full sync, but
>> that doesn't appear to be the case - so I do think that's worth trying.
>>
>> On 3/5/19 6:24 PM, Trey Palmer wrote:
>> > Casey,
>> >
>> > Thanks very much for the reply!
>> >
>> > We definitely have lots of errors on sync-disabled buckets and the
>> > workaround for that is obvious (most of them are empty anyway).
>> >
>> > Our second form of error is stale buckets.  We had dynamic resharding
>> > enabled but have now disabled it (having discovered it was on by
>> > default, and not supported in multisite).
>> >
>> > We removed several hundred stale buckets via 'radosgw-admin sharding
>> > stale-instances rm', but they are still giving us sync errors.
>> >
>> > I have found that these buckets do have entries in 'radosgw-admin
>> > datalog list', and my guess is this could be fixed by doing a
>> > 'radosgw-admin datalog trim' for each entry on the master zone.
>> >
>> > Does that sound right?  :-)
>> >
>> > Thanks again for the detailed explanation,
>> >
>> > Trey Palmer
>> >
>> > On Tue, Mar 5, 2019 at 5:55 PM Casey Bodley <cbod...@redhat.com
>> > <mailto:cbod...@redhat.com>> wrote:
>> >
>> >     Hi Christian,
>> >
>> >     I think you've correctly intuited that the issues are related to
>> >     the use
>> >     of 'bucket sync disable'. There was a bug fix for that feature in
>> >     http://tracker.ceph.com/issues/26895, and I recently found that a
>> >     block
>> >     of code was missing from its luminous backport. That missing code is
>> >     what handled those "ERROR: init sync on <bucket instance> failed,
>> >     retcode=-2" errors.
>> >
>> >     I included a fix for that in a later backport
>> >     (https://github.com/ceph/ceph/pull/26549), which I'm still working
>> to
>> >     get through qa. I'm afraid I can't really recommend a workaround
>> >     for the
>> >     issue in the meantime.
>> >
>> >     Looking forward though, we do plan to support something like s3's
>> >     cross
>> >     region replication so you can enable replication on a specific
>> bucket
>> >     without having to enable it globally.
>> >
>> >     Casey
>> >
>> >
>> >     On 3/5/19 2:32 PM, Christian Rice wrote:
>> >     >
>> >     > Much appreciated.  We’ll continue to poke around and certainly
>> will
>> >     > disable the dynamic resharding.
>> >     >
>> >     > We started with 12.2.8 in production.  We definitely did not
>> >     have it
>> >     > enabled in ceph.conf
>> >     >
>> >     > *From: *Matthew H <matthew.he...@hotmail.com
>> >     <mailto:matthew.he...@hotmail.com>>
>> >     > *Date: *Tuesday, March 5, 2019 at 11:22 AM
>> >     > *To: *Christian Rice <cr...@pandora.com
>> >     <mailto:cr...@pandora.com>>, ceph-users
>> >     > <ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com>>
>> >     > *Cc: *Trey Palmer <nerdmagic...@gmail.com
>> >     <mailto:nerdmagic...@gmail.com>>
>> >     > *Subject: *Re: radosgw sync falling behind regularly
>> >     >
>> >     > Hi Christian,
>> >     >
>> >     > To be on the safe side and future proof yourself will want to go
>> >     ahead
>> >     > and set the following in your ceph.conf file, and then issue a
>> >     restart
>> >     > to your RGW instances.
>> >     >
>> >     > rgw_dynamic_resharding = false
>> >     >
>> >     > There are a number of issues with dynamic resharding, multisite
>> rgw
>> >     > problems being just one of them. However I thought it was disabled
>> >     > automatically when multisite rgw is used (but I will have to
>> double
>> >     > check the code on that). What version of Ceph did you initially
>> >     > install the cluster with? Prior to v12.2.2 this feature was
>> >     enabled by
>> >     > default for all rgw use cases.
>> >     >
>> >     > Thanks,
>> >     >
>> >     >
>> >
>>  ------------------------------------------------------------------------
>> >     >
>> >     > *From:*Christian Rice <cr...@pandora.com <mailto:
>> cr...@pandora.com>>
>> >     > *Sent:* Tuesday, March 5, 2019 2:07 PM
>> >     > *To:* Matthew H; ceph-users
>> >     > *Subject:* Re: radosgw sync falling behind regularly
>> >     >
>> >     > Matthew, first of all, let me say we very much appreciate your
>> help!
>> >     >
>> >     > So I don’t think we turned dynamic resharding on, nor did we
>> >     manually
>> >     > reshard buckets. Seems like it defaults to on for luminous but the
>> >     > mimic docs say it’s not supported in multisite.  So do we need to
>> >     > disable it manually via tell and ceph.conf?
>> >     >
>> >     > Also, after running the command you suggested, all the stale
>> >     instances
>> >     > are gone…these from my examples were in output:
>> >     >
>> >     >         "bucket_instance":
>> >     >
>> >
>>  "sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18467.303",
>> >     >
>> >     >         "bucket_instance":
>> >     >
>> >
>>  "sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299",
>> >     >
>> >     >         "bucket_instance":
>> >     >
>> >
>>  "sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.301",
>> >     >
>> >     > Though we still get lots of log messages like so in rgw:
>> >     >
>> >     > 2019-03-05 11:01:09.526120 7f64120ae700  0 ERROR: failed to get
>> >     bucket
>> >     > instance info for
>> >     >
>> >
>>  
>> .bucket.meta.sysad_task:sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299
>> >     >
>> >     > 2019-03-05 11:01:09.528664 7f63e5016700  1 civetweb:
>> >     0x55976f1c2000:
>> >     > 172.17.136.17 - - [05/Mar/2019:10:54:06 -0800] "GET
>> >     >
>> >
>>  
>> /admin/metadata/bucket.instance/sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299?key=sysad_task%2Fsysad-task%3A1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299&rgwx-zonegroup=de6af748-1a2f-44a1-9d44-30799cf1313e
>> >
>> >     > HTTP/1.1" 404 0 - -
>> >     >
>> >     > 2019-03-05 11:01:09.529648 7f64130b0700  0 meta sync: ERROR: can't
>> >     > remove key:
>> >     >
>> >
>>  
>> bucket.instance:sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299
>> >
>> >     > ret=-2
>> >     >
>> >     > 2019-03-05 11:01:09.530324 7f64138b1700  0 ERROR: failed to get
>> >     bucket
>> >     > instance info for
>> >     >
>> >
>>  
>> .bucket.meta.sysad_task:sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299
>> >     >
>> >     > 2019-03-05 11:01:09.530345 7f6405094700  0 data sync: ERROR:
>> >     failed to
>> >     > retrieve bucket info for
>> >     >
>> >
>>  bucket=sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299
>> >     >
>> >     > 2019-03-05 11:01:09.531774 7f6405094700  0 data sync: WARNING:
>> >     > skipping data log entry for missing bucket
>> >     >
>> sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299
>> >     >
>> >     > 2019-03-05 11:01:09.571680 7f6405094700  0 data sync: ERROR:
>> >     init sync
>> >     > on
>> >     >
>> >     sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.302
>> >     > failed, retcode=-2
>> >     >
>> >     > 2019-03-05 11:01:09.573179 7f6405094700  0 data sync: WARNING:
>> >     > skipping data log entry for missing bucket
>> >     >
>> sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.302
>> >     >
>> >     > 2019-03-05 11:01:13.504308 7f63f903e700  1 civetweb:
>> >     0x55976f0f2000:
>> >     > 10.105.18.20 - - [05/Mar/2019:11:00:57 -0800] "GET
>> >     >
>> >
>>  
>> /admin/metadata/bucket.instance/sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299?key=sysad_task%2Fsysad-task%3A1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299&rgwx-zonegroup=de6af748-1a2f-44a1-9d44-30799cf1313e
>> >
>> >     > HTTP/1.1" 404 0 - -
>> >     >
>> >     > *From: *Matthew H <matthew.he...@hotmail.com
>> >     <mailto:matthew.he...@hotmail.com>>
>> >     > *Date: *Tuesday, March 5, 2019 at 10:03 AM
>> >     > *To: *Christian Rice <cr...@pandora.com
>> >     <mailto:cr...@pandora.com>>, ceph-users
>> >     > <ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com>>
>> >     > *Subject: *Re: radosgw sync falling behind regularly
>> >     >
>> >     > Hi Christian,
>> >     >
>> >     > You have stale bucket instances that need to be clean up, which is
>> >     > what 'radosgw-admin reshard stale-instances list' is showing
>> >     you. Have
>> >     > you or were you manually resharding your buckets? The errors you
>> >     are
>> >     > seeing in the logs are related to these stale instances being kept
>> >     > around.
>> >     >
>> >     > In v12.2.11 this command along with 'radosgw-admin reshard
>> >     > stale-instance rm' was introduced [1].
>> >     >
>> >     > Hopefully this helps.
>> >     >
>> >     > [1]
>> >     >
>> >     > https://ceph.com/releases/v12-2-11-luminous-released/
>> >     >
>> >     <
>> https://urldefense.proofpoint.com/v2/url?u=https-3A__ceph.com_releases_v12-2D2-2D11-2Dluminous-2Dreleased_&d=DwMF-g&c=gFTBenQ7Vj71sUi1A4CkFnmPzqwDo07QsHw-JRepxyw&r=NE1NbWtVhgG-K7YvLdoLZigfFx8zGPwOGk6HWpYK04I&m=vdtYIn6lEKaWD9wW297aHjQLpmQdHZrOVpOhmCBqkqo&s=nGCpS4p5jnaSpPUFlziSi3Y3pFijhVDy6e3867jA9BE&e=
>> >
>> >     >
>> >     > /"There have been fixes to RGW dynamic and manual resharding,
>> >     which no
>> >     > longer
>> >     > leaves behind stale bucket instances to be removed manually. For
>> >     > finding and
>> >     > cleaning up older instances from a reshard a radosgw-admin
>> >     command reshard
>> >     > stale-instances list and reshard stale-instances rm should do the
>> >     > necessary
>> >     > cleanup."/
>> >     >
>> >     >
>> >
>>  ------------------------------------------------------------------------
>> >     >
>> >     > *From:*Christian Rice <cr...@pandora.com <mailto:
>> cr...@pandora.com>>
>> >     > *Sent:* Tuesday, March 5, 2019 11:34 AM
>> >     > *To:* Matthew H; ceph-users
>> >     > *Subject:* Re: radosgw sync falling behind regularly
>> >     >
>> >     > The output of “radosgw-admin reshard stale-instances list” shows
>> >     242
>> >     > entries, which might embed too much proprietary info for me to
>> >     list,
>> >     > but here’s a tiny sample:
>> >     >
>> >     >
>> >
>>  "sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18467.303",
>> >     >
>> >     >
>> >
>>  "sysad_task/sysad_task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18467.281",
>> >     >
>> >     >
>> >
>>  "sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299",
>> >     >
>> >     >
>> >
>>  "sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.301",
>> >     >
>> >     > Some of appear repeatedly in the radosgw error logs like so:
>> >     >
>> >     > 2019-03-05 08:13:08.929206 7f6405094700  0 data sync: ERROR:
>> >     init sync
>> >     > on
>> >     >
>> >     sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.302
>> >     > failed, retcode=-2
>> >     >
>> >     > 2019-03-05 08:13:08.930581 7f6405094700  0 data sync: WARNING:
>> >     > skipping data log entry for missing bucket
>> >     >
>> sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.302
>> >     >
>> >     > 2019-03-05 08:13:08.972053 7f6405094700  0 data sync: ERROR:
>> >     init sync
>> >     > on
>> >     >
>> >     sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299
>> >     > failed, retcode=-2
>> >     >
>> >     > 2019-03-05 08:13:08.973442 7f6405094700  0 data sync: WARNING:
>> >     > skipping data log entry for missing bucket
>> >     >
>> sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299
>> >     >
>> >     > 2019-03-05 08:13:19.528295 7f6406897700  0 data sync: ERROR:
>> >     init sync
>> >     > on
>> >     >
>> >     sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299
>> >     > failed, retcode=-2
>> >     >
>> >     > Notably, “Sync is disabled for bucket sysad-task.”  We use “bucket
>> >     > sync disable” A LOT.  It’s the only way we’ve been able to use
>> >     > multisite and a single namespace and not replicate things that are
>> >     > unneeded to every zone.  Perhaps there’s a bug in the
>> >     implementation
>> >     > that’s tripping us up now, with the new sync multisite sync code
>> >     from
>> >     > 12.2.9 onward?
>> >     >
>> >     > What might we do with stale bucket instances, then?
>> >     >
>> >     > Of note, our master zone endpoint, which was timing out health
>> >     checks
>> >     > most of the day after the upgrade (was running but overworked by
>> >     > cluster confusion, so we couldn’t create new buckets or do user
>> >     ops),
>> >     > has returned to availability late last night.  There’s a lot of
>> >     data
>> >     > to look at, but in my estimation, due to lack of user complaints
>> >     (or
>> >     > their unawareness of specific issues), it seems the zones are
>> >     > nominally available, even with all the errors and warnings being
>> >     > logged.  We’ve tested simple zone replication by creating a few
>> >     files
>> >     > in one zone and seeing them show up elsewhere…
>> >     >
>> >     > here’s “period get” output:
>> >     >
>> >     > sv5-ceph-rgw1
>> >     >
>> >     > {
>> >     >
>> >     >     "id": "3d0d40ef-90de-40ea-8c44-caa20ea8dc53",
>> >     >
>> >     > "epoch": 16,
>> >     >
>> >     > "predecessor_uuid": "926c74c7-c1a7-46b1-9f25-eb5c392a7fbb",
>> >     >
>> >     > "sync_status": [],
>> >     >
>> >     > "period_map": {
>> >     >
>> >     > "id": "3d0d40ef-90de-40ea-8c44-caa20ea8dc53",
>> >     >
>> >     > "zonegroups": [
>> >     >
>> >     > {
>> >     >
>> >     > "id": "de6af748-1a2f-44a1-9d44-30799cf1313e",
>> >     >
>> >     > "name": "us",
>> >     >
>> >     > "api_name": "us",
>> >     >
>> >     > "is_master": "true",
>> >     >
>> >     > "endpoints": [
>> >     >
>> >     >                    "http://sv5-ceph-rgw1.savagebeast.com:8080";
>> >     >
>> >     > ],
>> >     >
>> >     > "hostnames": [],
>> >     >
>> >     > "hostnames_s3website": [],
>> >     >
>> >     > "master_zone": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04",
>> >     >
>> >     > "zones": [
>> >     >
>> >     > {
>> >     >
>> >     > "id": "107d29a0-b732-4bf1-a26e-1f64f820e839",
>> >     >
>> >     > "name": "dc11-prod",
>> >     >
>> >     > "endpoints": [
>> >     >
>> >     > "http://dc11-ceph-rgw1:8080";
>> >     >
>> >     >  ],
>> >     >
>> >     > "log_meta": "false",
>> >     >
>> >     > "log_data": "true",
>> >     >
>> >     > "bucket_index_max_shards": 0,
>> >     >
>> >     > "read_only": "false",
>> >     >
>> >     > "tier_type": "",
>> >     >
>> >     >   "sync_from_all": "true",
>> >     >
>> >     > "sync_from": []
>> >     >
>> >     > },
>> >     >
>> >     > {
>> >     >
>> >     > "id": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04",
>> >     >
>> >     > "name": "sv5-corp",
>> >     >
>> >     > "endpoints": [
>> >     >
>> >     > "http://sv5-ceph-rgw1.savagebeast.com:8080";
>> >     >
>> >     > ],
>> >     >
>> >     > "log_meta": "false",
>> >     >
>> >     > "log_data": "true",
>> >     >
>> >     > "bucket_index_max_shards": 0,
>> >     >
>> >     >                    "read_only": "false",
>> >     >
>> >     > "tier_type": "",
>> >     >
>> >     > "sync_from_all": "true",
>> >     >
>> >     > "sync_from": []
>> >     >
>> >     > },
>> >     >
>> >     > {
>> >     >
>> >     > "id": "331d3f1e-1b72-4c56-bb5a-d1d0fcf6d0b8",
>> >     >
>> >     > "name": "sv3-prod",
>> >     >
>> >     > "endpoints": [
>> >     >
>> >     > "http://sv3-ceph-rgw1:8080";
>> >     >
>> >     > ],
>> >     >
>> >     >     "log_meta": "false",
>> >     >
>> >     > "log_data": "true",
>> >     >
>> >     > "bucket_index_max_shards": 0,
>> >     >
>> >     > "read_only": "false",
>> >     >
>> >     > "tier_type": "",
>> >     >
>> >     > "sync_from_all": "true",
>> >     >
>> >     > "sync_from": []
>> >     >
>> >     > }
>> >     >
>> >     > ],
>> >     >
>> >     > "placement_targets": [
>> >     >
>> >     > {
>> >     >
>> >     > "name": "default-placement",
>> >     >
>> >     > "tags": []
>> >     >
>> >     > }
>> >     >
>> >     >                ],
>> >     >
>> >     > "default_placement": "default-placement",
>> >     >
>> >     > "realm_id": "b3e2afe7-2254-494a-9a34-ce50358779fd"
>> >     >
>> >     > }
>> >     >
>> >     >         ],
>> >     >
>> >     > "short_zone_ids": [
>> >     >
>> >     > {
>> >     >
>> >     > "key": "107d29a0-b732-4bf1-a26e-1f64f820e839",
>> >     >
>> >     > "val": 1720993486
>> >     >
>> >     > },
>> >     >
>> >     > {
>> >     >
>> >     > "key": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04",
>> >     >
>> >     > "val": 2301637458
>> >     >
>> >     > },
>> >     >
>> >     > {
>> >     >
>> >     > "key": "331d3f1e-1b72-4c56-bb5a-d1d0fcf6d0b8",
>> >     >
>> >     > "val": 1449486239
>> >     >
>> >     > }
>> >     >
>> >     >         ]
>> >     >
>> >     >     },
>> >     >
>> >     > "master_zonegroup": "de6af748-1a2f-44a1-9d44-30799cf1313e",
>> >     >
>> >     > "master_zone": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04",
>> >     >
>> >     > "period_config": {
>> >     >
>> >     > "bucket_quota": {
>> >     >
>> >     > "enabled": false,
>> >     >
>> >     > "check_on_raw": false,
>> >     >
>> >     > "max_size": -1,
>> >     >
>> >     > "max_size_kb": 0,
>> >     >
>> >     > "max_objects": -1
>> >     >
>> >     >         },
>> >     >
>> >     > "user_quota": {
>> >     >
>> >     > "enabled": false,
>> >     >
>> >     > "check_on_raw": false,
>> >     >
>> >     > "max_size": -1,
>> >     >
>> >     > "max_size_kb": 0,
>> >     >
>> >     > "max_objects": -1
>> >     >
>> >     >         }
>> >     >
>> >     >     },
>> >     >
>> >     > "realm_id": "b3e2afe7-2254-494a-9a34-ce50358779fd",
>> >     >
>> >     > "realm_name": "savagebucket",
>> >     >
>> >     > "realm_epoch": 2
>> >     >
>> >     > }
>> >     >
>> >     > sv3-ceph-rgw1
>> >     >
>> >     > {
>> >     >
>> >     >     "id": "3d0d40ef-90de-40ea-8c44-caa20ea8dc53",
>> >     >
>> >     > "epoch": 16,
>> >     >
>> >     > "predecessor_uuid": "926c74c7-c1a7-46b1-9f25-eb5c392a7fbb",
>> >     >
>> >     > "sync_status": [],
>> >     >
>> >     > "period_map": {
>> >     >
>> >     > "id": "3d0d40ef-90de-40ea-8c44-caa20ea8dc53",
>> >     >
>> >     > "zonegroups": [
>> >     >
>> >     > {
>> >     >
>> >     > "id": "de6af748-1a2f-44a1-9d44-30799cf1313e",
>> >     >
>> >     > "name": "us",
>> >     >
>> >     > "api_name": "us",
>> >     >
>> >     > "is_master": "true",
>> >     >
>> >     > "endpoints": [
>> >     >
>> >     >         "http://sv5-ceph-rgw1.savagebeast.com:8080";
>> >     >
>> >     > ],
>> >     >
>> >     > "hostnames": [],
>> >     >
>> >     > "hostnames_s3website": [],
>> >     >
>> >     > "master_zone": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04",
>> >     >
>> >     > "zones": [
>> >     >
>> >     >          {
>> >     >
>> >     > "id": "107d29a0-b732-4bf1-a26e-1f64f820e839",
>> >     >
>> >     > "name": "dc11-prod",
>> >     >
>> >     > "endpoints": [
>> >     >
>> >     > "http://dc11-ceph-rgw1:8080";
>> >     >
>> >     > ],
>> >     >
>> >     > "log_meta": "false",
>> >     >
>> >     > "log_data": "true",
>> >     >
>> >     > "bucket_index_max_shards": 0,
>> >     >
>> >     > "read_only": "false",
>> >     >
>> >     > "tier_type": "",
>> >     >
>> >     > "sync_from_all": "true",
>> >     >
>> >     > "sync_from": []
>> >     >
>> >     > },
>> >     >
>> >     > {
>> >     >
>> >     > "id": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04",
>> >     >
>> >     > "name": "sv5-corp",
>> >     >
>> >     > "endpoints": [
>> >     >
>> >     > "http://sv5-ceph-rgw1.savagebeast.com:8080";
>> >     >
>> >     > ],
>> >     >
>> >     > "log_meta": "false",
>> >     >
>> >     > "log_data": "true",
>> >     >
>> >     > "bucket_index_max_shards": 0,
>> >     >
>> >     >                "read_only": "false",
>> >     >
>> >     > "tier_type": "",
>> >     >
>> >     > "sync_from_all": "true",
>> >     >
>> >     > "sync_from": []
>> >     >
>> >     > },
>> >     >
>> >     > {
>> >     >
>> >     > "id": "331d3f1e-1b72-4c56-bb5a-d1d0fcf6d0b8",
>> >     >
>> >     > "name": "sv3-prod",
>> >     >
>> >     > "endpoints": [
>> >     >
>> >     > "http://sv3-ceph-rgw1:8080";
>> >     >
>> >     > ],
>> >     >
>> >     > "log_meta": "false",
>> >     >
>> >     > "log_data": "true",
>> >     >
>> >     > "bucket_index_max_shards": 0,
>> >     >
>> >     > "read_only": "false",
>> >     >
>> >     > "tier_type": "",
>> >     >
>> >     > "sync_from_all": "true",
>> >     >
>> >     >   "sync_from": []
>> >     >
>> >     > }
>> >     >
>> >     > ],
>> >     >
>> >     > "placement_targets": [
>> >     >
>> >     > {
>> >     >
>> >     > "name": "default-placement",
>> >     >
>> >     > "tags": []
>> >     >
>> >     > }
>> >     >
>> >     > ],
>> >     >
>> >     >          "default_placement": "default-placement",
>> >     >
>> >     > "realm_id": "b3e2afe7-2254-494a-9a34-ce50358779fd"
>> >     >
>> >     > }
>> >     >
>> >     >         ],
>> >     >
>> >     > "short_zone_ids": [
>> >     >
>> >     > {
>> >     >
>> >     > "key": "107d29a0-b732-4bf1-a26e-1f64f820e839",
>> >     >
>> >     >         "val": 1720993486
>> >     >
>> >     > },
>> >     >
>> >     > {
>> >     >
>> >     > "key": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04",
>> >     >
>> >     > "val": 2301637458
>> >     >
>> >     > },
>> >     >
>> >     > {
>> >     >
>> >     > "key": "331d3f1e-1b72-4c56-bb5a-d1d0fcf6d0b8",
>> >     >
>> >     > "val": 1449486239
>> >     >
>> >     > }
>> >     >
>> >     >         ]
>> >     >
>> >     >     },
>> >     >
>> >     > "master_zonegroup": "de6af748-1a2f-44a1-9d44-30799cf1313e",
>> >     >
>> >     > "master_zone": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04",
>> >     >
>> >     > "period_config": {
>> >     >
>> >     > "bucket_quota": {
>> >     >
>> >     > "enabled": false,
>> >     >
>> >     > "check_on_raw": false,
>> >     >
>> >     > "max_size": -1,
>> >     >
>> >     > "max_size_kb": 0,
>> >     >
>> >     > "max_objects": -1
>> >     >
>> >     >         },
>> >     >
>> >     > "user_quota": {
>> >     >
>> >     > "enabled": false,
>> >     >
>> >     > "check_on_raw": false,
>> >     >
>> >     > "max_size": -1,
>> >     >
>> >     > "max_size_kb": 0,
>> >     >
>> >     > "max_objects": -1
>> >     >
>> >     >         }
>> >     >
>> >     >     },
>> >     >
>> >     > "realm_id": "b3e2afe7-2254-494a-9a34-ce50358779fd",
>> >     >
>> >     > "realm_name": "savagebucket",
>> >     >
>> >     > "realm_epoch": 2
>> >     >
>> >     > }
>> >     >
>> >     > dc11-ceph-rgw1
>> >     >
>> >     > {
>> >     >
>> >     >     "id": "3d0d40ef-90de-40ea-8c44-caa20ea8dc53",
>> >     >
>> >     > "epoch": 16,
>> >     >
>> >     > "predecessor_uuid": "926c74c7-c1a7-46b1-9f25-eb5c392a7fbb",
>> >     >
>> >     > "sync_status": [],
>> >     >
>> >     > "period_map": {
>> >     >
>> >     > "id": "3d0d40ef-90de-40ea-8c44-caa20ea8dc53",
>> >     >
>> >     > "zonegroups": [
>> >     >
>> >     >         {
>> >     >
>> >     > "id": "de6af748-1a2f-44a1-9d44-30799cf1313e",
>> >     >
>> >     > "name": "us",
>> >     >
>> >     > "api_name": "us",
>> >     >
>> >     > "is_master": "true",
>> >     >
>> >     > "endpoints": [
>> >     >
>> >     > "http://sv5-ceph-rgw1.savagebeast.com:8080";
>> >     >
>> >     > ],
>> >     >
>> >     > "hostnames": [],
>> >     >
>> >     > "hostnames_s3website": [],
>> >     >
>> >     > "master_zone": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04",
>> >     >
>> >     > "zones": [
>> >     >
>> >     > {
>> >     >
>> >     > "id": "107d29a0-b732-4bf1-a26e-1f64f820e839",
>> >     >
>> >     > "name": "dc11-prod",
>> >     >
>> >     > "endpoints": [
>> >     >
>> >     > "http://dc11-ceph-rgw1:8080";
>> >     >
>> >     > ],
>> >     >
>> >     > "log_meta": "false",
>> >     >
>> >     >                       "log_data": "true",
>> >     >
>> >     > "bucket_index_max_shards": 0,
>> >     >
>> >     > "read_only": "false",
>> >     >
>> >     > "tier_type": "",
>> >     >
>> >     > "sync_from_all": "true",
>> >     >
>> >     > "sync_from": []
>> >     >
>> >     > },
>> >     >
>> >     > {
>> >     >
>> >     > "id": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04",
>> >     >
>> >     > "name": "sv5-corp",
>> >     >
>> >     > "endpoints": [
>> >     >
>> >     >           "http://sv5-ceph-rgw1.savagebeast.com:8080";
>> >     >
>> >     > ],
>> >     >
>> >     > "log_meta": "false",
>> >     >
>> >     > "log_data": "true",
>> >     >
>> >     > "bucket_index_max_shards": 0,
>> >     >
>> >     > "read_only": "false",
>> >     >
>> >     > "tier_type": "",
>> >     >
>> >     > "sync_from_all": "true",
>> >     >
>> >     > "sync_from": []
>> >     >
>> >     > },
>> >     >
>> >     > {
>> >     >
>> >     > "id": "331d3f1e-1b72-4c56-bb5a-d1d0fcf6d0b8",
>> >     >
>> >     > "name": "sv3-prod",
>> >     >
>> >     > "endpoints": [
>> >     >
>> >     > "http://sv3-ceph-rgw1:8080";
>> >     >
>> >     > ],
>> >     >
>> >     > "log_meta": "false",
>> >     >
>> >     > "log_data": "true",
>> >     >
>> >     > "bucket_index_max_shards": 0,
>> >     >
>> >     > "read_only": "false",
>> >     >
>> >     > "tier_type": "",
>> >     >
>> >     > "sync_from_all": "true",
>> >     >
>> >     > "sync_from": []
>> >     >
>> >     > }
>> >     >
>> >     > ],
>> >     >
>> >     > "placement_targets": [
>> >     >
>> >     > {
>> >     >
>> >     > "name": "default-placement",
>> >     >
>> >     > "tags": []
>> >     >
>> >     > }
>> >     >
>> >     > ],
>> >     >
>> >     > "default_placement": "default-placement",
>> >     >
>> >     > "realm_id": "b3e2afe7-2254-494a-9a34-ce50358779fd"
>> >     >
>> >     > }
>> >     >
>> >     >         ],
>> >     >
>> >     > "short_zone_ids": [
>> >     >
>> >     > {
>> >     >
>> >     > "key": "107d29a0-b732-4bf1-a26e-1f64f820e839",
>> >     >
>> >     > "val": 1720993486
>> >     >
>> >     > },
>> >     >
>> >     > {
>> >     >
>> >     > "key": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04",
>> >     >
>> >     > "val": 2301637458
>> >     >
>> >     > },
>> >     >
>> >     > {
>> >     >
>> >     > "key": "331d3f1e-1b72-4c56-bb5a-d1d0fcf6d0b8",
>> >     >
>> >     > "val": 1449486239
>> >     >
>> >     > }
>> >     >
>> >     >         ]
>> >     >
>> >     >     },
>> >     >
>> >     > "master_zonegroup": "de6af748-1a2f-44a1-9d44-30799cf1313e",
>> >     >
>> >     > "master_zone": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04",
>> >     >
>> >     > "period_config": {
>> >     >
>> >     > "bucket_quota": {
>> >     >
>> >     > "enabled": false,
>> >     >
>> >     > "check_on_raw": false,
>> >     >
>> >     >           "max_size": -1,
>> >     >
>> >     > "max_size_kb": 0,
>> >     >
>> >     > "max_objects": -1
>> >     >
>> >     >         },
>> >     >
>> >     > "user_quota": {
>> >     >
>> >     > "enabled": false,
>> >     >
>> >     > "check_on_raw": false,
>> >     >
>> >     > "max_size": -1,
>> >     >
>> >     > "max_size_kb": 0,
>> >     >
>> >     > "max_objects": -1
>> >     >
>> >     >         }
>> >     >
>> >     >     },
>> >     >
>> >     > "realm_id": "b3e2afe7-2254-494a-9a34-ce50358779fd",
>> >     >
>> >     > "realm_name": "savagebucket",
>> >     >
>> >     > "realm_epoch": 2
>> >     >
>> >     > }
>> >     >
>> >     > *From: *Matthew H <matthew.he...@hotmail.com
>> >     <mailto:matthew.he...@hotmail.com>>
>> >     > *Date: *Tuesday, March 5, 2019 at 4:31 AM
>> >     > *To: *Christian Rice <cr...@pandora.com
>> >     <mailto:cr...@pandora.com>>, ceph-users
>> >     > <ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com>>
>> >     > *Subject: *Re: radosgw sync falling behind regularly
>> >     >
>> >     > Hi Christian,
>> >     >
>> >     > You haven't resharded any of your buckets have you?  You can run
>> >     the
>> >     > command below in v12.2.11 to list stale bucket instances.
>> >     >
>> >     > radosgw-admin reshard stale-instances list
>> >     >
>> >     > Can you also send the output from the following command on each
>> rgw?
>> >     >
>> >     > radosgw-admin period get
>> >     >
>> >     >
>> >     > _______________________________________________
>> >     > ceph-users mailing list
>> >     > ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com>
>> >     > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> >     _______________________________________________
>> >     ceph-users mailing list
>> >     ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com>
>> >     http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> >
>>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to