Public

Hi All,

We are running a multisite setup running luminous on bluestore.  This setup 
worked perfectly since installation around the time 12.2.2 came out.  A few 
days ago we upgraded the clusters from 12.2.5 to 12.2.7 and now we noticed one 
of the clusters does not sync anymore.

Primary site:
# radosgw-admin sync status
          realm 87f16146-0729-4b37-9462-dd5e6d97b427 (pro)
      zonegroup 9fad4a8d-9a7b-4649-a54a-856450635808 (be)
           zone 4ed07bb2-a80b-4c69-aa15-fdc17ae6f5f2 (bccm-pro)
  metadata sync no sync (zone is master)
      data sync source: ad420c46-3ef3-430a-afef-bff78e26d410 (bccl-pro)
                        syncing
                        full sync: 0/128 shards
                        incremental sync: 128/128 shards
                        data is caught up with source

secondary site:
# radosgw-admin sync status
          realm 87f16146-0729-4b37-9462-dd5e6d97b427 (pro)
      zonegroup 9fad4a8d-9a7b-4649-a54a-856450635808 (be)
           zone ad420c46-3ef3-430a-afef-bff78e26d410 (bccl-pro)
  metadata sync syncing
                full sync: 0/64 shards
                incremental sync: 64/64 shards
                metadata is caught up with master
      data sync source: 4ed07bb2-a80b-4c69-aa15-fdc17ae6f5f2 (bccm-pro)
                        syncing
                        full sync: 80/128 shards
                        full sync: 4 buckets to sync
                        incremental sync: 48/128 shards
                        data is behind on 80 shards
                        behind shards: 
[0,6,25,44,51,52,53,54,55,56,57,58,59,60,61,62,63,64,65,66,67,68,69,70,71,72,73,74,75,76,77,78,79,80,81,82,83,84,85,86,87,88,89,90,91,92,93,94,95,96,97,98,99,100,101,102,103,104,105,106,107,108,109,110,111,112,113,115,116,117,118,119,120,121,122,123,124,125,126,127]
                        1 shards are recovering
                        recovering shards: [27]

When we first noticed the problem only 3 shards were behind and shard 27 was 
recovering.  One of the things we did was a radosgw-admin data sync init in an 
attempt to have everything syncing again.  Since then, 48 shards seem to have 
done the full sync and are now incremental syncing, but the rest just stays 
like this.  It seems the recovering of shard 27 blocks the rest of the sync?

Radosgw-admin sync error list shows a number of errors from during the upgrade, 
mostly "failed to sync object(5) Input/output error" and "failed to sync bucket 
instance: (5) Input/output error".  Does this mean radosgw was unable to write 
to the pool?

# radosgw-admin data sync status --shard-id 27 --source-zone bccm-pro
{
    "shard_id": 27,
    "marker": {
        "status": "incremental-sync",
        "marker": "1_1534494893.816775_131867195.1",
        "next_step_marker": "",
        "total_entries": 1,
        "pos": 0,
        "timestamp": "0.000000"
    },
    "pending_buckets": [],
    "recovering_buckets": [
        "pro-registry:4ed07bb2-a80b-4c69-aa15-fdc17ae6f5f2.314303.1:26"
    ]
}


How can we recover shard 27?  Any ideas on how we get this multisite setup 
healthy again?  I wanted to create an issue in the tracker for this, but it 
seems a normal user does not have permissions to do this anymore?

Many Thanks

Dieter





This e-mail and any attached files are confidential and may be legally 
privileged. If you are not the addressee, any disclosure, reproduction, 
copying, distribution, or other dissemination or use of this communication is 
strictly prohibited. If you have received this transmission in error please 
notify KBC immediately and then delete this e-mail.
KBC does not accept liability for the correct and complete transmission of the 
information, nor for any delay or interruption of the transmission, nor for 
damages arising from the use of or reliance on the information.
All e-mail messages addressed to, received or sent by KBC or KBC employees are 
deemed to be professional in nature. Accordingly, the sender or recipient of 
these messages agrees that they may be read by other KBC employees than the 
official recipient or sender in order to ensure the continuity of work-related 
activities and allow supervision thereof.

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to