Hi Orit!

I did some tests, and indeed the combination of Versioning/Lifecycle with 
Resharding is the problem:


  *   If I do not enable Versioning/Lifecycle, Autoresharding works fine.
  *   If I disable Autoresharding but enable Versioning+Lifecycle, pushing data 
works fine, until I manually reshard. This hangs also.

My lifecycle rule (which shall remove all versions older than 60 days):


{

    "Rules": [{

        "Status": "Enabled",

        "Prefix": "",

        "NoncurrentVersionExpiration": {

            "NoncurrentDays": 60

        },

        "Expiration": {

            "ExpiredObjectDeleteMarker": true

        },

        "ID": "expire-60days"

    }]

}

I am currently testing with an application containing customer data, but I am 
also creating some random test data to create logs I can share.
I will also test whether the versioning itself is the culprit, or if it is the 
lifecycle rule.

Regards,
Martin

Von: Orit Wasserman <owass...@redhat.com>
Datum: Dienstag, 16. Januar 2018 um 18:38
An: Martin Emrich <martin.emr...@empolis.com>
Cc: ceph-users <ceph-users@lists.ceph.com>
Betreff: Re: [ceph-users] Bug in RadosGW resharding? Hangs again...

Hi Martin,

On Mon, Jan 15, 2018 at 6:04 PM, Martin Emrich 
<martin.emr...@empolis.com<mailto:martin.emr...@empolis.com>> wrote:
Hi!

After having a completely broken radosgw setup due to damaged buckets, I 
completely deleted all rgw pools, and started from scratch.

But my problem is reproducible. After pushing ca. 100000 objects into a bucket, 
the resharding process appears to start, and the bucket is now unresponsive.

Sorry to hear that.
Can you share radosgw logs with --debug_rgw=20 --debug_ms=1?

I just see lots of these messages in all rgw logs:

2018-01-15 16:57:45.108826 7fd1779b1700  0 block_while_resharding ERROR: bucket 
is still resharding, please retry
2018-01-15 16:57:45.119184 7fd1779b1700  0 NOTICE: resharding operation on 
bucket index detected, blocking
2018-01-15 16:57:45.260751 7fd1120e6700  0 block_while_resharding ERROR: bucket 
is still resharding, please retry
2018-01-15 16:57:45.280410 7fd1120e6700  0 NOTICE: resharding operation on 
bucket index detected, blocking
2018-01-15 16:57:45.300775 7fd15b979700  0 block_while_resharding ERROR: bucket 
is still resharding, please retry
2018-01-15 16:57:45.300971 7fd15b979700  0 WARNING: set_req_state_err 
err_no=2300 resorting to 500
2018-01-15 16:57:45.301042 7fd15b979700  0 ERROR: 
RESTFUL_IO(s)->complete_header() returned err=Input/output error

One radosgw process and two OSDs housing the bucket index/metadata are still 
busy, but it seems to be stuck again.

How long is this resharding process supposed to take? I cannot believe that an 
application is supposed to block for more than half an hour...

I feel inclined to open a bug report, but I am yet unshure where the problem 
lies.

Some information:

* 3 RGW processes, 3 OSD hosts with 12 HDD OSDs and 6 SSD OSDs
* Ceph 12.2.2
* Auto-Resharding on, Bucket Versioning & Lifecycle rule enabled.

What life cycle rules do you use?

Regards,
Orit
Thanks,

Martin

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com<mailto:ceph-users@lists.ceph.com>
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to