There may be a mismatch between be auto-restarting and the omap warning code. Looks like you already have 349 shards, with 13 of them warning on size! You can increase a config value to shut that error up, but you may want to get somebody from RGW to look at how you’ve managed to exceed those default limits. On Wed, Jun 13, 2018 at 4:36 AM stephan schultchen < stephan.schultc...@gmail.com> wrote:
> Hello, > > i am running a ceph 13.2.0 cluster exclusively for radosrw / s3. > > i only have one big bucket. and the cluster is currently in warning state: > > cluster: > id: d605c463-9f1c-4d91-a390-a28eedb21650 > health: HEALTH_WARN > 13 large omap objects > > i tried to google it, but i was not able to find what to do about the > "large omap objects". > > as far as i understand ceph should automatically re shard an s3 bucket > when an omap is getting to big. or is this something i have to do? > > "radosgw-admin reshard list" tells that no resharding is ongoing right now. > > > radosgw-admin metadata get > bucket.instance:nuxeo_live:6f85d718-fd2e-4c1b-a21d-bafb04a8cfcc.4854.4 > { > "key": > "bucket.instance:nuxeo_live:6f85d718-fd2e-4c1b-a21d-bafb04a8cfcc.4854.4", > "ver": { > "tag": "Y2epzPoujRDfxM5CNMZgKPaA", > "ver": 6 > }, > "mtime": "2018-06-08 14:48:15.515840Z", > "data": { > "bucket_info": { > "bucket": { > "name": "nuxeo_live", > "marker": "6f85d718-fd2e-4c1b-a21d-bafb04a8cfcc.4848.1", > "bucket_id": "6f85d718-fd2e-4c1b-a21d-bafb04a8cfcc.4854.4", > "tenant": "", > "explicit_placement": { > "data_pool": "", > "data_extra_pool": "", > "index_pool": "" > } > }, > "creation_time": "2018-05-23 13:31:57.664398Z", > "owner": "nuxeo_live", > "flags": 0, > "zonegroup": "506cc27c-fef5-4b89-a9f3-4c928a74b955", > "placement_rule": "default-placement", > "has_instance_obj": "true", > "quota": { > "enabled": false, > "check_on_raw": false, > "max_size": -1, > "max_size_kb": 0, > "max_objects": -1 > }, > "num_shards": 349, > "bi_shard_hash_type": 0, > "requester_pays": "false", > "has_website": "false", > "swift_versioning": "false", > "swift_ver_location": "", > "index_type": 0, > "mdsearch_config": [], > "reshard_status": 2, > "new_bucket_instance_id": > "6f85d718-fd2e-4c1b-a21d-bafb04a8cfcc.176143.1" > }, > "attrs": [ > { > "key": "user.rgw.acl", > "val": > "AgKpAAAAAwIhAAAACgAAAG51eGVvX2xpdmUPAAAAbnV4ZW8gbGl2ZSB1c2VyBAN8AAAAAQEAAAAKAAAAbnV4ZW9fbGl2ZQ8AAAABAAAACgAAAG51eGVvX2xpdmUFA0UAAAACAgQAAAAAAAAACgAAAG51eGVvX2xpdmUAAAAAAAAAAAICBAAAAA8AAAAPAAAAbnV4ZW8gbGl2ZSB1c2VyAAAAAAAAAAAAAAAAAAAAAA==" > }, > { > "key": "user.rgw.idtag", > "val": "" > } > ] > > i also tried to manually trigger a resharding. but it failed with: > > > - NOTICE: operation will not remove old bucket index objects *** > - these will need to be removed manually *** > tenant: > bucket name: nuxeo_live > old bucket instance id: 6f85d718-fd2e-4c1b-a21d-bafb04a8cfcc.184670.1 > new bucket instance id: 6f85d718-fd2e-4c1b-a21d-bafb04a8cfcc.176197.1 > WARNING: RGWReshard::add failed to drop lock on > bucket_name:6f85d718-fd2e-4c1b-a21d-bafb04a8cfcc.184670.1 ret=-2 > > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com