As I wrote here:

http://lists.ceph.com/pipermail/ceph-users-ceph.com/2020-January/037909.html

I saw the same after an update from Luminous to Nautilus 14.2.6

Cheers, Massimo

On Tue, Jan 14, 2020 at 7:45 PM Liam Monahan <l...@umiacs.umd.edu> wrote:

> Hi,
>
> I am getting one inconsistent object on our cluster with an inconsistency
> error that I haven’t seen before.  This started happening during a rolling
> upgrade of the cluster from 14.2.3 -> 14.2.6, but I am not sure that’s
> related.
>
> I was hoping to know what the error means before trying a repair.
>
> [root@objmon04 ~]# ceph health detail
> HEALTH_ERR noout flag(s) set; 1 scrub errors; Possible data damage: 1 pg
> inconsistent
> OSDMAP_FLAGS noout flag(s) set
> OSD_SCRUB_ERRORS 1 scrub errors
> PG_DAMAGED Possible data damage: 1 pg inconsistent
>     pg 9.20e is active+clean+inconsistent, acting [509,674,659]
>
> rados list-inconsistent-obj 9.20e --format=json-pretty
> {
>     "epoch": 759019,
>     "inconsistents": [
>         {
>             "object": {
>                 "name":
> "2017-07-03-12-8b980d5b-23de-41f9-8b14-84a5bbc3f1c9.31293422.4-activedns-diff",
>                 "nspace": "",
>                 "locator": "",
>                 "snap": "head",
>                 "version": 692875
>             },
>             "errors": [
>                 "size_too_large"
>             ],
>             "union_shard_errors": [],
>             "selected_object_info": {
>                 "oid": {
>                     "oid":
> "2017-07-03-12-8b980d5b-23de-41f9-8b14-84a5bbc3f1c9.31293422.4-activedns-diff",
>                     "key": "",
>                     "snapid": -2,
>                     "hash": 3321413134,
>                     "max": 0,
>                     "pool": 9,
>                     "namespace": ""
>                 },
>                 "version": "281183'692875",
>                 "prior_version": "281183'692874",
>                 "last_reqid": "client.34042469.0:206759091",
>                 "user_version": 692875,
>                 "size": 146097278,
>                 "mtime": "2017-07-03 12:43:35.569986",
>                 "local_mtime": "2017-07-03 12:43:35.571196",
>                 "lost": 0,
>                 "flags": [
>                     "dirty",
>                     "data_digest",
>                     "omap_digest"
>                 ],
>                 "truncate_seq": 0,
>                 "truncate_size": 0,
>                 "data_digest": "0xf19c8035",
>                 "omap_digest": "0xffffffff",
>                 "expected_object_size": 0,
>                 "expected_write_size": 0,
>                 "alloc_hint_flags": 0,
>                 "manifest": {
>                     "type": 0
>                 },
>                 "watchers": {}
>             },
>             "shards": [
>                 {
>                     "osd": 509,
>                     "primary": true,
>                     "errors": [],
>                     "size": 146097278
>                 },
>                 {
>                     "osd": 659,
>                     "primary": false,
>                     "errors": [],
>                     "size": 146097278
>                 },
>                 {
>                     "osd": 674,
>                     "primary": false,
>                     "errors": [],
>                     "size": 146097278
>                 }
>             ]
>         }
>     ]
> }
>
> Thanks,
> Liam
> —
> Senior Developer
> Institute for Advanced Computer Studies
> University of Maryland
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to