Re: [ceph-users] Significant uptick in inconsistent pgs in Jewel 10.2.9

2017-09-08 Thread David Zafman


Robin,

Would you generate the values and keys for the various versions of at 
least one of the objects?   .dir.default.292886573.13181.12 is a good 
example because there are 3 variations for the same object.


If there isn't much activity to .dir.default.64449186.344176, you could 
do one osd at a time.  Otherwise, stop all 3 OSDs 1322, 990, 655 execute 
these for all 3.  I suspect you'll need to pipe to "od-cx" to get 
printable output.


I created a simple object with ascii omap.

$ ceph-objectstore-tool --data-path ... --pgid 5.3d40 
.dir.default.64449186.344176 get-omaphdr

obj_header

$ for i in $(ceph-objectstore-tool --data-path ... --pgid 5.3d40 
.dir.default.64449186.344176 list-omap)


do

echo -n "${i}: "

ceph-objectstore-tool --data-path ... .dir.default.292886573.13181.12 
get-omap $i


done

key1: val1
key2: val2
key3: val3

David


On 9/8/17 12:18 PM, David Zafman wrote:


Robin,

The only two changesets I can spot in Jewel that I think might be 
related are these:

1.
http://tracker.ceph.com/issues/20089
https://github.com/ceph/ceph/pull/15416


This should improve the repair functionality.


2.
http://tracker.ceph.com/issues/19404
https://github.com/ceph/ceph/pull/14204

This pull request fixes an issue that corrupted omaps.  It also finds 
and repairs them.  However, the repair process might resurrect deleted 
omaps which would show up as an omap digest error.


This could temporarily cause additional inconsistent PGs.  So if this 
has NOT been occurring longer than your deep-scrub interval since 
upgrading, I'd repair the pgs and monitor going forward to make sure 
the problem doesn't recur.


---

You have good example of repair scenarios:


.dir.default.292886573.13181.12   only has a omap_digest_mismatch and 
no shard errors.  The automatic repair won't be sure which is a good 
copy.


In this case we can see that osd 1327 doesn't match the other two.  To 
assist the repair process to repair the right one. Remove the copy on 
osd.1327


Stop osd 1327 and use "ceph-objectstore-tool --data-path .1327 
.dir.default.292886573.13181.12 remove"



.dir.default.64449186.344176 has selected_object_info with "od 
337cf025" so shards have "omap_digest_mismatch_oi" except for osd 990.


The pg repair code will use osd.990 to fix the other 2 copies without 
further handling.



David



On 9/8/17 11:16 AM, Robin H. Johnson wrote:

On Thu, Sep 07, 2017 at 08:24:04PM +, Robin H. Johnson wrote:

pg 5.3d40 is active+clean+inconsistent, acting [1322,990,655]
pg 5.f1c0 is active+clean+inconsistent, acting [631,1327,91]

Here is the output of 'rados list-inconsistent-obj' for the PGs:

$ sudo rados list-inconsistent-obj 5.f1c0 |json_pp -json_opt 
canonical,pretty

{
    "epoch" : 1221254,
    "inconsistents" : [
   {
  "errors" : [
 "omap_digest_mismatch"
  ],
  "object" : {
 "locator" : "",
 "name" : ".dir.default.292886573.13181.12",
 "nspace" : "",
 "snap" : "head",
 "version" : 483490
  },
  "selected_object_info" : 
"5:038f1cff:::.dir.default.292886573.13181.12:head(1221843'483490 
client.417313345.0:19515832 dirty|omap|data_digest s 0 uv 483490 dd 
 alloc_hint [0 0])",

  "shards" : [
 {
    "data_digest" : "0x",
    "errors" : [],
    "omap_digest" : "0x928b0c0b",
    "osd" : 91,
    "size" : 0
 },
 {
    "data_digest" : "0x",
    "errors" : [],
    "omap_digest" : "0x928b0c0b",
    "osd" : 631,
    "size" : 0
 },
 {
    "data_digest" : "0x",
    "errors" : [],
    "omap_digest" : "0x6556c868",
    "osd" : 1327,
    "size" : 0
 }
  ],
  "union_shard_errors" : []
   }
    ]
}
$ sudo rados list-inconsistent-obj 5.3d40  |json_pp -json_opt 
canonical,pretty

{
    "epoch" : 1210895,
    "inconsistents" : [
   {
  "errors" : [
 "omap_digest_mismatch"
  ],
  "object" : {
 "locator" : "",
 "name" : ".dir.default.64449186.344176",
 "nspace" : "",
 "snap" : "head",
 "version" : 1177199
  },
  "selected_object_info" : 
"5:02bc4def:::.dir.default.64449186.344176:head(1177700'1180639 
osd.1322.0:537914 dirty|omap|data_digest|omap_digest s 0 uv 1177199 
dd  od 337cf025 alloc_hint [0 0])",

  "shards" : [
 {
    "data_digest" : "0x",
    "errors" : [
   "omap_digest_mismatch_oi"
    ],
    "omap_digest" : "0x3242b04e",
    "osd" : 655,
    "size" : 0
 },
 {
    "data_digest" : "0x",

Re: [ceph-users] Significant uptick in inconsistent pgs in Jewel 10.2.9

2017-09-08 Thread David Zafman


Robin,

The only two changesets I can spot in Jewel that I think might be 
related are these:

1.
http://tracker.ceph.com/issues/20089
https://github.com/ceph/ceph/pull/15416


This should improve the repair functionality.


2.
http://tracker.ceph.com/issues/19404
https://github.com/ceph/ceph/pull/14204

This pull request fixes an issue that corrupted omaps.  It also finds 
and repairs them.  However, the repair process might resurrect deleted 
omaps which would show up as an omap digest error.


This could temporarily cause additional inconsistent PGs.  So if this 
has NOT been occurring longer than your deep-scrub interval since 
upgrading, I'd repair the pgs and monitor going forward to make sure the 
problem doesn't recur.


---

You have good example of repair scenarios:


.dir.default.292886573.13181.12   only has a omap_digest_mismatch and no 
shard errors.  The automatic repair won't be sure which is a good copy.


In this case we can see that osd 1327 doesn't match the other two.  To 
assist the repair process to repair the right one. Remove the copy on 
osd.1327


Stop osd 1327 and use "ceph-objectstore-tool --data-path .1327 
.dir.default.292886573.13181.12 remove"



.dir.default.64449186.344176 has selected_object_info with "od 337cf025" 
so shards have "omap_digest_mismatch_oi" except for osd 990.


The pg repair code will use osd.990 to fix the other 2 copies without 
further handling.



David



On 9/8/17 11:16 AM, Robin H. Johnson wrote:

On Thu, Sep 07, 2017 at 08:24:04PM +, Robin H. Johnson wrote:

pg 5.3d40 is active+clean+inconsistent, acting [1322,990,655]
pg 5.f1c0 is active+clean+inconsistent, acting [631,1327,91]

Here is the output of 'rados list-inconsistent-obj' for the PGs:

$ sudo rados list-inconsistent-obj 5.f1c0 |json_pp -json_opt canonical,pretty
{
"epoch" : 1221254,
"inconsistents" : [
   {
  "errors" : [
 "omap_digest_mismatch"
  ],
  "object" : {
 "locator" : "",
 "name" : ".dir.default.292886573.13181.12",
 "nspace" : "",
 "snap" : "head",
 "version" : 483490
  },
  "selected_object_info" : 
"5:038f1cff:::.dir.default.292886573.13181.12:head(1221843'483490 client.417313345.0:19515832 
dirty|omap|data_digest s 0 uv 483490 dd  alloc_hint [0 0])",
  "shards" : [
 {
"data_digest" : "0x",
"errors" : [],
"omap_digest" : "0x928b0c0b",
"osd" : 91,
"size" : 0
 },
 {
"data_digest" : "0x",
"errors" : [],
"omap_digest" : "0x928b0c0b",
"osd" : 631,
"size" : 0
 },
 {
"data_digest" : "0x",
"errors" : [],
"omap_digest" : "0x6556c868",
"osd" : 1327,
"size" : 0
 }
  ],
  "union_shard_errors" : []
   }
]
}
$ sudo rados list-inconsistent-obj 5.3d40  |json_pp -json_opt canonical,pretty
{
"epoch" : 1210895,
"inconsistents" : [
   {
  "errors" : [
 "omap_digest_mismatch"
  ],
  "object" : {
 "locator" : "",
 "name" : ".dir.default.64449186.344176",
 "nspace" : "",
 "snap" : "head",
 "version" : 1177199
  },
  "selected_object_info" : 
"5:02bc4def:::.dir.default.64449186.344176:head(1177700'1180639 osd.1322.0:537914 
dirty|omap|data_digest|omap_digest s 0 uv 1177199 dd  od 337cf025 alloc_hint [0 0])",
  "shards" : [
 {
"data_digest" : "0x",
"errors" : [
   "omap_digest_mismatch_oi"
],
"omap_digest" : "0x3242b04e",
"osd" : 655,
"size" : 0
 },
 {
"data_digest" : "0x",
"errors" : [],
"omap_digest" : "0x337cf025",
"osd" : 990,
"size" : 0
 },
 {
"data_digest" : "0x",
"errors" : [
   "omap_digest_mismatch_oi"
],
"omap_digest" : "0xc90d06a8",
"osd" : 1322,
"size" : 0
 }
  ],
  "union_shard_errors" : [
 "omap_digest_mismatch_oi"
  ]
   }
]
}





___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Significant uptick in inconsistent pgs in Jewel 10.2.9

2017-09-08 Thread Robin H. Johnson
On Thu, Sep 07, 2017 at 08:24:04PM +, Robin H. Johnson wrote:
> pg 5.3d40 is active+clean+inconsistent, acting [1322,990,655]
> pg 5.f1c0 is active+clean+inconsistent, acting [631,1327,91]
Here is the output of 'rados list-inconsistent-obj' for the PGs:

$ sudo rados list-inconsistent-obj 5.f1c0 |json_pp -json_opt canonical,pretty
{
   "epoch" : 1221254,
   "inconsistents" : [
  {
 "errors" : [
"omap_digest_mismatch"
 ],
 "object" : {
"locator" : "",
"name" : ".dir.default.292886573.13181.12",
"nspace" : "",
"snap" : "head",
"version" : 483490
 },
 "selected_object_info" : 
"5:038f1cff:::.dir.default.292886573.13181.12:head(1221843'483490 
client.417313345.0:19515832 dirty|omap|data_digest s 0 uv 483490 dd  
alloc_hint [0 0])",
 "shards" : [
{
   "data_digest" : "0x",
   "errors" : [],
   "omap_digest" : "0x928b0c0b",
   "osd" : 91,
   "size" : 0
},
{
   "data_digest" : "0x",
   "errors" : [],
   "omap_digest" : "0x928b0c0b",
   "osd" : 631,
   "size" : 0
},
{
   "data_digest" : "0x",
   "errors" : [],
   "omap_digest" : "0x6556c868",
   "osd" : 1327,
   "size" : 0
}
 ],
 "union_shard_errors" : []
  }
   ]
}
$ sudo rados list-inconsistent-obj 5.3d40  |json_pp -json_opt canonical,pretty
{
   "epoch" : 1210895,
   "inconsistents" : [
  {
 "errors" : [
"omap_digest_mismatch"
 ],
 "object" : {
"locator" : "",
"name" : ".dir.default.64449186.344176",
"nspace" : "",
"snap" : "head",
"version" : 1177199
 },
 "selected_object_info" : 
"5:02bc4def:::.dir.default.64449186.344176:head(1177700'1180639 
osd.1322.0:537914 dirty|omap|data_digest|omap_digest s 0 uv 1177199 dd  
od 337cf025 alloc_hint [0 0])",
 "shards" : [
{
   "data_digest" : "0x",
   "errors" : [
  "omap_digest_mismatch_oi"
   ],
   "omap_digest" : "0x3242b04e",
   "osd" : 655,
   "size" : 0
},
{
   "data_digest" : "0x",
   "errors" : [],
   "omap_digest" : "0x337cf025",
   "osd" : 990,
   "size" : 0
},
{
   "data_digest" : "0x",
   "errors" : [
  "omap_digest_mismatch_oi"
   ],
   "omap_digest" : "0xc90d06a8",
   "osd" : 1322,
   "size" : 0
}
 ],
 "union_shard_errors" : [
"omap_digest_mismatch_oi"
 ]
  }
   ]
}



-- 
Robin Hugh Johnson
Gentoo Linux: Dev, Infra Lead, Foundation Asst. Treasurer
E-Mail   : robb...@gentoo.org
GnuPG FP : 11ACBA4F 4778E3F6 E4EDF38E B27B944E 34884E85
GnuPG FP : 7D0B3CEB E9B85B1F 825BCECF EE05E6F6 A48F6136


signature.asc
Description: Digital signature
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] Significant uptick in inconsistent pgs in Jewel 10.2.9

2017-09-07 Thread Robin H. Johnson
Hi,

Our clusters were upgraded to v10.2.9, from ~v10.2.7 (actually a local
git snapshot that was not quite 10.2.7), and since then, we're seeing a
LOT more scrub errors than previously.

The digest logging on the scrub errors, in some cases, is also now maddeningly
short: it doesn't contain ANY information on what the mismatch was, and many of
the errors seem to also be 3-way mismatches in the digest :-(.

I'm wondering if other people have seen something similar rises in scrub errors
in the upgrade, and/or the lack of digest output. I did hear one anecdotal
report that 10.2.9 seemed much more likely to fail out marginal disks.

The only two changesets I can spot in Jewel that I think might be related are 
these:
1.
http://tracker.ceph.com/issues/20089
https://github.com/ceph/ceph/pull/15416
2.
http://tracker.ceph.com/issues/19404
https://github.com/ceph/ceph/pull/14204

Two example PGs that are inconsistent (chosen because they didn't convey any 
private information so I didn't have to redact anything except IP):
$ sudo ceph health detail |grep -e 5.3d40 -e 5.f1c0
pg 5.3d40 is active+clean+inconsistent, acting [1322,990,655]
pg 5.f1c0 is active+clean+inconsistent, acting [631,1327,91]

$ fgrep 5.3d40 /var/log/ceph/ceph.log
2017-09-07 19:50:16.231523 osd.1322 [REDACTED::8861]:6808/3479303 1736 : 
cluster [INF] osd.1322 pg 5.3d40 Deep scrub errors, upgrading scrub to 
deep-scrub
2017-09-07 19:50:16.231862 osd.1322 [REDACTED::8861]:6808/3479303 1737 : 
cluster [INF] 5.3d40 deep-scrub starts
2017-09-07 19:54:38.631232 osd.1322 [REDACTED::8861]:6808/3479303 1738 : 
cluster [ERR] 5.3d40 shard 655: soid 
5:02bc4def:::.dir.default.64449186.344176:head omap_digest 0x3242b04e != 
omap_digest 0x337cf025 from auth oi 
5:02bc4def:::.dir.default.64449186.344176:head(1177700'1180639 
osd.1322.0:537914 dirty|omap|data_digest|omap_digest s 0 uv 1177199 dd  
od 337cf025 alloc_hint [0 0])
2017-09-07 19:54:38.631332 osd.1322 [REDACTED::8861]:6808/3479303 1739 : 
cluster [ERR] 5.3d40 shard 1322: soid 
5:02bc4def:::.dir.default.64449186.344176:head omap_digest 0xc90d06a8 != 
omap_digest 0x3242b04e from shard 655, omap_digest 0xc90d06a8 != omap_digest 
0x337cf025 from auth oi 
5:02bc4def:::.dir.default.64449186.344176:head(1177700'1180639 
osd.1322.0:537914 dirty|omap|data_digest|omap_digest s 0 uv 1177199 dd  
od 337cf025 alloc_hint [0 0])
2017-09-07 20:03:54.721681 osd.1322 [REDACTED::8861]:6808/3479303 1740 : 
cluster [ERR] 5.3d40 deep-scrub 0 missing, 1 inconsistent objects
2017-09-07 20:03:54.721687 osd.1322 [REDACTED::8861]:6808/3479303 1741 : 
cluster [ERR] 5.3d40 deep-scrub 3 errors

$ fgrep 5.f1c0   /var/log/ceph/ceph.log
2017-09-07 11:11:36.773986 osd.631 [REDACTED::8877]:6813/4036028 4234 : cluster 
[INF] osd.631 pg 5.f1c0 Deep scrub errors, upgrading scrub to deep-scrub
2017-09-07 11:11:36.774127 osd.631 [REDACTED::8877]:6813/4036028 4235 : cluster 
[INF] 5.f1c0 deep-scrub starts
2017-09-07 11:25:26.231502 osd.631 [REDACTED::8877]:6813/4036028 4236 : cluster 
[ERR] 5.f1c0 deep-scrub 0 missing, 1 inconsistent objects
2017-09-07 11:25:26.231508 osd.631 [REDACTED::8877]:6813/4036028 4237 : cluster 
[ERR] 5.f1c0 deep-scrub 1 errors

-- 
Robin Hugh Johnson
Gentoo Linux: Dev, Infra Lead, Foundation Asst. Treasurer
E-Mail   : robb...@gentoo.org
GnuPG FP : 11ACBA4F 4778E3F6 E4EDF38E B27B944E 34884E85
GnuPG FP : 7D0B3CEB E9B85B1F 825BCECF EE05E6F6 A48F6136


signature.asc
Description: Digital signature
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com