Re: [ceph-users] Update to Mimic with prior Snapshots leads to MDS damaged metadata

2018-06-08 Thread Tobias Florek
Hi!

> We are fixing the release note. https://github.com/ceph/ceph/pull/22445

Thank you! It will help others.

Cheers,
 Tobias Florek


signature.asc
Description: signature
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Update to Mimic with prior Snapshots leads to MDS damaged metadata

2018-06-07 Thread Yan, Zheng
On Thu, Jun 7, 2018 at 2:44 PM, Tobias Florek  wrote:
> Hi!
>
> Thank you for your help! The cluster is running healthily for a day now.
>
> Regarding the problem, I just checked in the release notes [1] and on
> docs.ceph.com and did not find the right invocation after an upgrade.
> Maybe that ought to be fixed.
>

We are fixing the release note. https://github.com/ceph/ceph/pull/22445

>>> [upgrade from luminous to mimic with prior cephfs snapshots]
>> The correct commands should be:
>>
>> ceph daemon  scrub_path / force recursive repair
>> ceph daemon  scrub_path '~mdsdir' force recursive
>
> [1] https://ceph.com/releases/v13-2-0-mimic-released/
> [2]
> https://www.google.com/search?q=site%3Adocs.ceph.com+scrub_path+inurl%3Amimic=site%3Adocs.ceph.com+scrub_path+inurl%3Amimic
>
> Cheers,
>  Tobias Florek
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Update to Mimic with prior Snapshots leads to MDS damaged metadata

2018-06-07 Thread Tobias Florek
Hi!

Thank you for your help! The cluster is running healthily for a day now.

Regarding the problem, I just checked in the release notes [1] and on
docs.ceph.com and did not find the right invocation after an upgrade.
Maybe that ought to be fixed.

>> [upgrade from luminous to mimic with prior cephfs snapshots]
> The correct commands should be:
>
> ceph daemon  scrub_path / force recursive repair
> ceph daemon  scrub_path '~mdsdir' force recursive

[1] https://ceph.com/releases/v13-2-0-mimic-released/
[2]
https://www.google.com/search?q=site%3Adocs.ceph.com+scrub_path+inurl%3Amimic=site%3Adocs.ceph.com+scrub_path+inurl%3Amimic

Cheers,
 Tobias Florek
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Update to Mimic with prior Snapshots leads to MDS damaged metadata

2018-06-06 Thread Yan, Zheng
On Wed, Jun 6, 2018 at 3:25 PM, Tobias Florek  wrote:
> Hi,
>
> I upgraded a ceph cluster to mimic yesterday according to the release
> notes. Specifically I did stop all standby MDS and then restarted the
> only active MDS with the new version.
>
> The cluster was installed with luminous. Its cephfs volume had snapshots
> prior to the update, but only one active MDS.
>
> The post-installation steps failed though:
>  ceph daemon mds. scrub_path /
> returned an error, which I corrected with
>  ceph daemon mds. scrub_path / repair
>
> While
>  ceph daemon mds. scrub_path '~mdsdir'
> did not show any error.
>

The correct commands should be:

ceph daemon  scrub_path / force recursive repair
ceph daemon  scrub_path '~mdsdir' force recursive repair


>
> After some time, ceph health reported MDS damaged metadata:
>> ceph tell mds. damage ls | jq '.[].damage_type' | sort | uniq -c
> 398 "backtrace"
> 718 "dentry"
>
> Examples of damage:
>
> {
>   "damage_type": "dentry",
>   "id": 118195760,
>   "ino": 1099513350198,
>   "frag": "000100*",
>   "dname":
> "1524578400.M820820P705532.dovecot-15-hgjlx,S=425674,W=431250:2,RS",
>   "snap_id": "head",
>   "path":
> "/path/to/mails/user/Maildir/.Trash/cur/1524578400.M820820P705532.dovecot-15-hgjlx,S=425674,W=431250:2,RS"
> },
> {
>   "damage_type": "backtrace",
>   "id": 121083841,
>   "ino": 1099515215027,
>   "path":
> "/path/to/mails/other_user/Maildir/.Junk/cur/1528189963.M416032P698926.dovecot-15-xmpkh,S=4010,W=4100:2,Sab"
> },
>

'ceph daemon  scrub_path / force recursive repair'
should also fix above errors.

Regards
Yan, Zheng


>
> Directories with damage can still be listed by the kernel cephfs mount
> (4.16.7), but not the fuse mount, which stalls.
>
>
> Can anyone help? That's unfortunately a production cluster.
>
> Regards,
>  Tobias Florek
> ___
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] Update to Mimic with prior Snapshots leads to MDS damaged metadata

2018-06-06 Thread Tobias Florek
Hi,

I upgraded a ceph cluster to mimic yesterday according to the release
notes. Specifically I did stop all standby MDS and then restarted the
only active MDS with the new version.

The cluster was installed with luminous. Its cephfs volume had snapshots
prior to the update, but only one active MDS.

The post-installation steps failed though:
 ceph daemon mds. scrub_path /
returned an error, which I corrected with
 ceph daemon mds. scrub_path / repair

While
 ceph daemon mds. scrub_path '~mdsdir'
did not show any error.


After some time, ceph health reported MDS damaged metadata:
> ceph tell mds. damage ls | jq '.[].damage_type' | sort | uniq -c
398 "backtrace"
718 "dentry"

Examples of damage:

{
  "damage_type": "dentry",
  "id": 118195760,
  "ino": 1099513350198,
  "frag": "000100*",
  "dname":
"1524578400.M820820P705532.dovecot-15-hgjlx,S=425674,W=431250:2,RS",
  "snap_id": "head",
  "path":
"/path/to/mails/user/Maildir/.Trash/cur/1524578400.M820820P705532.dovecot-15-hgjlx,S=425674,W=431250:2,RS"
},
{
  "damage_type": "backtrace",
  "id": 121083841,
  "ino": 1099515215027,
  "path":
"/path/to/mails/other_user/Maildir/.Junk/cur/1528189963.M416032P698926.dovecot-15-xmpkh,S=4010,W=4100:2,Sab"
},


Directories with damage can still be listed by the kernel cephfs mount
(4.16.7), but not the fuse mount, which stalls.


Can anyone help? That's unfortunately a production cluster.

Regards,
 Tobias Florek
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com