Hello everyone and sorry. Maybe someone has already faced this problem. 
A day ago, we restored our Openshift cluster, however, at the moment, the PVCs 
cannot connect to the pod. We looked at the status of the ceph and found that 
our MDS were in standby mode, then found that the metadata was corrupted. After 
some manipulations, we were able to turn on our MDS daemons, but there is still 
no record on the cluster, the ceph status command shows the following.

sh-4.4$ ceph -s
  cluster:
    id:     9213604e-b0b6-49d5-bcb3-f55ab3d79119
    health: HEALTH_ERR
            1 MDSs report damaged metadata
            1 MDSs are read only
            6 daemons have recently crashed
  services:
    mon: 5 daemons, quorum bd,bj,bm,bn,bo (age 26h)
    mgr: a(active, since 25h)
    mds: 1/1 daemons up, 1 hot standby
    osd: 9 osds: 9 up (since 41h), 9 in (since 42h)
    rgw: 1 daemon active (1 hosts, 1 zones)
  data:
    volumes: 1/1 healthy
    pools:   10 pools, 225 pgs
    objects: 1.60M objects, 234 GiB
    usage:   606 GiB used, 594 GiB / 1.2 TiB avail
    pgs:     225 active+clean
  io:
    client:   852 B/s rd, 1 op/s rd, 0 op/s wr

Now we trying to follow this instructions:
https://docs.ceph.com/en/latest/cephfs/disaster-recovery-experts/#recovery-from-missing-metadata-objects

What else have we tried:

cephfs-journal-tool --rank=1:0 event recover_dentries summary
cephfs-journal-tool --rank=1:0 journal reset
cephfs-table-tool all reset session
ceph tell mds.gml--cephfs-a scrub start / recursive repair force
ceph tell mds.gml--cephfs-b scrub start / recursive repair force
ceph mds repaired 0

ceph tell mds.gml--cephfs-a damage ls

[
    {
        "damage_type": "dir_frag",
        "id": 26851730,
        "ino": 1100162409473,
        "frag": "*",
        "path": 
"/volumes/csi/csi-vol-5ad18c03-3205-11ed-9ba7-0a580a810206/e5664004-51e0-4bff-85c8-029944b431d8/store/096/096a1497-78ab-4802-a5a7-d09e011fd3a5/202301_1027796_1027796_0"
    },
………

    {
        "damage_type": "dir_frag",
        "id": 118336643,
        "ino": 1100162424469,
        "frag": "*",
        "path": 
"/volumes/csi/csi-vol-5ad18c03-3205-11ed-9ba7-0a580a810206/e5664004-51e0-4bff-85c8-029944b431d8/store/096/096a1497-78ab-4802-a5a7-d09e011fd3a5/202301_1027832_1027832_0"
    },

Now we trying: 

# Session table
cephfs-table-tool 0 reset session
# SnapServer
cephfs-table-tool 0 reset snap
# InoTable
cephfs-table-tool 0 reset inode
# Journal
cephfs-journal-tool --rank=0 journal reset
# Root inodes ("/" and MDS directory)
cephfs-data-scan init

cephfs-data-scan scan_extents <data pool>
cephfs-data-scan scan_inodes <data pool>
cephfs-data-scan scan_links

Is it right way and cant it be our salvation? 
Thank you!
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to