Hello,

First of all, thanks for reading my message. I set up a Ceph version 18.2.2 
cluster with 4 nodes, everything went fine for a while, but after copying some 
files, the storage showed a warning status and the following message : 
"HEALTH_WARN: 1 MDSs are read only mds.PVE-CZ235007SH(mds.0): MDS in read-only 
mode".

The logs are showing :

Jun 03 08:20:41 PVE-CZ235007SH ceph-mds[1329868]:  -9999> 
2024-06-03T07:57:17.589+0200 77250fc006c0 -1 log_channel(cluster) log [ERR] : 
failed to store backtrace on ino 0x1000000039c object, pool 5, errno -2
Jun 03 08:20:41 PVE-CZ235007SH ceph-mds[1329868]:  -9998> 
2024-06-03T07:57:17.589+0200 77250fc006c0 -1 mds.0.189541 unhandled write error 
(2) No such file or directory, force readonly...

After googling for a while, I did not find a hint to understand more precisely 
the root cause. Any help would we greatly appreciated, or even a link to post 
this request elsewhere if this is not the place to.

Please find below additional details if needed. Thanks a lot !

Nicolas

---

# ceph osd dump
[...]
pool 5 'cephfs_metadata' replicated size 3 min_size 2 crush_rule 0 object_hash 
rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 292 flags 
hashpspool stripe_width 0 pg_autoscale_bias 4 pg_num_min 16 recovery_priority 5 
application cephfs read_balance_score 4.51
[...]

# ceph osd lspools
1 .mgr
4 cephfs_data
5 cephfs_metadata
18 ec-pool-001-data
19 ec-pool-001-metadata


# ceph df
--- RAW STORAGE ---
CLASS     SIZE    AVAIL    USED  RAW USED  %RAW USED
hdd    633 TiB  633 TiB  61 GiB    61 GiB          0
TOTAL  633 TiB  633 TiB  61 GiB    61 GiB          0

--- POOLS ---
POOL                  ID  PGS   STORED  OBJECTS     USED  %USED  MAX AVAIL
.mgr                   1    1  119 MiB       31  357 MiB      0    200 TiB
cephfs_data            4   32   71 KiB    8.38k  240 KiB      0    200 TiB
cephfs_metadata        5   32  329 MiB    6.56k  987 MiB      0    200 TiB
ec-pool-001-data      18   32   42 GiB   15.99k   56 GiB      0    451 TiB
ec-pool-001-metadata  19   32      0 B        0      0 B      0    200 TiB



# ceph status
  cluster:
    id:     f16f53e1-7028-440f-bf48-f99912619c33
    health: HEALTH_WARN
            1 MDSs are read only

  services:
    mon: 4 daemons, quorum 
PVE-CZ235007SG,PVE-CZ2341016V,PVE-CZ235007SH,PVE-CZ2341016T (age 35h)
    mgr: PVE-CZ235007SG(active, since 2d), standbys: PVE-CZ235007SH, 
PVE-CZ2341016T, PVE-CZ2341016V
    mds: 1/1 daemons up, 3 standby
    osd: 48 osds: 48 up (since 2d), 48 in (since 3d)

  data:
    volumes: 1/1 healthy
    pools:   5 pools, 129 pgs
    objects: 30.97k objects, 42 GiB
    usage:   61 GiB used, 633 TiB / 633 TiB avail
    pgs:     129 active+clean
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to