Has it worked before or did it just stop working at some point? What's the exact command that fails (and error message if there is)?

For the "too many PGs per OSD" I suppose I have to add some other OSDs, right?

Either that or reduce the number of PGs. If you had only a few pools I'd suggest to leave it to the autoscaler, but not for 13 pools. You can paste 'ceph osd df' and 'ceph osd pool ls detail' if you need more input for that.

Zitat von Eugenio Tampieri <eugenio.tampi...@readydigital.it>:

Hi Eugen,
Sorry, but I had some trouble when I signed up and then I was away so I missed your reply.

ceph auth export client.migration
[client.migration]
        key = redacted
        caps mds = "allow rw fsname=repo"
        caps mon = "allow r fsname=repo"
        caps osd = "allow rw tag cephfs data=repo"

For the "too many PGs per OSD" I suppose I have to add some other OSDs, right?

Thanks,

Eugenio

-----Messaggio originale-----
Da: Eugen Block <ebl...@nde.ag>
Inviato: mercoledì 4 settembre 2024 10:07
A: ceph-users@ceph.io
Oggetto: [ceph-users] Re: CephFS troubleshooting

Hi, I already responded to your first attempt:

https://lists.ceph.io/hyperkitty/list/ceph-users@ceph.io/message/GS7KJRJP7BAOF66KJM255G27TJ4KG656/

Please provide the requested details.


Zitat von Eugenio Tampieri <eugenio.tampi...@readydigital.it>:

Hello,
I'm writing to troubleshoot an otherwise functional Ceph quincy
cluster that has issues with cephfs.
I cannot mount it with ceph-fuse (it gets stuck), and if I mount it
with NFS I can list the directories but I cannot read or write
anything.
Here's the output of ceph -s
  cluster:
    id:     3b92e270-1dd6-11ee-a738-000c2937f0ec
    health: HEALTH_WARN
            mon ceph-storage-a is low on available space
            1 daemons have recently crashed
            too many PGs per OSD (328 > max 250)

  services:
    mon:        5 daemons, quorum
ceph-mon-a,ceph-storage-a,ceph-mon-b,ceph-storage-c,ceph-storage-d
(age 105m)
    mgr:        ceph-storage-a.ioenwq(active, since 106m), standbys:
ceph-mon-a.tiosea
    mds:        1/1 daemons up, 2 standby
    osd:        4 osds: 4 up (since 104m), 4 in (since 24h)
    rbd-mirror: 2 daemons active (2 hosts)
    rgw:        2 daemons active (2 hosts, 1 zones)

  data:
    volumes: 1/1 healthy
    pools:   13 pools, 481 pgs
    objects: 231.83k objects, 648 GiB
    usage:   1.3 TiB used, 1.8 TiB / 3.1 TiB avail
    pgs:     481 active+clean

  io:
    client:   1.5 KiB/s rd, 8.6 KiB/s wr, 1 op/s rd, 0 op/s wr
Best regards,

Eugenio Tampieri
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an
email to ceph-users-le...@ceph.io


_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io


_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to