Hi,

We are looking at how to troubleshoot an issue with Ceph FS on k8s cluster.

This filesystem is provisioned via rook 0.9.2 and have following behavior:
- If ceph fs is mounted on K8S master, then it is writeable
- If ceph fs is mounted as PV to a POD, then we can write a 0-sized file to it, 
(or create empty file) but bigger writes do not work.
Following is reported as ceph -s :

# ceph -s
  cluster:
    id:     18f8d40e-1995-4de4-96dc-e905b097e643
    health: HEALTH_OK
 
  services:
    mon: 3 daemons, quorum a,b,d
    mgr: a(active)
    mds: workspace-storage-fs-1/1/1 up  {0=workspace-storage-fs-a=up:active}, 1 
up:standby-replay
    osd: 3 osds: 3 up, 3 in
 
  data:
    pools:   3 pools, 300 pgs
    objects: 212  objects, 181 MiB
    usage:   51 GiB used, 244 GiB / 295 GiB avail
    pgs:     300 active+clean
 
  io:
    client:   853 B/s rd, 2.7 KiB/s wr, 1 op/s rd, 0 op/s wr


I wonder what can be done for further diagnostics ?

With regards,
Ignat Zapolsky

Sent from Mail for Windows 10


-- 
This email and any files transmitted with it are confidential and intended 
solely for the use of the individual or entity to whom they are addressed. 
If you have received this email in error please notify the system manager. 
This message contains confidential information and is intended only for the 
individual named. If you are not the named addressee you should not 
disseminate, distribute or copy this e-mail.


_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to