Hi all,

I have a Ceph Cluster using squid 19.2.2.  I'm using CephFS storage with kernel mounts on clients and it works fine but for some older computers (still CentOS8 and one CentOS7 because it is running a commercial application) I need to use NFS so I've deployed Ganesha. But a few minutes after the first NFS client mount the volume (Just testing from the first client) and uses it, the Ceph NFS service falls in error on one node (not always the same node) and I have to reboot it.

# ceph orch ps --daemon-type nfs
NAME                                         HOST PORTS    STATUS         REFRESHED  AGE  MEM USE  MEM LIM VERSION    IMAGE ID      CONTAINER ID nfs.whitaker-nfs.0.2.whitaker04-ceph.itzyqf  whitaker04-ceph *:12049  running (49m)     5m ago   3h    26.3M        - 5.9        4892a7ef541b  42fd4ab7bba9 nfs.whitaker-nfs.1.0.whitaker05-ceph.vocxmr  whitaker05-ceph *:12049  running (63m)     5m ago  23h    85.7M        - 5.9        4892a7ef541b  ab1c06be2ce7 nfs.whitaker-nfs.2.0.whitaker01-ceph.rhwuth  whitaker01-ceph *:12049  running (33m)     5m ago  23h    24.6M        - 5.9        4892a7ef541b  eb93963f05cd nfs.whitaker-nfs.3.0.whitaker03-ceph.wqfnnq  whitaker03-ceph *:12049  error             5m ago  23h        -        - <unknown>  <unknown>     <unknown> nfs.whitaker-nfs.4.0.whitaker02-ceph.wjynvn  whitaker02-ceph *:12049  running (49m)     4m ago   3h    28.3M        - 5.9        4892a7ef541b  5cb71f9d8b89

and the client freeze on the mount point.

Any idea about this or how to track the problem ?

Thanks

Patrick

----------------------------------------------

Details of my NFS CEPH cluster, building commands and status:

----------------------------------------------

# *ceph nfs cluster create whitaker-nfs \
whitaker01-ceph,whitaker02-ceph,whitaker03-ceph,whitaker04-ceph,whitaker05-ceph \
  --ingress --virtual_ip 172.30.1.102/24 --ingress-mode haproxy-protocol*

*# ceph nfs cluster info whitaker-nfs*

{
  "whitaker-nfs": {
    "backend": [
      {
        "hostname": "whitaker01-ceph",
        "ip": "172.30.1.97",
        "port": 12049
      },
      {
        "hostname": "whitaker02-ceph",
        "ip": "172.30.1.98",
        "port": 12049
      },
      {
        "hostname": "whitaker03-ceph",
        "ip": "172.30.1.99",
        "port": 12049
      },
      {
        "hostname": "whitaker04-ceph",
        "ip": "172.30.1.100",
        "port": 12049
      },
      {
        "hostname": "whitaker05-ceph",
        "ip": "172.30.1.101",
        "port": 12049
      }
    ],
    "monitor_port": 9049,
    "port": 2049,
    "virtual_ip": "172.30.1.102"
  }
}

*# ceph nfs export create cephfs \
     --cluster-id whitaker-nfs \
     --pseudo-path /data \
     --fsname data \
     --client_addr 172.30.1.0/24*

*# ceph nfs export ls whitaker-nfs --detailed*
[
  {
    "access_type": "none",
    "clients": [
      {
        "access_type": "rw",
        "addresses": [
          "172.30.1.0/24"
        ],
        "squash": "none"
      }
    ],
    "cluster_id": "whitaker-nfs",
    "export_id": 1,
    "fsal": {
      "cmount_path": "/",
      "fs_name": "data",
      "name": "CEPH",
      "user_id": "nfs.whitaker-nfs.data.79da3503"
    },
    "path": "/",
    "protocols": [
      4
    ],
    "pseudo": "/data",
    "security_label": true,
    "squash": "none",
    "transports": [
      "TCP"
    ]
  }
]
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]

Reply via email to