Hi all,
I have a Ceph Cluster using squid 19.2.2. I'm using CephFS storage with
kernel mounts on clients and it works fine but for some older computers
(still CentOS8 and one CentOS7 because it is running a commercial
application) I need to use NFS so I've deployed Ganesha. But a few
minutes after the first NFS client mount the volume (Just testing from
the first client) and uses it, the Ceph NFS service falls in error on
one node (not always the same node) and I have to reboot it.
# ceph orch ps --daemon-type nfs
NAME HOST PORTS
STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID
CONTAINER ID
nfs.whitaker-nfs.0.2.whitaker04-ceph.itzyqf whitaker04-ceph *:12049
running (49m) 5m ago 3h 26.3M - 5.9 4892a7ef541b
42fd4ab7bba9
nfs.whitaker-nfs.1.0.whitaker05-ceph.vocxmr whitaker05-ceph *:12049
running (63m) 5m ago 23h 85.7M - 5.9 4892a7ef541b
ab1c06be2ce7
nfs.whitaker-nfs.2.0.whitaker01-ceph.rhwuth whitaker01-ceph *:12049
running (33m) 5m ago 23h 24.6M - 5.9 4892a7ef541b
eb93963f05cd
nfs.whitaker-nfs.3.0.whitaker03-ceph.wqfnnq whitaker03-ceph *:12049
error 5m ago 23h - - <unknown> <unknown>
<unknown>
nfs.whitaker-nfs.4.0.whitaker02-ceph.wjynvn whitaker02-ceph *:12049
running (49m) 4m ago 3h 28.3M - 5.9 4892a7ef541b
5cb71f9d8b89
and the client freeze on the mount point.
Any idea about this or how to track the problem ?
Thanks
Patrick
----------------------------------------------
Details of my NFS CEPH cluster, building commands and status:
----------------------------------------------
# *ceph nfs cluster create whitaker-nfs \
whitaker01-ceph,whitaker02-ceph,whitaker03-ceph,whitaker04-ceph,whitaker05-ceph
\
--ingress --virtual_ip 172.30.1.102/24 --ingress-mode haproxy-protocol*
*# ceph nfs cluster info whitaker-nfs*
{
"whitaker-nfs": {
"backend": [
{
"hostname": "whitaker01-ceph",
"ip": "172.30.1.97",
"port": 12049
},
{
"hostname": "whitaker02-ceph",
"ip": "172.30.1.98",
"port": 12049
},
{
"hostname": "whitaker03-ceph",
"ip": "172.30.1.99",
"port": 12049
},
{
"hostname": "whitaker04-ceph",
"ip": "172.30.1.100",
"port": 12049
},
{
"hostname": "whitaker05-ceph",
"ip": "172.30.1.101",
"port": 12049
}
],
"monitor_port": 9049,
"port": 2049,
"virtual_ip": "172.30.1.102"
}
}
*# ceph nfs export create cephfs \
--cluster-id whitaker-nfs \
--pseudo-path /data \
--fsname data \
--client_addr 172.30.1.0/24*
*# ceph nfs export ls whitaker-nfs --detailed*
[
{
"access_type": "none",
"clients": [
{
"access_type": "rw",
"addresses": [
"172.30.1.0/24"
],
"squash": "none"
}
],
"cluster_id": "whitaker-nfs",
"export_id": 1,
"fsal": {
"cmount_path": "/",
"fs_name": "data",
"name": "CEPH",
"user_id": "nfs.whitaker-nfs.data.79da3503"
},
"path": "/",
"protocols": [
4
],
"pseudo": "/data",
"security_label": true,
"squash": "none",
"transports": [
"TCP"
]
}
]
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]