On Thu, Apr 21, 2022 at 07:28:19PM +0000, Ryan Taylor wrote: > > Hi Luís, > > dmesg looks normal I think:
Yep, I don't see anything suspicious either. > > [ 265.269450] Key type ceph registered > [ 265.270914] libceph: loaded (mon/osd proto 15/24) > [ 265.303764] FS-Cache: Netfs 'ceph' registered for caching > [ 265.305460] ceph: loaded (mds proto 32) > [ 265.513616] libceph: mon0 (1)10.30.201.3:6789 session established > [ 265.520982] libceph: client3734313 fsid > 50004482-d5e3-4b76-9a4c-abd0626c9882 > [ 265.539710] ceph: mds0 rejected session > [ 265.544592] libceph: mon1 (1)10.30.202.3:6789 session established > [ 265.549564] libceph: client3698116 fsid > 50004482-d5e3-4b76-9a4c-abd0626c9882 > [ 265.552624] ceph: mds0 rejected session > [ 316.849402] libceph: mon0 (1)10.30.201.3:6789 session established > [ 316.855077] libceph: client3734316 fsid > 50004482-d5e3-4b76-9a4c-abd0626c9882 > [ 316.886834] ceph: mds0 rejected session > [ 372.064685] libceph: mon2 (1)10.30.203.3:6789 session established > [ 372.068731] libceph: client3708026 fsid > 50004482-d5e3-4b76-9a4c-abd0626c9882 > [ 372.071651] ceph: mds0 rejected session > [ 372.074641] libceph: mon0 (1)10.30.201.3:6789 session established > [ 372.080435] libceph: client3734319 fsid > 50004482-d5e3-4b76-9a4c-abd0626c9882 > [ 372.083270] ceph: mds0 rejected session > [ 443.855530] libceph: mon2 (1)10.30.203.3:6789 session established > [ 443.863231] libceph: client3708029 fsid > 50004482-d5e3-4b76-9a4c-abd0626c9882 > [ 555.889186] libceph: mon2 (1)10.30.203.3:6789 session established > [ 555.893677] libceph: client3708032 fsid > 50004482-d5e3-4b76-9a4c-abd0626c9882 > [ 1361.181405] libceph: mon0 (1)10.30.201.3:6789 session established > [ 1361.187230] libceph: client3734325 fsid > 50004482-d5e3-4b76-9a4c-abd0626c9882 > [ 1415.463391] libceph: mon2 (1)10.30.203.3:6789 session established > [ 1415.467663] libceph: client3708038 fsid > 50004482-d5e3-4b76-9a4c-abd0626c9882 > [ 2018.707478] libceph: mon0 (1)10.30.201.3:6789 session established > [ 2018.712834] libceph: client3734337 fsid > 50004482-d5e3-4b76-9a4c-abd0626c9882 > [ 2276.564841] libceph: mon1 (1)10.30.202.3:6789 session established > [ 2276.568899] libceph: client3698128 fsid > 50004482-d5e3-4b76-9a4c-abd0626c9882 > [ 2435.596579] libceph: mon2 (1)10.30.203.3:6789 session established > [ 2435.600599] libceph: client3708050 fsid > 50004482-d5e3-4b76-9a4c-abd0626c9882 > [89805.777644] libceph: mon0 (1)10.30.201.3:6789 session established > [89805.782455] libceph: client3740982 fsid > 50004482-d5e3-4b76-9a4c-abd0626c9882 > [89868.055719] libceph: mon1 (1)10.30.202.3:6789 session established > [89868.059600] libceph: client3704767 fsid > 50004482-d5e3-4b76-9a4c-abd0626c9882 > > Pretty sure "mds0 rejected session" was when I was accidentally trying to > mount the wrong share yesterday. > > Could it depend on the Ceph version (ours is v14.2.22) , or could it depend > on something Manila is doing? I thought that too, but then I compiled a 14.2.22 and I still couldn't reproduce it either (note: this was on a vstart cluster, not a *real* one). > Is there any other useful information I could collect? I guess you could try to get some more detailed kernel logs, but I'm not sure your kernel is compiled with the required options. To check if it is, just see if file "/sys/kernel/debug/dynamic_debug/control" exists. If it does, we're good to go! # enable kernel client debug: echo 'module ceph +p' > /sys/kernel/debug/dynamic_debug/control # run the mount command on the subdir # run df -h # disable kernel client debug: echo 'module ceph -p' > /sys/kernel/debug/dynamic_debug/control Note that the kernel logging can be quite verbose. Maybe a good idea is to just script it to be quick ;-) After that, please share the log (dmesg). Maybe it has some hint on what's going on. Cheers, -- Luís > > Thanks, > -rt > > Ryan Taylor > Research Computing Specialist > Research Computing Services, University Systems > University of Victoria > > ________________________________________ > From: Luís Henriques <lhenriq...@suse.de> > Sent: April 21, 2022 4:32 AM > To: Ryan Taylor > Cc: Hendrik Peyerl; Ramana Venkatesh Raja; ceph-users@ceph.io > Subject: Re: [ceph-users] Re: df shows wrong size of cephfs share when a > subdirectory is mounted > > Notice: This message was sent from outside the University of Victoria email > system. Please be cautious with links and sensitive information. > > > On Wed, Apr 20, 2022 at 07:05:37PM +0000, Ryan Taylor wrote: > > > > Hi Luís, > > > > The same cephx key is used for both mounts. It is a regular rw key which > > does not have permission to set any ceph xattrs (that was done > > separately with a different key). But it can read ceph xattrs and set > > user xattrs. > > Thank you for the very detail description. I'm still scratching my head > to figure out what's wrong as I can't reproduce this. Just out of > curiosity: are you seeing any errors/warnings in the kernel log? (dmesg) > > Cheers, > -- > Luís > > > > > I just did a test using the latest Fedora 35 kernel and reproduced the > > problem: > > > > [fedora@cephtest ~]$ sudo mkdir /mnt/ceph1 > > [fedora@cephtest ~]$ sudo mkdir /mnt/ceph2 > > [fedora@cephtest ~]$ sudo mount -t ceph > > 10.30.201.3:6789,10.30.202.3:6789,10.30.203.3:6789:/volumes/_nogroup/55e46a89-31ff-4878-9e2a-81b4226c3cb2 > > /mnt/ceph1 -o name=rwkey,secret=... > > [fedora@cephtest ~]$ sudo mkdir /mnt/ceph1/testsubdir > > [fedora@cephtest ~]$ sudo mount -t ceph > > 10.30.201.3:6789,10.30.202.3:6789,10.30.203.3:6789:/volumes/_nogroup/55e46a89-31ff-4878-9e2a-81b4226c3cb2/testsubdir > > /mnt/ceph2 -o name=rwkey,secret=... > > [fedora@cephtest ~]$ df | grep ceph > > 10.30.201.3:6789,10.30.202.3:6789,10.30.203.3:6789:/volumes/_nogroup/55e46a89-31ff-4878-9e2a-81b4226c3cb2 > > 5242880000 291385344 4951494656 > > 6% /mnt/ceph1 > > 10.30.201.3:6789,10.30.202.3:6789,10.30.203.3:6789:/volumes/_nogroup/55e46a89-31ff-4878-9e2a-81b4226c3cb2/testsubdir > > 4287562399744 295238516736 3992323883008 7% /mnt/ceph2 > > [fedora@cephtest ~]$ uname -r > > 5.16.20-200.fc35.x86_64 > > > > Furthermore I then repeated my earlier test regarding ceph.quota.max_bytes. > > The volume root already has the right quota based on the size of my Manila > > share in Openstack, and it matches the size reported by df (5000 GiB) > > > > [fedora@cephtest ~]$ getfattr -n ceph.quota.max_bytes /mnt/ceph1/ > > getfattr: Removing leading '/' from absolute path names > > # file: mnt/ceph1/ > > ceph.quota.max_bytes="5368709120000" > > > > And on a separate system with admin credentials I applied a max_bytes quota > > to the testsubdir: > > > > sudo setfattr -n ceph.quota.max_bytes -v 121212 > > /mnt/cephfs/volumes/_nogroup/55e46a89-31ff-4878-9e2a-81b4226c3cb2/testsubdir/ > > > > I unmounted and remounted testsubdir exactly as before, but even with > > ceph.quota.max_bytes applied on the subdir it still shows the wrong size: > > > > [fedora@cephtest ~]$ df | grep ceph > > 10.30.201.3:6789,10.30.202.3:6789,10.30.203.3:6789:/volumes/_nogroup/55e46a89-31ff-4878-9e2a-81b4226c3cb2 > > 5242880000 291385344 4951494656 > > 6% /mnt/ceph1 > > 10.30.201.3:6789,10.30.202.3:6789,10.30.203.3:6789:/volumes/_nogroup/55e46a89-31ff-4878-9e2a-81b4226c3cb2/testsubdir > > 4287544954880 295264587776 3992280367104 7% /mnt/ceph2 > > > > [fedora@cephtest ~]$ getfattr -n ceph.quota.max_bytes > > /mnt/ceph1/testsubdir/ > > getfattr: Removing leading '/' from absolute path names > > # file: mnt/ceph1/testsubdir/ > > ceph.quota.max_bytes="121212" > > > > [fedora@cephtest ~]$ getfattr -n ceph.quota.max_bytes /mnt/ceph2 > > getfattr: Removing leading '/' from absolute path names > > # file: mnt/ceph2 > > ceph.quota.max_bytes="121212" > > > > Thanks, > > -rt > > > > > > > > > > > > ________________________________________ > > From: Luís Henriques <lhenriq...@suse.de> > > Sent: April 20, 2022 7:16 AM > > To: Ryan Taylor > > Cc: Hendrik Peyerl; Ramana Venkatesh Raja; ceph-users@ceph.io > > Subject: Re: [ceph-users] Re: df shows wrong size of cephfs share when a > > subdirectory is mounted > > > > Notice: This message was sent from outside the University of Victoria email > > system. Please be cautious with links and sensitive information. > > > > > > On Tue, Apr 19, 2022 at 08:51:50PM +0000, Ryan Taylor wrote: > > > Thanks for the pointers! It does look like > > > https://tracker.ceph.com/issues/55090 > > > and I am not surprised Dan and I are hitting the same issue... > > > > Just a wild guess (already asked this on the tracker): > > > > Is it possible that you're using different credentials/keys so that the > > credentials used for mounting the subdir are not allowed to access the > > volume base directory? Would it be possible to get more details on the > > two mount commands being used? > > > > Cheers, > > -- > > Luís > > > > > > > > > > > I am using the latest available Almalinux 8, 4.18.0-348.20.1.el8_5.x86_64 > > > > > > Installing kernel-debuginfo-common-x86_64 > > > I see in > > > /usr/src/debug/kernel-4.18.0-348.2.1.el8_5/linux-4.18.0-348.2.1.el8_5.x86_64/fs/ceph/quota.c > > > for example: > > > > > > static inline bool ceph_has_realms_with_quotas(struct inode *inode) > > > { > > > struct super_block *sb = inode->i_sb; > > > struct ceph_mds_client *mdsc = ceph_sb_to_mdsc(sb); > > > struct inode *root = d_inode(sb->s_root); > > > > > > if (atomic64_read(&mdsc->quotarealms_count) > 0) > > > return true; > > > /* if root is the real CephFS root, we don't have quota realms */ > > > if (root && ceph_ino(root) == CEPH_INO_ROOT) > > > return false; > > > /* otherwise, we can't know for sure */ > > > return true; > > > } > > > > > > So this EL8.5 kernel already has at least some of the patches from > > > https://lore.kernel.org/all/20190301175752.17808-1-lhenriq...@suse.com/T/#u > > > for https://tracker.ceph.com/issues/38482 > > > That does not mention a specific commit, just says "Merged into 5.2-rc1." > > > > > > So it seems https://tracker.ceph.com/issues/55090 is either a new issue > > > or a regression of the previous issue. > > > > > > Thanks, > > > -rt > > > > > > Ryan Taylor > > > Research Computing Specialist > > > Research Computing Services, University Systems > > > University of Victoria > > > > > > ________________________________________ > > > From: Hendrik Peyerl <hpey...@plusline.net> > > > Sent: April 19, 2022 6:05 AM > > > To: Ramana Venkatesh Raja > > > Cc: Ryan Taylor; ceph-users@ceph.io > > > Subject: Re: [ceph-users] df shows wrong size of cephfs share when a > > > subdirectory is mounted > > > > > > Notice: This message was sent from outside the University of Victoria > > > email system. Please be cautious with links and sensitive information. > > > > > > > > > I did hit this issue aswell: https://tracker.ceph.com/issues/38482 > > > > > > you will need a kernel >= 5.2 that can handle the quotas on > > > subdirectories. > > > > > > > > > > On 19. Apr 2022, at 14:47, Ramana Venkatesh Raja <rr...@redhat.com> > > > > wrote: > > > > > > > > On Sat, Apr 16, 2022 at 10:15 PM Ramana Venkatesh Raja > > > > <rr...@redhat.com> wrote: > > > >> > > > >> On Thu, Apr 14, 2022 at 8:07 PM Ryan Taylor <rptay...@uvic.ca> wrote: > > > >>> > > > >>> Hello, > > > >>> > > > >>> > > > >>> I am using cephfs via Openstack Manila (Ussuri I think). > > > >>> > > > >>> The cephfs cluster is v14.2.22 and my client has kernel > > > >>> 4.18.0-348.20.1.el8_5.x86_64 > > > >>> > > > >>> > > > >>> I have a Manila share > > > >>> > > > >>> /volumes/_nogroup/55e46a89-31ff-4878-9e2a-81b4226c3cb2 > > > >>> > > > >>> > > > >>> that is 5000 GB in size. When I mount it the size is reported > > > >>> correctly: > > > >>> > > > >>> > > > >>> # df -h /cephfs > > > >>> Filesystem > > > >>> Size Used Avail Use% Mounted on > > > >>> 10.30.201.3:6789,10.30.202.3:6789,10.30.203.3:6789:/volumes/_nogroup/55e46a89-31ff-4878-9e2a-81b4226c3cb2 > > > >>> 4.9T 278G 4.7T 6% /cephfs > > > >>> > > > >>> > > > >>> However when I mount a subpath /test1 of my share, then both the size > > > >>> and usage are showing the size of the whole cephfs filesystem rather > > > >>> than my private share. > > > >>> > > > >>> > > > >>> # df -h /cephfs > > > >>> Filesystem > > > >>> Size Used Avail Use% > > > >>> Mounted on > > > >>> 10.30.201.3:6789,10.30.202.3:6789,10.30.203.3:6789:/volumes/_nogroup/55e46a89-31ff-4878-9e2a-81b4226c3cb2/test1 > > > >>> 4.0P 277T 3.7P 7% /cephfs > > > >>> > > > >> > > > >> What are the capabilities of the ceph client user ID that you used to > > > >> mount "/volumes/_nogroup/55e46a89-31ff-4878-9e2a-81b4226c3cb2/test1" ? > > > >> Maybe you're hitting this limitation in > > > >> https://docs.ceph.com/en/latest/cephfs/quota/#limitations , > > > >> "Quotas must be configured carefully when used with path-based mount > > > >> restrictions. The client needs to have access to the directory inode > > > >> on which quotas are configured in order to enforce them. If the client > > > >> has restricted access to a specific path (e.g., /home/user) based on > > > >> the MDS capability, and a quota is configured on an ancestor directory > > > >> they do not have access to (e.g., /home), the client will not enforce > > > >> it. When using path-based access restrictions be sure to configure the > > > >> quota on the directory the client is restricted too (e.g., /home/user) > > > >> or something nested beneath it. " > > > >> > > > > > > > > Hi Ryan, > > > > > > > > I think you maybe actually hitting this > > > > https://tracker.ceph.com/issues/55090 . Are you facing this issue with > > > > the FUSE client? > > > > > > > > -Ramana > > > > > > > >>> > > > >>> I tried setting the ceph.quota.max_bytes xattr on a subdirectory > > > >>> but it did not help. > > > >>> > > > >> > > > >> You can't set quota xattr if your ceph client user ID doesn't have 'p' > > > >> flag in its MDS capabilities, > > > >> https://docs.ceph.com/en/latest/cephfs/client-auth/#layout-and-quota-restriction-the-p-flag > > > >> . > > > >> > > > >> -Ramana > > > >> > > > >>> I'm not sure if the issue is in cephfs or Manila, but what would be > > > >>> required to get the right size and usage stats to be reported by df > > > >>> when a subpath of a share is mounted? > > > >>> > > > >>> > > > >>> Thanks! > > > >>> > > > >>> -rt > > > >>> > > > >>> > > > >>> Ryan Taylor > > > >>> Research Computing Specialist > > > >>> Research Computing Services, University Systems > > > >>> University of Victoria > > > >>> _______________________________________________ > > > >>> ceph-users mailing list -- ceph-users@ceph.io > > > >>> To unsubscribe send an email to ceph-users-le...@ceph.io > > > >>> > > > > > > > > _______________________________________________ > > > > ceph-users mailing list -- ceph-users@ceph.io > > > > To unsubscribe send an email to ceph-users-le...@ceph.io > > > > > > _______________________________________________ > > > ceph-users mailing list -- ceph-users@ceph.io > > > To unsubscribe send an email to ceph-users-le...@ceph.io > _______________________________________________ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io