Hi Hari, thank you very much for your answer. I will try to use the script mentioned above pointing to each backend bricks.
So, if I understand, since I have a gluster cluster composed by 3 nodes (with 12 bricks on each node), I have to execute the script 36 times. Right? You can find below the “df” command output executed on a cluster node: /dev/mapper/cl_s01-gluster 100G 33M 100G 1% /gluster /dev/mapper/gluster_vgd-gluster_lvd 9,0T 5,6T 3,5T 62% /gluster/mnt2 /dev/mapper/gluster_vge-gluster_lve 9,0T 5,7T 3,4T 63% /gluster/mnt3 /dev/mapper/gluster_vgj-gluster_lvj 9,0T 5,7T 3,4T 63% /gluster/mnt8 /dev/mapper/gluster_vgc-gluster_lvc 9,0T 5,6T 3,5T 62% /gluster/mnt1 /dev/mapper/gluster_vgl-gluster_lvl 9,0T 5,8T 3,3T 65% /gluster/mnt10 /dev/mapper/gluster_vgh-gluster_lvh 9,0T 5,7T 3,4T 64% /gluster/mnt6 /dev/mapper/gluster_vgf-gluster_lvf 9,0T 5,7T 3,4T 63% /gluster/mnt4 /dev/mapper/gluster_vgm-gluster_lvm 9,0T 5,4T 3,7T 60% /gluster/mnt11 /dev/mapper/gluster_vgn-gluster_lvn 9,0T 5,4T 3,7T 60% /gluster/mnt12 /dev/mapper/gluster_vgg-gluster_lvg 9,0T 5,7T 3,4T 64% /gluster/mnt5 /dev/mapper/gluster_vgi-gluster_lvi 9,0T 5,7T 3,4T 63% /gluster/mnt7 /dev/mapper/gluster_vgk-gluster_lvk 9,0T 5,8T 3,3T 65% /gluster/mnt9 I will execute the following command and I will put here the output. ./quota_fsck_new.py --full-logs --sub-dir /gluster/mnt{1..12} Thank you again for your support. Regards, Mauro > Il giorno 10 lug 2018, alle ore 11:02, Hari Gowtham <hgowt...@redhat.com> ha > scritto: > > Hi, > > There is no explicit command to backup all the quota limits as per my > understanding. need to look further about this. > But you can do the following to backup and set it. > Gluster volume quota volname list which will print all the quota > limits on that particular volume. > You will have to make a note of the directories with their respective limit > set. > Once noted down, you can disable quota on the volume and then enable it. > Once enabled, you will have to set each limit explicitly on the volume. > > Before doing this we suggest you can to try running the script > mentioned above with the backend brick path instead of the mount path. > you need to run this on the machines where the backend bricks are > located and not on the mount. > On Mon, Jul 9, 2018 at 9:01 PM Mauro Tridici <mauro.trid...@cmcc.it> wrote: >> >> Hi Sanoj, >> >> could you provide me the command that I need in order to backup all quota >> limits? >> If there is no solution for this kind of problem, I would like to try to >> follow your “backup” suggestion. >> >> Do you think that I should contact gluster developers too? >> >> Thank you very much. >> Regards, >> Mauro >> >> >> Il giorno 05 lug 2018, alle ore 09:56, Mauro Tridici <mauro.trid...@cmcc.it> >> ha scritto: >> >> Hi Sanoj, >> >> unfortunately the output of the command execution was not helpful. >> >> [root@s01 ~]# find /tier2/CSP/ans004 | xargs getfattr -d -m. -e hex >> [root@s01 ~]# >> >> Do you have some other idea in order to detect the cause of the issue? >> >> Thank you again, >> Mauro >> >> >> Il giorno 05 lug 2018, alle ore 09:08, Sanoj Unnikrishnan >> <sunni...@redhat.com> ha scritto: >> >> Hi Mauro, >> >> A script issue did not capture all necessary xattr. >> Could you provide the xattrs with.. >> find /tier2/CSP/ans004 | xargs getfattr -d -m. -e hex >> >> Meanwhile, If you are being impacted, you could do the following >> back up quota limits >> disable quota >> enable quota >> freshly set the limits. >> >> Please capture the xattr values first, so that we can get to know what went >> wrong. >> Regards, >> Sanoj >> >> >> On Tue, Jul 3, 2018 at 4:09 PM, Mauro Tridici <mauro.trid...@cmcc.it> wrote: >>> >>> Dear Sanoj, >>> >>> thank you very much for your support. >>> I just downloaded and executed the script you suggested. >>> >>> This is the full command I executed: >>> >>> ./quota_fsck_new.py --full-logs --sub-dir /tier2/CSP/ans004/ /gluster >>> >>> In attachment, you can find the logs generated by the script. >>> What can I do now? >>> >>> Thank you very much for your patience. >>> Mauro >>> >>> >>> >>> >>> Il giorno 03 lug 2018, alle ore 11:34, Sanoj Unnikrishnan >>> <sunni...@redhat.com> ha scritto: >>> >>> Hi Mauro, >>> >>> This may be an issue with update of backend xattrs. >>> To RCA further and provide resolution could you provide me with the logs by >>> running the following fsck script. >>> https://review.gluster.org/#/c/19179/6/extras/quota/quota_fsck.py >>> >>> Try running the script and revert with the logs generated. >>> >>> Thanks, >>> Sanoj >>> >>> >>> On Mon, Jul 2, 2018 at 2:21 PM, Mauro Tridici <mauro.trid...@cmcc.it> wrote: >>>> >>>> Dear Users, >>>> >>>> I just noticed that, after some data deletions executed inside >>>> "/tier2/CSP/ans004” folder, the amount of used disk reported by quota >>>> command doesn’t reflect the value indicated by du command. >>>> Surfing on the web, it seems that it is a bug of previous versions of >>>> Gluster FS and it was already fixed. >>>> In my case, the problem seems unfortunately still here. >>>> >>>> How can I solve this issue? Is it possible to do it without starting a >>>> downtime period? >>>> >>>> Thank you very much in advance, >>>> Mauro >>>> >>>> [root@s01 ~]# glusterfs -V >>>> glusterfs 3.10.5 >>>> Repository revision: git://git.gluster.org/glusterfs.git >>>> Copyright (c) 2006-2016 Red Hat, Inc. <https://www.gluster.org/> >>>> GlusterFS comes with ABSOLUTELY NO WARRANTY. >>>> It is licensed to you under your choice of the GNU Lesser >>>> General Public License, version 3 or any later version (LGPLv3 >>>> or later), or the GNU General Public License, version 2 (GPLv2), >>>> in all cases as published by the Free Software Foundation. >>>> >>>> [root@s01 ~]# gluster volume quota tier2 list /CSP/ans004 >>>> Path Hard-limit Soft-limit Used >>>> Available Soft-limit exceeded? Hard-limit exceeded? >>>> ------------------------------------------------------------------------------------------------------------------------------- >>>> /CSP/ans004 1.0TB 99%(1013.8GB) >>>> 3.9TB 0Bytes Yes Yes >>>> >>>> [root@s01 ~]# du -hs /tier2/CSP/ans004/ >>>> 295G /tier2/CSP/ans004/ >>>> >>>> >>>> >>>> >>>> _______________________________________________ >>>> Gluster-users mailing list >>>> Gluster-users@gluster.org >>>> http://lists.gluster.org/mailman/listinfo/gluster-users >>> >>> >>> >> >> >> >> _______________________________________________ >> Gluster-users mailing list >> Gluster-users@gluster.org >> https://lists.gluster.org/mailman/listinfo/gluster-users > > > > -- > Regards, > Hari Gowtham.
_______________________________________________ Gluster-users mailing list Gluster-users@gluster.org https://lists.gluster.org/mailman/listinfo/gluster-users