Hi, I've noticed that one of our 6 gluster 3.6.2 nodes has "glusterd" process using 50% of RAM, on the other nodes usage is about 5% This can be a bug? Should I restart glusterd daemon? Thank you A
From: Volnei Puttini [mailto:vol...@vcplinux.com.br] Sent: lunedì 9 febbraio 2015 18:06 To: RASTELLI Alessandro; gluster-users@gluster.org Subject: Re: [Gluster-users] cannot access to CIFS export Hi Alessandro, My system: CentOS 7 samba-vfs-glusterfs-4.1.1-37.el7_0.x86_64 samba-winbind-4.1.1-37.el7_0.x86_64 samba-libs-4.1.1-37.el7_0.x86_64 samba-common-4.1.1-37.el7_0.x86_64 samba-winbind-modules-4.1.1-37.el7_0.x86_64 samba-winbind-clients-4.1.1-37.el7_0.x86_64 samba-4.1.1-37.el7_0.x86_64 samba-client-4.1.1-37.el7_0.x86_64 glusterfs 3.6.2 built on Jan 22 2015 12:59:57 Try this, work fine for me: [GFSVOL] browseable = No comment = Gluster share of volume gfsvol path = / read only = No guest ok = Yes kernel share modes = No posix locking = No vfs objects = glusterfs glusterfs:loglevel = 7 glusterfs:logfile = /var/log/samba/glusterfs-gfstest.log glusterfs:volume = vgtest glusterfs:volfile_server = 192.168.2.21 On 09-02-2015 14:45, RASTELLI Alessandro wrote: Hi, I've created and started a new replica volume "downloadstat" with CIFS export enabled on GlusterFS 3.6.2. I can see the following piece has been added automatically to smb.conf: [gluster-downloadstat] comment = For samba share of volume downloadstat vfs objects = glusterfs glusterfs:volume = downloadstat glusterfs:logfile = /var/log/samba/glusterfs-downloadstat.%M.log glusterfs:loglevel = 7 path = / read only = no guest ok = yes I restarted smb service, without errors. When I try to access from Win7 client to "\\gluster01-mi\gluster-downloadstat<file:///\\gluster01-mi\gluster-downloadstat>" it asks me a login (which user do I need to put?) and then gives me error "The network path was not found" and on Gluster smb.log I see: [2015/02/09 17:21:13.111639, 0] smbd/vfs.c:173(vfs_init_custom) error probing vfs module 'glusterfs': NT_STATUS_UNSUCCESSFUL [2015/02/09 17:21:13.111709, 0] smbd/vfs.c:315(smbd_vfs_init) smbd_vfs_init: vfs_init_custom failed for glusterfs [2015/02/09 17:21:13.111741, 0] smbd/service.c:902(make_connection_snum) vfs_init failed for service gluster-downloadstat Can you explain how to fix? Thanks Alessandro From: gluster-users-boun...@gluster.org<mailto:gluster-users-boun...@gluster.org> [mailto:gluster-users-boun...@gluster.org] On Behalf Of David F. Robinson Sent: domenica 8 febbraio 2015 18:19 To: Gluster Devel; gluster-users@gluster.org<mailto:gluster-users@gluster.org> Subject: [Gluster-users] cannot delete non-empty directory I am seeing these messsages after I delete large amounts of data using gluster 3.6.2. cannot delete non-empty directory: old_shelf4/Aegis/!!!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_final >From the FUSE mount (as root), the directory shows up as empty: # pwd /backup/homegfs/backup.0/old_shelf4/Aegis/!!!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_final # ls -al total 5 d--------- 2 root root 4106 Feb 6 13:55 . drwxrws--- 3 601 dmiller 72 Feb 6 13:55 .. However, when you look at the bricks, the files are still there (none on brick01bkp, all files are on brick02bkp). All of the files are 0-length and have ------T permissions. Any suggestions on how to fix this and how to prevent it from happening? # ls -al /data/brick*/homegfs_bkp/backup.0/old_shelf4/Aegis/\!\!\!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_final /data/brick01bkp/homegfs_bkp/backup.0/old_shelf4/Aegis/!!!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_final: total 4 d---------+ 2 root root 10 Feb 6 13:55 . drwxrws---+ 3 601 raven 36 Feb 6 13:55 .. /data/brick02bkp/homegfs_bkp/backup.0/old_shelf4/Aegis/!!!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_final: total 8 d---------+ 3 root root 4096 Dec 31 1969 . drwxrws---+ 3 601 raven 36 Feb 6 13:55 .. ---------T 5 601 raven 0 Nov 20 00:08 read_inset.f.gz ---------T 5 601 raven 0 Nov 20 00:08 readbc.f.gz ---------T 5 601 raven 0 Nov 20 00:08 readcn.f.gz ---------T 5 601 raven 0 Nov 20 00:08 readinp.f.gz ---------T 5 601 raven 0 Nov 20 00:08 readinp_v1_2.f.gz ---------T 5 601 raven 0 Nov 20 00:08 readinp_v1_3.f.gz ---------T 5 601 raven 0 Nov 20 00:08 rotatept.f.gz d---------+ 2 root root 118 Feb 6 13:54 save1 ---------T 5 601 raven 0 Nov 20 00:08 sepvec.f.gz ---------T 5 601 raven 0 Nov 20 00:08 shadow.f.gz ---------T 5 601 raven 0 Nov 20 00:08 snksrc.f.gz ---------T 5 601 raven 0 Nov 20 00:08 source.f.gz ---------T 5 601 raven 0 Nov 20 00:08 step.f.gz ---------T 5 601 raven 0 Nov 20 00:08 stoprog.f.gz ---------T 5 601 raven 0 Nov 20 00:08 summer6.f.gz ---------T 5 601 raven 0 Nov 20 00:08 totforc.f.gz ---------T 5 601 raven 0 Nov 20 00:08 tritet.f.gz ---------T 5 601 raven 0 Nov 20 00:08 wallrsd.f.gz ---------T 5 601 raven 0 Nov 20 00:08 wheat.f.gz ---------T 5 601 raven 0 Nov 20 00:08 write_inset.f.gz This is using gluster 3.6.2 on a distributed gluster volume that resides on a single machine. Both of the bricks are on one machine consisting of 2x RAID-6 arrays. df -h | grep brick /dev/mapper/vg01-lvol1 88T 22T 66T 25% /data/brick01bkp /dev/mapper/vg02-lvol1 88T 22T 66T 26% /data/brick02bkp # gluster volume info homegfs_bkp Volume Name: homegfs_bkp Type: Distribute Volume ID: 96de8872-d957-4205-bf5a-076e3f35b294 Status: Started Number of Bricks: 2 Transport-type: tcp Bricks: Brick1: gfsib01bkp.corvidtec.com:/data/brick01bkp/homegfs_bkp Brick2: gfsib01bkp.corvidtec.com:/data/brick02bkp/homegfs_bkp Options Reconfigured: storage.owner-gid: 100 performance.io-thread-count: 32 server.allow-insecure: on network.ping-timeout: 10 performance.cache-size: 128MB performance.write-behind-window-size: 128MB server.manage-gids: on changelog.rollover-time: 15 changelog.fsync-interval: 3 =============================== David F. Robinson, Ph.D. President - Corvid Technologies 704.799.6944 x101 [office] 704.252.1310 [cell] 704.799.7974 [fax] david.robin...@corvidtec.com<mailto:david.robin...@corvidtec.com> http://www.corvidtechnologies.com _______________________________________________ Gluster-users mailing list Gluster-users@gluster.org<mailto:Gluster-users@gluster.org> http://www.gluster.org/mailman/listinfo/gluster-users
_______________________________________________ Gluster-users mailing list Gluster-users@gluster.org http://www.gluster.org/mailman/listinfo/gluster-users