OK, it works with your config. Thanks Alessandro From: RAGHAVENDRA TALUR [mailto:[email protected]] Sent: martedì 10 febbraio 2015 14:57 To: Volnei Puttini Cc: RASTELLI Alessandro; [email protected] Subject: Re: [Gluster-users] cannot access to CIFS export
On Mon, Feb 9, 2015 at 10:35 PM, Volnei Puttini <[email protected]<mailto:[email protected]>> wrote: Hi Alessandro, My system: CentOS 7 samba-vfs-glusterfs-4.1.1-37.el7_0.x86_64 This is the rpm that ships glusterfs.so which Samba searched for and did not find. samba-winbind-4.1.1-37.el7_0.x86_64 samba-libs-4.1.1-37.el7_0.x86_64 samba-common-4.1.1-37.el7_0.x86_64 samba-winbind-modules-4.1.1-37.el7_0.x86_64 samba-winbind-clients-4.1.1-37.el7_0.x86_64 samba-4.1.1-37.el7_0.x86_64 samba-client-4.1.1-37.el7_0.x86_64 glusterfs 3.6.2 built on Jan 22 2015 12:59:57 Try this, work fine for me: [GFSVOL] browseable = No comment = Gluster share of volume gfsvol path = / read only = No guest ok = Yes kernel share modes = No posix locking = No vfs objects = glusterfs glusterfs:loglevel = 7 glusterfs:logfile = /var/log/samba/glusterfs-gfstest.log glusterfs:volume = vgtest glusterfs:volfile_server = 192.168.2.21 On 09-02-2015 14:45, RASTELLI Alessandro wrote: Hi, I’ve created and started a new replica volume “downloadstat” with CIFS export enabled on GlusterFS 3.6.2. I can see the following piece has been added automatically to smb.conf: [gluster-downloadstat] comment = For samba share of volume downloadstat vfs objects = glusterfs glusterfs:volume = downloadstat glusterfs:logfile = /var/log/samba/glusterfs-downloadstat.%M.log glusterfs:loglevel = 7 path = / read only = no guest ok = yes I restarted smb service, without errors. When I try to access from Win7 client to “\\gluster01-mi\gluster-downloadstat<file:///\\gluster01-mi\gluster-downloadstat>” it asks me a login (which user do I need to put?) and then gives me error “The network path was not found” The user you use to login depends on your Samba setup, if it is in AD environment , you would use one of the AD users. If it is not, then you will have to use a user added to Samba on the server using smbpasswd -a <user>. Also, make sure that the same user exists on the node(s) and has rights on the bricks. and on Gluster smb.log I see: [2015/02/09 17:21:13.111639, 0] smbd/vfs.c:173(vfs_init_custom) error probing vfs module 'glusterfs': NT_STATUS_UNSUCCESSFUL [2015/02/09 17:21:13.111709, 0] smbd/vfs.c:315(smbd_vfs_init) smbd_vfs_init: vfs_init_custom failed for glusterfs [2015/02/09 17:21:13.111741, 0] smbd/service.c:902(make_connection_snum) vfs_init failed for service gluster-downloadstat Can you explain how to fix? Thanks Alessandro From: [email protected]<mailto:[email protected]> [mailto:[email protected]] On Behalf Of David F. Robinson Sent: domenica 8 febbraio 2015 18:19 To: Gluster Devel; [email protected]<mailto:[email protected]> Subject: [Gluster-users] cannot delete non-empty directory I am seeing these messsages after I delete large amounts of data using gluster 3.6.2. cannot delete non-empty directory: old_shelf4/Aegis/!!!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_final From the FUSE mount (as root), the directory shows up as empty: # pwd /backup/homegfs/backup.0/old_shelf4/Aegis/!!!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_final # ls -al total 5 d--------- 2 root root 4106 Feb 6 13:55 . drwxrws--- 3 601 dmiller 72 Feb 6 13:55 .. However, when you look at the bricks, the files are still there (none on brick01bkp, all files are on brick02bkp). All of the files are 0-length and have ------T permissions. Any suggestions on how to fix this and how to prevent it from happening? # ls -al /data/brick*/homegfs_bkp/backup.0/old_shelf4/Aegis/\!\!\!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_final /data/brick01bkp/homegfs_bkp/backup.0/old_shelf4/Aegis/!!!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_final: total 4 d---------+ 2 root root 10 Feb 6 13:55 . drwxrws---+ 3 601 raven 36 Feb 6 13:55 .. /data/brick02bkp/homegfs_bkp/backup.0/old_shelf4/Aegis/!!!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_final: total 8 d---------+ 3 root root 4096 Dec 31 1969 . drwxrws---+ 3 601 raven 36 Feb 6 13:55 .. ---------T 5 601 raven 0 Nov 20 00:08 read_inset.f.gz ---------T 5 601 raven 0 Nov 20 00:08 readbc.f.gz ---------T 5 601 raven 0 Nov 20 00:08 readcn.f.gz ---------T 5 601 raven 0 Nov 20 00:08 readinp.f.gz ---------T 5 601 raven 0 Nov 20 00:08 readinp_v1_2.f.gz ---------T 5 601 raven 0 Nov 20 00:08 readinp_v1_3.f.gz ---------T 5 601 raven 0 Nov 20 00:08 rotatept.f.gz d---------+ 2 root root 118 Feb 6 13:54 save1 ---------T 5 601 raven 0 Nov 20 00:08 sepvec.f.gz ---------T 5 601 raven 0 Nov 20 00:08 shadow.f.gz ---------T 5 601 raven 0 Nov 20 00:08 snksrc.f.gz ---------T 5 601 raven 0 Nov 20 00:08 source.f.gz ---------T 5 601 raven 0 Nov 20 00:08 step.f.gz ---------T 5 601 raven 0 Nov 20 00:08 stoprog.f.gz ---------T 5 601 raven 0 Nov 20 00:08 summer6.f.gz ---------T 5 601 raven 0 Nov 20 00:08 totforc.f.gz ---------T 5 601 raven 0 Nov 20 00:08 tritet.f.gz ---------T 5 601 raven 0 Nov 20 00:08 wallrsd.f.gz ---------T 5 601 raven 0 Nov 20 00:08 wheat.f.gz ---------T 5 601 raven 0 Nov 20 00:08 write_inset.f.gz This is using gluster 3.6.2 on a distributed gluster volume that resides on a single machine. Both of the bricks are on one machine consisting of 2x RAID-6 arrays. df -h | grep brick /dev/mapper/vg01-lvol1 88T 22T 66T 25% /data/brick01bkp /dev/mapper/vg02-lvol1 88T 22T 66T 26% /data/brick02bkp # gluster volume info homegfs_bkp Volume Name: homegfs_bkp Type: Distribute Volume ID: 96de8872-d957-4205-bf5a-076e3f35b294 Status: Started Number of Bricks: 2 Transport-type: tcp Bricks: Brick1: gfsib01bkp.corvidtec.com:/data/brick01bkp/homegfs_bkp Brick2: gfsib01bkp.corvidtec.com:/data/brick02bkp/homegfs_bkp Options Reconfigured: storage.owner-gid: 100 performance.io-thread-count: 32 server.allow-insecure: on network.ping-timeout: 10 performance.cache-size: 128MB performance.write-behind-window-size: 128MB server.manage-gids: on changelog.rollover-time: 15 changelog.fsync-interval: 3 =============================== David F. Robinson, Ph.D. President - Corvid Technologies 704.799.6944 x101 [office] 704.252.1310 [cell] 704.799.7974 [fax] [email protected]<mailto:[email protected]> http://www.corvidtechnologies.com _______________________________________________ Gluster-users mailing list [email protected]<mailto:[email protected]> http://www.gluster.org/mailman/listinfo/gluster-users _______________________________________________ Gluster-users mailing list [email protected]<mailto:[email protected]> http://www.gluster.org/mailman/listinfo/gluster-users -- Raghavendra Talur
_______________________________________________ Gluster-users mailing list [email protected] http://www.gluster.org/mailman/listinfo/gluster-users
