OK,
it works with your config.
Thanks
Alessandro

From: RAGHAVENDRA TALUR [mailto:[email protected]]
Sent: martedì 10 febbraio 2015 14:57
To: Volnei Puttini
Cc: RASTELLI Alessandro; [email protected]
Subject: Re: [Gluster-users] cannot access to CIFS export


On Mon, Feb 9, 2015 at 10:35 PM, Volnei Puttini 
<[email protected]<mailto:[email protected]>> wrote:
Hi Alessandro,

My system:

CentOS 7

samba-vfs-glusterfs-4.1.1-37.el7_0.x86_64
This is the rpm that ships glusterfs.so which Samba searched for and did not 
find.

samba-winbind-4.1.1-37.el7_0.x86_64
samba-libs-4.1.1-37.el7_0.x86_64
samba-common-4.1.1-37.el7_0.x86_64
samba-winbind-modules-4.1.1-37.el7_0.x86_64
samba-winbind-clients-4.1.1-37.el7_0.x86_64
samba-4.1.1-37.el7_0.x86_64
samba-client-4.1.1-37.el7_0.x86_64

glusterfs 3.6.2 built on Jan 22 2015 12:59:57

Try this, work fine for me:

[GFSVOL]
    browseable = No
    comment = Gluster share of volume gfsvol
    path = /
    read only = No
    guest ok = Yes
    kernel share modes = No
    posix locking = No
    vfs objects = glusterfs
    glusterfs:loglevel = 7
    glusterfs:logfile = /var/log/samba/glusterfs-gfstest.log
    glusterfs:volume = vgtest
    glusterfs:volfile_server = 192.168.2.21


On 09-02-2015 14:45, RASTELLI Alessandro wrote:
Hi,
I’ve created and started a new replica volume “downloadstat” with CIFS export 
enabled on GlusterFS 3.6.2.
I can see the following piece has been added automatically to smb.conf:
[gluster-downloadstat]
comment = For samba share of volume downloadstat
vfs objects = glusterfs
glusterfs:volume = downloadstat
glusterfs:logfile = /var/log/samba/glusterfs-downloadstat.%M.log
glusterfs:loglevel = 7
path = /
read only = no
guest ok = yes

I restarted smb service, without errors.
When I try to access from Win7 client to 
“\\gluster01-mi\gluster-downloadstat<file:///\\gluster01-mi\gluster-downloadstat>”
 it asks me a login (which user do I need to put?) and then gives me error “The 
network path was not found”
The user you use to login depends on your Samba setup, if it is in AD 
environment , you would use one of the AD users. If it is not, then you will 
have to use a user added to Samba on the server
using smbpasswd -a <user>. Also, make sure that the same user exists on the 
node(s) and has rights on the bricks.


and on Gluster smb.log I see:
[2015/02/09 17:21:13.111639,  0] smbd/vfs.c:173(vfs_init_custom)
  error probing vfs module 'glusterfs': NT_STATUS_UNSUCCESSFUL
[2015/02/09 17:21:13.111709,  0] smbd/vfs.c:315(smbd_vfs_init)
  smbd_vfs_init: vfs_init_custom failed for glusterfs
[2015/02/09 17:21:13.111741,  0] smbd/service.c:902(make_connection_snum)
  vfs_init failed for service gluster-downloadstat

Can you explain how to fix?
Thanks

Alessandro

From: 
[email protected]<mailto:[email protected]> 
[mailto:[email protected]] On Behalf Of David F. Robinson
Sent: domenica 8 febbraio 2015 18:19
To: Gluster Devel; [email protected]<mailto:[email protected]>
Subject: [Gluster-users] cannot delete non-empty directory

I am seeing these messsages after I delete large amounts of data using gluster 
3.6.2.
cannot delete non-empty directory: 
old_shelf4/Aegis/!!!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_final

From the FUSE mount (as root), the directory shows up as empty:

# pwd
/backup/homegfs/backup.0/old_shelf4/Aegis/!!!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_final

# ls -al
total 5
d--------- 2 root root    4106 Feb  6 13:55 .
drwxrws--- 3  601 dmiller   72 Feb  6 13:55 ..

However, when you look at the bricks, the files are still there (none on 
brick01bkp, all files are on brick02bkp).  All of the files are 0-length and 
have ------T permissions.
Any suggestions on how to fix this and how to prevent it from happening?

#  ls -al 
/data/brick*/homegfs_bkp/backup.0/old_shelf4/Aegis/\!\!\!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_final
/data/brick01bkp/homegfs_bkp/backup.0/old_shelf4/Aegis/!!!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_final:
total 4
d---------+ 2 root root  10 Feb  6 13:55 .
drwxrws---+ 3  601 raven 36 Feb  6 13:55 ..

/data/brick02bkp/homegfs_bkp/backup.0/old_shelf4/Aegis/!!!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_final:
total 8
d---------+ 3 root root  4096 Dec 31  1969 .
drwxrws---+ 3  601 raven   36 Feb  6 13:55 ..
---------T  5  601 raven    0 Nov 20 00:08 read_inset.f.gz
---------T  5  601 raven    0 Nov 20 00:08 readbc.f.gz
---------T  5  601 raven    0 Nov 20 00:08 readcn.f.gz
---------T  5  601 raven    0 Nov 20 00:08 readinp.f.gz
---------T  5  601 raven    0 Nov 20 00:08 readinp_v1_2.f.gz
---------T  5  601 raven    0 Nov 20 00:08 readinp_v1_3.f.gz
---------T  5  601 raven    0 Nov 20 00:08 rotatept.f.gz
d---------+ 2 root root   118 Feb  6 13:54 save1
---------T  5  601 raven    0 Nov 20 00:08 sepvec.f.gz
---------T  5  601 raven    0 Nov 20 00:08 shadow.f.gz
---------T  5  601 raven    0 Nov 20 00:08 snksrc.f.gz
---------T  5  601 raven    0 Nov 20 00:08 source.f.gz
---------T  5  601 raven    0 Nov 20 00:08 step.f.gz
---------T  5  601 raven    0 Nov 20 00:08 stoprog.f.gz
---------T  5  601 raven    0 Nov 20 00:08 summer6.f.gz
---------T  5  601 raven    0 Nov 20 00:08 totforc.f.gz
---------T  5  601 raven    0 Nov 20 00:08 tritet.f.gz
---------T  5  601 raven    0 Nov 20 00:08 wallrsd.f.gz
---------T  5  601 raven    0 Nov 20 00:08 wheat.f.gz
---------T  5  601 raven    0 Nov 20 00:08 write_inset.f.gz


This is using gluster 3.6.2 on a distributed gluster volume that resides on a 
single machine.  Both of the bricks are on one machine consisting of 2x RAID-6 
arrays.

df -h | grep brick
/dev/mapper/vg01-lvol1                       88T   22T   66T  25% 
/data/brick01bkp
/dev/mapper/vg02-lvol1                       88T   22T   66T  26% 
/data/brick02bkp

# gluster volume info homegfs_bkp
Volume Name: homegfs_bkp
Type: Distribute
Volume ID: 96de8872-d957-4205-bf5a-076e3f35b294
Status: Started
Number of Bricks: 2
Transport-type: tcp
Bricks:
Brick1: gfsib01bkp.corvidtec.com:/data/brick01bkp/homegfs_bkp
Brick2: gfsib01bkp.corvidtec.com:/data/brick02bkp/homegfs_bkp
Options Reconfigured:
storage.owner-gid: 100
performance.io-thread-count: 32
server.allow-insecure: on
network.ping-timeout: 10
performance.cache-size: 128MB
performance.write-behind-window-size: 128MB
server.manage-gids: on
changelog.rollover-time: 15
changelog.fsync-interval: 3



===============================
David F. Robinson, Ph.D.
President - Corvid Technologies
704.799.6944 x101 [office]
704.252.1310 [cell]
704.799.7974 [fax]
[email protected]<mailto:[email protected]>
http://www.corvidtechnologies.com




_______________________________________________

Gluster-users mailing list

[email protected]<mailto:[email protected]>

http://www.gluster.org/mailman/listinfo/gluster-users


_______________________________________________
Gluster-users mailing list
[email protected]<mailto:[email protected]>
http://www.gluster.org/mailman/listinfo/gluster-users



--
Raghavendra Talur

_______________________________________________
Gluster-users mailing list
[email protected]
http://www.gluster.org/mailman/listinfo/gluster-users

Reply via email to