Re: [Gluster-users] glusterfs client waiting on SYN_SENT to connect...

2010-12-03 Thread Liam Slusser
Ah the two different IPs are because I was changing my IPs for this mailing
list and I guess I forgot that one.  :)  Will try added a static route.
Also going to snoop traffic and see if the gluster client is actually
getting to the server or being blocked by the firewall.  Ill letcha all know
what I find.

Thanks for the ideas.

Liam
On Dec 3, 2010 6:32 PM,  wrote:
> On Fri, Dec 03, 2010 at 04:25:18PM -0800, Liam Slusser wrote:
>> [r...@client~]# netstat -pan|grep glus
>> tcp 0 1 10.8.10.107:1000 10.8.11.102:6996 SYN_SENT 3385/glusterfs
>>
>> from the gluster client log:
>>
>> However, the port is obviously open...
>>
>> [r...@client~]# telnet 10.8.11.102 6996
>> Trying 10.2.56.102...
>> Connected to glusterserverb (10.8.11.102).
>> Escape character is '^]'.
>> ^]
>> telnet> close
>> Connection closed.
>
> Looking further... why is your telnet trying 10.2.56.102 when you
> clearly specified 10.8.11.102? Also, what happens if you do a
> specific route for the 10.8.11.0/24 block thru the appropriate gw
> without relying on the default gw to route for you? In this way
> you dont end up in a situation where the client is mistakenly
> trying to go over the wrong interface. The telnet maybe switching
> to an alternate interface to see if it gets thru?
>
> Mohan
>
> ___
> Gluster-users mailing list
> Gluster-users@gluster.org
> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


Re: [Gluster-users] glusterfs client waiting on SYN_SENT to connect...

2010-12-03 Thread mki-glusterfs
On Fri, Dec 03, 2010 at 04:25:18PM -0800, Liam Slusser wrote:
> [r...@client~]# netstat -pan|grep glus
> tcp 0 1 10.8.10.107:1000 10.8.11.102:6996 SYN_SENT 3385/glusterfs
> 
> from the gluster client log:
> 
> However, the port is obviously open...
> 
> [r...@client~]# telnet 10.8.11.102 6996
> Trying 10.2.56.102...
> Connected to glusterserverb (10.8.11.102).
> Escape character is '^]'.
> ^]
> telnet> close
> Connection closed.

Looking further... why is your telnet trying 10.2.56.102 when you
clearly specified 10.8.11.102?  Also, what happens if you do a
specific route for the 10.8.11.0/24 block thru the appropriate gw
without relying on the default gw to route for you?  In this way
you dont end up in a situation where the client is mistakenly
trying to go over the wrong interface.  The telnet maybe switching
to an alternate interface to see if it gets thru?

Mohan

___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


Re: [Gluster-users] glusterfs client waiting on SYN_SENT to connect...

2010-12-03 Thread Liam Slusser
Telnet never fails.  Gluster client consistently fails however.
Server is using bonded NICs but as far as i can tell they're
configured correctly, both links are up and passing traffic.

On Fri, Dec 3, 2010 at 6:15 PM,   wrote:
> On Fri, Dec 03, 2010 at 06:03:19PM -0800, Liam Slusser wrote:
>> > This almost seems like a networking/firewall issue... ?Do you have
>> > any trunks setup between the switch that the client and/or server
>> > are on and the router? ?Perhaps one of those trunk legs is down
>> > causing random packets to get blackholed?
>>
>> I thought the exact same thing...but like i said i can telnet to the
>> host/port without any issue.  And there is no other issues on the
>> network that would indicate any not working correctly.  And all the
>> other clients on the same network/switch are working fine.  Its only
>> when a client crashes...
>
> Consistently?  If random telnets fail then that would explain your
> random SYN_SENT state stuck sockets.  Is the client or server using
> bonded nics?
>
> Mohan
> ___
> Gluster-users mailing list
> Gluster-users@gluster.org
> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
>
___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


Re: [Gluster-users] glusterfs client waiting on SYN_SENT to connect...

2010-12-03 Thread mki-glusterfs
On Fri, Dec 03, 2010 at 06:03:19PM -0800, Liam Slusser wrote:
> > This almost seems like a networking/firewall issue... ?Do you have
> > any trunks setup between the switch that the client and/or server
> > are on and the router? ?Perhaps one of those trunk legs is down
> > causing random packets to get blackholed?
>
> I thought the exact same thing...but like i said i can telnet to the
> host/port without any issue.  And there is no other issues on the
> network that would indicate any not working correctly.  And all the
> other clients on the same network/switch are working fine.  Its only
> when a client crashes...

Consistently?  If random telnets fail then that would explain your
random SYN_SENT state stuck sockets.  Is the client or server using
bonded nics?

Mohan
___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


Re: [Gluster-users] glusterfs client waiting on SYN_SENT to connect...

2010-12-03 Thread Liam Slusser
I thought the exact same thing...but like i said i can telnet to the
host/port without any issue.  And there is no other issues on the
network that would indicate any not working correctly.  And all the
other clients on the same network/switch are working fine.  Its only
when a client crashes...

liam

On Fri, Dec 3, 2010 at 4:34 PM,   wrote:
>> I've run into a weird problem.  I have a few client boxes that
>> occasionally crash due to a non-gluster related problem.  But once the
>> box comes back up i cannot get the Gluster client to reconnect to the
>> bricks.
>
> This almost seems like a networking/firewall issue...  Do you have
> any trunks setup between the switch that the client and/or server
> are on and the router?  Perhaps one of those trunk legs is down
> causing random packets to get blackholed?
>
> Mohan
>
___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


[Gluster-users] glusterfs client waiting on SYN_SENT to connect...

2010-12-03 Thread Liam Slusser
Hey all,

I've run into a weird problem.  I have a few client boxes that
occasionally crash due to a non-gluster related problem.  But once the
box comes back up i cannot get the Gluster client to reconnect to the
bricks.

Centos 5 64bit and Gluster 2.0.9

df shows:

df: `/mnt/mymount': Transport endpoint is not connected

[r...@client~]# netstat -pan|grep glus

tcp0  1 10.8.10.107:100010.8.11.102:6996
 SYN_SENT3385/glusterfs
tcp0  1 10.8.10.107:100110.8.11.102:6996
 SYN_SENT3385/glusterfs
tcp0  1 10.8.10.107:998 10.8.11.102:6996
 SYN_SENT3385/glusterfs
tcp0  1 10.8.10.107:996 10.8.11.102:6996
 SYN_SENT3385/glusterfs
tcp0  1 10.8.10.107:100310.8.11.101:6996
 SYN_SENT3385/glusterfs
tcp0  1 10.8.10.107:100210.8.11.101:6996
 SYN_SENT3385/glusterfs
tcp0  1 10.8.10.107:997 10.8.11.101:6996
 SYN_SENT3385/glusterfs
tcp0  1 10.8.10.107:999 10.8.11.101:6996
 SYN_SENT3385/glusterfs

from the gluster client log:

+--+
[2010-12-03 15:48:28] W [glusterfsd.c:526:_log_if_option_is_invalid]
readahead: option 'page-size' is not recognized
[2010-12-03 15:48:28] N [glusterfsd.c:1306:main] glusterfs: Successfully started
[2010-12-03 15:48:29] W [fuse-bridge.c:1892:fuse_statfs_cbk]
glusterfs-fuse: 2: ERR => -1 (Transport endpoint is not connected)
[2010-12-03 15:48:30] W [fuse-bridge.c:1892:fuse_statfs_cbk]
glusterfs-fuse: 3: ERR => -1 (Transport endpoint is not connected)
[2010-12-03 15:48:31] W [fuse-bridge.c:1892:fuse_statfs_cbk]
glusterfs-fuse: 4: ERR => -1 (Transport endpoint is not connected)
[2010-12-03 15:48:31] W [fuse-bridge.c:1892:fuse_statfs_cbk]
glusterfs-fuse: 5: ERR => -1 (Transport endpoint is not connected)
[2010-12-03 15:48:32] W [fuse-bridge.c:1892:fuse_statfs_cbk]
glusterfs-fuse: 6: ERR => -1 (Transport endpoint is not connected)
[2010-12-03 15:51:37] E [socket.c:745:socket_connect_finish] brick1a:
connection to  failed (Connection timed out)
[2010-12-03 15:51:37] E [socket.c:745:socket_connect_finish] brick1a:
connection to  failed (Connection timed out)
[2010-12-03 15:51:37] E [socket.c:745:socket_connect_finish] brick2a:
connection to  failed (Connection timed out)
[2010-12-03 15:51:37] E [socket.c:745:socket_connect_finish] brick2a:
connection to  failed (Connection timed out)
[2010-12-03 15:51:37] E [socket.c:745:socket_connect_finish] brick1b:
connection to  failed (Connection timed out)
[2010-12-03 15:51:37] E [socket.c:745:socket_connect_finish] brick1b:
connection to  failed (Connection timed out)
[2010-12-03 15:51:37] E [socket.c:745:socket_connect_finish] brick2b:
connection to  failed (Connection timed out)
[2010-12-03 15:51:37] E [socket.c:745:socket_connect_finish] brick2b:
connection to  failed (Connection timed out)
[2010-12-03 15:59:46] W [fuse-bridge.c:1892:fuse_statfs_cbk]
glusterfs-fuse: 7: ERR => -1 (Transport endpoint is not connected)
[2010-12-03 15:59:47] W [fuse-bridge.c:1892:fuse_statfs_cbk]
glusterfs-fuse: 8: ERR => -1 (Transport endpoint is not connected)
[2010-12-03 15:59:54] W [fuse-bridge.c:1892:fuse_statfs_cbk]
glusterfs-fuse: 9: ERR => -1 (Transport endpoint is not connected)
[2010-12-03 15:59:55] W [fuse-bridge.c:1892:fuse_statfs_cbk]
glusterfs-fuse: 10: ERR => -1 (Transport endpoint is not connected)
[2010-12-03 15:59:55] W [fuse-bridge.c:1892:fuse_statfs_cbk]
glusterfs-fuse: 11: ERR => -1 (Transport endpoint is not connected)
[2010-12-03 15:59:55] W [fuse-bridge.c:1892:fuse_statfs_cbk]
glusterfs-fuse: 12: ERR => -1 (Transport endpoint is not connected)
[2010-12-03 15:59:56] W [fuse-bridge.c:1892:fuse_statfs_cbk]
glusterfs-fuse: 13: ERR => -1 (Transport endpoint is not connected)

However, the port is obviously open...

[r...@client~]# telnet 10.8.11.102 6996
Trying 10.2.56.102...
Connected to glusterserverb (10.8.11.102).
Escape character is '^]'.
^]
telnet> close
Connection closed.

The gluster server log doesnt see ANY connection attempts from the
client however it DOES see my telnet tcp attempts.  I'm using IP
addresses in all my configuration files - no names.  I do have a
Juniper firewall between the two servers that is doing stateful
firewalling and i've set it up for the connections to never timeout -
and ive never had a problem once it finally connects.  And i can
create a new connection with telnet but not the client...

Anybody seen anything like this before?  Ideas?

thanks,
liam
___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


[Gluster-users] random file not founds (enoent) in a distribute+replicate setup

2010-12-03 Thread mki-glusterfs
Hi

In a distribute+replicate setup, I'm randomly seeing cases where a file 
can't be seen by processes.  A stat of the file using the full path 
results in an ENOENT (No such file or directory).  However if I cd to 
the directory that it's supposed to be in, and then do an ls it appears 
and is fine from there on.

Any ideas as to what might be happening here?  This is 3.0.4 btw, I'm
working to get it updated at the very least to 3.0.6 soonish but need
to get this other cruft cleaned up first, etc.

# ls -al 3/9/0/9/5/263259093
ls: cannot access 3/9/0/9/5/263259093: No such file or directory
# ls -al 3/9/0/9/5
... thousands of files spewed here ...
# ls -al 3/9/0/9/5/263259093
-rw-r--r-- 1 2000 2000 2780068 Nov 12 23:15 3/9/0/9/5/263259093

If a backend server goes offline and comes back on later, would that
cause this issue?  I thought that stat'ing a file would result in
the self heal and eventual consistency across all replicated pairs?
The configs are standard glusterfs-volgen generated raid1 configs
so no read-subvolume directives in there anywhere...

Thanks.

Mohan
___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


Re: [Gluster-users] NFS Mounted GlusterFS, secondary groups not working

2010-12-03 Thread Craig Miller
I confirm the same (bad) behavior with my install of 3.1.1 on CentOS  
5.5, with secondary group membership not having access.


Craig Miller



On 3-Dec-10, at 2:06 PM, Mike Hanby wrote:

The release notes for 3.1.1 don't mention bug 2045, however the bug  
itself indicates that it's fixed with a target of 3.1.1


did it actually make it into the build?

-Original Message-
From: gluster-users-boun...@gluster.org [mailto:gluster-users-boun...@gluster.org 
] On Behalf Of Mike Hanby

Sent: Friday, December 03, 2010 3:02 PM
To: gluster-users@gluster.org
Subject: Re: [Gluster-users] NFS Mounted GlusterFS, secondary groups  
not working


After upgrading my servers to 3.1.1 I'm still experiencing the issue  
with secondary groups not working with the client using NFS:


In this scenario root owns the directory and the admins non primary  
group (of which my id is a member) is the group:


$ groups
mikeh admins

$ ls -ld /research/project/
drwxrwx--- 10 root admins 73728 Nov 22 12:13 /gluster/project/

$ ls -l /gluster/project/
ls: /gluster/project/: Permission denied

$ mount|grep research
nas-01.mydom:/research-storage on /gluster type nfs  
(rw,addr=192.168.1.11)


Both Gluster servers and the client are CentOS 5 x86_64 systems. The  
servers had glusterfs-core / glusterfs-fuse 3.0.1-1 updated to  
3.1.1-1 via the RPMs. The storage is a simple distributed file  
system, no striping or replication.


Following the update I rebooted both servers and remounted the NFS  
file system on the client.


I'll add this info to the bug as well.

Mike

-Original Message-
From: gluster-users-boun...@gluster.org [mailto:gluster-users-boun...@gluster.org 
] On Behalf Of Mike Hanby

Sent: Thursday, November 11, 2010 1:53 PM
To: 'Shehjar Tikoo'
Cc: gluster-users@gluster.org
Subject: Re: [Gluster-users] NFS Mounted GlusterFS, secondary groups  
not working


Thanks, that looks to be the same issue. I'll add my comments to the  
bug.


Mike

-Original Message-
From: Shehjar Tikoo [mailto:shehj...@gluster.com]
Sent: Thursday, November 11, 2010 12:51 PM
To: Mike Hanby
Cc: gluster-users@gluster.org
Subject: Re: [Gluster-users] NFS Mounted GlusterFS, secondary groups  
not working


Hi,

It might be related to a bug filed at:
http://bugs.gluster.com/cgi-bin/bugzilla3/show_bug.cgi?id=2045

If you please update it there or file a new one, I'll take a look.  
Thanks.


- Original Message -

From: "Mike Hanby" 
To: gluster-users@gluster.org
Sent: Friday, November 12, 2010 12:00:23 AM
Subject: [Gluster-users] NFS Mounted GlusterFS, secondary groups  
not working

Howdy,

I have a GlusterFS 3.1 volume being mounted on a client using NFS.
From the client I created a directory under the mount point and set
the permissions to root:groupa 750

My user account is a member of groupa on the client, yet I am unable
to list the contents of the directory:

$ ls -l /gfs/dir1
ls: /gfs/dir1/: Permission denied

$ ls -ld /gfs/dir1
rwxr-x--- 9 root groupa 73728 Nov 9 09:44 /gfs/dir1/

$ groups
myuser groupa

I am able to list the directory as the user root. If I change the
group ownership to my primary group, myuser, then I can successfully
list the contents of the directory.

$ sudo chgrp myuser /gfs/dir1
$ ls -ld /gfs/dir1
rwxr-x--- 9 root myuser 73728 Nov 9 09:44 /gfs/dir1/

$ ls -l /gfs/dir1
drwxr-xr-x 5 root root 73728 Mar 26 2010 testdir1
drwxr-x--- 4 root root 73728 Apr 8 2010 testdir2
drwxr-x--- 2 root root 73728 Aug 4 21:23 testdir3

The volume is being exported using the builtin GlusterFS NFS server.
The servers and client are all CentOS 5.5 x86_64 boxes.

Thanks for any suggestions,

Mike

=
Mike Hanby
mha...@uab.edu
UAB School of Engineering
Information Systems Specialist II
IT HPCS / Research Computing


___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users

___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


[Gluster-users] GlusterFS - Abysmal Speed on Fresh Install

2010-12-03 Thread Christopher Michaelis

Hi there,

I'm new to the GlusterFS project - looks superb and very promising! I'm 
running into massive performance issues, however. I'm using the stock 
default configuration that GlusterFS put in place when I created the 
volume - it seems to reference io-cache, quick-read, etc in the volume 
files, which makes me think it's already pulling in these optimizations.


This is a replicate setup - I've tried with 2, 3, and 4 nodes, and 
performance remains awful on any of them - network communication seems 
fine, with average speeds at 9-15MB/sec.


# gluster --version
glusterfs 3.1.1 built on Nov 29 2010 10:07:45
Repository revision: v3.1.1

With no other activity on the filesystems on any of my nodes:
# time ls -al
total 10072
drwx--x--x 10 root   root   4096 Dec  3 16:50 ./
drwxr-xr-x  3 root   root   4096 Dec  3 14:19 ../
-rw-r--r--  1 root   root   1024 Dec  3 16:49 testfile

real0m1.347s
user0m0.000s
sys0m0.000s

# time cp testfile testfile2

real0m11.254s
user0m0.000s
sys0m0.000s
# time diff testfile*

real0m5.792s
user0m0.004s
sys0m0.000s

Read speed is marginally faster than write speed, but still horrible - 
e.g. if Apache is serving content off of a glusterfs mountpoint, it 
times out 95% of the time before it can read the files. I'm using mount 
-t glusterfs, with default mount options.


Can anyone point me in the right direction to getting things nice and 
speedy here? I'd appreciate any feedback or help! I can provide any 
configuration files necessary, or even root login access to the box(es) 
via private e-mail if you want to poke around (these are just test boxes 
presently).


Thanks,
--Chris
chris.michae...@uk2group.com
___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


Re: [Gluster-users] NFS Mounted GlusterFS, secondary groups not working

2010-12-03 Thread Mike Hanby
The release notes for 3.1.1 don't mention bug 2045, however the bug itself 
indicates that it's fixed with a target of 3.1.1

did it actually make it into the build?

-Original Message-
From: gluster-users-boun...@gluster.org 
[mailto:gluster-users-boun...@gluster.org] On Behalf Of Mike Hanby
Sent: Friday, December 03, 2010 3:02 PM
To: gluster-users@gluster.org
Subject: Re: [Gluster-users] NFS Mounted GlusterFS, secondary groups not working

After upgrading my servers to 3.1.1 I'm still experiencing the issue with 
secondary groups not working with the client using NFS:

In this scenario root owns the directory and the admins non primary group (of 
which my id is a member) is the group:

$ groups
mikeh admins

$ ls -ld /research/project/
drwxrwx--- 10 root admins 73728 Nov 22 12:13 /gluster/project/

$ ls -l /gluster/project/
ls: /gluster/project/: Permission denied

$ mount|grep research
nas-01.mydom:/research-storage on /gluster type nfs (rw,addr=192.168.1.11)

Both Gluster servers and the client are CentOS 5 x86_64 systems. The servers 
had glusterfs-core / glusterfs-fuse 3.0.1-1 updated to 3.1.1-1 via the RPMs. 
The storage is a simple distributed file system, no striping or replication.

Following the update I rebooted both servers and remounted the NFS file system 
on the client.

I'll add this info to the bug as well.

Mike

-Original Message-
From: gluster-users-boun...@gluster.org 
[mailto:gluster-users-boun...@gluster.org] On Behalf Of Mike Hanby
Sent: Thursday, November 11, 2010 1:53 PM
To: 'Shehjar Tikoo'
Cc: gluster-users@gluster.org
Subject: Re: [Gluster-users] NFS Mounted GlusterFS, secondary groups not working

Thanks, that looks to be the same issue. I'll add my comments to the bug.

Mike

-Original Message-
From: Shehjar Tikoo [mailto:shehj...@gluster.com] 
Sent: Thursday, November 11, 2010 12:51 PM
To: Mike Hanby
Cc: gluster-users@gluster.org
Subject: Re: [Gluster-users] NFS Mounted GlusterFS, secondary groups not working

Hi,

It might be related to a bug filed at:
http://bugs.gluster.com/cgi-bin/bugzilla3/show_bug.cgi?id=2045

If you please update it there or file a new one, I'll take a look. Thanks.

- Original Message -
> From: "Mike Hanby" 
> To: gluster-users@gluster.org
> Sent: Friday, November 12, 2010 12:00:23 AM
> Subject: [Gluster-users] NFS Mounted GlusterFS, secondary groups not working
> Howdy,
> 
> I have a GlusterFS 3.1 volume being mounted on a client using NFS.
> From the client I created a directory under the mount point and set
> the permissions to root:groupa 750
> 
> My user account is a member of groupa on the client, yet I am unable
> to list the contents of the directory:
> 
> $ ls -l /gfs/dir1
> ls: /gfs/dir1/: Permission denied
> 
> $ ls -ld /gfs/dir1
> rwxr-x--- 9 root groupa 73728 Nov 9 09:44 /gfs/dir1/
> 
> $ groups
> myuser groupa
> 
> I am able to list the directory as the user root. If I change the
> group ownership to my primary group, myuser, then I can successfully
> list the contents of the directory.
> 
> $ sudo chgrp myuser /gfs/dir1
> $ ls -ld /gfs/dir1
> rwxr-x--- 9 root myuser 73728 Nov 9 09:44 /gfs/dir1/
> 
> $ ls -l /gfs/dir1
> drwxr-xr-x 5 root root 73728 Mar 26 2010 testdir1
> drwxr-x--- 4 root root 73728 Apr 8 2010 testdir2
> drwxr-x--- 2 root root 73728 Aug 4 21:23 testdir3
> 
> The volume is being exported using the builtin GlusterFS NFS server.
> The servers and client are all CentOS 5.5 x86_64 boxes.
> 
> Thanks for any suggestions,
> 
> Mike
> 
> =
> Mike Hanby
> mha...@uab.edu
> UAB School of Engineering
> Information Systems Specialist II
> IT HPCS / Research Computing
> 
> 
> ___
> Gluster-users mailing list
> Gluster-users@gluster.org
> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


Re: [Gluster-users] NFS Mounted GlusterFS, secondary groups not working

2010-12-03 Thread Mike Hanby
After upgrading my servers to 3.1.1 I'm still experiencing the issue with 
secondary groups not working with the client using NFS:

In this scenario root owns the directory and the admins non primary group (of 
which my id is a member) is the group:

$ groups
mikeh admins

$ ls -ld /research/project/
drwxrwx--- 10 root admins 73728 Nov 22 12:13 /gluster/project/

$ ls -l /gluster/project/
ls: /gluster/project/: Permission denied

$ mount|grep research
nas-01.mydom:/research-storage on /gluster type nfs (rw,addr=192.168.1.11)

Both Gluster servers and the client are CentOS 5 x86_64 systems. The servers 
had glusterfs-core / glusterfs-fuse 3.0.1-1 updated to 3.1.1-1 via the RPMs. 
The storage is a simple distributed file system, no striping or replication.

Following the update I rebooted both servers and remounted the NFS file system 
on the client.

I'll add this info to the bug as well.

Mike

-Original Message-
From: gluster-users-boun...@gluster.org 
[mailto:gluster-users-boun...@gluster.org] On Behalf Of Mike Hanby
Sent: Thursday, November 11, 2010 1:53 PM
To: 'Shehjar Tikoo'
Cc: gluster-users@gluster.org
Subject: Re: [Gluster-users] NFS Mounted GlusterFS, secondary groups not working

Thanks, that looks to be the same issue. I'll add my comments to the bug.

Mike

-Original Message-
From: Shehjar Tikoo [mailto:shehj...@gluster.com] 
Sent: Thursday, November 11, 2010 12:51 PM
To: Mike Hanby
Cc: gluster-users@gluster.org
Subject: Re: [Gluster-users] NFS Mounted GlusterFS, secondary groups not working

Hi,

It might be related to a bug filed at:
http://bugs.gluster.com/cgi-bin/bugzilla3/show_bug.cgi?id=2045

If you please update it there or file a new one, I'll take a look. Thanks.

- Original Message -
> From: "Mike Hanby" 
> To: gluster-users@gluster.org
> Sent: Friday, November 12, 2010 12:00:23 AM
> Subject: [Gluster-users] NFS Mounted GlusterFS, secondary groups not working
> Howdy,
> 
> I have a GlusterFS 3.1 volume being mounted on a client using NFS.
> From the client I created a directory under the mount point and set
> the permissions to root:groupa 750
> 
> My user account is a member of groupa on the client, yet I am unable
> to list the contents of the directory:
> 
> $ ls -l /gfs/dir1
> ls: /gfs/dir1/: Permission denied
> 
> $ ls -ld /gfs/dir1
> rwxr-x--- 9 root groupa 73728 Nov 9 09:44 /gfs/dir1/
> 
> $ groups
> myuser groupa
> 
> I am able to list the directory as the user root. If I change the
> group ownership to my primary group, myuser, then I can successfully
> list the contents of the directory.
> 
> $ sudo chgrp myuser /gfs/dir1
> $ ls -ld /gfs/dir1
> rwxr-x--- 9 root myuser 73728 Nov 9 09:44 /gfs/dir1/
> 
> $ ls -l /gfs/dir1
> drwxr-xr-x 5 root root 73728 Mar 26 2010 testdir1
> drwxr-x--- 4 root root 73728 Apr 8 2010 testdir2
> drwxr-x--- 2 root root 73728 Aug 4 21:23 testdir3
> 
> The volume is being exported using the builtin GlusterFS NFS server.
> The servers and client are all CentOS 5.5 x86_64 boxes.
> 
> Thanks for any suggestions,
> 
> Mike
> 
> =
> Mike Hanby
> mha...@uab.edu
> UAB School of Engineering
> Information Systems Specialist II
> IT HPCS / Research Computing
> 
> 
> ___
> Gluster-users mailing list
> Gluster-users@gluster.org
> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


[Gluster-users] 3.1.1 crashing under moderate load

2010-12-03 Thread Lana Deere
I'm running GlusterFS 3.1.1, CentOS5.5 servers, CentOS5.4 clients, RDMA
transport, native/fuse access.

I have a directory which is shared on the gluster.  In fact, it is a clone
of /lib from one of the clients, shared so all can see it.

I have a script which does
find lib -type f -print0 | xargs -0 sum | md5sum

If I run this on my clients one at a time, they all yield the same md5sum:
for h in <>; do ssh $host script; done

If I run this on my clients concurrently, up to roughly 25 at a time they
still yield the same md5sum.
for h in <>; do ssh $host script& done

Beyond that the gluster share often, but not always, fails.  The errors vary.
- sometimes I get "sum: xxx.so not found"
- sometimes I get the wrong checksum without any error message
- sometimes the job simply hangs until I kill it


Some of the server logs show messages like these from the time of the
failures (other servers show nothing from around that time):

[2010-12-03 10:03:06.34328] E [rdma.c:4442:rdma_event_handler]
rpc-transport/rdma: rdma.RaidData-server: pollin received on tcp
socket (peer: 10.54.255.240:1022) after handshake is complete
[2010-12-03 10:03:06.34363] E [rpcsvc.c:1548:rpcsvc_submit_generic]
rpc-service: failed to submit message (XID: 0x55e82, Program:
GlusterFS-3.1.0, ProgVers: 310, Proc: 12) to rpc-transport
(rdma.RaidData-server)
[2010-12-03 10:03:06.34377] E [server.c:137:server_submit_reply] :
Reply submission failed
[2010-12-03 10:03:06.34464] E [rpcsvc.c:1548:rpcsvc_submit_generic]
rpc-service: failed to submit message (XID: 0x55e83, Program:
GlusterFS-3.1.0, ProgVers: 310, Proc: 12) to rpc-transport
(rdma.RaidData-server)
[2010-12-03 10:03:06.34520] E [server.c:137:server_submit_reply] :
Reply submission failed


On a client which had a failure I see messages like:

[2010-12-03 10:03:06.21290] E [rdma.c:4442:rdma_event_handler]
rpc-transport/rdma: RaidData-client-1: pollin received on tcp socket
(peer: 10.54.50.101:24009) after handshake is complete
[2010-12-03 10:03:06.21776] E [rpc-clnt.c:338:saved_frames_unwind]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_notify+0xb9) [0x3814a0f769]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_connection_cleanup+0x7e)
[0x3814a0ef1e] (-->/usr/lib64/libgfrpc.so.0(saved_frames_destroy+0xe)
[0x3814a0ee8e]))) rpc-clnt: forced unwinding frame type(GlusterFS 3.1)
op(READ(12)) called at 2010-12-03 10:03:06.20492
[2010-12-03 10:03:06.21821] E [rpc-clnt.c:338:saved_frames_unwind]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_notify+0xb9) [0x3814a0f769]
(-->/usr/lib64/libgfrpc.so.0(rpc_clnt_connection_cleanup+0x7e)
[0x3814a0ef1e] (-->/usr/lib64/libgfrpc.so.0(saved_frames_destroy+0xe)
[0x3814a0ee8e]))) rpc-clnt: forced unwinding frame type(GlusterFS 3.1)
op(READ(12)) called at 2010-12-03 10:03:06.20529
[2010-12-03 10:03:06.26827] I
[client-handshake.c:993:select_server_supported_programs]
RaidData-client-1: Using Program GlusterFS-3.1.0, Num (1298437),
Version (310)
[2010-12-03 10:03:06.27029] I
[client-handshake.c:829:client_setvolume_cbk] RaidData-client-1:
Connected to 10.54.50.101:24009, attached to remote volume '/data'.
[2010-12-03 10:03:06.27067] I
[client-handshake.c:698:client_post_handshake] RaidData-client-1: 2
fds open - Delaying child_up until they are re-opened


Anyone else seen anything like this and/or have suggestions about options I can
set to work around this?


.. Lana (lana.de...@gmail.com)
___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


Re: [Gluster-users] Who's using Fedora in production on Glusterfs storage servers?

2010-12-03 Thread Burnash, James
Thanks very much Mark.

I am going to put up a prototype using Fedora - being able to test btrfs as 
well is a nice plus.

For now, we apparently aren't going to be able to use Fedora in production 
because HP doesn't support it for their command line hardware reporting and 
configuration utilities (hpasmcli, hpacucli, etc).

James Burnash, Unix Engineering


-Original Message-
From: gluster-users-boun...@gluster.org 
[mailto:gluster-users-boun...@gluster.org] On Behalf Of Mark "Naoki" Rogers
Sent: Wednesday, December 01, 2010 10:20 PM
To: gluster-users@gluster.org
Subject: Re: [Gluster-users] Who's using Fedora in production on Glusterfs 
storage servers?

Hi James,

I'm using 3.1.1 on six bricks in dist+replicate all running F14+BTRFS,
the clients are on fedora12/13/14. I build the RPMs from source on a F14
machine. The cluster is running entirely on GbE (with some 10Gb lines
going in shortly), no RDMA/infiniband so I can't help there.

It's gone through a series of looped benchmarks for a while now (from
3.1.0 through a few qa releases) and have so far pushed/pulled over
110TB through it - I'm happy in the stability but not /entirely/ sure of
the performance just yet, just started up more testing under 3.1.1.

But back to your main question there really isn't enough difference
between the near-term releases of Fedora for it to make a huge
difference either way. I do think you're better off using the latest
Fedora release than an older one that will be end of life soon (f12
tomorrow). Being able to patch/maintain your system is more important
than an, often very arbitrary, vendor support list which is usually just
an outcome of what people have had time to look into, rather than any
measured reason a newer OS isn't supported. Besides the only thing you
ever have to /really/ care about is the kernel and glibc major versions,
so if it compiles you're pretty much ok (ldd it, that's all it needs).


On 12/02/2010 01:45 AM, Burnash, James wrote:
> How many people on the list are using Fedora 12 (or 13) in production for 
> Glusterfs storage servers? I know that Gluster Platform uses Fedora 12 as its 
> OS - I was thinking of building my new glusterfs storage servers using 
> Fedora, and was wondering whether Fedora 13 was tested by Gluster for v 3.1.1 
> and what other people's experiences were.
>
> One of the reasons for my interest was so that I could use ext4 as the 
> backend file store, instead of ext3.
>
> Thanks,
>
> James Burnash, Unix Engineering

___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


DISCLAIMER:
This e-mail, and any attachments thereto, is intended only for use by the 
addressee(s) named herein and may contain legally privileged and/or 
confidential information. If you are not the intended recipient of this e-mail, 
you are hereby notified that any dissemination, distribution or copying of this 
e-mail, and any attachments thereto, is strictly prohibited. If you have 
received this in error, please immediately notify me and permanently delete the 
original and any copy of any e-mail and any printout thereof. E-mail 
transmission cannot be guaranteed to be secure or error-free. The sender 
therefore does not accept liability for any errors or omissions in the contents 
of this message which arise as a result of e-mail transmission.
NOTICE REGARDING PRIVACY AND CONFIDENTIALITY Knight Capital Group may, at its 
discretion, monitor and review the content of all e-mail communications. 
http://www.knight.com
___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


[Gluster-users] Incomprehensive errors

2010-12-03 Thread Samuel Hassine
Hi there,

 

I have a lot (one each second) of these errors :

 

[2010-12-03 16:33:53.600610] W [fuse-bridge.c:2506:fuse_getxattr]
glusterfs-fuse: 248602: GETXATTR (null)/41872720 (security.capability)
(fuse_loc_fill() failed)

[2010-12-03 16:33:53.658075] W [fuse-bridge.c:2506:fuse_getxattr]
glusterfs-fuse: 248696: GETXATTR (null)/41872720 (security.capability)
(fuse_loc_fill() failed)

[2010-12-03 16:33:53.685461] W [fuse-bridge.c:2506:fuse_getxattr]
glusterfs-fuse: 248748: GETXATTR (null)/41872720 (security.capability)
(fuse_loc_fill() failed)

[2010-12-03 16:33:53.992316] W [fuse-bridge.c:2506:fuse_getxattr]
glusterfs-fuse: 248905: GETXATTR (null)/41872720 (security.capability)
(fuse_loc_fill() failed)

[2010-12-03 16:33:54.22034] W [fuse-bridge.c:2506:fuse_getxattr]
glusterfs-fuse: 248960: GETXATTR (null)/41872720 (security.capability)
(fuse_loc_fill() failed)

[2010-12-03 16:33:54.287419] W [fuse-bridge.c:2506:fuse_getxattr]
glusterfs-fuse: 249502: GETXATTR (null)/41872720 (security.capability)
(fuse_loc_fill() failed)

[2010-12-03 16:33:54.744890] W [fuse-bridge.c:2506:fuse_getxattr]
glusterfs-fuse: 249514: GETXATTR (null)/41872720 (security.capability)
(fuse_loc_fill() failed)

[2010-12-03 16:33:56.148194] W [fuse-bridge.c:2506:fuse_getxattr]
glusterfs-fuse: 250798: GETXATTR (null)/41872720 (security.capability)
(fuse_loc_fill() failed)

[2010-12-03 16:33:56.186352] W [fuse-bridge.c:2506:fuse_getxattr]
glusterfs-fuse: 250892: GETXATTR (null)/41872720 (security.capability)
(fuse_loc_fill() failed)

[2010-12-03 16:33:57.307090] W [fuse-bridge.c:2506:fuse_getxattr]
glusterfs-fuse: 251415: GETXATTR (null)/41872720 (security.capability)
(fuse_loc_fill() failed)

[2010-12-03 16:33:57.978952] W [fuse-bridge.c:2506:fuse_getxattr]
glusterfs-fuse: 252582: GETXATTR (null)/41872720 (security.capability)
(fuse_loc_fill() failed)

[2010-12-03 16:33:58.596328] W [fuse-bridge.c:2506:fuse_getxattr]
glusterfs-fuse: 252715: GETXATTR (null)/41872720 (security.capability)
(fuse_loc_fill() failed)

 

What does it mean?

 

Thanks for your answers.



Regards.

Samuel Hassine

___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


Re: [Gluster-users] RDMA Problems with GlusterFS 3.1.1

2010-12-03 Thread Jeremy Stout
That makes sense.

What software adjustments should I make to ensure there won't be any problems?

Thank you.

On Fri, Dec 3, 2010 at 7:53 AM, Raghavendra G  wrote:
> From the logs its evident that the reason for completion queue creation 
> failure is that the number of completion queue elements (in a completion 
> queue) we had requested in ibv_create_cq, (1024 * send_count) is less than 
> the maximum supported by the ib hardware (max_cqe = 131071).
>
> - Original Message -
> From: "Jeremy Stout" 
> To: "Raghavendra G" 
> Cc: gluster-users@gluster.org
> Sent: Friday, December 3, 2010 4:20:04 PM
> Subject: Re: [Gluster-users] RDMA Problems with GlusterFS 3.1.1
>
> I patched the source code and rebuilt GlusterFS. Here are the full logs:
> Server:
> [2010-12-03 07:08:55.945804] I [glusterd.c:275:init] management: Using
> /etc/glusterd as working directory
> [2010-12-03 07:08:55.947692] E [rdma.c:2047:rdma_create_cq]
> rpc-transport/rdma: max_mr_size = 18446744073709551615, max_cq =
> 65408, max_cqe = 131071, max_mr = 131056
> [2010-12-03 07:08:55.953226] E [rdma.c:2079:rdma_create_cq]
> rpc-transport/rdma: rdma.management: creation of send_cq failed
> [2010-12-03 07:08:55.953509] E [rdma.c:3785:rdma_get_device]
> rpc-transport/rdma: rdma.management: could not create CQ
> [2010-12-03 07:08:55.953582] E [rdma.c:3971:rdma_init]
> rpc-transport/rdma: could not create rdma device for mthca0
> [2010-12-03 07:08:55.953668] E [rdma.c:4803:init] rdma.management:
> Failed to initialize IB Device
> [2010-12-03 07:08:55.953691] E
> [rpc-transport.c:971:rpc_transport_load] rpc-transport: 'rdma'
> initialization failed
> [2010-12-03 07:08:55.953780] I [glusterd.c:96:glusterd_uuid_init]
> glusterd: generated UUID: 4eb47ca7-227c-49c4-97bd-25ac177b2f6a
> Given volfile:
> +--+
>  1: volume management
>  2:     type mgmt/glusterd
>  3:     option working-directory /etc/glusterd
>  4:     option transport-type socket,rdma
>  5:     option transport.socket.keepalive-time 10
>  6:     option transport.socket.keepalive-interval 2
>  7: end-volume
>  8:
>
> +--+
> [2010-12-03 07:09:10.244790] I
> [glusterd-handler.c:785:glusterd_handle_create_volume] glusterd:
> Received create volume req
> [2010-12-03 07:09:10.247646] I [glusterd-utils.c:232:glusterd_lock]
> glusterd: Cluster lock held by 4eb47ca7-227c-49c4-97bd-25ac177b2f6a
> [2010-12-03 07:09:10.247678] I
> [glusterd-handler.c:2835:glusterd_op_txn_begin] glusterd: Acquired
> local lock
> [2010-12-03 07:09:10.247708] I
> [glusterd3_1-mops.c:1091:glusterd3_1_cluster_lock] glusterd: Sent lock
> req to 0 peers
> [2010-12-03 07:09:10.248038] I
> [glusterd3_1-mops.c:1233:glusterd3_1_stage_op] glusterd: Sent op req
> to 0 peers
> [2010-12-03 07:09:10.251970] I
> [glusterd3_1-mops.c:1323:glusterd3_1_commit_op] glusterd: Sent op req
> to 0 peers
> [2010-12-03 07:09:10.252020] I
> [glusterd3_1-mops.c:1145:glusterd3_1_cluster_unlock] glusterd: Sent
> unlock req to 0 peers
> [2010-12-03 07:09:10.252036] I
> [glusterd-op-sm.c:4738:glusterd_op_txn_complete] glusterd: Cleared
> local lock
> [2010-12-03 07:09:22.11649] I
> [glusterd-handler.c:936:glusterd_handle_cli_start_volume] glusterd:
> Received start vol reqfor volume testdir
> [2010-12-03 07:09:22.11724] I [glusterd-utils.c:232:glusterd_lock]
> glusterd: Cluster lock held by 4eb47ca7-227c-49c4-97bd-25ac177b2f6a
> [2010-12-03 07:09:22.11734] I
> [glusterd-handler.c:2835:glusterd_op_txn_begin] glusterd: Acquired
> local lock
> [2010-12-03 07:09:22.11761] I
> [glusterd3_1-mops.c:1091:glusterd3_1_cluster_lock] glusterd: Sent lock
> req to 0 peers
> [2010-12-03 07:09:22.12120] I
> [glusterd3_1-mops.c:1233:glusterd3_1_stage_op] glusterd: Sent op req
> to 0 peers
> [2010-12-03 07:09:22.184403] I
> [glusterd-utils.c:971:glusterd_volume_start_glusterfs] : About to
> start glusterfs for brick pgh-submit-1:/mnt/gluster
> [2010-12-03 07:09:22.229143] I
> [glusterd3_1-mops.c:1323:glusterd3_1_commit_op] glusterd: Sent op req
> to 0 peers
> [2010-12-03 07:09:22.229198] I
> [glusterd3_1-mops.c:1145:glusterd3_1_cluster_unlock] glusterd: Sent
> unlock req to 0 peers
> [2010-12-03 07:09:22.229218] I
> [glusterd-op-sm.c:4738:glusterd_op_txn_complete] glusterd: Cleared
> local lock
> [2010-12-03 07:09:22.240157] I
> [glusterd-pmap.c:281:pmap_registry_remove] pmap: removing brick (null)
> on port 24009
>
>
> Client:
> [2010-12-03 07:09:00.82784] W [io-stats.c:1644:init] testdir: dangling
> volume. check volfile
> [2010-12-03 07:09:00.82824] W [dict.c:1204:data_to_str] dict: @data=(nil)
> [2010-12-03 07:09:00.82836] W [dict.c:1204:data_to_str] dict: @data=(nil)
> [2010-12-03 07:09:00.85980] E [rdma.c:2047:rdma_create_cq]
> rpc-transport/rdma: max_mr_size = 18446744073709551615, max_cq =
> 65408, max_cqe = 131071, max_mr = 131056
> [2010-12-03 07:09:00.92883] E [rdma.c:2079:rdma_create_cq]
> rpc-transp

Re: [Gluster-users] performance stops at 1Gb

2010-12-03 Thread Anand Avati
Do both read and write throughput peak at 1Gbit/s? What is the block size
used for performing I/O? Can you get the output of -

1. dd if=/dev/zero of=/mnt/stripe/file bs=1M count=1K

2. dd if=/mnt/stripe/file of=/dev/null bs=1M count=1K

Just one instance of dd is enough as the client network interface (10Gbit/s)
has enough juice to saturate 4x1Gbit servers.

Avati

On Fri, Dec 3, 2010 at 6:06 PM, Gotwalt, P.  wrote:

> Craig,
>
> Using multiple parallel bonnie++ benchmarks (4,8,16) does use several
> files. These file are 1GB each, and we take care there will be at least
> 32 of them. As we have multiple processes (4,8,16 bonnie++s) and each
> uses several files, we spread the io over different storage bricks. I
> can see this when monitoring network and disk activity on the bricks.
> For example: when bonnie++ does block read/writes on a striped (4
> bricks) volume I notice that the load of the client (network throughput)
> is evenly spread over the 4 nodes. These nodes have enough cpu, memory,
> network and disk resources left! The accoumulated throughput doesn't get
> over the 1 Gb.
> The 10Gb nic at the client is set to fixed 10Gb, full duplex, All the
> nics on the storage bricks are 1Gb, fixed, full duplex. The 10Gb client
> (dual quadcore, 16GB) has plenty of resources to run 16 bonnie++s
> parallel. We should be able to get more than this 1Gb throughput,
> especially with a striped volume.
>
> What kind of benchmarks do you run? And with what kind of setup?
>
> Peter
>
>
>
> > Peter -
> > Using Gluster the performance of any single file is going to be
> > limited to the performance of the server on which it exists, or in the
>
> > case of a striped volume of the server on which the segment of the
> file
> > you are accessing exists. If you were able to start 4 processes,
> > accessing different parts of the striped file, or lots of different
> > files in a distribute cluster you would see your performance increase
> > significantly.
>
> > Thanks,
>
> > Craig
>
> > -->
> > Craig Carl
> > Senior Systems Engineer
> > Gluster
> >
> >
> > On 11/26/2010 07:57 AM, Gotwalt, P. wrote:
> > > Hi All,
> > >
> > > I am doing some tests with gluster (3.1) and have a problem of not
> > > getting higher throughput than 1 Gb (yes bit!) with 4 storage
> bricks.
> > > My setup:
> > >
> > > 4 storage bricks (dualcore, 4GB mem) each with 3 sata 1Tb disks,
> > > connected to a switch with 1 Gb nics.  In my tests I only use 1 SATA
> > > disk as a volume, per brick.
> > > 1 client (2xquad core, 16 GB mem) with a 10Gb nic to the same switch
> as
> > > the bricks.
> > >
> > > When using striped of distributed configurations, with all 4 bricks
> > > configured to act as a server, the performance will never be higher
> than
> > > just below 1 Gb! I tested with 4, 8 and 16 parallel bonnie++ runs.
> > >
> > > The idea is that parallel bonnie's create enough files to get
> > > distributed over the storage bricks. And all this bonnie's wil
> deliver
> > > enough throughput to fill up this 10Gb line. I expect the throughput
> to
> > > be maximum 4Gb because that's the maximum the 4 storage bricks
> together
> > > can produce.
> > >
> > > I also tested the throughput of the network with iperf3 and got:
> > > - 5Gb to a second temporary client on another switch 200 Km from my
> > > site, connected with a 5Gb fiber
> > > - 908-920 Mb to the interfaces of the bricks.
> > > So the network seems ok.
> > >
> > > Can someone advise me on why I don't get 4Gb? Or can someone advise
> me
> > > on a better setup with the equipment I have?
> > >
> > >
> > > Peter Gotwalt
> ___
> Gluster-users mailing list
> Gluster-users@gluster.org
> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
>
___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


Re: [Gluster-users] RDMA Problems with GlusterFS 3.1.1

2010-12-03 Thread Raghavendra G
From the logs its evident that the reason for completion queue creation failure 
is that the number of completion queue elements (in a completion queue) we had 
requested in ibv_create_cq, (1024 * send_count) is less than the maximum 
supported by the ib hardware (max_cqe = 131071).

- Original Message -
From: "Jeremy Stout" 
To: "Raghavendra G" 
Cc: gluster-users@gluster.org
Sent: Friday, December 3, 2010 4:20:04 PM
Subject: Re: [Gluster-users] RDMA Problems with GlusterFS 3.1.1

I patched the source code and rebuilt GlusterFS. Here are the full logs:
Server:
[2010-12-03 07:08:55.945804] I [glusterd.c:275:init] management: Using
/etc/glusterd as working directory
[2010-12-03 07:08:55.947692] E [rdma.c:2047:rdma_create_cq]
rpc-transport/rdma: max_mr_size = 18446744073709551615, max_cq =
65408, max_cqe = 131071, max_mr = 131056
[2010-12-03 07:08:55.953226] E [rdma.c:2079:rdma_create_cq]
rpc-transport/rdma: rdma.management: creation of send_cq failed
[2010-12-03 07:08:55.953509] E [rdma.c:3785:rdma_get_device]
rpc-transport/rdma: rdma.management: could not create CQ
[2010-12-03 07:08:55.953582] E [rdma.c:3971:rdma_init]
rpc-transport/rdma: could not create rdma device for mthca0
[2010-12-03 07:08:55.953668] E [rdma.c:4803:init] rdma.management:
Failed to initialize IB Device
[2010-12-03 07:08:55.953691] E
[rpc-transport.c:971:rpc_transport_load] rpc-transport: 'rdma'
initialization failed
[2010-12-03 07:08:55.953780] I [glusterd.c:96:glusterd_uuid_init]
glusterd: generated UUID: 4eb47ca7-227c-49c4-97bd-25ac177b2f6a
Given volfile:
+--+
  1: volume management
  2: type mgmt/glusterd
  3: option working-directory /etc/glusterd
  4: option transport-type socket,rdma
  5: option transport.socket.keepalive-time 10
  6: option transport.socket.keepalive-interval 2
  7: end-volume
  8:

+--+
[2010-12-03 07:09:10.244790] I
[glusterd-handler.c:785:glusterd_handle_create_volume] glusterd:
Received create volume req
[2010-12-03 07:09:10.247646] I [glusterd-utils.c:232:glusterd_lock]
glusterd: Cluster lock held by 4eb47ca7-227c-49c4-97bd-25ac177b2f6a
[2010-12-03 07:09:10.247678] I
[glusterd-handler.c:2835:glusterd_op_txn_begin] glusterd: Acquired
local lock
[2010-12-03 07:09:10.247708] I
[glusterd3_1-mops.c:1091:glusterd3_1_cluster_lock] glusterd: Sent lock
req to 0 peers
[2010-12-03 07:09:10.248038] I
[glusterd3_1-mops.c:1233:glusterd3_1_stage_op] glusterd: Sent op req
to 0 peers
[2010-12-03 07:09:10.251970] I
[glusterd3_1-mops.c:1323:glusterd3_1_commit_op] glusterd: Sent op req
to 0 peers
[2010-12-03 07:09:10.252020] I
[glusterd3_1-mops.c:1145:glusterd3_1_cluster_unlock] glusterd: Sent
unlock req to 0 peers
[2010-12-03 07:09:10.252036] I
[glusterd-op-sm.c:4738:glusterd_op_txn_complete] glusterd: Cleared
local lock
[2010-12-03 07:09:22.11649] I
[glusterd-handler.c:936:glusterd_handle_cli_start_volume] glusterd:
Received start vol reqfor volume testdir
[2010-12-03 07:09:22.11724] I [glusterd-utils.c:232:glusterd_lock]
glusterd: Cluster lock held by 4eb47ca7-227c-49c4-97bd-25ac177b2f6a
[2010-12-03 07:09:22.11734] I
[glusterd-handler.c:2835:glusterd_op_txn_begin] glusterd: Acquired
local lock
[2010-12-03 07:09:22.11761] I
[glusterd3_1-mops.c:1091:glusterd3_1_cluster_lock] glusterd: Sent lock
req to 0 peers
[2010-12-03 07:09:22.12120] I
[glusterd3_1-mops.c:1233:glusterd3_1_stage_op] glusterd: Sent op req
to 0 peers
[2010-12-03 07:09:22.184403] I
[glusterd-utils.c:971:glusterd_volume_start_glusterfs] : About to
start glusterfs for brick pgh-submit-1:/mnt/gluster
[2010-12-03 07:09:22.229143] I
[glusterd3_1-mops.c:1323:glusterd3_1_commit_op] glusterd: Sent op req
to 0 peers
[2010-12-03 07:09:22.229198] I
[glusterd3_1-mops.c:1145:glusterd3_1_cluster_unlock] glusterd: Sent
unlock req to 0 peers
[2010-12-03 07:09:22.229218] I
[glusterd-op-sm.c:4738:glusterd_op_txn_complete] glusterd: Cleared
local lock
[2010-12-03 07:09:22.240157] I
[glusterd-pmap.c:281:pmap_registry_remove] pmap: removing brick (null)
on port 24009


Client:
[2010-12-03 07:09:00.82784] W [io-stats.c:1644:init] testdir: dangling
volume. check volfile
[2010-12-03 07:09:00.82824] W [dict.c:1204:data_to_str] dict: @data=(nil)
[2010-12-03 07:09:00.82836] W [dict.c:1204:data_to_str] dict: @data=(nil)
[2010-12-03 07:09:00.85980] E [rdma.c:2047:rdma_create_cq]
rpc-transport/rdma: max_mr_size = 18446744073709551615, max_cq =
65408, max_cqe = 131071, max_mr = 131056
[2010-12-03 07:09:00.92883] E [rdma.c:2079:rdma_create_cq]
rpc-transport/rdma: testdir-client-0: creation of send_cq failed
[2010-12-03 07:09:00.93156] E [rdma.c:3785:rdma_get_device]
rpc-transport/rdma: testdir-client-0: could not create CQ
[2010-12-03 07:09:00.93224] E [rdma.c:3971:rdma_init]
rpc-transport/rdma: could not create rdma device for mthca0
[2010-12-03 07:09:00.93313] E [rdma.c:4803:init] testdir-client-0:
Failed to init

Re: [Gluster-users] performance stops at 1Gb

2010-12-03 Thread Gotwalt, P.
Craig,

Using multiple parallel bonnie++ benchmarks (4,8,16) does use several
files. These file are 1GB each, and we take care there will be at least
32 of them. As we have multiple processes (4,8,16 bonnie++s) and each
uses several files, we spread the io over different storage bricks. I
can see this when monitoring network and disk activity on the bricks.
For example: when bonnie++ does block read/writes on a striped (4
bricks) volume I notice that the load of the client (network throughput)
is evenly spread over the 4 nodes. These nodes have enough cpu, memory,
network and disk resources left! The accoumulated throughput doesn't get
over the 1 Gb.
The 10Gb nic at the client is set to fixed 10Gb, full duplex, All the
nics on the storage bricks are 1Gb, fixed, full duplex. The 10Gb client
(dual quadcore, 16GB) has plenty of resources to run 16 bonnie++s
parallel. We should be able to get more than this 1Gb throughput,
especially with a striped volume.

What kind of benchmarks do you run? And with what kind of setup?

Peter 



> Peter -
> Using Gluster the performance of any single file is going to be 
> limited to the performance of the server on which it exists, or in the

> case of a striped volume of the server on which the segment of the
file 
> you are accessing exists. If you were able to start 4 processes, 
> accessing different parts of the striped file, or lots of different 
> files in a distribute cluster you would see your performance increase 
> significantly.

> Thanks,

> Craig

> -->
> Craig Carl
> Senior Systems Engineer
> Gluster
> 
> 
> On 11/26/2010 07:57 AM, Gotwalt, P. wrote:
> > Hi All,
> >
> > I am doing some tests with gluster (3.1) and have a problem of not
> > getting higher throughput than 1 Gb (yes bit!) with 4 storage
bricks.
> > My setup:
> >
> > 4 storage bricks (dualcore, 4GB mem) each with 3 sata 1Tb disks,
> > connected to a switch with 1 Gb nics.  In my tests I only use 1 SATA
> > disk as a volume, per brick.
> > 1 client (2xquad core, 16 GB mem) with a 10Gb nic to the same switch
as
> > the bricks.
> >
> > When using striped of distributed configurations, with all 4 bricks
> > configured to act as a server, the performance will never be higher
than
> > just below 1 Gb! I tested with 4, 8 and 16 parallel bonnie++ runs.
> >
> > The idea is that parallel bonnie's create enough files to get
> > distributed over the storage bricks. And all this bonnie's wil
deliver
> > enough throughput to fill up this 10Gb line. I expect the throughput
to
> > be maximum 4Gb because that's the maximum the 4 storage bricks
together
> > can produce.
> >
> > I also tested the throughput of the network with iperf3 and got:
> > - 5Gb to a second temporary client on another switch 200 Km from my
> > site, connected with a 5Gb fiber
> > - 908-920 Mb to the interfaces of the bricks.
> > So the network seems ok.
> >
> > Can someone advise me on why I don't get 4Gb? Or can someone advise
me
> > on a better setup with the equipment I have?
> >
> >
> > Peter Gotwalt
___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


Re: [Gluster-users] Start a new volume with pre-existing directories

2010-12-03 Thread Daniel Zander

Hi!

>Can you send the output of -
>
>`gluster volume info all`
>`gluster peer status`
>
>from a gluster storage server and
>
>`mount` from the client?

Certainly

--
r...@ekpfs8:~# gluster volume info all
Volume Name: heal_me
Type: Distribute
Status: Started
Number of Bricks: 2
Transport-type: tcp
Bricks:
Brick1: 192.168.101.246:/storage/8
Brick2: 192.168.101.247:/storage/7
--
r...@ekpfs8:~# gluster peer status
Number of Peers: 1

Hostname: 192.168.101.247
Uuid: b36ce6e3-fa14-4d7e-bc4a-170a59a6f4f5
State: Peer in Cluster (Connected)
--
[r...@ekpbelle ~]# mount
[ ... ]
glusterfs#192.168.101.246:/heal_me on /storage/gluster type fuse 
(rw,allow_other,default_permissions,max_read=131072)

--

Regards,
Daniel
___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


Re: [Gluster-users] RDMA Problems with GlusterFS 3.1.1

2010-12-03 Thread Jeremy Stout
I patched the source code and rebuilt GlusterFS. Here are the full logs:
Server:
[2010-12-03 07:08:55.945804] I [glusterd.c:275:init] management: Using
/etc/glusterd as working directory
[2010-12-03 07:08:55.947692] E [rdma.c:2047:rdma_create_cq]
rpc-transport/rdma: max_mr_size = 18446744073709551615, max_cq =
65408, max_cqe = 131071, max_mr = 131056
[2010-12-03 07:08:55.953226] E [rdma.c:2079:rdma_create_cq]
rpc-transport/rdma: rdma.management: creation of send_cq failed
[2010-12-03 07:08:55.953509] E [rdma.c:3785:rdma_get_device]
rpc-transport/rdma: rdma.management: could not create CQ
[2010-12-03 07:08:55.953582] E [rdma.c:3971:rdma_init]
rpc-transport/rdma: could not create rdma device for mthca0
[2010-12-03 07:08:55.953668] E [rdma.c:4803:init] rdma.management:
Failed to initialize IB Device
[2010-12-03 07:08:55.953691] E
[rpc-transport.c:971:rpc_transport_load] rpc-transport: 'rdma'
initialization failed
[2010-12-03 07:08:55.953780] I [glusterd.c:96:glusterd_uuid_init]
glusterd: generated UUID: 4eb47ca7-227c-49c4-97bd-25ac177b2f6a
Given volfile:
+--+
  1: volume management
  2: type mgmt/glusterd
  3: option working-directory /etc/glusterd
  4: option transport-type socket,rdma
  5: option transport.socket.keepalive-time 10
  6: option transport.socket.keepalive-interval 2
  7: end-volume
  8:

+--+
[2010-12-03 07:09:10.244790] I
[glusterd-handler.c:785:glusterd_handle_create_volume] glusterd:
Received create volume req
[2010-12-03 07:09:10.247646] I [glusterd-utils.c:232:glusterd_lock]
glusterd: Cluster lock held by 4eb47ca7-227c-49c4-97bd-25ac177b2f6a
[2010-12-03 07:09:10.247678] I
[glusterd-handler.c:2835:glusterd_op_txn_begin] glusterd: Acquired
local lock
[2010-12-03 07:09:10.247708] I
[glusterd3_1-mops.c:1091:glusterd3_1_cluster_lock] glusterd: Sent lock
req to 0 peers
[2010-12-03 07:09:10.248038] I
[glusterd3_1-mops.c:1233:glusterd3_1_stage_op] glusterd: Sent op req
to 0 peers
[2010-12-03 07:09:10.251970] I
[glusterd3_1-mops.c:1323:glusterd3_1_commit_op] glusterd: Sent op req
to 0 peers
[2010-12-03 07:09:10.252020] I
[glusterd3_1-mops.c:1145:glusterd3_1_cluster_unlock] glusterd: Sent
unlock req to 0 peers
[2010-12-03 07:09:10.252036] I
[glusterd-op-sm.c:4738:glusterd_op_txn_complete] glusterd: Cleared
local lock
[2010-12-03 07:09:22.11649] I
[glusterd-handler.c:936:glusterd_handle_cli_start_volume] glusterd:
Received start vol reqfor volume testdir
[2010-12-03 07:09:22.11724] I [glusterd-utils.c:232:glusterd_lock]
glusterd: Cluster lock held by 4eb47ca7-227c-49c4-97bd-25ac177b2f6a
[2010-12-03 07:09:22.11734] I
[glusterd-handler.c:2835:glusterd_op_txn_begin] glusterd: Acquired
local lock
[2010-12-03 07:09:22.11761] I
[glusterd3_1-mops.c:1091:glusterd3_1_cluster_lock] glusterd: Sent lock
req to 0 peers
[2010-12-03 07:09:22.12120] I
[glusterd3_1-mops.c:1233:glusterd3_1_stage_op] glusterd: Sent op req
to 0 peers
[2010-12-03 07:09:22.184403] I
[glusterd-utils.c:971:glusterd_volume_start_glusterfs] : About to
start glusterfs for brick pgh-submit-1:/mnt/gluster
[2010-12-03 07:09:22.229143] I
[glusterd3_1-mops.c:1323:glusterd3_1_commit_op] glusterd: Sent op req
to 0 peers
[2010-12-03 07:09:22.229198] I
[glusterd3_1-mops.c:1145:glusterd3_1_cluster_unlock] glusterd: Sent
unlock req to 0 peers
[2010-12-03 07:09:22.229218] I
[glusterd-op-sm.c:4738:glusterd_op_txn_complete] glusterd: Cleared
local lock
[2010-12-03 07:09:22.240157] I
[glusterd-pmap.c:281:pmap_registry_remove] pmap: removing brick (null)
on port 24009


Client:
[2010-12-03 07:09:00.82784] W [io-stats.c:1644:init] testdir: dangling
volume. check volfile
[2010-12-03 07:09:00.82824] W [dict.c:1204:data_to_str] dict: @data=(nil)
[2010-12-03 07:09:00.82836] W [dict.c:1204:data_to_str] dict: @data=(nil)
[2010-12-03 07:09:00.85980] E [rdma.c:2047:rdma_create_cq]
rpc-transport/rdma: max_mr_size = 18446744073709551615, max_cq =
65408, max_cqe = 131071, max_mr = 131056
[2010-12-03 07:09:00.92883] E [rdma.c:2079:rdma_create_cq]
rpc-transport/rdma: testdir-client-0: creation of send_cq failed
[2010-12-03 07:09:00.93156] E [rdma.c:3785:rdma_get_device]
rpc-transport/rdma: testdir-client-0: could not create CQ
[2010-12-03 07:09:00.93224] E [rdma.c:3971:rdma_init]
rpc-transport/rdma: could not create rdma device for mthca0
[2010-12-03 07:09:00.93313] E [rdma.c:4803:init] testdir-client-0:
Failed to initialize IB Device
[2010-12-03 07:09:00.93332] E [rpc-transport.c:971:rpc_transport_load]
rpc-transport: 'rdma' initialization failed
Given volfile:
+--+
  1: volume testdir-client-0
  2: type protocol/client
  3: option remote-host submit-1
  4: option remote-subvolume /mnt/gluster
  5: option transport-type rdma
  6: end-volume
  7:
  8: volume testdir-write-behind
  9: type performance/write-be

Re: [Gluster-users] Anormal Gluster shutdown

2010-12-03 Thread Anand Avati
Release 3.0.x has the same behavior what you describe here, which is not
really an issue. Please verify if some other component in your system
changed (kernel or fuse kernel module) which has changed the behavior for
you.

Avati

On Thu, Dec 2, 2010 at 7:29 PM, Samuel Hassine  wrote:

> Hi all,
>
>
>
> GlusterFS partition automatically shutdown when umounting a binded mount
> point
>
> with "-f" option (without it works).
>
>
>
> How to reproduce:
>
>
>
> mounted Gluster partition on /gluster (any config):
>
>
>
> df: localhost:/gluster4.5T  100G  4.4T   3% /gluster
>
> mount: localhost:/gluster on /gluster type fuse.glusterfs
>
> (rw,allow_other,default_permissions,max_read=131072)
>
>
>
> commands:
>
>
>
> mkdir /test
>
> mount -n --bind /gluster /test
>
> ls /test (verify you have the Gluster)
>
>
>
> and:
>
>
>
> umount -f /test
>
>
>
> ===>
>
>
>
> df: `/gluster': Transport endpoint is not connected
>
> [2010-12-02 14:48:56.38309] I [fuse-bridge.c:3138:fuse_thread_proc] fuse:
>
> unmounting /gluster
>
> [2010-12-02 14:48:56.38364] I [glusterfsd.c:672:cleanup_and_exit]
> glusterfsd:
>
> shutting down
>
>
>
> Before 3.1.x I did not have this bug.
>
>
>
> Regards.
>
> Sam
>
>
> ___
> Gluster-users mailing list
> Gluster-users@gluster.org
> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
>
>
___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


Re: [Gluster-users] Start a new volume with pre-existing directories

2010-12-03 Thread Craig Carl

Can you send the output of -

`gluster volume info all`
`gluster peer status`

from a gluster storage server and

`mount` from the client?

Craig


On 12/03/2010 02:50 AM, Daniel Zander wrote:

Dear Craig,

I'm afraid I wasn't able to start a self heal the way you suggested. I 
tested the following:


--
WHAT I DID
--
I created on fs7:/storage/7
user_7_1  user_7_2  user_7_3

and on fs8:/storage/8
user_8_1  user_8_2  user_8_3

and filled all of the directories with some small files and 
subdirectories.


Then, on fs8:
gluster volume create heal_me transport tcp 192.168.101.246:/storage/8 
192.168.101.247:/storage/7


Then on fs8 and afterwards on fs7:
mount -t glusterfs localhost:/heal_me /tempmount/
cd /tempmount
find . | xargs stat >>/dev/null 2>&1
umount /tempmount

All went well, no error messages or anything. The output of `find . | 
xargs stat` is probably too long to post it here, but there were no 
error messages or anything else that would look suspicious to me.


---
RESULTS
---
ls fs8:storage/8
user_8_1  user_8_2  user_8_3

ls fs7:/storage/7
user_7_1  user_7_2  user_7_3  user_8_1  user_8_2  user_8_3

ls client:/storage/gluster
user_8_1  user_8_2  user_8_3

ls fs7:/tempmount
user_8_1  user_8_2  user_8_3

ls fs8:/tempmount
user_8_1  user_8_2  user_8_3

Unmounting and remounting has no effect.

Servers are both Ubuntu Server 10.4, client is CentOS 5, 64bits all 
around.


Thanks and regards,
Daniel


On 12/03/2010 10:10 AM, Craig Carl wrote:

Daniel -
If you want to export existing data you will need to run the self heal
process so extended attributes can get written. While this should work
without any issues it isn't an officially supported process, please make
sure you have complete and up to date backups.

After you have setup and started the Gluster volume mount it locally on
one of the servers using `mount -t glusterfs localhost:/ /`. CD into the root of the mount point and run `find . |
xargs stat >>/dev/null 2>&1` to start a self heal.

Also the command you used to create the volume should not have worked,
it is missing a volume name - gluster volume create  transport
tcp fs7:/storage/7, fs8:/storage/8, typo maybe?

Please let us know how it goes, and please let me know if you have any
other questions.

Thanks,

Craig

-->
Craig Carl
Senior Systems Engineer; Gluster, Inc.
Cell - (408) 829-9953 (California, USA)
Office - (408) 770-1884
Gtalk - craig.c...@gmail.com
Twitter - @gluster
http://rackerhacker.com/2010/08/11/one-month-with-glusterfs-in-production/ 






On 12/02/2010 11:38 PM, Daniel Zander wrote:

Dear all,

at our institute, we currently have 6 file servers, each one of them
individually mounted via NFS on ~ 20 clients. The structure on the
servers and the clients is the following:

/storage/1/ (NFS export from FS1)
/storage/2/ (NFS export from FS2)
etc ...

Recently, we decided that we would like to migrate this to glusterFS,
so that we can have one big storage directory on the clients. Let's
call it

/gluster/

I tried to set up a gluster volume with two empty fileservers and it
worked without any problems. I could easily mount it on a client and
use it (using the native glusterFS mount).

If we now want to migrate the entire institute, it would be very
convenient, if existing folders could be easily included into a new
volume. I tried to do this, but I did not succeed.

Here's a short description of what I tried:

Existing folders:
on fs7: /storage/7/user_1,user_2
on fs8: /storage/8/user_3,user_4

gluster volume create transport tcp fs7:/storage/7, fs8:/storage/8

I hoped to see on the client:
/gluster/user_1
/gluster/user_2
/gluster/user_3
/gluster/user_4

The creation was successful, the volume could be started and mounted.
On the client, however, I could only find (via "ls /gluster") the
directories user_1 and user_2. But when I tried "cd /gluster/user_3",
it succeeded! Now "ls /gluster" showed me user_1, user_2 and user_3.
Unfortunately, user_3's subdirectories and files were still invisible,
but with the above mentioned trick, I could make them visible.

This is however not an option, as there are too much users and too
complicated file structures to do this manually. It anyhow seems like
Voodoo to me.

Is it possible to include all of the existing directories in the new
glusterFS volume? If yes: how?

Thank you in advance for your efforts,
Regards,
Daniel
___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users

___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users

___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users

___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-

Re: [Gluster-users] Start a new volume with pre-existing directories

2010-12-03 Thread Daniel Zander

Dear Craig,

I'm afraid I wasn't able to start a self heal the way you suggested. I 
tested the following:


--
WHAT I DID
--
I created on fs7:/storage/7
user_7_1  user_7_2  user_7_3

and on fs8:/storage/8
user_8_1  user_8_2  user_8_3

and filled all of the directories with some small files and subdirectories.

Then, on fs8:
gluster volume create heal_me transport tcp 192.168.101.246:/storage/8 
192.168.101.247:/storage/7


Then on fs8 and afterwards on fs7:
mount -t glusterfs localhost:/heal_me /tempmount/
cd /tempmount
find . | xargs stat >>/dev/null 2>&1
umount /tempmount

All went well, no error messages or anything. The output of `find . | 
xargs stat` is probably too long to post it here, but there were no 
error messages or anything else that would look suspicious to me.


---
RESULTS
---
ls fs8:storage/8
user_8_1  user_8_2  user_8_3

ls fs7:/storage/7
user_7_1  user_7_2  user_7_3  user_8_1  user_8_2  user_8_3

ls client:/storage/gluster
user_8_1  user_8_2  user_8_3

ls fs7:/tempmount
user_8_1  user_8_2  user_8_3

ls fs8:/tempmount
user_8_1  user_8_2  user_8_3

Unmounting and remounting has no effect.

Servers are both Ubuntu Server 10.4, client is CentOS 5, 64bits all around.

Thanks and regards,
Daniel


On 12/03/2010 10:10 AM, Craig Carl wrote:

Daniel -
If you want to export existing data you will need to run the self heal
process so extended attributes can get written. While this should work
without any issues it isn't an officially supported process, please make
sure you have complete and up to date backups.

After you have setup and started the Gluster volume mount it locally on
one of the servers using `mount -t glusterfs localhost:/ /`. CD into the root of the mount point and run `find . |
xargs stat >>/dev/null 2>&1` to start a self heal.

Also the command you used to create the volume should not have worked,
it is missing a volume name - gluster volume create  transport
tcp fs7:/storage/7, fs8:/storage/8, typo maybe?

Please let us know how it goes, and please let me know if you have any
other questions.

Thanks,

Craig

-->
Craig Carl
Senior Systems Engineer; Gluster, Inc.
Cell - (408) 829-9953 (California, USA)
Office - (408) 770-1884
Gtalk - craig.c...@gmail.com
Twitter - @gluster
http://rackerhacker.com/2010/08/11/one-month-with-glusterfs-in-production/




On 12/02/2010 11:38 PM, Daniel Zander wrote:

Dear all,

at our institute, we currently have 6 file servers, each one of them
individually mounted via NFS on ~ 20 clients. The structure on the
servers and the clients is the following:

/storage/1/ (NFS export from FS1)
/storage/2/ (NFS export from FS2)
etc ...

Recently, we decided that we would like to migrate this to glusterFS,
so that we can have one big storage directory on the clients. Let's
call it

/gluster/

I tried to set up a gluster volume with two empty fileservers and it
worked without any problems. I could easily mount it on a client and
use it (using the native glusterFS mount).

If we now want to migrate the entire institute, it would be very
convenient, if existing folders could be easily included into a new
volume. I tried to do this, but I did not succeed.

Here's a short description of what I tried:

Existing folders:
on fs7: /storage/7/user_1,user_2
on fs8: /storage/8/user_3,user_4

gluster volume create transport tcp fs7:/storage/7, fs8:/storage/8

I hoped to see on the client:
/gluster/user_1
/gluster/user_2
/gluster/user_3
/gluster/user_4

The creation was successful, the volume could be started and mounted.
On the client, however, I could only find (via "ls /gluster") the
directories user_1 and user_2. But when I tried "cd /gluster/user_3",
it succeeded! Now "ls /gluster" showed me user_1, user_2 and user_3.
Unfortunately, user_3's subdirectories and files were still invisible,
but with the above mentioned trick, I could make them visible.

This is however not an option, as there are too much users and too
complicated file structures to do this manually. It anyhow seems like
Voodoo to me.

Is it possible to include all of the existing directories in the new
glusterFS volume? If yes: how?

Thank you in advance for your efforts,
Regards,
Daniel
___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users

___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users

___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


Re: [Gluster-users] Anormal Gluster shutdown

2010-12-03 Thread Craig Carl

Samuel -
   I was able to recreate the failure and have updated the bug you filed.

Thanks,

Craig

-->
Craig Carl
Senior Systems Engineer
Gluster

On 12/03/2010 01:24 AM, Samuel Hassine wrote:

Craig,

I am using Debian Lenny (Proxmox 1.7)

r...@on-003:/# uname -a
Linux on-003 2.6.32-3-pve #1 SMP Fri Sep 17 17:56:13 CEST 2010 x86_64
GNU/Linux

On all Gluster nodes and gluster clients.

For hardware, it is sata disks with an LVM partition of 2.2To, in
distributed-replicated Gluster.

I tested this morning and the problem is still here.

Regards.
Sam

-Message d'origine-
De : gluster-users-boun...@gluster.org
[mailto:gluster-users-boun...@gluster.org] De la part de Craig Carl
Envoyé : vendredi 3 décembre 2010 09:02
À : gluster-users@gluster.org
Objet : Re: [Gluster-users] Anormal Gluster shutdown

Samuel -
 I can't reproduce this issue locally, can you send me operating system
and hardware details for both the Gluster servers and the client?

Thanks,

Craig

-->
Craig Carl
Senior Systems Engineer
Gluster



On 12/02/2010 05:59 AM, Samuel Hassine wrote:

Hi all,



GlusterFS partition automatically shutdown when umounting a binded
mount point

with "-f" option (without it works).



How to reproduce:



mounted Gluster partition on /gluster (any config):



df: localhost:/gluster4.5T  100G  4.4T   3% /gluster

mount: localhost:/gluster on /gluster type fuse.glusterfs

(rw,allow_other,default_permissions,max_read=131072)



commands:



mkdir /test

mount -n --bind /gluster /test

ls /test (verify you have the Gluster)



and:



umount -f /test



===>



df: `/gluster': Transport endpoint is not connected

[2010-12-02 14:48:56.38309] I [fuse-bridge.c:3138:fuse_thread_proc] fuse:

unmounting /gluster

[2010-12-02 14:48:56.38364] I [glusterfsd.c:672:cleanup_and_exit]
glusterfsd:

shutting down



Before 3.1.x I did not have this bug.



Regards.

Sam




___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users

___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


Re: [Gluster-users] Start a new volume with pre-existing directories

2010-12-03 Thread Daniel Zander

Hi!

> Also the command you used to create the volume should not have worked,
> it is missing a volume name - gluster volume create  transport
> tcp fs7:/storage/7, fs8:/storage/8, typo maybe?

Yes, typo. Sorry ...

Unfortunately, we do not have the storage capacity for a complete 
backup. If we should decide to take the risk, I will let you know how it 
goes.


Thanks for your help,
Daniel



On 12/03/2010 10:10 AM, Craig Carl wrote:

Daniel -
If you want to export existing data you will need to run the self heal
process so extended attributes can get written. While this should work
without any issues it isn't an officially supported process, please make
sure you have complete and up to date backups.

After you have setup and started the Gluster volume mount it locally on
one of the servers using `mount -t glusterfs localhost:/ /`. CD into the root of the mount point and run `find . |
xargs stat >>/dev/null 2>&1` to start a self heal.

Also the command you used to create the volume should not have worked,
it is missing a volume name - gluster volume create  transport
tcp fs7:/storage/7, fs8:/storage/8, typo maybe?

Please let us know how it goes, and please let me know if you have any
other questions.

Thanks,

Craig

-->
Craig Carl
Senior Systems Engineer; Gluster, Inc.
Cell - (408) 829-9953 (California, USA)
Office - (408) 770-1884
Gtalk - craig.c...@gmail.com
Twitter - @gluster
http://rackerhacker.com/2010/08/11/one-month-with-glusterfs-in-production/




On 12/02/2010 11:38 PM, Daniel Zander wrote:

Dear all,

at our institute, we currently have 6 file servers, each one of them
individually mounted via NFS on ~ 20 clients. The structure on the
servers and the clients is the following:

/storage/1/ (NFS export from FS1)
/storage/2/ (NFS export from FS2)
etc ...

Recently, we decided that we would like to migrate this to glusterFS,
so that we can have one big storage directory on the clients. Let's
call it

/gluster/

I tried to set up a gluster volume with two empty fileservers and it
worked without any problems. I could easily mount it on a client and
use it (using the native glusterFS mount).

If we now want to migrate the entire institute, it would be very
convenient, if existing folders could be easily included into a new
volume. I tried to do this, but I did not succeed.

Here's a short description of what I tried:

Existing folders:
on fs7: /storage/7/user_1,user_2
on fs8: /storage/8/user_3,user_4

gluster volume create transport tcp fs7:/storage/7, fs8:/storage/8

I hoped to see on the client:
/gluster/user_1
/gluster/user_2
/gluster/user_3
/gluster/user_4

The creation was successful, the volume could be started and mounted.
On the client, however, I could only find (via "ls /gluster") the
directories user_1 and user_2. But when I tried "cd /gluster/user_3",
it succeeded! Now "ls /gluster" showed me user_1, user_2 and user_3.
Unfortunately, user_3's subdirectories and files were still invisible,
but with the above mentioned trick, I could make them visible.

This is however not an option, as there are too much users and too
complicated file structures to do this manually. It anyhow seems like
Voodoo to me.

Is it possible to include all of the existing directories in the new
glusterFS volume? If yes: how?

Thank you in advance for your efforts,
Regards,
Daniel
___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users

___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users

___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


Re: [Gluster-users] Anormal Gluster shutdown

2010-12-03 Thread Samuel Hassine
Craig,

I am using Debian Lenny (Proxmox 1.7)

r...@on-003:/# uname -a
Linux on-003 2.6.32-3-pve #1 SMP Fri Sep 17 17:56:13 CEST 2010 x86_64
GNU/Linux

On all Gluster nodes and gluster clients.

For hardware, it is sata disks with an LVM partition of 2.2To, in
distributed-replicated Gluster.

I tested this morning and the problem is still here.

Regards.
Sam

-Message d'origine-
De : gluster-users-boun...@gluster.org
[mailto:gluster-users-boun...@gluster.org] De la part de Craig Carl
Envoyé : vendredi 3 décembre 2010 09:02
À : gluster-users@gluster.org
Objet : Re: [Gluster-users] Anormal Gluster shutdown

Samuel -
I can't reproduce this issue locally, can you send me operating system
and hardware details for both the Gluster servers and the client?

Thanks,

Craig

-->
Craig Carl
Senior Systems Engineer
Gluster



On 12/02/2010 05:59 AM, Samuel Hassine wrote:
> Hi all,
>
>
>
> GlusterFS partition automatically shutdown when umounting a binded 
> mount point
>
> with "-f" option (without it works).
>
>
>
> How to reproduce:
>
>
>
> mounted Gluster partition on /gluster (any config):
>
>
>
> df: localhost:/gluster4.5T  100G  4.4T   3% /gluster
>
> mount: localhost:/gluster on /gluster type fuse.glusterfs
>
> (rw,allow_other,default_permissions,max_read=131072)
>
>
>
> commands:
>
>
>
> mkdir /test
>
> mount -n --bind /gluster /test
>
> ls /test (verify you have the Gluster)
>
>
>
> and:
>
>
>
> umount -f /test
>
>
>
> ===>
>
>
>
> df: `/gluster': Transport endpoint is not connected
>
> [2010-12-02 14:48:56.38309] I [fuse-bridge.c:3138:fuse_thread_proc] fuse:
>
> unmounting /gluster
>
> [2010-12-02 14:48:56.38364] I [glusterfsd.c:672:cleanup_and_exit]
> glusterfsd:
>
> shutting down
>
>
>
> Before 3.1.x I did not have this bug.
>
>
>
> Regards.
>
> Sam
>
>
>
>
> ___
> Gluster-users mailing list
> Gluster-users@gluster.org
> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users

___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


Re: [Gluster-users] Storage Platform 3.1.1?

2010-12-03 Thread Craig Carl

Christian -
We're working on it everyday but we don't have a release date yet. 
As soon as we have a date we will send a note to the list. Thanks for 
your interest in Gluster and please let me know if you have any other 
questions.


Thanks again,

Craig

-->
Craig Carl
Senior Systems Engineer; Gluster, Inc.
Cell - (408) 829-9953 (California, USA)
Office - (408) 770-1884
Gtalk - craig.c...@gmail.com
Twitter - @gluster
http://rackerhacker.com/2010/08/11/one-month-with-glusterfs-in-production/



On 12/03/2010 12:48 AM, Christian Fischer wrote:

Sorry to bother you.
Is there a planned release date for Storage Platform 3.1.1?

Thanks
Christian
___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users

___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


Re: [Gluster-users] Start a new volume with pre-existing directories

2010-12-03 Thread Craig Carl

Daniel -
   If you want to export existing data you will need to run the self 
heal process so extended attributes can get written. While this should 
work without any issues it isn't an officially supported process, please 
make sure you have complete and up to date backups.


After you have setup and started the Gluster volume mount it locally on 
one of the servers using `mount -t  glusterfs localhost:/ 
/`. CD into the root of the mount point and run  
`find . | xargs stat >>/dev/null 2>&1` to start a self heal.


Also the command you used to create the volume should not have worked, 
it is missing a volume name - gluster volume create  transport 
tcp fs7:/storage/7, fs8:/storage/8, typo maybe?


Please let us know how it goes, and please let me know if you have any 
other questions.


Thanks,

Craig

-->
Craig Carl
Senior Systems Engineer; Gluster, Inc.
Cell - (408) 829-9953 (California, USA)
Office - (408) 770-1884
Gtalk - craig.c...@gmail.com
Twitter - @gluster
http://rackerhacker.com/2010/08/11/one-month-with-glusterfs-in-production/




On 12/02/2010 11:38 PM, Daniel Zander wrote:

Dear all,

at our institute, we currently have 6 file servers, each one of them 
individually mounted via NFS on ~ 20 clients. The structure on the 
servers and the clients is the following:


/storage/1/ (NFS export from FS1)
/storage/2/ (NFS export from FS2)
etc ...

Recently, we decided that we would like to migrate this to glusterFS, 
so that we can have one big storage directory on the clients. Let's 
call it


/gluster/

I tried to set up a gluster volume with two empty fileservers and it 
worked without any problems. I could easily mount it on a client and 
use it (using the native glusterFS mount).


If we now want to migrate the entire institute, it would be very 
convenient, if existing folders could be easily included into a new 
volume. I tried to do this, but I did not succeed.


Here's a short description of what I tried:

Existing folders:
on fs7: /storage/7/user_1,user_2
on fs8: /storage/8/user_3,user_4

gluster volume create transport tcp fs7:/storage/7, fs8:/storage/8

I hoped to see on the client:
/gluster/user_1
/gluster/user_2
/gluster/user_3
/gluster/user_4

The creation was successful, the volume could be started and mounted. 
On the client, however, I could only find (via "ls /gluster") the 
directories user_1 and user_2. But when I tried  "cd /gluster/user_3", 
it succeeded! Now "ls /gluster" showed me user_1, user_2 and user_3. 
Unfortunately, user_3's subdirectories and files were still invisible, 
but with the above mentioned trick, I could make them visible.


This is however not an option, as there are too much users and too 
complicated file structures to do this manually. It anyhow seems like 
Voodoo to me.


Is it possible to include all of the existing directories in the new 
glusterFS volume? If yes: how?


Thank you in advance for your efforts,
Regards,
Daniel
___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users

___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


Re: [Gluster-users] Who's using Fedora in production on Glusterfs storage servers?

2010-12-03 Thread Mark "Naoki" Rogers

Hi James,

I'm using 3.1.1 on six bricks in dist+replicate all running F14+BTRFS, 
the clients are on fedora12/13/14. I build the RPMs from source on a F14 
machine. The cluster is running entirely on GbE (with some 10Gb lines 
going in shortly), no RDMA/infiniband so I can't help there.


It's gone through a series of looped benchmarks for a while now (from 
3.1.0 through a few qa releases) and have so far pushed/pulled over 
110TB through it - I'm happy in the stability but not /entirely/ sure of 
the performance just yet, just started up more testing under 3.1.1.


But back to your main question there really isn't enough difference 
between the near-term releases of Fedora for it to make a huge 
difference either way. I do think you're better off using the latest 
Fedora release than an older one that will be end of life soon (f12 
tomorrow). Being able to patch/maintain your system is more important 
than an, often very arbitrary, vendor support list which is usually just 
an outcome of what people have had time to look into, rather than any 
measured reason a newer OS isn't supported. Besides the only thing you 
ever have to /really/ care about is the kernel and glibc major versions, 
so if it compiles you're pretty much ok (ldd it, that's all it needs).



On 12/02/2010 01:45 AM, Burnash, James wrote:

How many people on the list are using Fedora 12 (or 13) in production for 
Glusterfs storage servers? I know that Gluster Platform uses Fedora 12 as its 
OS - I was thinking of building my new glusterfs storage servers using Fedora, 
and was wondering whether Fedora 13 was tested by Gluster for v 3.1.1 and what 
other people's experiences were.

One of the reasons for my interest was so that I could use ext4 as the backend 
file store, instead of ext3.

Thanks,

James Burnash, Unix Engineering


___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


Re: [Gluster-users] Who's using Fedora in production on Glusterfs storage servers?

2010-12-03 Thread Mark "Naoki" Rogers

Hi James,

I'm using 3.1.1 on six bricks in dist+replicate all running F14+BTRFS, 
the clients are on fedora12/13/14. I build the RPMs from source on a F14 
machine. The cluster is running entirely on GbE (with some 10Gb lines 
going in shortly), no RDMA/infiniband so I can't help there.


It's gone through a series of looped benchmarks for a while now (from 
3.1.0 through a few qa releases) and have so far pushed/pulled over 
110TB through it - I'm happy in the stability but not /entirely/ sure of 
the performance just yet, just started up more testing under 3.1.1.


But back to your main question there really isn't enough difference 
between the near-term releases of Fedora for it to make a huge 
difference either way. I do think you're better off using the latest 
Fedora release than an older one that will be end of life soon (f12 
tomorrow). Being able to patch/maintain your system is more important 
than an, often very arbitrary, vendor support list which is usually just 
an outcome of what people have had time to look into, rather than any 
measured reason a newer OS isn't supported. Besides the only thing you 
ever have to /really/ care about is the kernel and glibc major versions, 
so if it compiles you're pretty much ok (ldd it, that's all it needs).



On 12/02/2010 01:45 AM, Burnash, James wrote:

How many people on the list are using Fedora 12 (or 13) in production for 
Glusterfs storage servers? I know that Gluster Platform uses Fedora 12 as its 
OS - I was thinking of building my new glusterfs storage servers using Fedora, 
and was wondering whether Fedora 13 was tested by Gluster for v 3.1.1 and what 
other people's experiences were.

One of the reasons for my interest was so that I could use ext4 as the backend 
file store, instead of ext3.

Thanks,

James Burnash, Unix Engineering


___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


Re: [Gluster-users] Who's using Fedora in production on Glusterfs storage servers?

2010-12-03 Thread Mark "Naoki" Rogers
I'm using 3.1.1 on six bricks running F14+BTRFS. Been running looped 
benchmarks for a while and pushed/pulled over 110TB so far. I'm 
convinced it's stable but not entirely sure of the performance just yet.


On 12/02/2010 01:45 AM, Burnash, James wrote:

How many people on the list are using Fedora 12 (or 13) in production for 
Glusterfs storage servers? I know that Gluster Platform uses Fedora 12 as its 
OS - I was thinking of building my new glusterfs storage servers using Fedora, 
and was wondering whether Fedora 13 was tested by Gluster for v 3.1.1 and what 
other people's experiences were.

One of the reasons for my interest was so that I could use ext4 as the backend 
file store, instead of ext3.


___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


[Gluster-users] Storage Platform 3.1.1?

2010-12-03 Thread Christian Fischer
Sorry to bother you.
Is there a planned release date for Storage Platform 3.1.1?

Thanks
Christian
___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


Re: [Gluster-users] Anormal Gluster shutdown

2010-12-03 Thread Craig Carl

Samuel -
   I can't reproduce this issue locally, can you send me operating 
system and hardware details for both the Gluster servers and the client?


Thanks,

Craig

-->
Craig Carl
Senior Systems Engineer
Gluster



On 12/02/2010 05:59 AM, Samuel Hassine wrote:

Hi all,



GlusterFS partition automatically shutdown when umounting a binded mount
point

with "-f" option (without it works).



How to reproduce:



mounted Gluster partition on /gluster (any config):



df: localhost:/gluster4.5T  100G  4.4T   3% /gluster

mount: localhost:/gluster on /gluster type fuse.glusterfs

(rw,allow_other,default_permissions,max_read=131072)



commands:



mkdir /test

mount -n --bind /gluster /test

ls /test (verify you have the Gluster)



and:



umount -f /test



===>



df: `/gluster': Transport endpoint is not connected

[2010-12-02 14:48:56.38309] I [fuse-bridge.c:3138:fuse_thread_proc] fuse:

unmounting /gluster

[2010-12-02 14:48:56.38364] I [glusterfsd.c:672:cleanup_and_exit]
glusterfsd:

shutting down



Before 3.1.x I did not have this bug.



Regards.

Sam




___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users