Re: [Gluster-users] hanging of mounted filesystem (3.3.1)

2013-01-31 Thread Michael Colonno
To close out this thread I rebuilt the entire volume with ZFS
instead of ext4 and it now appears to work well. I have not loaded it up yet
but the ls / ll hanging problem is gone and I can move data to and from the
mounted filesystem via a glusterfs mount. 

Thanks,
~Mike C. 

-Original Message-
From: gluster-users-boun...@gluster.org
[mailto:gluster-users-boun...@gluster.org] On Behalf Of Robert Hajime
Lanning
Sent: Thursday, January 31, 2013 7:12 PM
To: gluster-users@gluster.org
Subject: Re: [Gluster-users] hanging of mounted filesystem (3.3.1)

On 01/31/13 19:01, Robert Hajime Lanning wrote:
> On 01/31/13 17:58, Michael Colonno wrote:
>> Doing more searching, it appears there are several posts on various 
>> forums describing the same issue with Gluster deployments of 4+ nodes 
>> (e.g.
>> http://serverfault.com/questions/453345/certain-operations-like-ls-ha
>> ng-in-4
>>
>> -node-gluster-setup). None of these posts are answered which doesn't 
>> fill me with confidence. Can anyone confirm a 4+ node cluster with 
>> replication that does not show this (indefinite hang on ls or ll) 
>> behavior?
>>
>
> Volume Name: vol01
> Type: Distributed-Replicate
> Volume ID: 3d5d1791-044e-4cf3-b17b-79c74578e590
> Status: Started
> Number of Bricks: 24 x 2 = 48
> Transport-type: tcp
>
> 4 bricks per server * 12 servers
> All bricks are 37TB logical volumes formated with XFS.
>
> Filesystem Size Used Avail Use% Mounted on
> gfs-vol:/vol01 888T 119G 888T 1% /glusterfs/vol01
>
> No issues.
>

Sorry, should have included:
CentOS 6.3
GlusterFS 3.3.1

--
Mr. Flibble
King of the Potato People
___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] hanging of mounted filesystem (3.3.1)

2013-01-31 Thread Robert Hajime Lanning

On 01/31/13 19:01, Robert Hajime Lanning wrote:

On 01/31/13 17:58, Michael Colonno wrote:

Doing more searching, it appears there are several posts on various
forums describing the same issue with Gluster deployments of 4+ nodes
(e.g.
http://serverfault.com/questions/453345/certain-operations-like-ls-hang-in-4

-node-gluster-setup). None of these posts are answered which doesn't
fill me
with confidence. Can anyone confirm a 4+ node cluster with replication
that
does not show this (indefinite hang on ls or ll) behavior?



Volume Name: vol01
Type: Distributed-Replicate
Volume ID: 3d5d1791-044e-4cf3-b17b-79c74578e590
Status: Started
Number of Bricks: 24 x 2 = 48
Transport-type: tcp

4 bricks per server * 12 servers
All bricks are 37TB logical volumes formated with XFS.

Filesystem Size Used Avail Use% Mounted on
gfs-vol:/vol01 888T 119G 888T 1% /glusterfs/vol01

No issues.



Sorry, should have included:
CentOS 6.3
GlusterFS 3.3.1

--
Mr. Flibble
King of the Potato People
___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] hanging of mounted filesystem (3.3.1)

2013-01-31 Thread Robert Hajime Lanning

On 01/31/13 17:58, Michael Colonno wrote:

Doing more searching, it appears there are several posts on various
forums describing the same issue with Gluster deployments of 4+ nodes (e.g.
http://serverfault.com/questions/453345/certain-operations-like-ls-hang-in-4
-node-gluster-setup). None of these posts are answered which doesn't fill me
with confidence. Can anyone confirm a 4+ node cluster with replication that
does not show this (indefinite hang on ls or ll) behavior?



Volume Name: vol01
Type: Distributed-Replicate
Volume ID: 3d5d1791-044e-4cf3-b17b-79c74578e590
Status: Started
Number of Bricks: 24 x 2 = 48
Transport-type: tcp

4 bricks per server * 12 servers
All bricks are 37TB logical volumes formated with XFS.

FilesystemSize  Used Avail Use% Mounted on
gfs-vol:/vol01888T  119G  888T   1% /glusterfs/vol01

No issues.

--
Mr. Flibble
King of the Potato People
___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] hanging of mounted filesystem (3.3.1)

2013-01-31 Thread Anand Avati
That's probably because in a 2-node replicated environment, there is no
d_off transformation (at the DHT level) and therefore the bug is "avoided".

Avati

On Thu, Jan 31, 2013 at 5:58 PM, Michael Colonno wrote:

> Doing more searching, it appears there are several posts on various
> forums describing the same issue with Gluster deployments of 4+ nodes (e.g.
>
> http://serverfault.com/questions/453345/certain-operations-like-ls-hang-in-4
> -node-gluster-setup). None of these posts are answered which doesn't fill
> me
> with confidence. Can anyone confirm a 4+ node cluster with replication that
> does not show this (indefinite hang on ls or ll) behavior?
>
> Thanks,
> ~Mike C.
>
> -Original Message-
> From: gluster-users-boun...@gluster.org
> [mailto:gluster-users-boun...@gluster.org] On Behalf Of Michael Colonno
> Sent: Thursday, January 31, 2013 5:48 PM
> To: gluster-users@gluster.org
> Subject: Re: [Gluster-users] hanging of mounted filesystem (3.3.1)
>
> I rebuilt the volume with only TCP transport; same issues occurring
> so RDMA must not be the issue here. With a glusterfs mount (which is
> successful) I can copy files into the mounted volume but any command to
> list
> the contents (ls or ll) hangs forever while ls (for example uses ~ 5% of
> the
> system's CPU and glusterfs uses ~ 50% of the system's CPU). I've never seen
> this behavior before with my other Gluster deployments, even much larger /
> more complicated ones. This one is pretty ordinary. Any advice appreciated.
>
> Thanks,
> ~Mike C.
>
> -Original Message-
> From: gluster-users-boun...@gluster.org
> [mailto:gluster-users-boun...@gluster.org] On Behalf Of Michael Colonno
> Sent: Thursday, January 31, 2013 4:50 PM
> To: gluster-users@gluster.org
> Subject: Re: [Gluster-users] hanging of mounted filesystem (3.3.1)
>
> Tried the proto=tcp setting, same error unfortunately...
>
> Thanks,
> ~Mike C.
>
> -Original Message-
> From: Robert Hajime Lanning [mailto:lann...@lanning.cc]
> Sent: Thursday, January 31, 2013 4:28 PM
> To: Michael Colonno
> Cc: gluster-users@gluster.org
> Subject: Re: [Gluster-users] hanging of mounted filesystem (3.3.1)
>
> On 01/31/13 15:54, Michael Colonno wrote:
> > Updating this thread: I tried to switch over to NFS mounting. When I
> > try to mount the volume with the following line in fstab:
> >
> > node1:/Volume /tmp/mnt nfs defaults,_netdev,vers=3 0 0
>
> node1:/Volume /tmp/mnt nfs defaults,_netdev,vers=3,proto=tcp 0 0
>
> --
> Mr. Flibble
> King of the Potato People
>
> ___
> Gluster-users mailing list
> Gluster-users@gluster.org
> http://supercolony.gluster.org/mailman/listinfo/gluster-users
>
> ___
> Gluster-users mailing list
> Gluster-users@gluster.org
> http://supercolony.gluster.org/mailman/listinfo/gluster-users
>
> ___
> Gluster-users mailing list
> Gluster-users@gluster.org
> http://supercolony.gluster.org/mailman/listinfo/gluster-users
>
___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] hanging of mounted filesystem (3.3.1)

2013-01-31 Thread Michael Colonno
Many thanks - I'll give this a try shortly and report results
for the thread. 

 

~Mike C. 

 

From: Anand Avati [mailto:anand.av...@gmail.com] 
Sent: Thursday, January 31, 2013 5:55 PM
To: Michael Colonno
Cc: gluster-users@gluster.org
Subject: Re: [Gluster-users] hanging of mounted filesystem (3.3.1)

 

Looks like you are using a recent kernel with ext4 as the brick filesystem.
It is a known issue with ext4
(http://joejulian.name/blog/glusterfs-bit-by-ext4-structure-change/). Please
use XFS as the brick filesystem for now.

 

Avati

On Thu, Jan 31, 2013 at 5:47 PM, Michael Colonno 
wrote:

I rebuilt the volume with only TCP transport; same issues occurring
so RDMA must not be the issue here. With a glusterfs mount (which is
successful) I can copy files into the mounted volume but any command to list
the contents (ls or ll) hangs forever while ls (for example uses ~ 5% of the
system's CPU and glusterfs uses ~ 50% of the system's CPU). I've never seen
this behavior before with my other Gluster deployments, even much larger /
more complicated ones. This one is pretty ordinary. Any advice appreciated.


Thanks,
~Mike C.

-Original Message-

From: gluster-users-boun...@gluster.org
[mailto:gluster-users-boun...@gluster.org] On Behalf Of Michael Colonno

Sent: Thursday, January 31, 2013 4:50 PM
To: gluster-users@gluster.org
Subject: Re: [Gluster-users] hanging of mounted filesystem (3.3.1)

Tried the proto=tcp setting, same error unfortunately...

Thanks,
~Mike C.

-Original Message-
From: Robert Hajime Lanning [mailto:lann...@lanning.cc]
Sent: Thursday, January 31, 2013 4:28 PM
To: Michael Colonno
Cc: gluster-users@gluster.org
Subject: Re: [Gluster-users] hanging of mounted filesystem (3.3.1)

On 01/31/13 15:54, Michael Colonno wrote:
> Updating this thread: I tried to switch over to NFS mounting. When I
> try to mount the volume with the following line in fstab:
>
> node1:/Volume /tmp/mnt nfs defaults,_netdev,vers=3 0 0

node1:/Volume /tmp/mnt nfs defaults,_netdev,vers=3,proto=tcp 0 0

--
Mr. Flibble
King of the Potato People

___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

 

___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] hanging of mounted filesystem (3.3.1)

2013-01-31 Thread Michael Colonno
Doing more searching, it appears there are several posts on various
forums describing the same issue with Gluster deployments of 4+ nodes (e.g.
http://serverfault.com/questions/453345/certain-operations-like-ls-hang-in-4
-node-gluster-setup). None of these posts are answered which doesn't fill me
with confidence. Can anyone confirm a 4+ node cluster with replication that
does not show this (indefinite hang on ls or ll) behavior? 

Thanks,
~Mike C. 

-Original Message-
From: gluster-users-boun...@gluster.org
[mailto:gluster-users-boun...@gluster.org] On Behalf Of Michael Colonno
Sent: Thursday, January 31, 2013 5:48 PM
To: gluster-users@gluster.org
Subject: Re: [Gluster-users] hanging of mounted filesystem (3.3.1)

I rebuilt the volume with only TCP transport; same issues occurring
so RDMA must not be the issue here. With a glusterfs mount (which is
successful) I can copy files into the mounted volume but any command to list
the contents (ls or ll) hangs forever while ls (for example uses ~ 5% of the
system's CPU and glusterfs uses ~ 50% of the system's CPU). I've never seen
this behavior before with my other Gluster deployments, even much larger /
more complicated ones. This one is pretty ordinary. Any advice appreciated. 

Thanks,
~Mike C. 

-Original Message-
From: gluster-users-boun...@gluster.org
[mailto:gluster-users-boun...@gluster.org] On Behalf Of Michael Colonno
Sent: Thursday, January 31, 2013 4:50 PM
To: gluster-users@gluster.org
Subject: Re: [Gluster-users] hanging of mounted filesystem (3.3.1)

Tried the proto=tcp setting, same error unfortunately...

Thanks,
~Mike C. 

-Original Message-
From: Robert Hajime Lanning [mailto:lann...@lanning.cc]
Sent: Thursday, January 31, 2013 4:28 PM
To: Michael Colonno
Cc: gluster-users@gluster.org
Subject: Re: [Gluster-users] hanging of mounted filesystem (3.3.1)

On 01/31/13 15:54, Michael Colonno wrote:
> Updating this thread: I tried to switch over to NFS mounting. When I 
> try to mount the volume with the following line in fstab:
>
> node1:/Volume /tmp/mnt nfs defaults,_netdev,vers=3 0 0

node1:/Volume /tmp/mnt nfs defaults,_netdev,vers=3,proto=tcp 0 0

--
Mr. Flibble
King of the Potato People

___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] hanging of mounted filesystem (3.3.1)

2013-01-31 Thread Anand Avati
Looks like you are using a recent kernel with ext4 as the brick filesystem.
It is a known issue with ext4 (
http://joejulian.name/blog/glusterfs-bit-by-ext4-structure-change/). Please
use XFS as the brick filesystem for now.

Avati

On Thu, Jan 31, 2013 at 5:47 PM, Michael Colonno wrote:

> I rebuilt the volume with only TCP transport; same issues occurring
> so RDMA must not be the issue here. With a glusterfs mount (which is
> successful) I can copy files into the mounted volume but any command to
> list
> the contents (ls or ll) hangs forever while ls (for example uses ~ 5% of
> the
> system's CPU and glusterfs uses ~ 50% of the system's CPU). I've never seen
> this behavior before with my other Gluster deployments, even much larger /
> more complicated ones. This one is pretty ordinary. Any advice appreciated.
>
> Thanks,
> ~Mike C.
>
> -Original Message-
> From: gluster-users-boun...@gluster.org
> [mailto:gluster-users-boun...@gluster.org] On Behalf Of Michael Colonno
> Sent: Thursday, January 31, 2013 4:50 PM
> To: gluster-users@gluster.org
> Subject: Re: [Gluster-users] hanging of mounted filesystem (3.3.1)
>
> Tried the proto=tcp setting, same error unfortunately...
>
> Thanks,
> ~Mike C.
>
> -Original Message-
> From: Robert Hajime Lanning [mailto:lann...@lanning.cc]
> Sent: Thursday, January 31, 2013 4:28 PM
> To: Michael Colonno
> Cc: gluster-users@gluster.org
> Subject: Re: [Gluster-users] hanging of mounted filesystem (3.3.1)
>
> On 01/31/13 15:54, Michael Colonno wrote:
> > Updating this thread: I tried to switch over to NFS mounting. When I
> > try to mount the volume with the following line in fstab:
> >
> > node1:/Volume /tmp/mnt nfs defaults,_netdev,vers=3 0 0
>
> node1:/Volume /tmp/mnt nfs defaults,_netdev,vers=3,proto=tcp 0 0
>
> --
> Mr. Flibble
> King of the Potato People
>
> ___
> Gluster-users mailing list
> Gluster-users@gluster.org
> http://supercolony.gluster.org/mailman/listinfo/gluster-users
>
> ___
> Gluster-users mailing list
> Gluster-users@gluster.org
> http://supercolony.gluster.org/mailman/listinfo/gluster-users
>
___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] hanging of mounted filesystem (3.3.1)

2013-01-31 Thread Michael Colonno
I rebuilt the volume with only TCP transport; same issues occurring
so RDMA must not be the issue here. With a glusterfs mount (which is
successful) I can copy files into the mounted volume but any command to list
the contents (ls or ll) hangs forever while ls (for example uses ~ 5% of the
system's CPU and glusterfs uses ~ 50% of the system's CPU). I've never seen
this behavior before with my other Gluster deployments, even much larger /
more complicated ones. This one is pretty ordinary. Any advice appreciated. 

Thanks,
~Mike C. 

-Original Message-
From: gluster-users-boun...@gluster.org
[mailto:gluster-users-boun...@gluster.org] On Behalf Of Michael Colonno
Sent: Thursday, January 31, 2013 4:50 PM
To: gluster-users@gluster.org
Subject: Re: [Gluster-users] hanging of mounted filesystem (3.3.1)

Tried the proto=tcp setting, same error unfortunately...

Thanks,
~Mike C. 

-Original Message-
From: Robert Hajime Lanning [mailto:lann...@lanning.cc]
Sent: Thursday, January 31, 2013 4:28 PM
To: Michael Colonno
Cc: gluster-users@gluster.org
Subject: Re: [Gluster-users] hanging of mounted filesystem (3.3.1)

On 01/31/13 15:54, Michael Colonno wrote:
> Updating this thread: I tried to switch over to NFS mounting. When I 
> try to mount the volume with the following line in fstab:
>
> node1:/Volume /tmp/mnt nfs defaults,_netdev,vers=3 0 0

node1:/Volume /tmp/mnt nfs defaults,_netdev,vers=3,proto=tcp 0 0

--
Mr. Flibble
King of the Potato People

___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] NFS availability

2013-01-31 Thread Joe Julian

On 01/31/2013 03:57 PM, Stephan von Krawczynski wrote:

On Thu, 31 Jan 2013 16:00:38 -0500
Jeff Darcy  wrote:


Most common network errors are not a matter of design, but of dead
iron.

It's usually both - a design that is insufficiently tolerant of
component failure, plus a combination of component failures that exceeds
that tolerance.  You seem to have a very high standard for filesystems
continuing to maintain 100% functionality - and I suppose 100%
performance as well - if there's any possibility whatsoever that they
could do so.  Why don't you apply that same standard to the part of the
system that you're responsible for designing?  Running any distributed
system on top of a deficient network infrastructure will lead only to
disappointment.

I am sorry that glusterfs is part of the design and your critics.
This sentence is incomprehensible. GlusterFS is part of the design and 
his/our critics? I haven't seen anything from GlusterFS criticizing anyone.

Everyone working sufficiently long with networks of all kinds of sizes and
components can tell you that in the end you want a design for a file service
that works as long as possible. This means it should survive even if there is
only one client and server and network path left.
Most users, if they're down to one client and one server, they're out of 
business anyway. I know we can't run our whole company on one or two 
computers, and we're not even all that big.

At least that is what is expected from glusterfs. Unfortunately sometimes you
get disappointed. We saw just about everything happening when switching off
all but one reliable network path including network hangs and server hangs
(the last one) (read the list for examples by others).
After taking a power hit to the building which was able to get through 
our commercial ups and line conditioners (which also had to be replaced 
after that), our switches all had to be replaced. Part of my testing was 
to do just that. I pulled the plugs on all the switches, leaving only 
one data path then changing that path. All the while I had full activity 
on all my volumes, including innodb transactions, vm image activity, 
web, samba, and workstation home directories. I ran multiple dd tests 
from multiple clients during these tests. Not once did they even slow down.

On the other end of the story clients see servers go offline if you increase
the non-gluster traffic on the network. Main (but not only) reason is the very
low default ping time (read the list for examples by others).
42 seconds is low? I've never (and I used to have really crappy dual 
32bit xeon servers) saturated my systems to the point where it took 42 
seconds for a server to respond. If you're switches can't handle that 
kind of traffic, perhaps you're using the wrong hardware.

All these seen effects show clearly that noone ever tested this to an extent I
would have done writing this kind of software. After all this is a piece of
software whose merely only purpose is surviving dead servers and networks.
It is no question of design, because on paper everything looks promising.

Sometimes your arguments let me believe you want glusterfs working like a ford
car. A lot of technical gameplay built in but the idea that a car should be a
good car in the first place got lost on the way somewhere. Quite a lot of the
features built in lately have the quality of an mp3-player in your ford. Nice
to have but does not help you a lot driving 200 and a rabbit crossing.
And this is why I am requesting the equivalent of a BMW.

Using your own analogy, what good is a BMW if you've got no roads to 
drive it on?


You're talking in terms of single entities. Most (if not all) of the 
sysadmins I work with on a daily basis, my peers in the industry, 
members of LOPSA... we work in systems. We know how to build in 
redundancy and plan for and survive failures. There's not a week that 
goes by where something in my system hasn't encountered some issue, yet 
our company has not lost any productivity because of it. We have the 
best availability of systems of all our competitors (and I do monitor 
their systems).


I'm not saying you're doing it wrong. Be as asssured as you feel is 
appropriate for your system requirements. Just be aware that the 
majority of the industry does not share those requirements. I'm not 
speaking from my "gut instinct" but I am a member of several 
professional organizations. I attend functions on a weekly basis 
attended by members of organizations like Expedia, Ebay, Amazon, Google, 
Boeing, Starbucks, etc, etc. and we talk about this stuff. Fault 
tolerance and recovery is a big part of what we do, probably the 
biggest, and I still advise the way I do, not through just my own 
experiences, but through the experiences of my peers.


Offer advice and back it up with facts, anecdotes, and/or tests, but 
accept that there are as many ways of managing systems as there are 
systems to be managed. Accept that there are professionals in th

Re: [Gluster-users] hanging of mounted filesystem (3.3.1)

2013-01-31 Thread Michael Colonno
Tried the proto=tcp setting, same error unfortunately...

Thanks,
~Mike C. 

-Original Message-
From: Robert Hajime Lanning [mailto:lann...@lanning.cc] 
Sent: Thursday, January 31, 2013 4:28 PM
To: Michael Colonno
Cc: gluster-users@gluster.org
Subject: Re: [Gluster-users] hanging of mounted filesystem (3.3.1)

On 01/31/13 15:54, Michael Colonno wrote:
> Updating this thread: I tried to switch over to NFS mounting. When I 
> try to mount the volume with the following line in fstab:
>
> node1:/Volume /tmp/mnt nfs defaults,_netdev,vers=3 0 0

node1:/Volume /tmp/mnt nfs defaults,_netdev,vers=3,proto=tcp 0 0

--
Mr. Flibble
King of the Potato People

___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] hanging of mounted filesystem (3.3.1)

2013-01-31 Thread Joe Julian

On 01/31/2013 04:00 PM, Michael Colonno wrote:


Related: any official word on full RDMA support in Gluster 3.x?

Thanks,

~Mike C.


The last official word I got from Raghavendra Bhat defined the state of 
RDMA in 3.3.1 as "Tech Preview". As I understand it, this is a Red Hat 
term for "it hasn't gone through sufficient QA testing for us to support 
it commercially". There's no less support in 3.3 than there was in 3.1 
or 3.2.


The people that I've seen use RDMA successfully used bleeding edge 
kernels and drivers, fwiw.
___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] hanging of mounted filesystem (3.3.1)

2013-01-31 Thread Robert Hajime Lanning

On 01/31/13 15:54, Michael Colonno wrote:

Updating this thread: I tried to switch over to NFS mounting. When I try
to mount the volume with the following line in fstab:

node1:/Volume /tmp/mnt nfs defaults,_netdev,vers=3 0 0


node1:/Volume /tmp/mnt nfs defaults,_netdev,vers=3,proto=tcp 0 0

--
Mr. Flibble
King of the Potato People
___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] hanging of mounted filesystem (3.3.1)

2013-01-31 Thread Michael Colonno
Do I need to blow up and rebuild the brick to make that happen
or can this be set on the fly? Possibly relevant fact: I do not have my IB
fabric in place yet but I'm happy to use IPoIB for this deployment when I
do. I included it as a placeholder. 

 

Related: any official word on full RDMA support in Gluster 3.x? 

 

Thanks,

~Mike C. 

 

From: gluster-users-boun...@gluster.org
[mailto:gluster-users-boun...@gluster.org] On Behalf Of Bryan Whitehead
Sent: Thursday, January 31, 2013 3:52 PM
To: Michael Colonno
Cc: gluster-users
Subject: Re: [Gluster-users] hanging of mounted filesystem (3.3.1)

 

remove the transport rdma and try again. When using RDMA I've also had
extremely bad CPU eating issues.

 

I currently run gluster with IPoIB to get the speed of infiniband and the
non-crazy cpu usage of rdma gluster.

 

On Thu, Jan 31, 2013 at 9:20 AM, Michael Colonno
 wrote:

Hi All ~

I created an eight-brick gluster 3.3.1 volume (2x replication) on
eight CentOS 6.3 x64 systems. I was able to form and start the volume
without issue. I was also able to mount it through /etc/fstab as a native
glusterfs mount. I have a couple questions issues at this point:

- On a client machine, "glusterfs" is not recognized as a valid type
unless gluster-server is installed. This seems to contradict the
documentation - wanted to make sure I'm not doing something wrong. More a
clarification than issue here.

- The glusterfs process is taking between 50% and 80% CPU on both
the brick and client systems (these are fairly powerful, brand new servers).


- No doubt linked to the above, the mounted filesystem hangs
indefinitely when accessed. I tried an "ls -a" on the mounted filesystem,
for example, which hangs forever. I tested this by mounting a brick system
to itself and to a client which is not a brick and the same behavior was
observed. Both were glusterfs mounts.

There is nothing special about my deployment except for the use of
transport = tcp,rdma. I am running on Ethernet now but will be migrating to
Infiniband after this is debugged.

Thanks for any advice,
~Mike C.




___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

 

___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] NFS availability

2013-01-31 Thread Stephan von Krawczynski
On Thu, 31 Jan 2013 16:00:38 -0500
Jeff Darcy  wrote:

> > Most common network errors are not a matter of design, but of dead
> > iron.
> 
> It's usually both - a design that is insufficiently tolerant of
> component failure, plus a combination of component failures that exceeds
> that tolerance.  You seem to have a very high standard for filesystems
> continuing to maintain 100% functionality - and I suppose 100%
> performance as well - if there's any possibility whatsoever that they
> could do so.  Why don't you apply that same standard to the part of the
> system that you're responsible for designing?  Running any distributed
> system on top of a deficient network infrastructure will lead only to
> disappointment.

I am sorry that glusterfs is part of the design and your critics. 
Everyone working sufficiently long with networks of all kinds of sizes and
components can tell you that in the end you want a design for a file service
that works as long as possible. This means it should survive even if there is
only one client and server and network path left.
At least that is what is expected from glusterfs. Unfortunately sometimes you
get disappointed. We saw just about everything happening when switching off
all but one reliable network path including network hangs and server hangs
(the last one) (read the list for examples by others).
On the other end of the story clients see servers go offline if you increase
the non-gluster traffic on the network. Main (but not only) reason is the very
low default ping time (read the list for examples by others).
All these seen effects show clearly that noone ever tested this to an extent I
would have done writing this kind of software. After all this is a piece of
software whose merely only purpose is surviving dead servers and networks.
It is no question of design, because on paper everything looks promising.

Sometimes your arguments let me believe you want glusterfs working like a ford
car. A lot of technical gameplay built in but the idea that a car should be a
good car in the first place got lost on the way somewhere. Quite a lot of the
features built in lately have the quality of an mp3-player in your ford. Nice
to have but does not help you a lot driving 200 and a rabbit crossing.
And this is why I am requesting the equivalent of a BMW.

-- 
Regards,
Stephan
___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] hanging of mounted filesystem (3.3.1)

2013-01-31 Thread Michael Colonno
Updating this thread: I tried to switch over to NFS mounting.
When I try to mount the volume with the following line in fstab:

 

node1:/Volume  /tmp/mntnfs
defaults,_netdev,vers=3 0 0

 

I get:

 

mount.nfs: mounting node1:/Volume failed, reason given by
server:

  No such file or directory

 

It seems from the documentation that the mount folder is
supposed to be the gluster volume name. I also assume I don't need an
exports entry since there is nothing in the documentation about this and the
volume options seem to indicate that My gluster volume itself seems to start
and stop without any issue but can't seem to mount via any method(?)
Hopefully I am just doing something simple wrong in config; I have tried to
meticulously follow the admin guide. For reference here is my volume info:

 

Volume Name: Volume

Type: Distributed-Replicate

Volume ID: 409d4096-feda-4b5f-b01b-3383d39af1c6

Status: Started

Number of Bricks: 4 x 2 = 8

Transport-type: tcp,rdma

Bricks:

Brick1: node1:/volume

Brick2: node2:/volume

Brick3: node3:/volume

Brick4: node4:/volume

Brick5: node5:/volume

Brick6: node6:/volume

Brick7: node7:/volume

Brick8: node8:/volume

 

Thanks,

~Mike C.

 

From: gluster-users-boun...@gluster.org
[mailto:gluster-users-boun...@gluster.org] On Behalf Of Michael Colonno
Sent: Thursday, January 31, 2013 2:41 PM
To: gluster-users@gluster.org
Subject: Re: [Gluster-users] hanging of mounted filesystem (3.3.1)

 

 

Hi All ~

 

I created an eight-brick gluster 3.3.1 volume (2x replication)
on eight CentOS 6.3 x64 systems. I was able to form and start the volume
without issue. I was also able to mount it through /etc/fstab as a native
glusterfs mount. I have a couple questions issues at this point:

 

- On a client machine, "glusterfs" is not recognized as a valid
type unless gluster-server is installed. This seems to contradict the
documentation - wanted to make sure I'm not doing something wrong. More a
clarification than issue here. 

 

- The glusterfs process is taking between 50% and 80% CPU on
both the brick and client systems (these are fairly powerful new servers).

 

- No doubt linked to the above, the mounted filesystem hangs
indefinitely when accessed. I tried an "ls -a" on the mounted filesystem,
for example, which hangs forever. I tested this by mounting a brick system
to itself and to a client which is not a brick and the same behavior was
observed. Both were glusterfs mounts. 

 

There is nothing special about my deployment except for the use
of transport = tcp,rdma. I am running on Ethernet now but will be migrating
to Infiniband after this is debugged. 

 

Thanks for any advice,

~Mike C.

___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] hanging of mounted filesystem (3.3.1)

2013-01-31 Thread Bryan Whitehead
remove the transport rdma and try again. When using RDMA I've also had
extremely bad CPU eating issues.

I currently run gluster with IPoIB to get the speed of infiniband and the
non-crazy cpu usage of rdma gluster.


On Thu, Jan 31, 2013 at 9:20 AM, Michael Colonno  wrote:

> Hi All ~
>
> I created an eight-brick gluster 3.3.1 volume (2x replication) on
> eight CentOS 6.3 x64 systems. I was able to form and start the volume
> without issue. I was also able to mount it through /etc/fstab as a native
> glusterfs mount. I have a couple questions issues at this point:
>
> - On a client machine, "glusterfs" is not recognized as a valid
> type
> unless gluster-server is installed. This seems to contradict the
> documentation - wanted to make sure I'm not doing something wrong. More a
> clarification than issue here.
>
> - The glusterfs process is taking between 50% and 80% CPU on both
> the brick and client systems (these are fairly powerful, brand new
> servers).
>
>
> - No doubt linked to the above, the mounted filesystem hangs
> indefinitely when accessed. I tried an "ls -a" on the mounted filesystem,
> for example, which hangs forever. I tested this by mounting a brick system
> to itself and to a client which is not a brick and the same behavior was
> observed. Both were glusterfs mounts.
>
> There is nothing special about my deployment except for the use of
> transport = tcp,rdma. I am running on Ethernet now but will be migrating to
> Infiniband after this is debugged.
>
> Thanks for any advice,
> ~Mike C.
>
>
>
>
> ___
> Gluster-users mailing list
> Gluster-users@gluster.org
> http://supercolony.gluster.org/mailman/listinfo/gluster-users
>
___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

[Gluster-users] Striped Replicated Volumes: create files error.

2013-01-31 Thread axel.weber
Hi there,
each time I copy (or dd or similar) a file to a striped replicated volume I get 
an error: the argument is not valid.
An empty file is created.
If I now run the copy, it works.
This is in independed of the client platform.
We are using version 3.3.1


Mit freundlichen Grüßen / Kind regards


Axel Weber


___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

[Gluster-users] hanging of mounted filesystem (3.3.1)

2013-01-31 Thread Michael Colonno
Hi All ~

 

I created an eight-brick gluster 3.3.1 volume (2x replication)
on eight CentOS 6.3 x64 systems. I was able to form and start the volume
without issue. I was also able to mount it through /etc/fstab as a native
glusterfs mount. I have a couple questions issues at this point:

 

- On a client machine, "glusterfs" is not recognized as a valid
type unless gluster-server is installed. This seems to contradict the
documentation - wanted to make sure I'm not doing something wrong. More a
clarification than issue here. 

 

- The glusterfs process is taking between 50% and 80% CPU on
both the brick and client systems (these are fairly powerful new servers).

 

- No doubt linked to the above, the mounted filesystem hangs
indefinitely when accessed. I tried an "ls -a" on the mounted filesystem,
for example, which hangs forever. I tested this by mounting a brick system
to itself and to a client which is not a brick and the same behavior was
observed. Both were glusterfs mounts. 

 

There is nothing special about my deployment except for the use
of transport = tcp,rdma. I am running on Ethernet now but will be migrating
to Infiniband after this is debugged. 

 

Thanks for any advice,

~Mike C.

___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] hanging of mounted filesystem (3.3.1)

2013-01-31 Thread Michael Colonno
 

Hi All ~

 

I created an eight-brick gluster 3.3.1 volume (2x replication)
on eight CentOS 6.3 x64 systems. I was able to form and start the volume
without issue. I was also able to mount it through /etc/fstab as a native
glusterfs mount. I have a couple questions issues at this point:

 

- On a client machine, "glusterfs" is not recognized as a valid
type unless gluster-server is installed. This seems to contradict the
documentation - wanted to make sure I'm not doing something wrong. More a
clarification than issue here. 

 

- The glusterfs process is taking between 50% and 80% CPU on
both the brick and client systems (these are fairly powerful new servers).

 

- No doubt linked to the above, the mounted filesystem hangs
indefinitely when accessed. I tried an "ls -a" on the mounted filesystem,
for example, which hangs forever. I tested this by mounting a brick system
to itself and to a client which is not a brick and the same behavior was
observed. Both were glusterfs mounts. 

 

There is nothing special about my deployment except for the use
of transport = tcp,rdma. I am running on Ethernet now but will be migrating
to Infiniband after this is debugged. 

 

Thanks for any advice,

~Mike C.

___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] New version of UFO - is there a new HOWTO?

2013-01-31 Thread Shawn Heisey

On 1/31/2013 12:36 PM, Kaleb Keithley wrote:

Not sure if you saw this in #gluster on IRC.

The other work-around for F18 is to delete 
/etc/swift/{account,container,object}-server.conf before starting UFO.

With that my UFO set-up works as it did in F17.


Still no joy.

http://fpaste.org/SOYM/

Same thing happens if I remove all the packages, get rid of /etc/swift, 
and reinstall.  I can confirm that the file referenced in the last part 
of the paste (/etc/swift/object.ring.gz) did not exist at any point.  A 
previous installation of 3.3.1-2 on F17 has both .builder and .ring.gz 
files.


http://fpaste.org/jIm5/

Thanks,
Shawn
___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] NFS availability

2013-01-31 Thread Jeff Darcy

On 01/31/2013 02:57 PM, Stephan von Krawczynski wrote:

You are asking in the wrong direction. The simple question is: is
there any dynamic configuration equally safe than a local config
file?


Yes, because inconsistent configurations are a danger too.


There is no probability, either you are a dead client or a working
one.


That's a total non sequitur.  As soon as there's more than one state,
there are probabilities of being in each one.  As it turns out, there is
also a third possibility - that you are a malfunctioning or incorrectly
configured client.  Fail-stop errors are the easy ones.


It wouldn't be useful to release a network filesystem that drops dead
in case of network errors.


Every network filesystem, or other distributed system of any kind, will
fail if there are *sufficient* errors.  Some are better than others at
making that number larger, some are better than others with respect to
the consequence of exceeding that threshold, but all will fail
eventually.  GlusterFS does handle a great many kinds of failure that
you probably can't even imagine, but neither it nor any other system can
handle all possible failures.


Most common network errors are not a matter of design, but of dead
iron.


It's usually both - a design that is insufficiently tolerant of
component failure, plus a combination of component failures that exceeds
that tolerance.  You seem to have a very high standard for filesystems
continuing to maintain 100% functionality - and I suppose 100%
performance as well - if there's any possibility whatsoever that they
could do so.  Why don't you apply that same standard to the part of the
system that you're responsible for designing?  Running any distributed
system on top of a deficient network infrastructure will lead only to
disappointment.

___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] NFS availability

2013-01-31 Thread Stephan von Krawczynski
On Thu, 31 Jan 2013 14:17:32 -0500
Jeff Darcy  wrote:

> There is *always* at least one situation, however unlikely, where you're 
> busted.  Designing reliable systems is always about probabilities.  If 
> none of the solutions mentioned so far suffice for you, there are still 
> others that don't involve sacrificing the advantages of dynamic 
> configuration.  If your network is so FUBAR that you have trouble 
> reaching any server to fetch a volfile, then it probably wouldn't do you 
> any good to have one locally because you wouldn't be able to reach those 
> servers for I/O anyway.  You'd be just asking for split brain and other 
> problems.  Redesigning the mount is likely to yield less benefit than 
> redesigning the network that's susceptible to such failures.

You are asking in the wrong direction. The simple question is: is there any
dynamic configuration equally safe than a local config file?
If your local fs is dead, then you are really dead. But if it is alive you
have a config. And that's about it. You need no working DNS, no poisoned cache
and no special server that must not fail.
Everything with less security is inacceptable. There is no probability, either
you are a dead client or a working one.
And if you really want to include the network as a question. I would expect
the gluster client-server and server-server protocols to accept network
failure as a default case. It wouldn't be useful to release a network
filesystem that drops dead in case of network errors. If there is some chance
to survive it should be able to do so and still work.
Most common network errors are not a matter of design, but of dead iron. 

-- 
Regards,
Stephan
___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] New version of UFO - is there a new HOWTO?

2013-01-31 Thread Kaleb Keithley
Not sure if you saw this in #gluster on IRC.

The other work-around for F18 is to delete 
/etc/swift/{account,container,object}-server.conf before starting UFO.

With that my UFO set-up works as it did in F17.

- Original Message -
From: "Kaleb S. KEITHLEY" 

On 01/29/2013 11:02 PM, Shawn Heisey wrote:
> This is in the proxy-server.conf file:
>
> [pipeline:main]
> pipeline = healthcheck cache authtoken keystone proxy-server

proxy-server.conf-gluster has tempauth. That comes from the gluster-ufo 
rpm. Keystone auth is coming in a future release. Soon hopefully

And yes, the conflict between glusterfs-ufo and glusterfs-swift-* is a 
bug. You can work around it with:
% yumdownloader glusterfs-ufo
% rpm -i --force glustefs-ufo-3.3.1-8.fc18.noarch.rpm

>
> One thing to note: I was not able to install glusterfs-ufo - it has
> config file conflicts with the other packages - one of which
> (glusterfs-swift) it actually depends on.  Does the -ufo package have
> the config file with tempauth?
>
> Thanks,
> Shawn
>
> On 1/29/2013 8:43 PM, Kaleb Keithley wrote:
>> 3.3.1-8 in f18 still uses tempauth.
>>
>> - Original Message -
>> From: "Shawn Heisey" 
>> To: gluster-users@gluster.org
>> Sent: Tuesday, January 29, 2013 8:21:11 PM
>> Subject: [Gluster-users] New version of UFO - is there a new HOWTO?
>>
>> I just installed glusterfs-swift 3.3.1 on a couple of Fedora 18 servers.
>>This is based on swift 1.7.4 and has keystone in the config.  I had
>> experimented with the one based on swift 1.4.8 and tempauth and had some
>> problems with it.  The HOWTO I can find is still for the old one.  Is
>> there an updated one?
>>
>> I would also need to find some instructions on setting up keystone from
>> scratch for use with gluster-swift.
>>
>> Thanks,
>> Shawn
>> ___
>> Gluster-users mailing list
>> Gluster-users@gluster.org
>> http://supercolony.gluster.org/mailman/listinfo/gluster-users
>> ___
>> Gluster-users mailing list
>> Gluster-users@gluster.org
>> http://supercolony.gluster.org/mailman/listinfo/gluster-users
>
> ___
> Gluster-users mailing list
> Gluster-users@gluster.org
> http://supercolony.gluster.org/mailman/listinfo/gluster-users


-- 

Kaleb
___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users
___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] NFS availability

2013-01-31 Thread Jeff Darcy

On 01/31/2013 01:59 PM, Stephan von Krawczynski wrote:

You don't want to use DNS in an environment where security is your first rule.
If your DNS drops dead your setup is dead. Not very promising ...
The basic goal of glusterfs has been to secure data by replicating it.
Data distribution is really not interesting for us. Now you say "go and
replicate your data for security, but use DNS to secure your setup".
???
You really seem to like Domino-setups. DNS dead => everything dead.


This vulnerability can be reduced by using multiple DNS servers, and 
even further by using local caches/proxies which can still do 
round-robin for you.  There have also been solutions mentioned that 
don't rely on DNS, so talking about DNS unreliability is a total red 
herring.



   * The mount option backupvolfile-server. An fstab entry like
 "server1:myvol /mnt/myvol glusterfs backupvolfile-server=server2 0
 0" will allow the mount command to try server2 if server1 does not
 mount successfully.


And how many backup servers do you want to name in your fstab? In fact you
have to name all your servers because else there will always be at least one
situation you are busted.


There is *always* at least one situation, however unlikely, where you're 
busted.  Designing reliable systems is always about probabilities.  If 
none of the solutions mentioned so far suffice for you, there are still 
others that don't involve sacrificing the advantages of dynamic 
configuration.  If your network is so FUBAR that you have trouble 
reaching any server to fetch a volfile, then it probably wouldn't do you 
any good to have one locally because you wouldn't be able to reach those 
servers for I/O anyway.  You'd be just asking for split brain and other 
problems.  Redesigning the mount is likely to yield less benefit than 
redesigning the network that's susceptible to such failures.


___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] NFS availability

2013-01-31 Thread Stephan von Krawczynski
On Thu, 31 Jan 2013 09:07:50 -0800
Joe Julian  wrote:

> On 01/31/2013 08:38 AM, Stephan von Krawczynski wrote:
> > On Thu, 31 Jan 2013 12:47:30 +
> > Brian Candler  wrote:
> >
> >> On Thu, Jan 31, 2013 at 09:18:26AM +0100, Stephan von Krawczynski wrote:
>  The client will still fail (in most cases) since host1 (if I follow you) 
>  is
>  part of the gluster groupset. Certainly if it's a distributed-only, 
>  maybe not
>  if it's a dist/repl gluster.  But if host1 goes down, the client will 
>  not be
>  able to find a gluster vol to mount.
> >>> For sure it will not fail if replication is used.
> >> Aside: it will *fail* if the client reboots, and /etc/fstab has
> >> server1:/volname, and server1 is the one which failed.
> > Well, this is exactly the reason we generally deny to fetch the volfile from
> > the server. This whole idea is obvious nonsense for exactly the reason you
> > described.
> >
> That doesn't lend me much confidence in your expertise with regard to 
> your other recommendations, Stephan.
> 
> There are two good ways to make this work even if a server is down:
> 
>   * Round robin DNS. A hostname (ie. glusterfs.domain.dom) with multiple
> A records that point to all your servers. Using that hostname in
> fstab will allow the client to roll over to the additional servers
> in the event the first one it gets is not available (ie.
> "glusterfs.domain.dom:myvol /mnt/myvol glusterfs defaults 0 0").

You don't want to use DNS in an environment where security is your first rule.
If your DNS drops dead your setup is dead. Not very promising ...
The basic goal of glusterfs has been to secure data by replicating it.
Data distribution is really not interesting for us. Now you say "go and
replicate your data for security, but use DNS to secure your setup".
???
You really seem to like Domino-setups. DNS dead => everything dead.

>   * The mount option backupvolfile-server. An fstab entry like
> "server1:myvol /mnt/myvol glusterfs backupvolfile-server=server2 0
> 0" will allow the mount command to try server2 if server1 does not
> mount successfully.

And how many backup servers do you want to name in your fstab? In fact you
have to name all your servers because else there will always be at least one
situation you are busted. 
 
> This whole idea is obvious experience and forethought, not nonsense. By 
> having a management service that provides configuration, on-the-fly 
> configuration changes are possible. If one "denies to fetch the volfile" 
> one cripples their cluster's flexibility.

I don't know what kind of setups you drive. In our environment we don't want
to fiddle around with fs configs. We want them to work as expected even if
other parts of the total setup fall apart. Flexibility in our world means you
can do widespread types of configurations. It does not mean we switch the
running configs every day only because gluster is so flexible.

-- 
Regards,
Stephan

___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] Extremely high CPU load on one server

2013-01-31 Thread Dan Bretherton

On 01/31/2013 05:37 PM, gluster-users-requ...@gluster.org wrote:

Date: Thu, 31 Jan 2013 15:28:25 + (UTC)
From: Adri? Garc?a-Alz?rriz 
To: gluster-users@gluster.org
Subject: Re: [Gluster-users] Extremely high CPU load on one server
Message-ID: 
Content-Type: text/plain; charset="us-ascii"

El dia Thu, 31 Jan 2013 15:02:25 +, en/na Dan Bretherton va escriure:

Can anyone suggest a way to troubleshoot this problem?  The rebalance
logs don't show anything unusual but glustershd.log has a lot of
metadata split-brain warnings.   The brick logs are full of scary
looking warnings but none flagged 'E' or 'C'.  The trouble is that I see
messages like these on all the servers, and I can find nothing unusual
about the server with a CPU load of 70.  Users are complaining about
very poor performance, which has been going on or several weeks, so I
must at least find a work-around that allows people to work normally.

-Dan.

What's the output for your gluster volume info command?
What kind of disks are you using for?
How is the wait parameter in top?

Hello- Here is the information you requested; thanks.


What's the output for your gluster volume info command?
There are ten volumes in the cluster, and the one I'm having the most 
trouble with is shown below.


[root@bdan11 ~]# gluster volume info atmos

Volume Name: atmos
Type: Distributed-Replicate
Volume ID: a4a7774e-cf4d-4a3e-9477-5c3d1b0efd93
Status: Started
Number of Bricks: 16 x 2 = 32
Transport-type: tcp
Bricks:
Brick1: bdan0.nerc-essc.ac.uk:/local/glusterfs
Brick2: bdan1.nerc-essc.ac.uk:/local/glusterfs
Brick3: bdan2.nerc-essc.ac.uk:/local/glusterfs
Brick4: bdan3.nerc-essc.ac.uk:/local/glusterfs
Brick5: bdan4.nerc-essc.ac.uk:/atmos/glusterfs
Brick6: bdan5.nerc-essc.ac.uk:/atmos/glusterfs
Brick7: bdan6.nerc-essc.ac.uk:/local/glusterfs
Brick8: bdan7.nerc-essc.ac.uk:/local/glusterfs
Brick9: bdan8.nerc-essc.ac.uk:/atmos/glusterfs
Brick10: bdan9.nerc-essc.ac.uk:/atmos/glusterfs
Brick11: bdan10.nerc-essc.ac.uk:/atmos/glusterfs
Brick12: bdan11.nerc-essc.ac.uk:/atmos/glusterfs
Brick13: bdan12.nerc-essc.ac.uk:/atmos/glusterfs
Brick14: bdan13.nerc-essc.ac.uk:/atmos/glusterfs
Brick15: bdan10.nerc-essc.ac.uk:/atmos2/glusterfs
Brick16: bdan11.nerc-essc.ac.uk:/atmos2/glusterfs
Brick17: bdan12.nerc-essc.ac.uk:/atmos2/glusterfs
Brick18: bdan13.nerc-essc.ac.uk:/atmos2/glusterfs
Brick19: bdan4.nerc-essc.ac.uk:/atmos2/glusterfs
Brick20: bdan5.nerc-essc.ac.uk:/atmos2/glusterfs
Brick21: bdan12.nerc-essc.ac.uk:/atmos3/glusterfs
Brick22: bdan13.nerc-essc.ac.uk:/atmos3/glusterfs
Brick23: bdan12.nerc-essc.ac.uk:/atmos4/glusterfs
Brick24: bdan13.nerc-essc.ac.uk:/atmos4/glusterfs
Brick25: bdan12.nerc-essc.ac.uk:/atmos5/glusterfs
Brick26: bdan13.nerc-essc.ac.uk:/atmos5/glusterfs
Brick27: pegasus.nerc-essc.ac.uk:/atmos/glusterfs
Brick28: bdan14.nerc-essc.ac.uk:/atmos/glusterfs
Brick29: bdan15.nerc-essc.ac.uk:/atmos/glusterfs
Brick30: bdan16.nerc-essc.ac.uk:/atmos/glusterfs
Brick31: bdan15.nerc-essc.ac.uk:/atmos2/glusterfs
Brick32: bdan16.nerc-essc.ac.uk:/atmos2/glusterfs
Options Reconfigured:
nfs.enable-ino32: on
server.allow-insecure: on
performance.stat-prefetch: off
performance.quick-read: off
cluster.min-free-disk: 338GB
nfs.rpc-auth-allow: 192.171.166.*,134.225.100.*
features.quota: off


What kind of disks are you using for?
They are Hitachi Deskstar 7200 RPM 2TB SATA drives, attached to an Areca 
ARC1880 RAID controller.  I'm aware that it's an unusual choice of drive 
but the server vendor swears by them.  I have another pair of servers 
from the same vendor with 1TB Hitachi Deskstars and they have not given 
any GlusterFS related trouble.


The bricks are logical volumes on top of RAID-6 provided by the Areca.  
On this particular server they are formatted ext4, but I recently 
started using XFS for more recently added bricks on newer servers.  The 
bricks vary in size from 3.3TB to 500GB, depending on which volume they 
belong to.  I use smaller bricks for smaller volumes so that the brick 
size is an appropriate increment of volume growth.



How is the wait parameter in top?
There isn't much I/O wait as far as I can gather.  Here is a view 
showing the top few processes.


top - 18:22:58 up  1:59,  1 user,  load average: 64.25, 64.36, 64.14
Tasks: 218 total,   9 running, 209 sleeping,   0 stopped,   0 zombie
Cpu(s): 78.2%us, 20.7%sy,  0.0%ni,  0.9%id,  0.0%wa,  0.0%hi, 0.1%si,  
0.0%st

Mem:  12290244k total,  6639128k used,  5651116k free,  2752944k buffers
Swap: 14352376k total,0k used, 14352376k free,  1441712k cached

  PID USER  PR  NI  VIRT  RES  SHR S %CPU %MEMTIME+ COMMAND
 3661 root  15   0  289m  42m 2168 R 85.0  0.4 101:54.33 glusterfsd
 3697 root  15   0  304m  37m 2112 S 83.1  0.3  75:53.43 glusterfsd
 3655 root  15   0  289m  42m 2160 S 79.4  0.4  97:11.52 glusterfsd
 3685 root  15   0  550m  62m 2100 S 77.5  0.5 109:07.53 glusterfsd
 3691 root  15   0  483m  40m 2100 S 75.6  0.3  76:10.93 glusterfsd
 3679 root  

Re: [Gluster-users] NFS availability

2013-01-31 Thread James
On Thu, 2013-01-31 at 09:07 -0800, Joe Julian wrote:
> That doesn't lend me much confidence in your expertise with regard to 
> your other recommendations, Stephan.
> 
> There are two good ways to make this work even if a server is down:
I'd just like to add one more option to the already excellent list that
Joe has provided. It happens to be my favourite: VRRP. You can use
something like Keepalived or even cman+rgmanager+an ip resource, to
manager a "virtual IP". Each of your servers has an IP, and then this
extra VIP moves around automatically based on what is actually up, and
where you want the mount to come from.

If you want to be particularly fastidious, then I suppose you could even
create a second VIP (which defaults to a different "base" node, and
which you use the below "backupvolfile-server" option.

VRRP has the added benefit that the failover happens within seconds, and
can be set manually if for some reason you want to choose which volfile
server you have.

HTH,
James


> 
>   * Round robin DNS. A hostname (ie. glusterfs.domain.dom) with
> multiple
> A records that point to all your servers. Using that hostname in
> fstab will allow the client to roll over to the additional servers
> in the event the first one it gets is not available (ie.
> "glusterfs.domain.dom:myvol /mnt/myvol glusterfs defaults 0 0").
>   * The mount option backupvolfile-server. An fstab entry like
> "server1:myvol /mnt/myvol glusterfs backupvolfile-server=server2 0
> 0" will allow the mount command to try server2 if server1 does not
> mount successfully.
> 
> This whole idea is obvious experience and forethought, not nonsense.
> By 
> having a management service that provides configuration, on-the-fly 
> configuration changes are possible. If one "denies to fetch the
> volfile" 
> one cripples their cluster's flexibility.



signature.asc
Description: This is a digitally signed message part
___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

[Gluster-users] hanging of mounted filesystem (3.3.1)

2013-01-31 Thread Michael Colonno
Hi All ~

I created an eight-brick gluster 3.3.1 volume (2x replication) on
eight CentOS 6.3 x64 systems. I was able to form and start the volume
without issue. I was also able to mount it through /etc/fstab as a native
glusterfs mount. I have a couple questions issues at this point:

- On a client machine, "glusterfs" is not recognized as a valid type
unless gluster-server is installed. This seems to contradict the
documentation - wanted to make sure I'm not doing something wrong. More a
clarification than issue here. 

- The glusterfs process is taking between 50% and 80% CPU on both
the brick and client systems (these are fairly powerful, brand new servers).


- No doubt linked to the above, the mounted filesystem hangs
indefinitely when accessed. I tried an "ls -a" on the mounted filesystem,
for example, which hangs forever. I tested this by mounting a brick system
to itself and to a client which is not a brick and the same behavior was
observed. Both were glusterfs mounts. 

There is nothing special about my deployment except for the use of
transport = tcp,rdma. I am running on Ethernet now but will be migrating to
Infiniband after this is debugged. 

Thanks for any advice,
~Mike C. 




___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] NFS availability

2013-01-31 Thread Joe Julian

On 01/31/2013 08:38 AM, Stephan von Krawczynski wrote:

On Thu, 31 Jan 2013 12:47:30 +
Brian Candler  wrote:


On Thu, Jan 31, 2013 at 09:18:26AM +0100, Stephan von Krawczynski wrote:

The client will still fail (in most cases) since host1 (if I follow you) is
part of the gluster groupset. Certainly if it's a distributed-only, maybe not
if it's a dist/repl gluster.  But if host1 goes down, the client will not be
able to find a gluster vol to mount.

For sure it will not fail if replication is used.

Aside: it will *fail* if the client reboots, and /etc/fstab has
server1:/volname, and server1 is the one which failed.

Well, this is exactly the reason we generally deny to fetch the volfile from
the server. This whole idea is obvious nonsense for exactly the reason you
described.

That doesn't lend me much confidence in your expertise with regard to 
your other recommendations, Stephan.


There are two good ways to make this work even if a server is down:

 * Round robin DNS. A hostname (ie. glusterfs.domain.dom) with multiple
   A records that point to all your servers. Using that hostname in
   fstab will allow the client to roll over to the additional servers
   in the event the first one it gets is not available (ie.
   "glusterfs.domain.dom:myvol /mnt/myvol glusterfs defaults 0 0").
 * The mount option backupvolfile-server. An fstab entry like
   "server1:myvol /mnt/myvol glusterfs backupvolfile-server=server2 0
   0" will allow the mount command to try server2 if server1 does not
   mount successfully.

This whole idea is obvious experience and forethought, not nonsense. By 
having a management service that provides configuration, on-the-fly 
configuration changes are possible. If one "denies to fetch the volfile" 
one cripples their cluster's flexibility.


___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] NFS availability

2013-01-31 Thread Stephan von Krawczynski
On Thu, 31 Jan 2013 12:47:30 +
Brian Candler  wrote:

> On Thu, Jan 31, 2013 at 09:18:26AM +0100, Stephan von Krawczynski wrote:
> > > The client will still fail (in most cases) since host1 (if I follow you) 
> > > is 
> > > part of the gluster groupset. Certainly if it's a distributed-only, maybe 
> > > not 
> > > if it's a dist/repl gluster.  But if host1 goes down, the client will not 
> > > be 
> > > able to find a gluster vol to mount.
> > 
> > For sure it will not fail if replication is used. 
> 
> Aside: it will *fail* if the client reboots, and /etc/fstab has
> server1:/volname, and server1 is the one which failed.

Well, this is exactly the reason we generally deny to fetch the volfile from
the server. This whole idea is obvious nonsense for exactly the reason you
described.

-- 
Regards,
Stephan
___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] Extremely high CPU load on one server

2013-01-31 Thread Adrià García-Alzórriz
El dia Thu, 31 Jan 2013 15:02:25 +, en/na Dan Bretherton va escriure:

> Can anyone suggest a way to troubleshoot this problem?  The rebalance
> logs don't show anything unusual but glustershd.log has a lot of
> metadata split-brain warnings.   The brick logs are full of scary
> looking warnings but none flagged 'E' or 'C'.  The trouble is that I see
> messages like these on all the servers, and I can find nothing unusual
> about the server with a CPU load of 70.  Users are complaining about
> very poor performance, which has been going on or several weeks, so I
> must at least find a work-around that allows people to work normally.
> 
> -Dan.

What's the output for your gluster volume info command?
What kind of disks are you using for?
How is the wait parameter in top?

signature.asc
Description: PGP signature
___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

[Gluster-users] Extremely high CPU load on one server

2013-01-31 Thread Dan Bretherton

Dear All-
I originally asked this question in another thread, but as it is a 
separate problem I thought it deserved its own thread.


I had to extend two volumes recently and the layout fixes for both of 
them are now running at the same time.  One server now has a load of 
over 70 most of the time (mostly glusterfsd), but none of the others 
seem to be particularly busy.  I restarted the server in question but 
the CPU load quickly went up to 70 again.  I can't see any particular 
reason why this one server should be so badly affected by the layout 
fixing processes.  It isn't a particularly big server, with only five 
3TB bricks involved in the two volumes that were extended.  One 
possibility is that a lot of batch jobs on our compute cluster are 
accessing the same set of files that just happen to be on this one 
server, which could potentially happen because I have not been able to 
rebalance the files in the storage volumes for a long time after the 
last attempt resulted in data corruption (in GlusterFS version 3.2).  
However, poorly distributed files certainly isn't the whole story, 
because even when there is not much running on the compute cluster the 
storage server load remains extremely high.


Can anyone suggest a way to troubleshoot this problem?  The rebalance 
logs don't show anything unusual but glustershd.log has a lot of 
metadata split-brain warnings.   The brick logs are full of scary 
looking warnings but none flagged 'E' or 'C'.  The trouble is that I see 
messages like these on all the servers, and I can find nothing unusual 
about the server with a CPU load of 70.  Users are complaining about 
very poor performance, which has been going on or several weeks, so I 
must at least find a work-around that allows people to work normally.


-Dan.

--
Mr. D.A. Bretherton
Computer System Manager
Environmental Systems Science Centre (ESSC)
Department of Meteorology
Harry Pitt Building
3 Earley Gate
University of Reading
Reading, RG6 7BE (or RG6 6AL for postal service deliveries)
UK
Tel. +44 118 378 5205, Fax: +44 118 378 6413

___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] NFS availability

2013-01-31 Thread Brian Candler
On Thu, Jan 31, 2013 at 09:18:26AM +0100, Stephan von Krawczynski wrote:
> > The client will still fail (in most cases) since host1 (if I follow you) is 
> > part of the gluster groupset. Certainly if it's a distributed-only, maybe 
> > not 
> > if it's a dist/repl gluster.  But if host1 goes down, the client will not 
> > be 
> > able to find a gluster vol to mount.
> 
> For sure it will not fail if replication is used. 

Aside: it will *fail* if the client reboots, and /etc/fstab has
server1:/volname, and server1 is the one which failed.
___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] NFS availability

2013-01-31 Thread Stephan von Krawczynski
On Wed, 30 Jan 2013 20:44:52 -0800
harry mangalam  wrote:

> On Thursday, January 31, 2013 11:28:04 AM glusterzhxue wrote:
> > Hi all,
> > As is known to us all, gluster provides NFS mount. However, if the mount
> > point fails, clients will lose connection to Gluster. While if we use
> > gluster native client, this fail will have no effect on clients. For
> > example:
>  mount -t glusterfs host1:/vol1  /mnt
> > 
> > If host1 goes down for some reason, client works still, it has no sense
> > about the failure(suppose we have multiple gluster servers).  
> 
> The client will still fail (in most cases) since host1 (if I follow you) is 
> part of the gluster groupset. Certainly if it's a distributed-only, maybe not 
> if it's a dist/repl gluster.  But if host1 goes down, the client will not be 
> able to find a gluster vol to mount.

For sure it will not fail if replication is used. 
 
> > However, if
> > we use the following:
>  mount -t nfs -o vers=3   host1:/vol1 /mnt
> 
> > If host1 failed, client will lose connection to gluster servers.
> 
> If the client was mounting the glusterfs via a re-export from an intermediate 
> host, you might be able to failover to another intermediate NFS server, but 
> if 
> it was a gluster host, it would fail due to the reasons above.

kernel-nfs _may_ failover from server A to server B if B takes the original
server IP and some requirements are met.
You don't need an intermediate (re-exporting) server for this.

-- 
Regards,
Stephan 
___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users