Re: [Gluster-users] Dose Gluster 3.1 support authorisation control and how to do

2011-01-10 Thread W.C Lee
Hi, HU

Thank for your help.
I tried to use your example(1 server ,1 Client) to test authentication 
function, it's work.

But I tried to test it in replication mode (multi-node),FUSE mounting work, but 
NFS didn't.
Any node can mount volume via NFS. 

And 
Following is my config.

 26: volume gluster-new-volume-server
 27: type protocol/server
 28: option transport-type tcp
 29: option auth.addr./mnt/gluster1.allow 
10.18.14.240,10.18.14.248,10.18.14.241,10.18.14.242,10.18.14.243
 30: subvolumes /mnt/gluster1
 31: end-volume


After starting volume, log showed below:

+--+
[2011-01-11 01:07:54.188695] E [authenticate.c:235:gf_authenticate] auth: no 
authentication module is interested in accepting remote-client (null)
[2011-01-11 01:07:54.188716] E [server-handshake.c:545:server_setvolume] 
gluster-new-volume-server: Cannot authenticate client from 127.0.0.1:1017
[2011-01-11 01:07:55.264728] I [server-handshake.c:535:server_setvolume] 
gluster-new-volume-server: accepted client from 10.18.14.241:995
[2011-01-11 01:07:55.267990] I [server-handshake.c:535:server_setvolume] 
gluster-new-volume-server: accepted client from 10.18.14.242:1012
[2011-01-11 01:07:55.272025] I [server-handshake.c:535:server_setvolume] 
gluster-new-volume-server: accepted client from 10.18.14.243:996


Do you know is it necessary to set 127.0.0.1 to allow list?
And it can't use host real ip (10.18.14.240) ?

But even if I used 127.0.0.1 to replace 10.18.14.240, NFS authentication 
control still not work. 



-Original message-
From:HU Zhong hz02...@gmail.com
To:wei.ch...@m2k.com.tw
Cc:gluster-users gluster-users@gluster.org
Date:Mon, 10 Jan 2011 11:36:00 +0800
Subject:Re: [Gluster-users] Dose Gluster 3.1 support authorisation control and 
how to do


Hi, Cheng

I think you did the configuration in the wrong place. Instead of
/etc/glusterd/nfs/nfs-server.vol, you need to modify files
under /etc/glusterd/vols/.

As a simple example, consider a one-server-one-client system, both
server and client are one machine(localhost, ip:192.168.4.112), and
export directory /home/huz/share for sharing, the client wants to mount
it on /home/huz/mnt.

if i modify default
configuration 
/etc/glusterd/vols/testvol/testvol.192.168.4.112.home-huz-share.vol

from
..
26 volume testvol-server
27 type protocol/server
28 option transport-type tcp
29 option auth.addr./home/huz/share.allow *
30 subvolumes /home/huz/share
31 end-volume

to
..
26 volume testvol-server
27 type protocol/server
28 option transport-type tcp
29 option auth.addr./home/huz/share.reject *
30 subvolumes /home/huz/share
31 end-volume

the mount command will fail:
$sudo mount -o mountproto=tcp -t nfs localhost:/testvol /home/huz/mnt
mount.nfs: mounting localhost:/testvol failed, reason given by server:
  No such file or directory

and the log shows that the authentication error.
11-01-10 11:09:58.203600] E
[client-handshake.c:786:client_setvolume_cbk] testvol-client-0:
SETVOLUME on remote-host failed: Authentication failed

change reject to allow, the mount operation will be ok.

you can configure you own ip rule. As for how to use ip auth and
usrname/password auth, you can check the attachment. It's a
documentation file under the directory doc of glusterfs src project.

On Sun, 2011-01-09 at 22:31 +0800, 第二信箱 wrote:
 Hi, HU:
 Thanks for your help.
 
 I have the following environment:
 Gluster 3.1.1
 Volume Name: gluster-volume
 Type: Distributed-Replicate
 Status: Started
 Number of Bricks: 2 x 2 = 4
 Transport-type: tcp
 Bricks:
 Brick1: gluster1:/mnt/gluster1
 Brick2: gluster2:/mnt/gluster2
 Brick3: gluster3:/mnt/gluster3
 Brick4: gluster4:/mnt/gluster4
 
 
 I want to use authenticate module by your suggestion.
 The way I used below:
 1. Stop Volume
 2. Edit /etc/glusterd/nfs/nfs-server.vol on Brick1(Gluster1)
 3. Modify and Add  From
volume nfs-server
 type nfs/server
 option nfs.dynamic-volumes on
 option rpc-auth.addr.gluster-volume.allow *
 option nfs3.gluster-volume.volume-id 907941d9-6950-425b-
 b3d5-4e43dd420d9e
 subvolumes gluster-volume
 end-volume
 
 to 
 
 volume nfs-server
 type nfs/server
 option nfs.dynamic-volumes on
 option rpc-auth.addr.gluster-volume.allow  10.18.14.1
 option auth.addr.gluster-volume.allow 10.18.14.1
 option nfs3.gluster-volume.volume-id
 907941d9-6950-425b-b3d5-4e43dd420d9e
 subvolumes gluster-volume
 end-volume
 
 4.Start Volume
 
 -- But I still be able to mount volume from 10.18.14.2 by NFS.
 
 Anything I missed or be wrong?
 
 And I find 
 
 A. After I started volume , nfs-server.vol was initialed to option
 rpc-auth.addr.gluster-volume.allow * .
 B. 4 nodes all have /etc/glusterd/nfs/nfs-server.vol , Should I Edit
 every .vol file on 4 nodes?
 
 
 
 
 
 
 -Original message-
 From:HU 

Re: [Gluster-users] Read-only volumes - timeframe for features/filter?

2011-01-10 Thread Burnash, James
Hello.

Has anyone seen error messages like this in /var/log/glusterfs/nfs.log:

tail /var/log/glusterfs/nfs.log
[2011-01-10 14:22:55.859066] I 
[afr-self-heal-common.c:1526:afr_self_heal_completion_cbk] pfs-ro1-replicate-3: 
background  meta-data data self-heal completed on /
[2011-01-10 14:22:55.859084] I 
[afr-self-heal-common.c:1526:afr_self_heal_completion_cbk] pfs-ro1-replicate-5: 
background  meta-data data self-heal completed on /
[2011-01-10 14:22:57.786088] E [nfs3.c:4346:nfs3_fsstat] nfs-nfsv3: Failed to 
map FH to vol
[2011-01-10 14:22:59.355112] E [nfs3.c:4346:nfs3_fsstat] nfs-nfsv3: Failed to 
map FH to vol
[2011-01-10 14:22:59.415732] E [nfs3.c:675:nfs3_getattr] nfs-nfsv3: Failed to 
map FH to vol
[2011-01-10 14:22:59.455029] E [nfs3.c:675:nfs3_getattr] nfs-nfsv3: Failed to 
map FH to vol
[2011-01-10 14:23:01.800751] E [nfs3.c:675:nfs3_getattr] nfs-nfsv3: Failed to 
map FH to vol
[2011-01-10 14:23:02.127233] E [nfs3.c:675:nfs3_getattr] nfs-nfsv3: Failed to 
map FH to vol
[2011-01-10 14:23:07.834044] E [nfs3.c:4346:nfs3_fsstat] nfs-nfsv3: Failed to 
map FH to vol
[2011-01-10 14:23:09.478852] E [nfs3.c:4346:nfs3_fsstat] nfs-nfsv3: Failed to 
map FH to vol
[2011-01-10 14:40:18.558072] E 
[afr-self-heal-metadata.c:524:afr_sh_metadata_fix] pfs-ro1-replicate-5: Unable 
to self-heal permissions/ownership of '/' (possible split-brain). Please fix 
the file on all backend volumes

Mount is done with this command:
mount -v -t nfs -o soft,rsize=16384,wsize=16384 jc1lpfsnfsro:/pfs-ro1 /pfs1

Command line being executed is:

rsync -av  --progress /pfs1/online_archive/2010 .

This CentOS 5.5 x86_64, GlusterFS 3.1.1. Currently configured:

gluster volume info

Volume Name: pfs-ro1
Type: Distributed-Replicate
Status: Started
Number of Bricks: 10 x 2 = 20
Transport-type: tcp
Bricks:
Brick1: jc1letgfs17-pfs1:/export/read-only/g01
Brick2: jc1letgfs18-pfs1:/export/read-only/g01
Brick3: jc1letgfs17-pfs1:/export/read-only/g02
Brick4: jc1letgfs18-pfs1:/export/read-only/g02
Brick5: jc1letgfs17-pfs1:/export/read-only/g03
Brick6: jc1letgfs18-pfs1:/export/read-only/g03
Brick7: jc1letgfs17-pfs1:/export/read-only/g04
Brick8: jc1letgfs18-pfs1:/export/read-only/g04
Brick9: jc1letgfs17-pfs1:/export/read-only/g05
Brick10: jc1letgfs18-pfs1:/export/read-only/g05
Brick11: jc1letgfs17-pfs1:/export/read-only/g06
Brick12: jc1letgfs18-pfs1:/export/read-only/g06
Brick13: jc1letgfs17-pfs1:/export/read-only/g07
Brick14: jc1letgfs18-pfs1:/export/read-only/g07
Brick15: jc1letgfs17-pfs1:/export/read-only/g08
Brick16: jc1letgfs18-pfs1:/export/read-only/g08
Brick17: jc1letgfs17-pfs1:/export/read-only/g09
Brick18: jc1letgfs18-pfs1:/export/read-only/g09
Brick19: jc1letgfs17-pfs1:/export/read-only/g10
Brick20: jc1letgfs18-pfs1:/export/read-only/g10

Thanks.

James Burnash
Unix Engineering.


DISCLAIMER:
This e-mail, and any attachments thereto, is intended only for use by the 
addressee(s) named herein and may contain legally privileged and/or 
confidential information. If you are not the intended recipient of this e-mail, 
you are hereby notified that any dissemination, distribution or copying of this 
e-mail, and any attachments thereto, is strictly prohibited. If you have 
received this in error, please immediately notify me and permanently delete the 
original and any copy of any e-mail and any printout thereof. E-mail 
transmission cannot be guaranteed to be secure or error-free. The sender 
therefore does not accept liability for any errors or omissions in the contents 
of this message which arise as a result of e-mail transmission.
NOTICE REGARDING PRIVACY AND CONFIDENTIALITY Knight Capital Group may, at its 
discretion, monitor and review the content of all e-mail communications. 
http://www.knight.com
___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


[Gluster-users] How can i turn on the Gluster Debug mode?

2011-01-10 Thread 공용준(yongjoon kong)/Cloud Comput ing 사업담당
Hello, all

How can I turn on the Debug mode on 3.1

When I start the volume by command line, is there any way to turn the debug 
message on brick side?

___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


[Gluster-users] infiniband speed

2011-01-10 Thread Christopher Hawkins
I am testing Infiniband for the first time. It seems that I should be able to 
get a lot more speed than I am with some pretty basic tests. Maybe someone 
running Infiniband can confirm that what I am seeing is way out of line, and/or 
help diagnose? 

I have two systems connected using 3.1.2qa3. With 3.1.1 infiniband wouldn't 
even start, it gave an error about unable to intialize rdma. But with the 
latest version and an upgrade to OFED 1.5.2, everything starts up with no 
errors and I can create a volume and mount it. 

The underlying Infiniband seems ok, and a basic ibv_rc_pingpong test shows I 
can move data pretty fast:
8192 bytes in 0.23 seconds = 2858.45 Mbit/sec
1 iters in 0.23 seconds = 22.93 usec/iter

So now I have two volumes created, one that uses tcp over a gig-e link and one 
that uses rdma. I mount them and do some file copy tests... And they are almost 
exactly the same? What? 

gluster volume info

Volume Name: test2_volume
Type: Replicate
Status: Started
Number of Bricks: 2
Transport-type: tcp
Bricks:
Brick1: bravo:/cluster/shadow/test2
Brick2: backup:/cluster/shadow/test2

Volume Name: test_volume
Type: Replicate
Status: Started
Number of Bricks: 2
Transport-type: rdma
Bricks:
Brick1: bravo:/cluster/shadow/test
Brick2: backup:/cluster/shadow/test

mount:
glusterfs#localhost:/test_volume on /mnt/test type fuse 
(rw,allow_other,default_permissions,max_read=131072)
glusterfs#localhost:/test2_volume on /mnt/test2 type fuse 
(rw,allow_other,default_permissions,max_read=131072)


time cp files.tar /mnt/test2/

real0m11.159s
user0m0.123s
sys 0m1.244s

files.tar is single file, 390MB, so this about 35MB/s. Fine for gig-e. 


time cp files.tar /mnt/test/

real0m5.656s
user0m0.116s
sys 0m0.962s

69MB/s... ehhh. Faster at least. On a few runs, this was not any faster at all. 
Maybe a cache effect? 


time cp -av /usr/src/kernels /mnt/test2/
real0m49.605s
user0m0.681s
sys 0m2.593s

kernels dir is 34MB of small files. The low latency of IB should really show an 
improvement here I thought. 
-

time cp -av /usr/src/kernels /mnt/test/

real0m56.046s
user0m0.625s
sys 0m2.675s

It took LONGER? That can't be right. 
--

And finally, this error is appearing in the rdma mount log every 3 seconds on 
both nodes:

[2011-01-10 19:46:56.728127] E [rdma.c:4428:tcp_connect_finish] 
test_volume-client-1: tcp connect to  failed (Connection refused)
[2011-01-10 19:46:59.738291] E [rdma.c:4428:tcp_connect_finish] 
test_volume-client-1: tcp connect to  failed (Connection refused)
[2011-01-10 19:47:02.748260] E [rdma.c:4428:tcp_connect_finish] 
test_volume-client-1: tcp connect to  failed (Connection refused)
[2011-01-10 19:47:05.758256] E [rdma.c:4428:tcp_connect_finish] 
test_volume-client-1: tcp connect to  failed (Connection refused)
[2011-01-10 19:47:08.768299] E [rdma.c:4428:tcp_connect_finish] 
test_volume-client-1: tcp connect to  failed (Connection refused)
[2011-01-10 19:47:11.778308] E [rdma.c:4428:tcp_connect_finish] 
test_volume-client-1: tcp connect to  failed (Connection refused)
[2011-01-10 19:47:14.788356] E [rdma.c:4428:tcp_connect_finish] 
test_volume-client-1: tcp connect to  failed (Connection refused)
[2011-01-10 19:47:17.798381] E [rdma.c:4428:tcp_connect_finish] 
test_volume-client-1: tcp connect to  failed (Connection refused)
[2011-01-10 19:47:20.808413] E [rdma.c:4428:tcp_connect_finish] 
test_volume-client-1: tcp connect to  failed (Connection refused)

But there are no restrictions in the config. Everything is allow *. So my 
questions are, can anyone else tell me what kind of basic file copy performance 
they see using IB? And what can I do to troubleshoot?

Thanks List and Devs, 

Chris
___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


Re: [Gluster-users] Dose Gluster 3.1 support authorisation control and how to do

2011-01-10 Thread HU Zhong
Hi

It seems that the node 10.18.14.240 runs both server and client. 
If not, write the server list and the client list here.
As you can see in the log, the node other than above are all accepted by
the server, so you can add both 10.18.14.240 and 127.0.0.1 to the
ip-allowed list to see whether it works or not.


On Tue, 2011-01-11 at 01:25 +0800, W.C Lee wrote: 
 Hi, HU
 
 Thank for your help.
 I tried to use your example(1 server ,1 Client) to test authentication 
 function, it's work.
 
 But I tried to test it in replication mode (multi-node),FUSE mounting work, 
 but NFS didn't.
 Any node can mount volume via NFS. 
 
 And 
 Following is my config.
 
  26: volume gluster-new-volume-server
  27: type protocol/server
  28: option transport-type tcp
  29: option auth.addr./mnt/gluster1.allow 
 10.18.14.240,10.18.14.248,10.18.14.241,10.18.14.242,10.18.14.243
  30: subvolumes /mnt/gluster1
  31: end-volume
 
 
 After starting volume, log showed below:
 
 +--+
 [2011-01-11 01:07:54.188695] E [authenticate.c:235:gf_authenticate] auth: no 
 authentication module is interested in accepting remote-client (null)
 [2011-01-11 01:07:54.188716] E [server-handshake.c:545:server_setvolume] 
 gluster-new-volume-server: Cannot authenticate client from 127.0.0.1:1017
 [2011-01-11 01:07:55.264728] I [server-handshake.c:535:server_setvolume] 
 gluster-new-volume-server: accepted client from 10.18.14.241:995
 [2011-01-11 01:07:55.267990] I [server-handshake.c:535:server_setvolume] 
 gluster-new-volume-server: accepted client from 10.18.14.242:1012
 [2011-01-11 01:07:55.272025] I [server-handshake.c:535:server_setvolume] 
 gluster-new-volume-server: accepted client from 10.18.14.243:996
 
 
 Do you know is it necessary to set 127.0.0.1 to allow list?
 And it can't use host real ip (10.18.14.240) ?
 
 But even if I used 127.0.0.1 to replace 10.18.14.240, NFS authentication 
 control still not work. 
 
 
 
 -Original message-
 From:HU Zhong hz02...@gmail.com
 To:wei.ch...@m2k.com.tw
 Cc:gluster-users gluster-users@gluster.org
 Date:Mon, 10 Jan 2011 11:36:00 +0800
 Subject:Re: [Gluster-users] Dose Gluster 3.1 support authorisation control 
 and how to do
 
 
 Hi, Cheng
 
 I think you did the configuration in the wrong place. Instead of
 /etc/glusterd/nfs/nfs-server.vol, you need to modify files
 under /etc/glusterd/vols/.
 
 As a simple example, consider a one-server-one-client system, both
 server and client are one machine(localhost, ip:192.168.4.112), and
 export directory /home/huz/share for sharing, the client wants to mount
 it on /home/huz/mnt.
 
 if i modify default
 configuration 
 /etc/glusterd/vols/testvol/testvol.192.168.4.112.home-huz-share.vol
 
 from
 ..
 26 volume testvol-server
 27 type protocol/server
 28 option transport-type tcp
 29 option auth.addr./home/huz/share.allow *
 30 subvolumes /home/huz/share
 31 end-volume
 
 to
 ..
 26 volume testvol-server
 27 type protocol/server
 28 option transport-type tcp
 29 option auth.addr./home/huz/share.reject *
 30 subvolumes /home/huz/share
 31 end-volume
 
 the mount command will fail:
 $sudo mount -o mountproto=tcp -t nfs localhost:/testvol /home/huz/mnt
 mount.nfs: mounting localhost:/testvol failed, reason given by server:
   No such file or directory
 
 and the log shows that the authentication error.
 11-01-10 11:09:58.203600] E
 [client-handshake.c:786:client_setvolume_cbk] testvol-client-0:
 SETVOLUME on remote-host failed: Authentication failed
 
 change reject to allow, the mount operation will be ok.
 
 you can configure you own ip rule. As for how to use ip auth and
 usrname/password auth, you can check the attachment. It's a
 documentation file under the directory doc of glusterfs src project.
 
 On Sun, 2011-01-09 at 22:31 +0800, 第二信箱 wrote:
  Hi, HU:
  Thanks for your help.
  
  I have the following environment:
  Gluster 3.1.1
  Volume Name: gluster-volume
  Type: Distributed-Replicate
  Status: Started
  Number of Bricks: 2 x 2 = 4
  Transport-type: tcp
  Bricks:
  Brick1: gluster1:/mnt/gluster1
  Brick2: gluster2:/mnt/gluster2
  Brick3: gluster3:/mnt/gluster3
  Brick4: gluster4:/mnt/gluster4
  
  
  I want to use authenticate module by your suggestion.
  The way I used below:
  1. Stop Volume
  2. Edit /etc/glusterd/nfs/nfs-server.vol on Brick1(Gluster1)
  3. Modify and Add  From
 volume nfs-server
  type nfs/server
  option nfs.dynamic-volumes on
  option rpc-auth.addr.gluster-volume.allow *
  option nfs3.gluster-volume.volume-id 907941d9-6950-425b-
  b3d5-4e43dd420d9e
  subvolumes gluster-volume
  end-volume
  
  to 
  
  volume nfs-server
  type nfs/server
  option nfs.dynamic-volumes on
  option rpc-auth.addr.gluster-volume.allow  10.18.14.1
  option auth.addr.gluster-volume.allow 10.18.14.1
  

[Gluster-users] 4 node replica 2 crash

2011-01-10 Thread rickytato rickytato
Hi,
I'm using for about 4 weeks a simple 2 node replica 2 cluster; I'm
using glusterfs 3.1.1 built on Dec  9 2010 15:41:32 Repository revision:
v3.1.1 .
I use it to provide images trought Nginx.
All works well.

Today i've added 2 new brick, and rebalance volume. For about 4 hours work,
after the Nginx hang; i've rebooted all server but nothings to do.

When I removed two brick all returns ok (I've manually copied file from
old brick to the original).


What's wrong?
___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


Re: [Gluster-users] Infiniband

2011-01-10 Thread Joe Landman

On 01/09/2011 10:06 PM, Bryan McGuire wrote:

Hello,

I am looking into GlusterFS as a high availability solution for our
email servers. I am new to Infiniband but find it could possibly
provide us with the necessary speed.


Hi Bryan

 We've done this for various ISP/email hosting customers.



Could someone describe what I would need in the way of Infiniband
hardware / software to complete the following.

Two to 4 front end email servers with each being a client and server
for the GlusterFS file system performing replication of the data.


Depends on what sort of machine you use for your front end, and what the 
software is.  Are these servers for IMAP/POP or are these postfix/exim etc?



I think I would need the necessary Infiniband cards in each server
along with an Infiniband switch. But do not have any background to
determine which or even if this is correct.


Simplest architecture is a small IB switch, IB HCA's in each node, an IB 
stack (OFED) in each node, a subnet manager (OpenSM) daemon, IB cables, 
and then Gluster built against your stack.  Make sure your time daemon 
is up, running, and correct between the nodes.  Give serious 
consideration to really fast disk in each node (fast IOP, so SSD, in 
RAID10).


Once you have that, you are ready to build your volumes (probably 
replicated distributed, so 4 way at least).


Do beware that there are some gotchas in configurating MTA/MUA software 
for Gluster.  Its doable, but such software often abhors shared storage 
of any sort.  You need to make sure that you turn down some of the 
caching.




Thanks in advance.

Bryan McGuire Senior Network Engineer NewNet 66

918.231.8063 bmcgu...@newnet66.org


Regards,

Joe

--
Joseph Landman, Ph.D
Founder and CEO
Scalable Informatics, Inc.
email: land...@scalableinformatics.com
web  : http://scalableinformatics.com
   http://scalableinformatics.com/sicluster
phone: +1 734 786 8423 x121
fax  : +1 866 888 3112
cell : +1 734 612 4615
___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


Re: [Gluster-users] infiniband speed

2011-01-10 Thread Joe Landman

On 01/10/2011 07:43 PM, Christopher Hawkins wrote:

I am testing Infiniband for the first time. It seems that I should be
able to get a lot more speed than I am with some pretty basic tests.
Maybe someone running Infiniband can confirm that what I am seeing is
way out of line, and/or help diagnose?


What is the underlying disk speed?

Try this:

# (on bravo)
dd if=/dev/zero of=/cluster/shadow/big.file bs=1M count=20k

This will write a 20GB file to the same partition.  We need to see how 
fast that write is (outside of cache)


Do the same test on the other machine.

Infiniband won't make slow disks fast.  So unless you are already 
getting a gigabyte or two per second (we do on our units), adding 
Infiniband won't get you any significant additional performance.


Put another way, you are bound by the slowest aspect of your storage system.

What are your underlying disks?  Are these individual drives?  RAID? 
What sort of RAID if it is a RAID?  How is it organized?


We tune our systems pretty hard, so we start with 2+GB/s for TB sized 
files before we ever touch the next stack up.  Each additional stack you 
traverse takes performance away (you lose it in stack inefficiency).  So 
you should start out with as fast a system as you possibly can, so you 
don't lose too much performance.


--
Joseph Landman, Ph.D
Founder and CEO
Scalable Informatics, Inc.
email: land...@scalableinformatics.com
web  : http://scalableinformatics.com
   http://scalableinformatics.com/sicluster
phone: +1 734 786 8423 x121
fax  : +1 866 888 3112
cell : +1 734 612 4615
___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


Re: [Gluster-users] Read-only volumes - timeframe for features/filter?

2011-01-10 Thread Shehjar Tikoo


Are any apps on the mount point erroring out with:

Invalid argument

or

Stale NFS file handle?

Burnash, James wrote:

Hello.

Has anyone seen error messages like this in /var/log/glusterfs/nfs.log:

tail /var/log/glusterfs/nfs.log
[2011-01-10 14:22:55.859066] I 
[afr-self-heal-common.c:1526:afr_self_heal_completion_cbk] pfs-ro1-replicate-3: 
background  meta-data data self-heal completed on /
[2011-01-10 14:22:55.859084] I 
[afr-self-heal-common.c:1526:afr_self_heal_completion_cbk] pfs-ro1-replicate-5: 
background  meta-data data self-heal completed on /
[2011-01-10 14:22:57.786088] E [nfs3.c:4346:nfs3_fsstat] nfs-nfsv3: Failed to 
map FH to vol
[2011-01-10 14:22:59.355112] E [nfs3.c:4346:nfs3_fsstat] nfs-nfsv3: Failed to 
map FH to vol
[2011-01-10 14:22:59.415732] E [nfs3.c:675:nfs3_getattr] nfs-nfsv3: Failed to 
map FH to vol
[2011-01-10 14:22:59.455029] E [nfs3.c:675:nfs3_getattr] nfs-nfsv3: Failed to 
map FH to vol
[2011-01-10 14:23:01.800751] E [nfs3.c:675:nfs3_getattr] nfs-nfsv3: Failed to 
map FH to vol
[2011-01-10 14:23:02.127233] E [nfs3.c:675:nfs3_getattr] nfs-nfsv3: Failed to 
map FH to vol
[2011-01-10 14:23:07.834044] E [nfs3.c:4346:nfs3_fsstat] nfs-nfsv3: Failed to 
map FH to vol
[2011-01-10 14:23:09.478852] E [nfs3.c:4346:nfs3_fsstat] nfs-nfsv3: Failed to 
map FH to vol
[2011-01-10 14:40:18.558072] E 
[afr-self-heal-metadata.c:524:afr_sh_metadata_fix] pfs-ro1-replicate-5: Unable 
to self-heal permissions/ownership of '/' (possible split-brain). Please fix 
the file on all backend volumes

Mount is done with this command:
mount -v -t nfs -o soft,rsize=16384,wsize=16384 jc1lpfsnfsro:/pfs-ro1 /pfs1

Command line being executed is:

rsync -av  --progress /pfs1/online_archive/2010 .

This CentOS 5.5 x86_64, GlusterFS 3.1.1. Currently configured:

gluster volume info

Volume Name: pfs-ro1
Type: Distributed-Replicate
Status: Started
Number of Bricks: 10 x 2 = 20
Transport-type: tcp
Bricks:
Brick1: jc1letgfs17-pfs1:/export/read-only/g01
Brick2: jc1letgfs18-pfs1:/export/read-only/g01
Brick3: jc1letgfs17-pfs1:/export/read-only/g02
Brick4: jc1letgfs18-pfs1:/export/read-only/g02
Brick5: jc1letgfs17-pfs1:/export/read-only/g03
Brick6: jc1letgfs18-pfs1:/export/read-only/g03
Brick7: jc1letgfs17-pfs1:/export/read-only/g04
Brick8: jc1letgfs18-pfs1:/export/read-only/g04
Brick9: jc1letgfs17-pfs1:/export/read-only/g05
Brick10: jc1letgfs18-pfs1:/export/read-only/g05
Brick11: jc1letgfs17-pfs1:/export/read-only/g06
Brick12: jc1letgfs18-pfs1:/export/read-only/g06
Brick13: jc1letgfs17-pfs1:/export/read-only/g07
Brick14: jc1letgfs18-pfs1:/export/read-only/g07
Brick15: jc1letgfs17-pfs1:/export/read-only/g08
Brick16: jc1letgfs18-pfs1:/export/read-only/g08
Brick17: jc1letgfs17-pfs1:/export/read-only/g09
Brick18: jc1letgfs18-pfs1:/export/read-only/g09
Brick19: jc1letgfs17-pfs1:/export/read-only/g10
Brick20: jc1letgfs18-pfs1:/export/read-only/g10

Thanks.

James Burnash
Unix Engineering.


DISCLAIMER:
This e-mail, and any attachments thereto, is intended only for use by the 
addressee(s) named herein and may contain legally privileged and/or 
confidential information. If you are not the intended recipient of this e-mail, 
you are hereby notified that any dissemination, distribution or copying of this 
e-mail, and any attachments thereto, is strictly prohibited. If you have 
received this in error, please immediately notify me and permanently delete the 
original and any copy of any e-mail and any printout thereof. E-mail 
transmission cannot be guaranteed to be secure or error-free. The sender 
therefore does not accept liability for any errors or omissions in the contents 
of this message which arise as a result of e-mail transmission.
NOTICE REGARDING PRIVACY AND CONFIDENTIALITY Knight Capital Group may, at its 
discretion, monitor and review the content of all e-mail communications. 
http://www.knight.com
___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


Re: [Gluster-users] How can i turn on the Gluster Debug mode?

2011-01-10 Thread Lakshmipathi
Hi,
For bricks use - 
gluster volume set volume-name diagnostics.brick-log-level DEBUG
For more tunable options refer 
http://www.gluster.com/community/documentation/index.php/Gluster_3.1:_Setting_Volume_Options

-- 

Cheers,
Lakshmipathi.G
FOSS Programmer.

- Original Message -
From: 공용준(yongjoon kong)/Cloud Computing 사업담당 andrew.k...@sk.com
To: gluster-users@gluster.org
Sent: Tuesday, January 11, 2011 6:17:49 AM
Subject: [Gluster-users] How can i turn on the Gluster Debug mode?

Hello, all

How can I turn on the Debug mode on 3.1

When I start the volume by command line, is there any way to turn the debug 
message on brick side?


___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users

___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


Re: [Gluster-users] infiniband speed

2011-01-10 Thread Christopher Hawkins
Thanks Joe, you nailed it. These are just test machines and in each case, just 
a single 10k scsi drive. That is the throughput bottleneck... I was not able to 
get more than 70MB/s sustained.

Chris

- Joe Landman land...@scalableinformatics.com wrote:

 On 01/10/2011 07:43 PM, Christopher Hawkins wrote:
  I am testing Infiniband for the first time. It seems that I should
 be
  able to get a lot more speed than I am with some pretty basic
 tests.
  Maybe someone running Infiniband can confirm that what I am seeing
 is
  way out of line, and/or help diagnose?
 
 What is the underlying disk speed?
 
 Try this:
 
   # (on bravo)
   dd if=/dev/zero of=/cluster/shadow/big.file bs=1M count=20k
 
 This will write a 20GB file to the same partition.  We need to see how
 
 fast that write is (outside of cache)
 
 Do the same test on the other machine.
 
 Infiniband won't make slow disks fast.  So unless you are already 
 getting a gigabyte or two per second (we do on our units), adding 
 Infiniband won't get you any significant additional performance.
 
 Put another way, you are bound by the slowest aspect of your storage
 system.
 
 What are your underlying disks?  Are these individual drives?  RAID? 
 What sort of RAID if it is a RAID?  How is it organized?
 
 We tune our systems pretty hard, so we start with 2+GB/s for TB sized
 
 files before we ever touch the next stack up.  Each additional stack
 you 
 traverse takes performance away (you lose it in stack inefficiency). 
 So 
 you should start out with as fast a system as you possibly can, so you
 
 don't lose too much performance.
 
 -- 
 Joseph Landman, Ph.D
 Founder and CEO
 Scalable Informatics, Inc.
 email: land...@scalableinformatics.com
 web  : http://scalableinformatics.com
 http://scalableinformatics.com/sicluster
 phone: +1 734 786 8423 x121
 fax  : +1 866 888 3112
 cell : +1 734 612 4615
 ___
 Gluster-users mailing list
 Gluster-users@gluster.org
 http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


Re: [Gluster-users] infiniband speed

2011-01-10 Thread Christopher Hawkins
Out of curiosity, what is a typical RAID / spindle count / rpm configuration 
for you that yields 2 GB/s?

- Christopher Hawkins chawk...@bplinux.com wrote:

 Thanks Joe, you nailed it. These are just test machines and in each
 case, just a single 10k scsi drive. That is the throughput
 bottleneck... I was not able to get more than 70MB/s sustained.
 
 Chris
 
 - Joe Landman land...@scalableinformatics.com wrote:
 
  On 01/10/2011 07:43 PM, Christopher Hawkins wrote:
   I am testing Infiniband for the first time. It seems that I
 should
  be
   able to get a lot more speed than I am with some pretty basic
  tests.
   Maybe someone running Infiniband can confirm that what I am
 seeing
  is
   way out of line, and/or help diagnose?
  
  What is the underlying disk speed?
  
  Try this:
  
  # (on bravo)
  dd if=/dev/zero of=/cluster/shadow/big.file bs=1M count=20k
  
  This will write a 20GB file to the same partition.  We need to see
 how
  
  fast that write is (outside of cache)
  
  Do the same test on the other machine.
  
  Infiniband won't make slow disks fast.  So unless you are already 
  getting a gigabyte or two per second (we do on our units), adding 
  Infiniband won't get you any significant additional performance.
  
  Put another way, you are bound by the slowest aspect of your
 storage
  system.
  
  What are your underlying disks?  Are these individual drives?  RAID?
 
  What sort of RAID if it is a RAID?  How is it organized?
  
  We tune our systems pretty hard, so we start with 2+GB/s for TB
 sized
  
  files before we ever touch the next stack up.  Each additional
 stack
  you 
  traverse takes performance away (you lose it in stack inefficiency).
 
  So 
  you should start out with as fast a system as you possibly can, so
 you
  
  don't lose too much performance.
  
  -- 
  Joseph Landman, Ph.D
  Founder and CEO
  Scalable Informatics, Inc.
  email: land...@scalableinformatics.com
  web  : http://scalableinformatics.com
  http://scalableinformatics.com/sicluster
  phone: +1 734 786 8423 x121
  fax  : +1 866 888 3112
  cell : +1 734 612 4615
  ___
  Gluster-users mailing list
  Gluster-users@gluster.org
  http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
 ___
 Gluster-users mailing list
 Gluster-users@gluster.org
 http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


Re: [Gluster-users] infiniband speed

2011-01-10 Thread Rudi Ahlers
On Tue, Jan 11, 2011 at 5:26 AM, Joe Landman
land...@scalableinformatics.com wrote:
 We tune our systems pretty hard, so we start with 2+GB/s for TB sized files
 before we ever touch the next stack up.  Each additional stack you traverse
 takes performance away (you lose it in stack inefficiency).  So you should
 start out with as fast a system as you possibly can, so you don't lose too
 much performance.

 --
 Joseph Landman, Ph.D
 Founder and CEO
 Scalable Informatics, Inc.
 email: land...@scalableinformatics.com
 web  : http://scalableinformatics.com
       http://scalableinformatics.com/sicluster
 phone: +1 734 786 8423 x121
 fax  : +1 866 888 3112
 cell : +1 734 612 4615
 ___



Wow, 2GB/s is pretty impressive, would you mind telling me how you achieve this?


-- 
Kind Regards
Rudi Ahlers
SoftDux

Website: http://www.SoftDux.com
Technical Blog: http://Blog.SoftDux.com
Office: 087 805 9573
Cell: 082 554 7532
___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users