Re: [Gluster-users] Interesting split-brain...

2017-06-14 Thread Mohammed Rafi K C
Can you please explain How we ended up in this scenario. I think that
will help to understand more about this scenarios and why gluster
recommend replica 3 or arbiter volume.

Regards

Rafi KC


On 06/15/2017 10:46 AM, Karthik Subrahmanya wrote:
> Hi Ludwig,
>
> There is no way to resolve gfid split-brains with type mismatch. You
> have to do it manually by following the steps in [1].
> In case of type mismatch it is recommended to resolve it manually. But
> for only gfid mismatch in 3.11 we have a way to
> resolve it by using the *favorite-child-policy*.
> Since the file is not important, you can go with deleting that.
>
> [1]
> https://gluster.readthedocs.io/en/latest/Troubleshooting/split-brain/#fixing-directory-entry-split-brain
>
> HTH,
> Karthik
>
> On Thu, Jun 15, 2017 at 8:23 AM, Ludwig Gamache  > wrote:
>
> I am new to gluster but already like it. I did a maintenance last
> week where I shutdown both nodes (one after each others). I had
> many files that needed to be healed after that. Everything worked
> well, except for 1 file. It is in split-brain, with 2 different
> GFID. I read the documentation but it only covers the cases where
> the GFID is the same on both bricks. BTW, I am running Gluster 3.10.
>
> Here are some details...
>
> [root@NAS-01 .glusterfs]# gluster volume heal data01 info
>
> Brick 192.168.186.11:/mnt/DATA/data
>
> /abc/.zsh_history 
>
> /abc - Is in split-brain
>
>
> Status: Connected
>
> Number of entries: 2
>
>
> Brick 192.168.186.12:/mnt/DATA/data
>
> /abc - Is in split-brain
>
>
> /abc/.zsh_history 
>
> Status: Connected
>
> Number of entries: 2
>
>
> On brick 1:
>
> [root@NAS-01 abc]# ls -lart
>
> total 75
>
> drwxr-xr-x.  2 root  root  2 Jun  8 13:26 .zsh_history
>
> drwxr-xr-x.  3 12078 root  3 Jun 12 11:36 .
>
> drwxrwxrwt. 17 root  root 17 Jun 12 12:20 ..
>
>
> On brick 2:
>
> [root@DC-MTL-NAS-02 abc]# ls -lart
>
> total 66
>
> -rw-rw-r--.  2 12078 12078 1085 Jun 12 04:42 .zsh_history
>
> drwxr-xr-x.  2 12078 root 3 Jun 12 10:36 .
>
> drwxrwxrwt. 17 root  root17 Jun 12 11:20 ..
>
>
> Notice that on one brick, it is a file and on the other one it is
> a directory.
>
> On brick 1:
>
> [root@NAS-01 abc]# getfattr -d -m . -e hex
> /mnt/DATA/data/abc/.zsh_history
>
> getfattr: Removing leading '/' from absolute path names
>
> # file: mnt/DATA/data/abc/.zsh_history
>
> 
> security.selinux=0x73797374656d5f753a6f626a6563745f723a756e6c6162656c65645f743a733000
>
> trusted.afr.data01-client-0=0x
>
> trusted.afr.data01-client-1=0x0002
>
> trusted.gfid=0xdee43407139d41f091d13e106a51f262
>
> trusted.glusterfs.dht=0x0001
>
>
> On brick 2:
>
> root@NAS-02 abc]# getfattr -d -m . -e hex
> /mnt/DATA/data/abc/.zsh_history 
>
> getfattr: Removing leading '/' from absolute path names
>
> # file: mnt/DATA/data/abc/.zsh_history
>
> 
> security.selinux=0x73797374656d5f753a6f626a6563745f723a756e6c6162656c65645f743a733000
>
> trusted.afr.data01-client-0=0x00170002
>
> trusted.afr.data01-client-1=0x
>
> trusted.bit-rot.version=0x060059397acd0005dadd
>
> trusted.gfid=0xa70ae9af887a4a37875f5c7c81ebc803
>
>
> Any recommendation on how to recover from that? BTW, the file is
> not important and I could easily get rid of it without impact. So,
> if this is an easy solution...
>
> Regards,
>
> -- 
> Ludwig Gamache
>
>
> ___
> Gluster-users mailing list
> Gluster-users@gluster.org 
> http://lists.gluster.org/mailman/listinfo/gluster-users
> 
>
>
>
>
> ___
> Gluster-users mailing list
> Gluster-users@gluster.org
> http://lists.gluster.org/mailman/listinfo/gluster-users

___
Gluster-users mailing list
Gluster-users@gluster.org
http://lists.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] Interesting split-brain...

2017-06-14 Thread Karthik Subrahmanya
Hi Ludwig,

There is no way to resolve gfid split-brains with type mismatch. You have
to do it manually by following the steps in [1].
In case of type mismatch it is recommended to resolve it manually. But for
only gfid mismatch in 3.11 we have a way to
resolve it by using the *favorite-child-policy*.
Since the file is not important, you can go with deleting that.

[1]
https://gluster.readthedocs.io/en/latest/Troubleshooting/split-brain/#fixing-directory-entry-split-brain

HTH,
Karthik

On Thu, Jun 15, 2017 at 8:23 AM, Ludwig Gamache 
wrote:

> I am new to gluster but already like it. I did a maintenance last week
> where I shutdown both nodes (one after each others). I had many files that
> needed to be healed after that. Everything worked well, except for 1 file.
> It is in split-brain, with 2 different GFID. I read the documentation but
> it only covers the cases where the GFID is the same on both bricks. BTW, I
> am running Gluster 3.10.
>
> Here are some details...
>
> [root@NAS-01 .glusterfs]# gluster volume heal data01 info
>
> Brick 192.168.186.11:/mnt/DATA/data
>
> /abc/.zsh_history
>
> /abc - Is in split-brain
>
>
> Status: Connected
>
> Number of entries: 2
>
>
> Brick 192.168.186.12:/mnt/DATA/data
>
> /abc - Is in split-brain
>
>
> /abc/.zsh_history
>
> Status: Connected
>
> Number of entries: 2
>
> On brick 1:
>
> [root@NAS-01 abc]# ls -lart
>
> total 75
>
> drwxr-xr-x.  2 root  root  2 Jun  8 13:26 .zsh_history
>
> drwxr-xr-x.  3 12078 root  3 Jun 12 11:36 .
>
> drwxrwxrwt. 17 root  root 17 Jun 12 12:20 ..
>
> On brick 2:
>
> [root@DC-MTL-NAS-02 abc]# ls -lart
>
> total 66
>
> -rw-rw-r--.  2 12078 12078 1085 Jun 12 04:42 .zsh_history
>
> drwxr-xr-x.  2 12078 root 3 Jun 12 10:36 .
>
> drwxrwxrwt. 17 root  root17 Jun 12 11:20 ..
>
> Notice that on one brick, it is a file and on the other one it is a
> directory.
>
> On brick 1:
>
> [root@NAS-01 abc]# getfattr -d -m . -e hex /mnt/DATA/data/abc/.zsh_history
>
> getfattr: Removing leading '/' from absolute path names
>
> # file: mnt/DATA/data/abc/.zsh_history
>
> security.selinux=0x73797374656d5f753a6f626a6563
> 745f723a756e6c6162656c65645f743a733000
>
> trusted.afr.data01-client-0=0x
>
> trusted.afr.data01-client-1=0x0002
>
> trusted.gfid=0xdee43407139d41f091d13e106a51f262
>
> trusted.glusterfs.dht=0x0001
>
> On brick 2:
>
> root@NAS-02 abc]# getfattr -d -m . -e hex /mnt/DATA/data/abc/.zsh_history
>
> getfattr: Removing leading '/' from absolute path names
>
> # file: mnt/DATA/data/abc/.zsh_history
>
> security.selinux=0x73797374656d5f753a6f626a6563
> 745f723a756e6c6162656c65645f743a733000
>
> trusted.afr.data01-client-0=0x00170002
>
> trusted.afr.data01-client-1=0x
>
> trusted.bit-rot.version=0x060059397acd0005dadd
>
> trusted.gfid=0xa70ae9af887a4a37875f5c7c81ebc803
>
> Any recommendation on how to recover from that? BTW, the file is not
> important and I could easily get rid of it without impact. So, if this is
> an easy solution...
>
> Regards,
>
> --
> Ludwig Gamache
>
>
> ___
> Gluster-users mailing list
> Gluster-users@gluster.org
> http://lists.gluster.org/mailman/listinfo/gluster-users
>
___
Gluster-users mailing list
Gluster-users@gluster.org
http://lists.gluster.org/mailman/listinfo/gluster-users

[Gluster-users] Gluster Documentation Feedback

2017-06-14 Thread Nithya Balachandran
Hi,

We are looking at improving our documentation (http://gluster.readthedocs.
io/en/latest/) and would like your feedback.

Please let us know what would make the documentation more useful by
answering a few questions:

   - Which guides do you use (admin, developer)?
   - How easy is it to find the information you need?
   - What existing topics need more explanation?
   - What topics are missing?
   - What would make the documentation easier to use?
   - Any other feedback.


Thanks,
Nithya
___
Gluster-users mailing list
Gluster-users@gluster.org
http://lists.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] Transport Endpoint Not connected while running sysbench on Gluster Volume

2017-06-14 Thread Ben Turner
- Original Message -
> From: "Julio Guevara" 
> To: gluster-users@gluster.org
> Sent: Tuesday, June 13, 2017 4:43:06 PM
> Subject: [Gluster-users] Transport Endpoint Not connected while running   
> sysbench on Gluster Volume
> 
> I'm having a hard time trying to get a gluster volume up and running. I have
> setup other gluster volumes on other systems without much problems but this
> one is killing me.
> 
> The gluster vol was created with the command:
> gluster volume create mariadb_gluster_volume
> laeft-dccdb01p:/export/mariadb/brick
> 
> I had to lower frame-timeout since the system would become unresponsive until
> the frame failed by timeout:
> gluster volume set mariadb_gluster_volume networking.frame-timeout 5
> 
> running gluster version: glusterfs 3.8.12
> 
> The workload i'm using is: sysbench --test=fileio --file-total-size=4G
> --file-num=64 prepare
> 
> sysbench version: sysbench 0.4.12-5.el6
> 
> kernel version: 2.6.32-696.1.1.el6
> 
> centos: 6.8
> 
> Issue: Whenever I run the sysbench over the mount /var/lib/mysql_backups I
> get the error that is shown on the log output.
> 
> It is a constant issue, I can reproduce it when I start increasing the
> --file-num for sysbench above 3.

It looks like you may be seeing a crash.  If you look at /var/log/messages on 
all of the clients / servers do you see any crashes / seg faults / ABRT 
messages in the log?  If so can you open a BZ with the core / other info here?  
Here is an example of a crash on one of the bricks:

http://lists.gluster.org/pipermail/gluster-users/2016-February/025460.html

My guess is something is happening client sidesince we don't see anything in 
the server logs, check the client mount 
log(/var/log/glusterfs/.log and the messages file on your client.  
Also check messages on the servers.  If you see anything shoot us out the info 
and lets get a BZ open, if not maybe someone else on the list has some other 
ideas.

-b

> 
> 
> 
> ___
> Gluster-users mailing list
> Gluster-users@gluster.org
> http://lists.gluster.org/mailman/listinfo/gluster-users
___
Gluster-users mailing list
Gluster-users@gluster.org
http://lists.gluster.org/mailman/listinfo/gluster-users


[Gluster-users] Transport Endpoint Not connected while running sysbench on Gluster Volume

2017-06-14 Thread Julio Guevara
I'm having a hard time trying to get a gluster volume up and running. I
have setup other gluster volumes on other systems without much problems but
this one is killing me.

The gluster vol was created with the command:
gluster volume create mariadb_gluster_volume
laeft-dccdb01p:/export/mariadb/brick

I had to lower frame-timeout since the system would become unresponsive
until the frame failed by timeout:
gluster volume set mariadb_gluster_volume networking.frame-timeout 5

running gluster version: glusterfs 3.8.12

The workload i'm using is: sysbench --test=fileio --file-total-size=4G
--file-num=64 prepare

sysbench version: sysbench 0.4.12-5.el6

kernel version: 2.6.32-696.1.1.el6

centos: 6.8

Issue: Whenever I run the sysbench over the mount /var/lib/mysql_backups I
get the error that is shown on the log output.

It is a constant issue, I can reproduce it when I start increasing the
--file-num for sysbench above 3.


all.logs
Description: Binary data
___
Gluster-users mailing list
Gluster-users@gluster.org
http://lists.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] How to expand Replicated Volume

2017-06-14 Thread GiangCoi Mr
Hi Nag Pavan Chilakam

Can I use this command "gluster vol add-brick vol1 replica 2
file01g:/brick3/data/vol1 file02g:/brick4/data/vol1" in both file server 01
and 02 exited without add new servers. Is it ok for expanding volume?
Thanks for your support

Regards,
Giang



2017-06-14 22:26 GMT+07:00 Nag Pavan Chilakam :

> Hi,
> You can use add-brick command , this would make the volume a distributed
> replicated volume. Eg: gluster vol add-brick  rep 2
>   
>
> Regards,
> Nag
>
> On Jun 14, 2017 7:31 PM, "GiangCoi Mr"  wrote:
>
>> Hi Team.
>>
>> I have a issue about expand storage for replicated volume:
>>   - I have 2 server installed glusterfs. Each brick on server have total
>> 1 TB
>>   - I created volume named vol1 using replica 2 (gluster vol create vol1
>> replica 2 file01g:/brick1/data/vol1 file02g:/brick2/data/vol1)
>>
>> My issue: now total storage used ~ 900GB and I have to expand storage to
>> 2 TB. So how I can expand my storage to 2 TB. Please help me to fix this
>> issue. Thanks so much
>>
>> Regards,
>> Giang
>>
>>
>>
>> ___
>> Gluster-users mailing list
>> Gluster-users@gluster.org
>> http://lists.gluster.org/mailman/listinfo/gluster-users
>>
>
___
Gluster-users mailing list
Gluster-users@gluster.org
http://lists.gluster.org/mailman/listinfo/gluster-users

[Gluster-users] Interesting split-brain...

2017-06-14 Thread Ludwig Gamache
I am new to gluster but already like it. I did a maintenance last week
where I shutdown both nodes (one after each others). I had many files that
needed to be healed after that. Everything worked well, except for 1 file.
It is in split-brain, with 2 different GFID. I read the documentation but
it only covers the cases where the GFID is the same on both bricks. BTW, I
am running Gluster 3.10.

Here are some details...

[root@NAS-01 .glusterfs]# gluster volume heal data01 info

Brick 192.168.186.11:/mnt/DATA/data

/abc/.zsh_history

/abc - Is in split-brain


Status: Connected

Number of entries: 2


Brick 192.168.186.12:/mnt/DATA/data

/abc - Is in split-brain


/abc/.zsh_history

Status: Connected

Number of entries: 2

On brick 1:

[root@NAS-01 abc]# ls -lart

total 75

drwxr-xr-x.  2 root  root  2 Jun  8 13:26 .zsh_history

drwxr-xr-x.  3 12078 root  3 Jun 12 11:36 .

drwxrwxrwt. 17 root  root 17 Jun 12 12:20 ..

On brick 2:

[root@DC-MTL-NAS-02 abc]# ls -lart

total 66

-rw-rw-r--.  2 12078 12078 1085 Jun 12 04:42 .zsh_history

drwxr-xr-x.  2 12078 root 3 Jun 12 10:36 .

drwxrwxrwt. 17 root  root17 Jun 12 11:20 ..

Notice that on one brick, it is a file and on the other one it is a
directory.

On brick 1:

[root@NAS-01 abc]# getfattr -d -m . -e hex /mnt/DATA/data/abc/.zsh_history

getfattr: Removing leading '/' from absolute path names

# file: mnt/DATA/data/abc/.zsh_history

security.selinux=0x73797374656d5f753a6f626a6563745f723a756e6c6162656c65645f743a733000

trusted.afr.data01-client-0=0x

trusted.afr.data01-client-1=0x0002

trusted.gfid=0xdee43407139d41f091d13e106a51f262

trusted.glusterfs.dht=0x0001

On brick 2:

root@NAS-02 abc]# getfattr -d -m . -e hex /mnt/DATA/data/abc/.zsh_history

getfattr: Removing leading '/' from absolute path names

# file: mnt/DATA/data/abc/.zsh_history

security.selinux=0x73797374656d5f753a6f626a6563745f723a756e6c6162656c65645f743a733000

trusted.afr.data01-client-0=0x00170002

trusted.afr.data01-client-1=0x

trusted.bit-rot.version=0x060059397acd0005dadd

trusted.gfid=0xa70ae9af887a4a37875f5c7c81ebc803

Any recommendation on how to recover from that? BTW, the file is not
important and I could easily get rid of it without impact. So, if this is
an easy solution...

Regards,

-- 
Ludwig Gamache
___
Gluster-users mailing list
Gluster-users@gluster.org
http://lists.gluster.org/mailman/listinfo/gluster-users

[Gluster-users] gluster peer probe failing

2017-06-14 Thread Guy Cukierman
Hi,
I have a gluster (version 3.10.2) server running on a 3 node (centos7) cluster.
Firewalld and SELinux are disabled, and I see I can telnet from each node to 
the other on port 24007.

When I try to create the first peering by running on node1 the command:
gluster peer probe 

I get the error:
"Connection failed. Please check if gluster daemon is operational."

And Glusterd.log shows:

[2017-06-14 14:46:09.927510] I [MSGID: 106487] 
[glusterd-handler.c:1242:__glusterd_handle_cli_probe] 0-glusterd: Received CLI 
probe req 192.168.1.17 24007
[2017-06-14 14:46:09.928560] I [MSGID: 106129] 
[glusterd-handler.c:3690:glusterd_probe_begin] 0-glusterd: Unable to find 
peerinfo for host: 192.168.1.17 (24007)
[2017-06-14 14:46:09.930783] W [MSGID: 106062] 
[glusterd-handler.c:3466:glusterd_transport_inet_options_build] 0-glusterd: 
Failed to get tcp-user-timeout
[2017-06-14 14:46:09.930837] I [rpc-clnt.c:1059:rpc_clnt_connection_init] 
0-management: setting frame-timeout to 600
pending frames:
frame : type(0) op(0)
patchset: git://git.gluster.org/glusterfs.git
signal received: 11
time of crash:
2017-06-14 14:46:09
configuration details:
argp 1
backtrace 1
dlfcn 1
libpthread 1
llistxattr 1
setfsid 1
spinlock 1
epoll.h 1
xattr.h 1
st_atim.tv_nsec 1
package-string: glusterfs 3.10.3
/lib64/libglusterfs.so.0(_gf_msg_backtrace_nomem+0xa0)[0x7f69625da4d0]
/lib64/libglusterfs.so.0(gf_print_trace+0x324)[0x7f69625e3dd4]
/lib64/libc.so.6(+0x35250)[0x7f6960cbe250]
/lib64/libglusterfs.so.0(gf_ports_reserved+0x15c)[0x7f69625e444c]
/lib64/libglusterfs.so.0(gf_process_reserved_ports+0xbe)[0x7f69625e470e]
/usr/lib64/glusterfs/3.10.3/rpc-transport/socket.so(+0xd158)[0x7f6954568158]
/usr/lib64/glusterfs/3.10.3/rpc-transport/socket.so(client_bind+0x93)[0x7f69545685a3]
/usr/lib64/glusterfs/3.10.3/rpc-transport/socket.so(+0xa875)[0x7f6954565875]
/lib64/libgfrpc.so.0(rpc_clnt_reconnect+0xc9)[0x7f69623a3f89]
/lib64/libgfrpc.so.0(rpc_clnt_start+0x39)[0x7f69623a4049]
/usr/lib64/glusterfs/3.10.3/xlator/mgmt/glusterd.so(+0x24218)[0x7f69570f9218]
/usr/lib64/glusterfs/3.10.3/xlator/mgmt/glusterd.so(+0x24843)[0x7f69570f9843]
/usr/lib64/glusterfs/3.10.3/xlator/mgmt/glusterd.so(+0x24ae0)[0x7f69570f9ae0]
/usr/lib64/glusterfs/3.10.3/xlator/mgmt/glusterd.so(+0x27890)[0x7f69570fc890]
/usr/lib64/glusterfs/3.10.3/xlator/mgmt/glusterd.so(+0x27e20)[0x7f69570fce20]
/usr/lib64/glusterfs/3.10.3/xlator/mgmt/glusterd.so(+0x20f5e)[0x7f69570f5f5e]
/lib64/libglusterfs.so.0(synctask_wrap+0x10)[0x7f6962611750]
/lib64/libc.so.6(+0x46cf0)[0x7f6960ccfcf0]

And a file is create under /var/lib/glusterd/peers/ which 
contains:
uuid=----
state=0
hostname1=192.168.1.17

and the glusterd daemon exits and I cannot restart it until I delete this file 
from the peers folder.

Any idea what is wrong?
thanks!
___
Gluster-users mailing list
Gluster-users@gluster.org
http://lists.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] ganesha with gluster not starting

2017-06-14 Thread Kaleb S. KEITHLEY
On 06/14/2017 02:05 PM, Tony Schreiner wrote:
> I am getting startup errors for nfs ganesha
> [snip]
> 
> I note that when glusterd starts up, (independent of ganesha), there is
> this item in the glusterd.log
> 
> [2017-06-14 17:58:41.779797] W [MSGID: 101095]
> [xlator.c:162:xlator_volopt_dynload] 0-xlator:
> /usr/lib64/glusterfs/3.10.2/xlator/features/ganesha.so: cannot open
> shared object file: No such file or directory
> 
> which seems relevant, but I'm not sure if it is. If this is the problem,
> which package is supposed to supply this file? 

None. It's a benign error. There is no ganesha.so and there's not
supposed to be.

As for nfs-ganesha not starting check /var/log/ganesha.log and/or enable
core dumps and paste the bt from the core dump.

-- 

Kaleb
___
Gluster-users mailing list
Gluster-users@gluster.org
http://lists.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] Transport Endpoint Not connected while running sysbench on Gluster Volume

2017-06-14 Thread Julio Guevara
Also, this is the profile output of this Volume:

gluster> volume profile mariadb_gluster_volume info cumulative
Brick: laeft-dccdb01p.core.epay.us.loc:/export/mariadb_backup/brick
---
Cumulative Stats:
   Block Size:  16384b+   32768b+
65536b+
 No. of Reads:0 0
  0
No. of Writes:83391465750
 102911

   Block Size: 131072b+
 No. of Reads:   33
No. of Writes: 8551
 %-latency   Avg-latency   Min-Latency   Max-Latency   No. of calls
Fop
 -   ---   ---   ---   
 
  0.00   0.00 us   0.00 us   0.00 us 26
 FORGET
  0.00   0.00 us   0.00 us   0.00 us 33
RELEASE
  0.00   0.00 us   0.00 us   0.00 us 92
 RELEASEDIR
  0.00 102.25 us  70.00 us 155.00 us  4
 SETXATTR
  0.00 159.80 us  97.00 us 207.00 us  5
RMDIR
  0.00 266.75 us 121.00 us 675.00 us  4
SETATTR
  0.00 100.62 us  29.00 us 700.00 us 16
INODELK
  0.00  84.33 us  29.00 us 477.00 us 33
FLUSH
  0.00  68.16 us  34.00 us 165.00 us 92
OPENDIR
  0.01  88.35 us  21.00 us 608.00 us 92
 STAT
  0.01 754.04 us 262.00 us   10104.00 us 25
 CREATE
  0.02 169.04 us  27.00 us 997.00 us179
 READDIRP
  0.04 150.90 us  43.00 us1867.00 us365
 LOOKUP
  0.04   16330.75 us 297.00 us   46360.00 us  4
MKDIR
  0.187896.70 us  71.00 us  256814.00 us 33
 READ
  0.68 466.93 us  19.00 us1848.00 us   2119
 STATFS
  2.39  151339.17 us 227.00 us  540998.00 us 23
 UNLINK
  5.49  320155.28 us2035.00 us 1273394.00 us 25
FSYNC
 31.28  69.00 us  30.00 us   11447.00 us 660603
WRITE
 59.84 10899586.88 us 5827342.00 us 13921169.00 us  8
 OPEN

Duration: 71425 seconds
   Data Read: 4325376 bytes
Data Written: 29195534336 bytes


As you can see OPEN Fop take the most amount of time and they normally
timeout, even with default value for networking.frame-timeout

Thanks

On Tue, Jun 13, 2017 at 3:43 PM, Julio Guevara 
wrote:

> I'm having a hard time trying to get a gluster volume up and running. I
> have setup other gluster volumes on other systems without much problems but
> this one is killing me.
>
> The gluster vol was created with the command:
> gluster volume create mariadb_gluster_volume laeft-dccdb01p:/export/
> mariadb/brick
>
> I had to lower frame-timeout since the system would become unresponsive
> until the frame failed by timeout:
> gluster volume set mariadb_gluster_volume networking.frame-timeout 5
>
> running gluster version: glusterfs 3.8.12
>
> The workload i'm using is: sysbench --test=fileio --file-total-size=4G
> --file-num=64 prepare
>
> sysbench version: sysbench 0.4.12-5.el6
>
> kernel version: 2.6.32-696.1.1.el6
>
> centos: 6.8
>
> Issue: Whenever I run the sysbench over the mount /var/lib/mysql_backups I
> get the error that is shown on the log output.
>
> It is a constant issue, I can reproduce it when I start increasing the
> --file-num for sysbench above 3.
>
>
>
___
Gluster-users mailing list
Gluster-users@gluster.org
http://lists.gluster.org/mailman/listinfo/gluster-users

[Gluster-users] How to expand Replicated Volume

2017-06-14 Thread GiangCoi Mr
Hi Team.

I have a issue about expand storage for replicated volume:
  - I have 2 server installed glusterfs. Each brick on server have total 1
TB
  - I created volume named vol1 using replica 2 (gluster vol create vol1
replica 2 file01g:/brick1/data/vol1 file02g:/brick2/data/vol1)

My issue: now total storage used ~ 900GB and I have to expand storage to 2
TB. So how I can expand my storage to 2 TB. Please help me to fix this
issue. Thanks so much

Regards,
Giang
___
Gluster-users mailing list
Gluster-users@gluster.org
http://lists.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] About starting nfs-ganesha

2017-06-14 Thread te-yamauchi
Kaleb,
Thank you.
If you disable gnfs you can only use nfs - ganesha + glusterfs.

I will refer to the URL.

--
Tetsuo

> > When using nfs-ganesha with GlusterFS, is it necessary to enable it with the
> following command?
> > # gluster nfs-ganesha enable
> > It is said that it is necessary to enable nfs-ganesha option when setting
> ganesha.enable on for volume.
> > It is said that setting of ganesha-ha.conf is necessary, but is it 
> > necessary to
> set up HA to use nfs-ganesha?
> > I would be pleased if you could tell me about HA setting example if 
> > possible.
> >
> > Please wait .. This will take a few minutes to complete.
> > nfs-ganesha: failed: Failed to set up HA config for NFS - Ganesha.
> > Please check the log file for details
> >
> > By the way, if you start up nfs - ganesha service without doing the
> > above and do export setting, NFS connection is possible, but is not this an
> official connection method?
> >
> 
> You may use nfs-ganesha+glusterfs without HA.  You need to disable legacy
> gluster NFS (gnfs).
> 
> The "walk" part of my article at
> http://blog.gluster.org/2015/10/linux-scale-out-nfsv4-using-nfs-ganesha-an
> d-glusterfs-one-step-at-a-time
> shows how to do this.
> 
> --
> 
> Kaleb

___
Gluster-users mailing list
Gluster-users@gluster.org
http://lists.gluster.org/mailman/listinfo/gluster-users


[Gluster-users] No NFS connection due to GlusterFS CPU load

2017-06-14 Thread te-yamauchi

When executing the load test with the FIO tool, execute the following job from 
the client
When executed, the load of 2 cores is high for the CPU. Up to 100%.
At that time, if another client is performing NFS mounting, the df command
I can not connect NFS without coming back. The log will continue to be output 
below.
I believe that if the CPU utilization is distributed, the load will be 
eliminated.

Will not improve by tuning the following parameters?
Most of the parameters remain as defaults.

server.event-threads : 1
client.event-threads : 2
server.outstanding-rpc-limit : 64
nfs.outstanding-rpc-limit : 16
performance.io-thread-count : 16

/var/log/glusterfs/nfs.log
[2017-06-14 10:02:03.964405] I [MSGID: 108006] 
[afr-common.c:4941:afr_local_init] 0-gvol01-replicate-0: no subvolumes up
[2017-06-14 10:02:04.026299] E [rpc-clnt.c:365:saved_frames_unwind] (--> 
/lib64/libglusterfs.so.0(_gf_log_callingfn+0x13b)[0x7f2729b3ae8b] (--> 
/lib64/libgfrpc.so.0(saved_frames_unwind+0x1de)[0x7f27299018ee] (--> 
/lib64/libgfrpc.so.0(saved_frames_destroy+0xe)[0x7f27299019fe] (--> 
/lib64/libgfrpc.so.0(rpc_clnt_connection_cleanup+0x90)[0x7f2729903170] (--> 
/lib64/libgfrpc.so.0(rpc_clnt_notify+0x2a0)[0x7f2729903c20] ) 
0-gvol01-client-1: forced unwinding frame type(GlusterFS 3.3) op(WRITE(13)) 
called at 2017-06-14 09:58:00.107164 (xid=0x8dc455)
[2017-06-14 10:02:06.967780] E [rpc-clnt.c:365:saved_frames_unwind] (--> 
/lib64/libglusterfs.so.0(_gf_log_callingfn+0x13b)[0x7f2729b3ae8b] (--> 
/lib64/libgfrpc.so.0(saved_frames_unwind+0x1de)[0x7f27299018ee] (--> 
/lib64/libgfrpc.so.0(saved_frames_destroy+0xe)[0x7f27299019fe] (--> 
/lib64/libgfrpc.so.0(rpc_clnt_connection_cleanup+0x90)[0x7f2729903170] (--> 
/lib64/libgfrpc.so.0(rpc_clnt_notify+0x2a0)[0x7f2729903c20] ) 
0-gvol01-client-0: forced unwinding frame type(GlusterFS 3.3) op(FSYNC(16)) 
called at 2017-06-14 09:55:20.342725 (xid=0x663a73)
The message "I [MSGID: 108006] [afr-common.c:4941:afr_local_init] 
0-gvol01-replicate-0: no subvolumes up" repeated 37 times between [2017-06-14 
10:02:03.964405] and [2017-06-14 10:02:06.880464]
[2017-06-14 10:02:06.967820] W [MSGID: 114031] 
[client-rpc-fops.c:972:client3_3_fsync_cbk] 0-gvol01-client-0: remote operation 
failed [Communication end point is not connected]
[2017-06-14 10:02:06.967890] W [MSGID: 108035] 
[afr-transaction.c:2243:afr_changelog_fsync_cbk] 0-gvol01-replicate-0: 
fsync(08ed1905-d81e-4ad3-9de2-1395f2c4667e) failed on subvolume 
gvol01-client-0. Transaction was WRITE [Communication end point is not 
connected]
___
Gluster-users mailing list
Gluster-users@gluster.org
http://lists.gluster.org/mailman/listinfo/gluster-users