[Gluster-users] Announcing GlusterFS-3.7.18

2016-12-12 Thread Kaushal M
Hi all,

GlusterFS-3.7.18 has been released. This is a regular bug fix release.
This release fixes 13 bugs. The release-notes can be found at [1].

Packages have been built at the CentOS Storage SIG [2] and
download.gluster.org [3]. The tarball can be downloaded from [3].

The next release might be delayed (more than normal) due to it being
the end of the year. The tracker for 3.7.19 is at [4], mark any bugs
that need to be fixed as dependencies.

See you all in the new year.

~kaushal

[1] 
https://github.com/gluster/glusterfs/blob/release-3.7/doc/release-notes/3.7.18.md
[2] https://wiki.centos.org/SpecialInterestGroup/Storage
[3] https://download.gluster.org/pub/gluster/glusterfs/3.7/3.7.18/
[4] https://bugzilla.redhat.com/show_bug.cgi?id=glusterfs-3.7.19
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users


[Gluster-users] Call for Papers, Red Hat Summit

2016-12-12 Thread Amye Scavarda
One of the big events for the year is the Red Hat Summit, which has a
dedicated open source community track for upstream projects. At the
very least this is an opportunity to give a "state of the union" talk
for the various communities, including Gluster.

There are a significant number of downstream-related talks for Summit,
but if you would like to submit a Gluster-related talk for the open
source community track[1] before the CFP for Summit closes on Dec. 16,
please do so!
Thank you!
- amye

[1] https://rh2017.smarteventscloud.com/portal/cfp/


-- 
Amye Scavarda | a...@redhat.com | Gluster Community Lead
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users


[Gluster-users] glusterd fails to start as systemd service when system starts in...

2016-12-12 Thread lejeczek
...emergency/recovery mode, because, eg. a storage device 
could not be mounted during boot, (not a gluster storage nor 
a storage which gluster uses)


I login in rescue mode, I fix that storage/mount problem 
(whatever that might be, does not matter) and I do

$ system default
and I find glusterd failed to start. (system boots fine into 
default). I can re/start it manually yes, but by then all 
gluster fstab mounts are not mounted.
What I want to highlight here is that that storage, gluster 
has nothing to do with. Or maybe it completely has nothing 
to do with a failed mount but simply if system boots into 
rescure/recovery then systemd's gluster fails.


Did you experience such a problem?
thanks,
L.

___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users


[Gluster-users] File operation failure on simple distributed volume

2016-12-12 Thread yonex
Hi,

When my application moves a file from it's local disk to FUSE-mounted
GlusterFS volume, the client outputs many warnings and errors not
always but occasionally. The volume is a simple distributed volume.

A sample of logs pasted: http://pastebin.com/axkTCRJX

It seems to come from something like a network disconnection
("Transport endpoint is not connected") at a glance, but other
networking applications on the same machine don't observe such a
thing. So I guess there may be a problem somewhere in GlusterFS stack.

It ended in failing to rename a file, logging PHP Warning like below:

PHP Warning:  rename(/glusterfs01/db1/stack/f0/13a9a2f0): failed
to open stream: Input/output error in [snipped].php on line 278
PHP Warning:
rename(/var/stack/13a9a2f0,/glusterfs01/db1/stack/f0/13a9a2f0):
Input/output error in [snipped].php on line 278

Conditions:

- GlusterFS 3.8.5 installed via yum CentOS-Gluster-3.8.repo
- Volume info and status pasted: http://pastebin.com/JPt2KeD8
- Client machines' OS: Scientific Linux 6 or CentOS 6.
- Server machines' OS: CentOS 6.
- Kernel version is 2.6.32-642.6.2.el6.x86_64 on all machines.
- The number of connected FUSE clients is 260.
- No firewall between connected machines.
- Neither remounting volumes nor rebooting client machines take effect.
- It is caused by not only rename() but also copy() and filesize() operation.
- No outputs in brick logs when it happens.

Any ideas? I'd appreciate any help.

Regards.
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] Can I remove a brick for dispersed volume on 3.9?

2016-12-12 Thread Cedric Lemarchand
I am not an expert, but I think bricks could not be removed on dispersed 
volume. Maybe “volume replace-brick” could be more accurate in this situation ?


> On 10 Dec 2016, at 12:13, Zhang Hang  wrote:
> 
> Hi,
> I am new to glusterFS and now I am trying it with the proxmox VE under lab 
> situation. I have 5 nodes on proxmox VE and made a dispersed volume of  1 x 
> (4 + 1) = 5. however, the node02 has failed to work now. so i plan to 
> reinstall proxmox VE on node02 so i need to remove the brick on node02 but i 
> failed with this message:
> gluster> volume remove-brick DispersedVol node02:/gluster/node02/brick01 force
> Removing brick(s) can result in data loss. Do you want to Continue? (y/n) y
> volume remove-brick commit force: failed: Remove brick incorrect brick count 
> of 1 for disperse 5
> 
> is that possible to remove a brick from the dispersed volume in this kind of 
> circumstance?
> 
> here is some information of my volume: 
> root@node04:~# gluster 
> gluster> peer status
> Number of Peers: 4
> 
> Hostname: node02
> State: Peer in Cluster (Disconnected)
> 
> Hostname: node05
> State: Peer in Cluster (Connected)
> 
> Hostname: node03
> State: Peer in Cluster (Connected)
> 
> Hostname: node01
> State: Peer in Cluster (Connected)
> 
> gluster> volume info
>  
> Volume Name: DispersedVol
> Type: Disperse
> Status: Started
> Snapshot Count: 0
> Number of Bricks: 1 x (4 + 1) = 5
> Transport-type: tcp
> Bricks:
> Brick1: node01:/gluster/node01/brick01
> Brick2: node02:/gluster/node02/brick01
> Brick3: node03:/gluster/node03/brick01
> Brick4: node04:/gluster/node04/brick01
> Brick5: node05:/gluster/node05/brick01
> Options Reconfigured:
> nfs.disable: on
> performance.readdir-ahead: on
> transport.address-family: inet
> 
> Anybody can help? Thanks a lot!
> ___
> Gluster-users mailing list
> Gluster-users@gluster.org
> http://www.gluster.org/mailman/listinfo/gluster-users

___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

[Gluster-users] [Glusterfs-client] Glusterfs client crashed

2016-12-12 Thread Binh Pham Van
Hi,

We have a problem with glusterfs client. This is detail information:
- OS: CentOS Linux release 7.2.1511
- glusterfs server and client version: 3.7.11
- the error when glusterfs-client carsh:

[2016-12-07 02:40:55.219009] E [MSGID: 108008]
[afr-transaction.c:1989:afr_transaction] 0-cloudbkf01-replicate-0: Failing
WRITE on gfid 9cb4e123-d742-46be-8d9f-584fbd57f24f: split-brain observed.
[Input/output error]
The message "E [MSGID: 108008] [afr-transaction.c:1989:afr_transaction]
0-cloudbkf01-replicate-0: Failing WRITE on gfid
9cb4e123-d742-46be-8d9f-584fbd57f24f: split-brain observed. [Input/output
error]" repeated 25 times between [2016-12-07 02:40:55.219009] and
[2016-12-07 02:41:20.068359]
pending frames:
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(1) op(FLUSH)
frame : type(1) op(OPEN)
frame : type(0) op(0)
frame : type(1) op(FLUSH)
frame : type(1) op(FLUSH)
frame : type(1) op(OPEN)
frame : type(0) op(0)
frame : type(1) op(FLUSH)
frame : type(1) op(OPEN)
frame : type(0) op(0)
frame : type(1) op(FLUSH)
frame : type(1) op(OPEN)
frame : type(0) op(0)
frame : type(1) op(OPEN)
frame : type(1) op(OPEN)
frame : type(0) op(0)
frame : type(1) op(OPEN)
frame : type(1) op(OPEN)
frame : type(0) op(0)
frame : type(1) op(FLUSH)
frame : type(1) op(OPEN)
frame : type(1) op(FLUSH)
frame : type(1) op(LOOKUP)
frame : type(1) op(OPEN)
frame : type(1) op(OPEN)
frame : type(1) op(OPEN)
frame : type(1) op(OPEN)
frame : type(1) op(OPEN)
frame : type(1) op(OPEN)
frame : type(1) op(OPEN)
frame : type(1) op(OPEN)
frame : type(1) op(FLUSH)
frame : type(1) op(OPEN)
frame : type(1) op(OPEN)
frame : type(1) op(OPEN)
frame : type(1) op(OPEN)
frame : type(0) op(0)
patchset: git://git.gluster.com/glusterfs.git
signal received: 6
time of crash:
2016-12-07 02:45:17
configuration details:
argp 1
backtrace 1
dlfcn 1
libpthread 1
llistxattr 1
setfsid 1
spinlock 1
epoll.h 1
xattr.h 1
st_atim.tv_nsec 1
package-string: glusterfs 3.7.11
/lib64/libglusterfs.so.0(_gf_msg_backtrace_nomem+0xc2)[0x7f9adf1e7072]
/lib64/libglusterfs.so.0(gf_print_trace+0x31d)[0x7f9adf20366d]
/lib64/libc.so.6(+0x35670)[0x7f9add8d5670]
/lib64/libc.so.6(gsignal+0x37)[0x7f9add8d55f7]
/lib64/libc.so.6(abort+0x148)[0x7f9add8d6ce8]
/lib64/libc.so.6(+0x75317)[0x7f9add915317]
/lib64/libc.so.6(+0x7d023)[0x7f9add91d023]
/lib64/libglusterfs.so.0(data_destroy+0x55)[0x7f9adf1dee85]
/lib64/libglusterfs.so.0(dict_destroy+0x40)[0x7f9adf1df600]
/usr/lib64/glusterfs/3.7.11/xlator/cluster/replicate.so(afr_local_cleanup+0x7b)[0x7f9ad128839b]
/usr/lib64/glusterfs/3.7.11/xlator/cluster/replicate.so(+0x4b72f)[0x7f9ad129072f]
/usr/lib64/glusterfs/3.7.11/xlator/cluster/replicate.so(afr_lookup_metadata_heal_check+0x224)[0x7f9ad1291024]
/usr/lib64/glusterfs/3.7.11/xlator/cluster/replicate.so(afr_lookup_cbk+0x24d)[0x7f9ad1291a9d]
/usr/lib64/glusterfs/3.7.11/xlator/protocol/client.so(client3_3_lookup_cbk+0x707)[0x7f9ad14da417]
/lib64/libgfrpc.so.0(rpc_clnt_handle_reply+0x90)[0x7f9adefb5ab0]
/lib64/libgfrpc.so.0(rpc_clnt_notify+0x1bf)[0x7f9adefb5d6f]
/lib64/libgfrpc.so.0(rpc_transport_notify+0x23)[0x7f9adefb1853]
/usr/lib64/glusterfs/3.7.11/rpc-transport/socket.so(+0x)[0x7f9ad3bc1666]
/usr/lib64/glusterfs/3.7.11/rpc-transport/socket.so(+0x92c4)[0x7f9ad3bc42c4]
/lib64/libglusterfs.so.0(+0x87a4a)[0x7f9adf248a4a]
/lib64/libpthread.so.0(+0x7dc5)[0x7f9ade04fdc5]
/lib64/libc.so.6(clone+0x6d)[0x7f9add99628d]
-

​We appreciate for your reply and hope you can help us fix this error.

Thanks and best regards,
​

-- 

*Phạm Văn Bình (Mr.)*

*System Engineer – GS Department*

*VNG Corporation*

182 Le Dai Hanh St., Ward 15, Dist. 11, Ho Chi Minh City, Vietnam

*T:* (+84 8) 3962 3888 (Ext: 6355) *M:* (+84) 1678 092 361

*F:* (+84 8) 3962 4666  *E:* bin...@vng.com.vn

*“*"*GS là team hỗ trợ kỹ thuật cho phát triển và vận hành game"**”*
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] [Gluster-devel] Release 3.10: Feature list frozen

2016-12-12 Thread Shyam
All proposed 3.10 features are now marked against the 3.10 release, and 
the release board is also updated to reflect the same, see [1]. If any 
feature you proposed/requested does not appear in this board, do shout out.


The branching for this release would happen on, 17th Jan, 2017 (~4 weeks 
prior to the release date of 14th Feb, 2017).


We will do a checkpoint with feature owners, around 3rd/4th Jan, 2017, 
on state of the feature, and help sought, to move the feature to 
completion (reviews, testing, what not). This would provide us with 
about a week to push the required work towards feature completion, so 
that branching can happen on time.


Thank you,
Shyam
[1] Release 3.10 Project board: 
https://github.com/gluster/glusterfs/projects/1

___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users


[Gluster-users] Can I remove a brick for dispersed volume on 3.9?

2016-12-12 Thread Zhang Hang
Hi,
I am new to glusterFS and now I am trying it with the proxmox VE under lab 
situation. I have 5 nodes on proxmox VE and made a dispersed volume of  1 x (4 
+ 1) = 5. however, the node02 has failed to work now. so i plan to reinstall 
proxmox VE on node02 so i need to remove the brick on node02 but i failed with 
this message:
gluster> volume remove-brick DispersedVol node02:/gluster/node02/brick01 force
Removing brick(s) can result in data loss. Do you want to Continue? (y/n) y
volume remove-brick commit force: failed: Remove brick incorrect brick count of 
1 for disperse 5

is that possible to remove a brick from the dispersed volume in this kind of 
circumstance?

here is some information of my volume: 
root@node04:~# gluster 
gluster> peer status
Number of Peers: 4

Hostname: node02
State: Peer in Cluster (Disconnected)

Hostname: node05
State: Peer in Cluster (Connected)

Hostname: node03
State: Peer in Cluster (Connected)

Hostname: node01
State: Peer in Cluster (Connected)

gluster> volume info
 
Volume Name: DispersedVol
Type: Disperse
Status: Started
Snapshot Count: 0
Number of Bricks: 1 x (4 + 1) = 5
Transport-type: tcp
Bricks:
Brick1: node01:/gluster/node01/brick01
Brick2: node02:/gluster/node02/brick01
Brick3: node03:/gluster/node03/brick01
Brick4: node04:/gluster/node04/brick01
Brick5: node05:/gluster/node05/brick01
Options Reconfigured:
nfs.disable: on
performance.readdir-ahead: on
transport.address-family: inet

Anybody can help? Thanks a lot!___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] Geo-replication failed to delete from slave file partially written to master volume.

2016-12-12 Thread Viktor Nosov
Hi Kotresh,

Logs from geo-replication test that I used to report the problem were 
overwritten.
So I run the test again. Each time the problem is reproduced with 100% 
probability.
This time I took all glusterfs  logs from Master and Slave systems. Logs are 
attached to this message.

All steps to duplicate the problem are reported at the text file 
"geo-rep-test.txt". This file is attached too.
Records at the file "geo-rep-test.txt" are in chronological order. 
Names of volumes, bricks, nodes are the same as before.

The test starts when geo-replication synchronization is working properly.
After that attempt is made to write file "big.file.1" to the master that has 
not enough space to handle the whole file.
Linux writes the file partially and geo-replication syncs this partial file 
"big.file.1" with the slave volume.

The next write of new file "test.file" to the master fails but it creates file 
handler "test.file" on Master for file size zero. This new file "test.file" is 
not replicated to the slave
volume.

The next step is to delete partially written file "big.file.1" from the master 
to make free space on the master. The delete is successful,
but it never is sync to the slave. This is the problem. File "big.file.1" is 
still on the slave volume.

The  next step is to repeat write for file "test.file". We let system to 
overwrite file on the master volume.
Now file "test.file" has some contents. But this change did not sync to the 
slave. This is other flavor of the same problem.

Finally new file "test.file.1" was written to the master. The file was 
successfully replicated to the slave.

Best regards,

Viktor Nosov



-Original Message-
From: Kotresh Hiremath Ravishankar [mailto:khire...@redhat.com] 
Sent: Friday, December 09, 2016 3:03 AM
To: Viktor Nosov
Cc: gluster-users@gluster.org
Subject: Re: [Gluster-users] Geo-replication failed to delete from slave file 
partially written to master volume.

Hi Viktor,

I went through the slave logs and there are no errors w.r.t deletion of the 
files.
I suspect the changelog has missed recording deletion of file because of which 
the delete is not processed.

Please share the following logs from master volume to root cause the issue.

1. Master geo-replication logs: 
/var/log/glusterfs/geo-replication//*.log
2. Master brick logs: /var/log/glusterfs/bricks/*.log 3. Also changelogs from 
Master volume: 
/exports/nas-segment-0012/master-for-183-0003/.glusterfs/changelogs/*

Thanks and Regards,
Kotresh H R

- Original Message -
> From: "Viktor Nosov" 
> To: "Kotresh Hiremath Ravishankar" 
> Cc: gluster-users@gluster.org
> Sent: Wednesday, December 7, 2016 10:48:52 PM
> Subject: RE: [Gluster-users] Geo-replication failed to delete from slave file 
> partially written to master volume.
> 
> Hi Kotresh,
> 
> Thanks for looking into this issue!
> I'm attaching log files from the slave node from 
> /var/log/glusterfs/geo-replication-slaves/
> 
> [root@SC-183 log]# cp
> /var/log/glusterfs/geo-replication-slaves/84501a83-b07c-4768-bfaa-418b
> 038e1a9e\:gluster%3A%2F%2F127.0.0.1%3Arem-volume-0001.gluster.log
> /home/vnosov/
> [root@SC-183 log]# cp 
> /var/log/glusterfs/geo-replication-slaves/slave.log
> /home/vnosov/
> [root@SC-183 log]# cp
> /var/log/glusterfs/geo-replication-slaves/mbr/84501a83-b07c-4768-bfaa-
> 418b038e1a9e\:gluster%3A%2F%2F127.0.0.1%3Arem-volume-0001.log
> /home/vnosov/
> 
> Best regards,
> 
> Viktor Nosov
> 
> 
> -Original Message-
> From: Kotresh Hiremath Ravishankar [mailto:khire...@redhat.com]
> Sent: Tuesday, December 06, 2016 9:25 PM
> To: Viktor Nosov
> Cc: gluster-users@gluster.org
> Subject: Re: [Gluster-users] Geo-replication failed to delete from 
> slave file partially written to master volume.
> 
> Hi Viktor,
> 
> Please share geo-replication-slave mount logs from slave nodes.
> 
> Thanks and Regards,
> Kotresh H R
> 
> - Original Message -
> > From: "Viktor Nosov" 
> > To: gluster-users@gluster.org
> > Cc: vno...@stonefly.com
> > Sent: Tuesday, December 6, 2016 7:13:22 AM
> > Subject: [Gluster-users] Geo-replication failed to delete from slave file
> > partially written to master volume.
> > 
> > Hi,
> > 
> > I hit problem while testing geo-replication. Anybody knows how to 
> > fix it except deleting and recreating geo-replication?
> > 
> > Geo-replication failed to delete from slave file partially written 
> > to master volume.
> > 
> > Have geo-replication between two nodes that are running glusterfs
> > 3.7.16
> > 
> > with master volume:
> > 
> > [root@SC-182 log]# gluster volume info master-for-183-0003
> > 
> > Volume Name: master-for-183-0003
> > Type: Distribute
> > Volume ID: 84501a83-b07c-4768-bfaa-418b038e1a9e
> > Status: Started
> > Number of Bricks: 1
> > Transport-type: tcp
> > Bricks:
> > Brick1: 10.10.60.182:/exports/nas-segment-0012/master-for-183-0003
> > Options Reconfigured:
> > changelog.changelog: on
> > 

[Gluster-users] Glustertests with 3 Nodes and dispersed Volumes are slow

2016-12-12 Thread Bode , Jörg
Hi everbofy there, I am Jörg from Germany and love to build good solution
with high intelligent Software in a simple way.

I come on your grate software by evaluating my environment and think that I
have lot to much hardware , So I try to move my NFS storage for my KVM
Servers to the Servers themselves. It there are easy qcow2 images. I want
the to run from local (cluster) disk. So, I start building a lab for
testing my idea. The first few tests were very fine wie distributed storage
and also replicated. I came a long to begin loving your work and start more
tests. I read about erasure code - a raid5 style setup. So I install 3
Cents mashines and build a build a dispersed volume over 3 servers with 1
Brick on each. What I see on my simple tests was very hard. Only 40 MB/s
writespeed on the local mounted glustervolume. I tested a lot - also on
high perfomance hardware. But I didn't find a solution.

Have you any idea why the performance is so slow? Even with a 4 Server...
All are connected over 1 Gbit/s Network.

Thanks for your answer...
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users