[Gluster-users] Announcing GlusterFS-3.7.18
Hi all, GlusterFS-3.7.18 has been released. This is a regular bug fix release. This release fixes 13 bugs. The release-notes can be found at [1]. Packages have been built at the CentOS Storage SIG [2] and download.gluster.org [3]. The tarball can be downloaded from [3]. The next release might be delayed (more than normal) due to it being the end of the year. The tracker for 3.7.19 is at [4], mark any bugs that need to be fixed as dependencies. See you all in the new year. ~kaushal [1] https://github.com/gluster/glusterfs/blob/release-3.7/doc/release-notes/3.7.18.md [2] https://wiki.centos.org/SpecialInterestGroup/Storage [3] https://download.gluster.org/pub/gluster/glusterfs/3.7/3.7.18/ [4] https://bugzilla.redhat.com/show_bug.cgi?id=glusterfs-3.7.19 ___ Gluster-users mailing list Gluster-users@gluster.org http://www.gluster.org/mailman/listinfo/gluster-users
[Gluster-users] Call for Papers, Red Hat Summit
One of the big events for the year is the Red Hat Summit, which has a dedicated open source community track for upstream projects. At the very least this is an opportunity to give a "state of the union" talk for the various communities, including Gluster. There are a significant number of downstream-related talks for Summit, but if you would like to submit a Gluster-related talk for the open source community track[1] before the CFP for Summit closes on Dec. 16, please do so! Thank you! - amye [1] https://rh2017.smarteventscloud.com/portal/cfp/ -- Amye Scavarda | a...@redhat.com | Gluster Community Lead ___ Gluster-users mailing list Gluster-users@gluster.org http://www.gluster.org/mailman/listinfo/gluster-users
[Gluster-users] glusterd fails to start as systemd service when system starts in...
...emergency/recovery mode, because, eg. a storage device could not be mounted during boot, (not a gluster storage nor a storage which gluster uses) I login in rescue mode, I fix that storage/mount problem (whatever that might be, does not matter) and I do $ system default and I find glusterd failed to start. (system boots fine into default). I can re/start it manually yes, but by then all gluster fstab mounts are not mounted. What I want to highlight here is that that storage, gluster has nothing to do with. Or maybe it completely has nothing to do with a failed mount but simply if system boots into rescure/recovery then systemd's gluster fails. Did you experience such a problem? thanks, L. ___ Gluster-users mailing list Gluster-users@gluster.org http://www.gluster.org/mailman/listinfo/gluster-users
[Gluster-users] File operation failure on simple distributed volume
Hi, When my application moves a file from it's local disk to FUSE-mounted GlusterFS volume, the client outputs many warnings and errors not always but occasionally. The volume is a simple distributed volume. A sample of logs pasted: http://pastebin.com/axkTCRJX It seems to come from something like a network disconnection ("Transport endpoint is not connected") at a glance, but other networking applications on the same machine don't observe such a thing. So I guess there may be a problem somewhere in GlusterFS stack. It ended in failing to rename a file, logging PHP Warning like below: PHP Warning: rename(/glusterfs01/db1/stack/f0/13a9a2f0): failed to open stream: Input/output error in [snipped].php on line 278 PHP Warning: rename(/var/stack/13a9a2f0,/glusterfs01/db1/stack/f0/13a9a2f0): Input/output error in [snipped].php on line 278 Conditions: - GlusterFS 3.8.5 installed via yum CentOS-Gluster-3.8.repo - Volume info and status pasted: http://pastebin.com/JPt2KeD8 - Client machines' OS: Scientific Linux 6 or CentOS 6. - Server machines' OS: CentOS 6. - Kernel version is 2.6.32-642.6.2.el6.x86_64 on all machines. - The number of connected FUSE clients is 260. - No firewall between connected machines. - Neither remounting volumes nor rebooting client machines take effect. - It is caused by not only rename() but also copy() and filesize() operation. - No outputs in brick logs when it happens. Any ideas? I'd appreciate any help. Regards. ___ Gluster-users mailing list Gluster-users@gluster.org http://www.gluster.org/mailman/listinfo/gluster-users
Re: [Gluster-users] Can I remove a brick for dispersed volume on 3.9?
I am not an expert, but I think bricks could not be removed on dispersed volume. Maybe “volume replace-brick” could be more accurate in this situation ? > On 10 Dec 2016, at 12:13, Zhang Hangwrote: > > Hi, > I am new to glusterFS and now I am trying it with the proxmox VE under lab > situation. I have 5 nodes on proxmox VE and made a dispersed volume of 1 x > (4 + 1) = 5. however, the node02 has failed to work now. so i plan to > reinstall proxmox VE on node02 so i need to remove the brick on node02 but i > failed with this message: > gluster> volume remove-brick DispersedVol node02:/gluster/node02/brick01 force > Removing brick(s) can result in data loss. Do you want to Continue? (y/n) y > volume remove-brick commit force: failed: Remove brick incorrect brick count > of 1 for disperse 5 > > is that possible to remove a brick from the dispersed volume in this kind of > circumstance? > > here is some information of my volume: > root@node04:~# gluster > gluster> peer status > Number of Peers: 4 > > Hostname: node02 > State: Peer in Cluster (Disconnected) > > Hostname: node05 > State: Peer in Cluster (Connected) > > Hostname: node03 > State: Peer in Cluster (Connected) > > Hostname: node01 > State: Peer in Cluster (Connected) > > gluster> volume info > > Volume Name: DispersedVol > Type: Disperse > Status: Started > Snapshot Count: 0 > Number of Bricks: 1 x (4 + 1) = 5 > Transport-type: tcp > Bricks: > Brick1: node01:/gluster/node01/brick01 > Brick2: node02:/gluster/node02/brick01 > Brick3: node03:/gluster/node03/brick01 > Brick4: node04:/gluster/node04/brick01 > Brick5: node05:/gluster/node05/brick01 > Options Reconfigured: > nfs.disable: on > performance.readdir-ahead: on > transport.address-family: inet > > Anybody can help? Thanks a lot! > ___ > Gluster-users mailing list > Gluster-users@gluster.org > http://www.gluster.org/mailman/listinfo/gluster-users ___ Gluster-users mailing list Gluster-users@gluster.org http://www.gluster.org/mailman/listinfo/gluster-users
[Gluster-users] [Glusterfs-client] Glusterfs client crashed
Hi, We have a problem with glusterfs client. This is detail information: - OS: CentOS Linux release 7.2.1511 - glusterfs server and client version: 3.7.11 - the error when glusterfs-client carsh: [2016-12-07 02:40:55.219009] E [MSGID: 108008] [afr-transaction.c:1989:afr_transaction] 0-cloudbkf01-replicate-0: Failing WRITE on gfid 9cb4e123-d742-46be-8d9f-584fbd57f24f: split-brain observed. [Input/output error] The message "E [MSGID: 108008] [afr-transaction.c:1989:afr_transaction] 0-cloudbkf01-replicate-0: Failing WRITE on gfid 9cb4e123-d742-46be-8d9f-584fbd57f24f: split-brain observed. [Input/output error]" repeated 25 times between [2016-12-07 02:40:55.219009] and [2016-12-07 02:41:20.068359] pending frames: frame : type(0) op(0) frame : type(0) op(0) frame : type(1) op(FLUSH) frame : type(1) op(OPEN) frame : type(0) op(0) frame : type(1) op(FLUSH) frame : type(1) op(FLUSH) frame : type(1) op(OPEN) frame : type(0) op(0) frame : type(1) op(FLUSH) frame : type(1) op(OPEN) frame : type(0) op(0) frame : type(1) op(FLUSH) frame : type(1) op(OPEN) frame : type(0) op(0) frame : type(1) op(OPEN) frame : type(1) op(OPEN) frame : type(0) op(0) frame : type(1) op(OPEN) frame : type(1) op(OPEN) frame : type(0) op(0) frame : type(1) op(FLUSH) frame : type(1) op(OPEN) frame : type(1) op(FLUSH) frame : type(1) op(LOOKUP) frame : type(1) op(OPEN) frame : type(1) op(OPEN) frame : type(1) op(OPEN) frame : type(1) op(OPEN) frame : type(1) op(OPEN) frame : type(1) op(OPEN) frame : type(1) op(OPEN) frame : type(1) op(OPEN) frame : type(1) op(FLUSH) frame : type(1) op(OPEN) frame : type(1) op(OPEN) frame : type(1) op(OPEN) frame : type(1) op(OPEN) frame : type(0) op(0) patchset: git://git.gluster.com/glusterfs.git signal received: 6 time of crash: 2016-12-07 02:45:17 configuration details: argp 1 backtrace 1 dlfcn 1 libpthread 1 llistxattr 1 setfsid 1 spinlock 1 epoll.h 1 xattr.h 1 st_atim.tv_nsec 1 package-string: glusterfs 3.7.11 /lib64/libglusterfs.so.0(_gf_msg_backtrace_nomem+0xc2)[0x7f9adf1e7072] /lib64/libglusterfs.so.0(gf_print_trace+0x31d)[0x7f9adf20366d] /lib64/libc.so.6(+0x35670)[0x7f9add8d5670] /lib64/libc.so.6(gsignal+0x37)[0x7f9add8d55f7] /lib64/libc.so.6(abort+0x148)[0x7f9add8d6ce8] /lib64/libc.so.6(+0x75317)[0x7f9add915317] /lib64/libc.so.6(+0x7d023)[0x7f9add91d023] /lib64/libglusterfs.so.0(data_destroy+0x55)[0x7f9adf1dee85] /lib64/libglusterfs.so.0(dict_destroy+0x40)[0x7f9adf1df600] /usr/lib64/glusterfs/3.7.11/xlator/cluster/replicate.so(afr_local_cleanup+0x7b)[0x7f9ad128839b] /usr/lib64/glusterfs/3.7.11/xlator/cluster/replicate.so(+0x4b72f)[0x7f9ad129072f] /usr/lib64/glusterfs/3.7.11/xlator/cluster/replicate.so(afr_lookup_metadata_heal_check+0x224)[0x7f9ad1291024] /usr/lib64/glusterfs/3.7.11/xlator/cluster/replicate.so(afr_lookup_cbk+0x24d)[0x7f9ad1291a9d] /usr/lib64/glusterfs/3.7.11/xlator/protocol/client.so(client3_3_lookup_cbk+0x707)[0x7f9ad14da417] /lib64/libgfrpc.so.0(rpc_clnt_handle_reply+0x90)[0x7f9adefb5ab0] /lib64/libgfrpc.so.0(rpc_clnt_notify+0x1bf)[0x7f9adefb5d6f] /lib64/libgfrpc.so.0(rpc_transport_notify+0x23)[0x7f9adefb1853] /usr/lib64/glusterfs/3.7.11/rpc-transport/socket.so(+0x)[0x7f9ad3bc1666] /usr/lib64/glusterfs/3.7.11/rpc-transport/socket.so(+0x92c4)[0x7f9ad3bc42c4] /lib64/libglusterfs.so.0(+0x87a4a)[0x7f9adf248a4a] /lib64/libpthread.so.0(+0x7dc5)[0x7f9ade04fdc5] /lib64/libc.so.6(clone+0x6d)[0x7f9add99628d] - We appreciate for your reply and hope you can help us fix this error. Thanks and best regards, -- *Phạm Văn Bình (Mr.)* *System Engineer – GS Department* *VNG Corporation* 182 Le Dai Hanh St., Ward 15, Dist. 11, Ho Chi Minh City, Vietnam *T:* (+84 8) 3962 3888 (Ext: 6355) *M:* (+84) 1678 092 361 *F:* (+84 8) 3962 4666 *E:* bin...@vng.com.vn *“*"*GS là team hỗ trợ kỹ thuật cho phát triển và vận hành game"**”* ___ Gluster-users mailing list Gluster-users@gluster.org http://www.gluster.org/mailman/listinfo/gluster-users
Re: [Gluster-users] [Gluster-devel] Release 3.10: Feature list frozen
All proposed 3.10 features are now marked against the 3.10 release, and the release board is also updated to reflect the same, see [1]. If any feature you proposed/requested does not appear in this board, do shout out. The branching for this release would happen on, 17th Jan, 2017 (~4 weeks prior to the release date of 14th Feb, 2017). We will do a checkpoint with feature owners, around 3rd/4th Jan, 2017, on state of the feature, and help sought, to move the feature to completion (reviews, testing, what not). This would provide us with about a week to push the required work towards feature completion, so that branching can happen on time. Thank you, Shyam [1] Release 3.10 Project board: https://github.com/gluster/glusterfs/projects/1 ___ Gluster-users mailing list Gluster-users@gluster.org http://www.gluster.org/mailman/listinfo/gluster-users
[Gluster-users] Can I remove a brick for dispersed volume on 3.9?
Hi, I am new to glusterFS and now I am trying it with the proxmox VE under lab situation. I have 5 nodes on proxmox VE and made a dispersed volume of 1 x (4 + 1) = 5. however, the node02 has failed to work now. so i plan to reinstall proxmox VE on node02 so i need to remove the brick on node02 but i failed with this message: gluster> volume remove-brick DispersedVol node02:/gluster/node02/brick01 force Removing brick(s) can result in data loss. Do you want to Continue? (y/n) y volume remove-brick commit force: failed: Remove brick incorrect brick count of 1 for disperse 5 is that possible to remove a brick from the dispersed volume in this kind of circumstance? here is some information of my volume: root@node04:~# gluster gluster> peer status Number of Peers: 4 Hostname: node02 State: Peer in Cluster (Disconnected) Hostname: node05 State: Peer in Cluster (Connected) Hostname: node03 State: Peer in Cluster (Connected) Hostname: node01 State: Peer in Cluster (Connected) gluster> volume info Volume Name: DispersedVol Type: Disperse Status: Started Snapshot Count: 0 Number of Bricks: 1 x (4 + 1) = 5 Transport-type: tcp Bricks: Brick1: node01:/gluster/node01/brick01 Brick2: node02:/gluster/node02/brick01 Brick3: node03:/gluster/node03/brick01 Brick4: node04:/gluster/node04/brick01 Brick5: node05:/gluster/node05/brick01 Options Reconfigured: nfs.disable: on performance.readdir-ahead: on transport.address-family: inet Anybody can help? Thanks a lot!___ Gluster-users mailing list Gluster-users@gluster.org http://www.gluster.org/mailman/listinfo/gluster-users
Re: [Gluster-users] Geo-replication failed to delete from slave file partially written to master volume.
Hi Kotresh, Logs from geo-replication test that I used to report the problem were overwritten. So I run the test again. Each time the problem is reproduced with 100% probability. This time I took all glusterfs logs from Master and Slave systems. Logs are attached to this message. All steps to duplicate the problem are reported at the text file "geo-rep-test.txt". This file is attached too. Records at the file "geo-rep-test.txt" are in chronological order. Names of volumes, bricks, nodes are the same as before. The test starts when geo-replication synchronization is working properly. After that attempt is made to write file "big.file.1" to the master that has not enough space to handle the whole file. Linux writes the file partially and geo-replication syncs this partial file "big.file.1" with the slave volume. The next write of new file "test.file" to the master fails but it creates file handler "test.file" on Master for file size zero. This new file "test.file" is not replicated to the slave volume. The next step is to delete partially written file "big.file.1" from the master to make free space on the master. The delete is successful, but it never is sync to the slave. This is the problem. File "big.file.1" is still on the slave volume. The next step is to repeat write for file "test.file". We let system to overwrite file on the master volume. Now file "test.file" has some contents. But this change did not sync to the slave. This is other flavor of the same problem. Finally new file "test.file.1" was written to the master. The file was successfully replicated to the slave. Best regards, Viktor Nosov -Original Message- From: Kotresh Hiremath Ravishankar [mailto:khire...@redhat.com] Sent: Friday, December 09, 2016 3:03 AM To: Viktor Nosov Cc: gluster-users@gluster.org Subject: Re: [Gluster-users] Geo-replication failed to delete from slave file partially written to master volume. Hi Viktor, I went through the slave logs and there are no errors w.r.t deletion of the files. I suspect the changelog has missed recording deletion of file because of which the delete is not processed. Please share the following logs from master volume to root cause the issue. 1. Master geo-replication logs: /var/log/glusterfs/geo-replication//*.log 2. Master brick logs: /var/log/glusterfs/bricks/*.log 3. Also changelogs from Master volume: /exports/nas-segment-0012/master-for-183-0003/.glusterfs/changelogs/* Thanks and Regards, Kotresh H R - Original Message - > From: "Viktor Nosov"> To: "Kotresh Hiremath Ravishankar" > Cc: gluster-users@gluster.org > Sent: Wednesday, December 7, 2016 10:48:52 PM > Subject: RE: [Gluster-users] Geo-replication failed to delete from slave file > partially written to master volume. > > Hi Kotresh, > > Thanks for looking into this issue! > I'm attaching log files from the slave node from > /var/log/glusterfs/geo-replication-slaves/ > > [root@SC-183 log]# cp > /var/log/glusterfs/geo-replication-slaves/84501a83-b07c-4768-bfaa-418b > 038e1a9e\:gluster%3A%2F%2F127.0.0.1%3Arem-volume-0001.gluster.log > /home/vnosov/ > [root@SC-183 log]# cp > /var/log/glusterfs/geo-replication-slaves/slave.log > /home/vnosov/ > [root@SC-183 log]# cp > /var/log/glusterfs/geo-replication-slaves/mbr/84501a83-b07c-4768-bfaa- > 418b038e1a9e\:gluster%3A%2F%2F127.0.0.1%3Arem-volume-0001.log > /home/vnosov/ > > Best regards, > > Viktor Nosov > > > -Original Message- > From: Kotresh Hiremath Ravishankar [mailto:khire...@redhat.com] > Sent: Tuesday, December 06, 2016 9:25 PM > To: Viktor Nosov > Cc: gluster-users@gluster.org > Subject: Re: [Gluster-users] Geo-replication failed to delete from > slave file partially written to master volume. > > Hi Viktor, > > Please share geo-replication-slave mount logs from slave nodes. > > Thanks and Regards, > Kotresh H R > > - Original Message - > > From: "Viktor Nosov" > > To: gluster-users@gluster.org > > Cc: vno...@stonefly.com > > Sent: Tuesday, December 6, 2016 7:13:22 AM > > Subject: [Gluster-users] Geo-replication failed to delete from slave file > > partially written to master volume. > > > > Hi, > > > > I hit problem while testing geo-replication. Anybody knows how to > > fix it except deleting and recreating geo-replication? > > > > Geo-replication failed to delete from slave file partially written > > to master volume. > > > > Have geo-replication between two nodes that are running glusterfs > > 3.7.16 > > > > with master volume: > > > > [root@SC-182 log]# gluster volume info master-for-183-0003 > > > > Volume Name: master-for-183-0003 > > Type: Distribute > > Volume ID: 84501a83-b07c-4768-bfaa-418b038e1a9e > > Status: Started > > Number of Bricks: 1 > > Transport-type: tcp > > Bricks: > > Brick1: 10.10.60.182:/exports/nas-segment-0012/master-for-183-0003 > > Options Reconfigured: > > changelog.changelog: on > >
[Gluster-users] Glustertests with 3 Nodes and dispersed Volumes are slow
Hi everbofy there, I am Jörg from Germany and love to build good solution with high intelligent Software in a simple way. I come on your grate software by evaluating my environment and think that I have lot to much hardware , So I try to move my NFS storage for my KVM Servers to the Servers themselves. It there are easy qcow2 images. I want the to run from local (cluster) disk. So, I start building a lab for testing my idea. The first few tests were very fine wie distributed storage and also replicated. I came a long to begin loving your work and start more tests. I read about erasure code - a raid5 style setup. So I install 3 Cents mashines and build a build a dispersed volume over 3 servers with 1 Brick on each. What I see on my simple tests was very hard. Only 40 MB/s writespeed on the local mounted glustervolume. I tested a lot - also on high perfomance hardware. But I didn't find a solution. Have you any idea why the performance is so slow? Even with a 4 Server... All are connected over 1 Gbit/s Network. Thanks for your answer... ___ Gluster-users mailing list Gluster-users@gluster.org http://www.gluster.org/mailman/listinfo/gluster-users