[Gluster-users] Gluster processes remaining after stopping glusterd
Hi, I noticed that when i stop my gluster server via systemctl stop glusterd command , one glusterfs process is still up. Which is the correct way to stop all gluster processes in my host? That's we see after run the command: *** [root@xx ~]# ps -ef | grep -i glu root 1825 1 0 Oct05 ?00:05:07 /usr/sbin/glusterfsd -s dvihcasc0s --volfile-id advdemo.dvihcasc0s.opt-glusterfs-advdemo -p /var/lib/glusterd/vols/advdemo/run/dvihcasc0s-opt-glusterfs-advdemo.pid -S /var/run/gluster/b7cbd8cac308062ef1ad823a3abf54f5.socket --brick-name /opt/glusterfs/advdemo -l /var/log/glusterfs/bricks/opt-glusterfs-advdemo.log --xlator-option *-posix.glusterd-uuid=30865b77-4da5-4f40-9945-0bd2cf55ac2a --brick-port 49152 --xlator-option advdemo-server.listen-port=49152 root 2058 1 0 Oct05 ?00:00:28 /usr/sbin/glusterfs -s localhost --volfile-id gluster/glustershd -p /var/lib/glusterd/glustershd/run/glustershd.pid -l /var/log/glusterfs/glustershd.log -S /var/run/gluster/fe4d4b13937be47e8fef6fd69be60899.socket --xlator-option *replicate*.node-uuid=30865b77-4da5-4f40-9945-0bd2cf55ac2a root 40044 39906 0 11:55 pts/000:00:00 grep --color=auto -i glu * Thanks Ismael ___ Gluster-users mailing list Gluster-users@gluster.org http://lists.gluster.org/mailman/listinfo/gluster-users
Re: [Gluster-users] Gluster CLI Feedback
Hi, 1. How often do you use the Gluster CLI? Is it a preferred method to manage Gluster? It's the only way we manage our volumes. 2. What operations do you commonly perform using the CLI? Create, replace, set, and healing info. 3. How intuitive/easy to use do you find the CLI ? it's easy enough and not confusing. 4. Is the help/information displayed sufficient? Yes, the help is suffecint enough. 5. Are there operations that are difficult to perform? no. -- Respectfully Mahdi A. Mahdi From: gluster-users-boun...@gluster.orgon behalf of Nithya Balachandran Sent: Monday, October 16, 2017 12:57 PM To: gluster-users Cc: Gluster Devel Subject: Re: [Gluster-users] Gluster CLI Feedback Gentle reminder. Thanks to those who have already responded. Nithya On 11 October 2017 at 14:38, Nithya Balachandran > wrote: Hi, As part of our initiative to improve Gluster usability, we would like feedback on the current Gluster CLI. Gluster 4.0 upstream development is currently in progress and it is an ideal time to consider CLI changes. Answers to the following would be appreciated: 1. How often do you use the Gluster CLI? Is it a preferred method to manage Gluster? 2. What operations do you commonly perform using the CLI? 3. How intuitive/easy to use do you find the CLI ? 4. Is the help/information displayed sufficient? 5. Are there operations that are difficult to perform? Regards, Nithya ___ Gluster-users mailing list Gluster-users@gluster.org http://lists.gluster.org/mailman/listinfo/gluster-users
[Gluster-users] gluster + synology
Hi, Does anyone have any experience of using Synology NAS servers as bricks in a gluster setup? The ops team at my work prefers Synology since that is what they are already using and some of the nice out-of-the-box admin features. From what I can tell Synology runs a custom linux flavor so it should be possible to compile gluster on it. Any first hand experience with it? Thanks, Ben ___ Gluster-users mailing list Gluster-users@gluster.org http://lists.gluster.org/mailman/listinfo/gluster-users
Re: [Gluster-users] gfid entries in volume heal info that do not heal
Attached is the heal log for the volume as well as the shd log. >> Run these commands on all the bricks of the replica pair to get the attrs >> set on the backend. [root@tpc-cent-glus1-081017 ~]# getfattr -d -e hex -m . /exp/b1/gv0/.glusterfs/10/86/108694db-c039-4b7c-bd3d-ad6a15d811a2 getfattr: Removing leading '/' from absolute path names # file: exp/b1/gv0/.glusterfs/10/86/108694db-c039-4b7c-bd3d-ad6a15d811a2 security.selinux=0x73797374656d5f753a6f626a6563745f723a756e6c6162656c65645f743a733000 trusted.afr.dirty=0x trusted.afr.gv0-client-2=0x0001 trusted.gfid=0x108694dbc0394b7cbd3dad6a15d811a2 trusted.gfid2path.9a2f5ada22eb9c45=0x38633262623330322d323466332d346463622d393630322d3839356136396461363131662f435f564f4c2d623030312d693637342d63642d63772e6d6435 [root@tpc-cent-glus2-081017 ~]# getfattr -d -e hex -m . /exp/b1/gv0/.glusterfs/10/86/108694db-c039-4b7c-bd3d-ad6a15d811a2 getfattr: Removing leading '/' from absolute path names # file: exp/b1/gv0/.glusterfs/10/86/108694db-c039-4b7c-bd3d-ad6a15d811a2 security.selinux=0x73797374656d5f753a6f626a6563745f723a756e6c6162656c65645f743a733000 trusted.afr.dirty=0x trusted.afr.gv0-client-2=0x0001 trusted.gfid=0x108694dbc0394b7cbd3dad6a15d811a2 trusted.gfid2path.9a2f5ada22eb9c45=0x38633262623330322d323466332d346463622d393630322d3839356136396461363131662f435f564f4c2d623030312d693637342d63642d63772e6d6435 [root@tpc-arbiter1-100617 ~]# getfattr -d -e hex -m . /exp/b1/gv0/.glusterfs/10/86/108694db-c039-4b7c-bd3d-ad6a15d811a2 getfattr: /exp/b1/gv0/.glusterfs/10/86/108694db-c039-4b7c-bd3d-ad6a15d811a2: No such file or directory [root@tpc-cent-glus1-081017 ~]# getfattr -d -e hex -m . /exp/b4/gv0/.glusterfs/e0/c5/e0c56bf7-8bfe-46ca-bde1-e46b92d33df3 getfattr: Removing leading '/' from absolute path names # file: exp/b4/gv0/.glusterfs/e0/c5/e0c56bf7-8bfe-46ca-bde1-e46b92d33df3 security.selinux=0x73797374656d5f753a6f626a6563745f723a756e6c6162656c65645f743a733000 trusted.afr.dirty=0x trusted.afr.gv0-client-11=0x0001 trusted.gfid=0xe0c56bf78bfe46cabde1e46b92d33df3 trusted.gfid2path.be3ba24c3ef95ff2=0x63323366353834652d353566652d343033382d393131622d3866373063656334616136662f435f564f4c2d623030332d69313331342d63642d636d2d63722e6d6435 [root@tpc-cent-glus2-081017 ~]# getfattr -d -e hex -m . /exp/b4/gv0/.glusterfs/e0/c5/e0c56bf7-8bfe-46ca-bde1-e46b92d33df3 getfattr: Removing leading '/' from absolute path names # file: exp/b4/gv0/.glusterfs/e0/c5/e0c56bf7-8bfe-46ca-bde1-e46b92d33df3 security.selinux=0x73797374656d5f753a6f626a6563745f723a756e6c6162656c65645f743a733000 trusted.afr.dirty=0x trusted.afr.gv0-client-11=0x0001 trusted.gfid=0xe0c56bf78bfe46cabde1e46b92d33df3 trusted.gfid2path.be3ba24c3ef95ff2=0x63323366353834652d353566652d343033382d393131622d3866373063656334616136662f435f564f4c2d623030332d69313331342d63642d636d2d63722e6d6435 [root@tpc-arbiter1-100617 ~]# getfattr -d -e hex -m . /exp/b4/gv0/.glusterfs/e0/c5/e0c56bf7-8bfe-46ca-bde1-e46b92d33df3 getfattr: /exp/b4/gv0/.glusterfs/e0/c5/e0c56bf7-8bfe-46ca-bde1-e46b92d33df3: No such file or directory >> And the output of "gluster volume heal info split-brain" [root@tpc-cent-glus1-081017 ~]# gluster volume heal gv0 info split-brain Brick tpc-cent-glus1-081017:/exp/b1/gv0 Status: Connected Number of entries in split-brain: 0 Brick tpc-cent-glus2-081017:/exp/b1/gv0 Status: Connected Number of entries in split-brain: 0 Brick tpc-arbiter1-100617:/exp/b1/gv0 Status: Connected Number of entries in split-brain: 0 Brick tpc-cent-glus1-081017:/exp/b2/gv0 Status: Connected Number of entries in split-brain: 0 Brick tpc-cent-glus2-081017:/exp/b2/gv0 Status: Connected Number of entries in split-brain: 0 Brick tpc-arbiter1-100617:/exp/b2/gv0 Status: Connected Number of entries in split-brain: 0 Brick tpc-cent-glus1-081017:/exp/b3/gv0 Status: Connected Number of entries in split-brain: 0 Brick tpc-cent-glus2-081017:/exp/b3/gv0 Status: Connected Number of entries in split-brain: 0 Brick tpc-arbiter1-100617:/exp/b3/gv0 Status: Connected Number of entries in split-brain: 0 Brick tpc-cent-glus1-081017:/exp/b4/gv0 Status: Connected Number of entries in split-brain: 0 Brick tpc-cent-glus2-081017:/exp/b4/gv0 Status: Connected Number of entries in split-brain: 0 Brick tpc-arbiter1-100617:/exp/b4/gv0 Status: Connected Number of entries in split-brain: 0 -Matt From: Karthik Subrahmanya [mailto:ksubr...@redhat.com] Sent: Tuesday, October 17, 2017 1:26 AM To: Matt WaymackCc: gluster-users Subject: Re: [Gluster-users] gfid entries in volume heal info that do not heal Hi Matt, Run these commands on all the bricks of the replica pair to get the attrs set on the backend. On the bricks of first replica set: getfattr -d -e hex -m . /.glusterfs/10/86/108694db-c039-4b7c-bd3d-ad6a15d811a2 On the fourth
Re: [Gluster-users] Distribute rebalance issues
On 17 October 2017 at 16:39, Stephen Remdewrote: > Nithya, > > Is there any way to increase the logging level of the brick? There is > nothing obvious (to me) in the log (see below for the same time period as > the latest rebalance failure). This is the only brick on that server that > has disconnects like this. > You can use gluster volume set brick-log-level DEBUG or gluster volume set brick-log-level TRACE @Milind and Raghavendra G , can you take a look at this to see why there are so many disconnects? Regards, Nithya > > Steve > > [2017-10-17 02:22:13.453575] I [MSGID: 115029] > [server-handshake.c:692:server_setvolume] 0-video-server: accepted client > from node-dc4-03-5825-2017/08/30-20:45:55:170091-video-client-4-2-318 > (version: 3.8.15) > [2017-10-17 02:22:31.353286] I [MSGID: 115036] > [server.c:548:server_rpc_notify] 0-video-server: disconnecting connection > from node-dc4-02-29040-2017/08/04-09:31:22:842268-video-client-4-7-403 > [2017-10-17 02:22:31.353326] I [MSGID: 101055] > [client_t.c:415:gf_client_unref] 0-video-server: Shutting down connection > node-dc4-02-29040-2017/08/04-09:31:22:842268-video-client-4-7-403 > [2017-10-17 02:22:42.288856] I [MSGID: 115029] > [server-handshake.c:692:server_setvolume] 0-video-server: accepted client > from node-dc4-02-29040-2017/08/04-09:31:22:842268-video-client-4-7-404 > (version: 3.8.13) > [2017-10-17 02:29:04.889303] I [MSGID: 115036] > [server.c:548:server_rpc_notify] 0-video-server: disconnecting connection > from node-dc4-02-29040-2017/08/04-09:31:22:842268-video-client-4-7-404 > [2017-10-17 02:29:04.889347] I [MSGID: 101055] > [client_t.c:415:gf_client_unref] 0-video-server: Shutting down connection > node-dc4-02-29040-2017/08/04-09:31:22:842268-video-client-4-7-404 > [2017-10-17 02:29:15.327604] I [MSGID: 115029] > [server-handshake.c:692:server_setvolume] 0-video-server: accepted client > from node-dc4-02-29040-2017/08/04-09:31:22:842268-video-client-4-7-405 > (version: 3.8.13) > [2017-10-17 02:33:30.745314] I [MSGID: 115036] > [server.c:548:server_rpc_notify] 0-video-server: disconnecting connection > from node-dc4-01-6174-2017/07/13-10:46:48:503667-video-client-4-7-594 > [2017-10-17 02:33:30.745360] I [MSGID: 115013] > [server-helpers.c:293:do_fd_cleanup] 0-video-server: fd cleanup on /xx > [2017-10-17 02:33:30.745396] I [MSGID: 101055] > [client_t.c:415:gf_client_unref] 0-video-server: Shutting down connection > node-dc4-01-6174-2017/07/13-10:46:48:503667-video-client-4-7-594 > [2017-10-17 02:33:41.563748] I [MSGID: 115029] > [server-handshake.c:692:server_setvolume] 0-video-server: accepted client > from node-dc4-01-6174-2017/07/13-10:46:48:503667-video-client-4-7-595 > (version: 3.8.13) > [2017-10-17 02:36:43.833304] I [MSGID: 115036] > [server.c:548:server_rpc_notify] 0-video-server: disconnecting connection > from node-dc4-01-6174-2017/07/13-10:46:48:503667-video-client-4-7-595 > [2017-10-17 02:36:43.833342] I [MSGID: 115013] > [server-helpers.c:293:do_fd_cleanup] 0-video-server: fd cleanup on /xx > [2017-10-17 02:36:43.833371] I [MSGID: 101055] > [client_t.c:415:gf_client_unref] 0-video-server: Shutting down connection > node-dc4-01-6174-2017/07/13-10:46:48:503667-video-client-4-7-595 > [2017-10-17 02:36:54.569836] I [MSGID: 115029] > [server-handshake.c:692:server_setvolume] 0-video-server: accepted client > from node-dc4-01-6174-2017/07/13-10:46:48:503667-video-client-4-7-596 > (version: 3.8.13) > [2017-10-17 02:38:16.697306] I [MSGID: 115036] > [server.c:548:server_rpc_notify] 0-video-server: disconnecting connection > from node-dc4-01-6174-2017/07/13-10:46:48:503667-video-client-4-7-596 > [2017-10-17 02:38:16.697370] I [MSGID: 115013] > [server-helpers.c:293:do_fd_cleanup] 0-video-server: fd cleanup on /xx > [2017-10-17 02:38:16.697432] I [MSGID: 101055] > [client_t.c:415:gf_client_unref] 0-video-server: Shutting down connection > node-dc4-01-6174-2017/07/13-10:46:48:503667-video-client-4-7-596 > [2017-10-17 02:38:34.591506] I [MSGID: 115029] > [server-handshake.c:692:server_setvolume] 0-video-server: accepted client > from node-dc4-01-6174-2017/07/13-10:46:48:503667-video-client-4-7-597 > (version: 3.8.13) > [2017-10-17 02:55:56.473306] I [MSGID: 115036] > [server.c:548:server_rpc_notify] 0-video-server: disconnecting connection > from titan-17527-2017/09/18-19:57:41:611709-video-client-4-0-19 > [2017-10-17 02:55:56.473366] I [MSGID: 101055] > [client_t.c:415:gf_client_unref] 0-video-server: Shutting down connection > titan-17527-2017/09/18-19:57:41:611709-video-client-4-0-19 > [2017-10-17 02:56:07.161790] I [MSGID: 115029] > [server-handshake.c:692:server_setvolume] 0-video-server: accepted client > from titan-17527-2017/09/18-19:57:41:611709-video-client-4-0-20 (version: > 3.8.8) > [2017-10-17 03:15:13.529281] I [MSGID: 115036] > [server.c:548:server_rpc_notify] 0-video-server: disconnecting connection > from
Re: [Gluster-users] Distribute rebalance issues
Nithya, Is there any way to increase the logging level of the brick? There is nothing obvious (to me) in the log (see below for the same time period as the latest rebalance failure). This is the only brick on that server that has disconnects like this. Steve [2017-10-17 02:22:13.453575] I [MSGID: 115029] [server-handshake.c:692:server_setvolume] 0-video-server: accepted client from node-dc4-03-5825-2017/08/30-20:45:55:170091-video-client-4-2-318 (version: 3.8.15) [2017-10-17 02:22:31.353286] I [MSGID: 115036] [server.c:548:server_rpc_notify] 0-video-server: disconnecting connection from node-dc4-02-29040-2017/08/04-09:31:22:842268-video-client-4-7-403 [2017-10-17 02:22:31.353326] I [MSGID: 101055] [client_t.c:415:gf_client_unref] 0-video-server: Shutting down connection node-dc4-02-29040-2017/08/04-09:31:22:842268-video-client-4-7-403 [2017-10-17 02:22:42.288856] I [MSGID: 115029] [server-handshake.c:692:server_setvolume] 0-video-server: accepted client from node-dc4-02-29040-2017/08/04-09:31:22:842268-video-client-4-7-404 (version: 3.8.13) [2017-10-17 02:29:04.889303] I [MSGID: 115036] [server.c:548:server_rpc_notify] 0-video-server: disconnecting connection from node-dc4-02-29040-2017/08/04-09:31:22:842268-video-client-4-7-404 [2017-10-17 02:29:04.889347] I [MSGID: 101055] [client_t.c:415:gf_client_unref] 0-video-server: Shutting down connection node-dc4-02-29040-2017/08/04-09:31:22:842268-video-client-4-7-404 [2017-10-17 02:29:15.327604] I [MSGID: 115029] [server-handshake.c:692:server_setvolume] 0-video-server: accepted client from node-dc4-02-29040-2017/08/04-09:31:22:842268-video-client-4-7-405 (version: 3.8.13) [2017-10-17 02:33:30.745314] I [MSGID: 115036] [server.c:548:server_rpc_notify] 0-video-server: disconnecting connection from node-dc4-01-6174-2017/07/13-10:46:48:503667-video-client-4-7-594 [2017-10-17 02:33:30.745360] I [MSGID: 115013] [server-helpers.c:293:do_fd_cleanup] 0-video-server: fd cleanup on /xx [2017-10-17 02:33:30.745396] I [MSGID: 101055] [client_t.c:415:gf_client_unref] 0-video-server: Shutting down connection node-dc4-01-6174-2017/07/13-10:46:48:503667-video-client-4-7-594 [2017-10-17 02:33:41.563748] I [MSGID: 115029] [server-handshake.c:692:server_setvolume] 0-video-server: accepted client from node-dc4-01-6174-2017/07/13-10:46:48:503667-video-client-4-7-595 (version: 3.8.13) [2017-10-17 02:36:43.833304] I [MSGID: 115036] [server.c:548:server_rpc_notify] 0-video-server: disconnecting connection from node-dc4-01-6174-2017/07/13-10:46:48:503667-video-client-4-7-595 [2017-10-17 02:36:43.833342] I [MSGID: 115013] [server-helpers.c:293:do_fd_cleanup] 0-video-server: fd cleanup on /xx [2017-10-17 02:36:43.833371] I [MSGID: 101055] [client_t.c:415:gf_client_unref] 0-video-server: Shutting down connection node-dc4-01-6174-2017/07/13-10:46:48:503667-video-client-4-7-595 [2017-10-17 02:36:54.569836] I [MSGID: 115029] [server-handshake.c:692:server_setvolume] 0-video-server: accepted client from node-dc4-01-6174-2017/07/13-10:46:48:503667-video-client-4-7-596 (version: 3.8.13) [2017-10-17 02:38:16.697306] I [MSGID: 115036] [server.c:548:server_rpc_notify] 0-video-server: disconnecting connection from node-dc4-01-6174-2017/07/13-10:46:48:503667-video-client-4-7-596 [2017-10-17 02:38:16.697370] I [MSGID: 115013] [server-helpers.c:293:do_fd_cleanup] 0-video-server: fd cleanup on /xx [2017-10-17 02:38:16.697432] I [MSGID: 101055] [client_t.c:415:gf_client_unref] 0-video-server: Shutting down connection node-dc4-01-6174-2017/07/13-10:46:48:503667-video-client-4-7-596 [2017-10-17 02:38:34.591506] I [MSGID: 115029] [server-handshake.c:692:server_setvolume] 0-video-server: accepted client from node-dc4-01-6174-2017/07/13-10:46:48:503667-video-client-4-7-597 (version: 3.8.13) [2017-10-17 02:55:56.473306] I [MSGID: 115036] [server.c:548:server_rpc_notify] 0-video-server: disconnecting connection from titan-17527-2017/09/18-19:57:41:611709-video-client-4-0-19 [2017-10-17 02:55:56.473366] I [MSGID: 101055] [client_t.c:415:gf_client_unref] 0-video-server: Shutting down connection titan-17527-2017/09/18-19:57:41:611709-video-client-4-0-19 [2017-10-17 02:56:07.161790] I [MSGID: 115029] [server-handshake.c:692:server_setvolume] 0-video-server: accepted client from titan-17527-2017/09/18-19:57:41:611709-video-client-4-0-20 (version: 3.8.8) [2017-10-17 03:15:13.529281] I [MSGID: 115036] [server.c:548:server_rpc_notify] 0-video-server: disconnecting connection from node-dc4-01-6174-2017/07/13-10:46:48:503667-video-client-4-7-597 [2017-10-17 03:15:13.529330] I [MSGID: 115013] [server-helpers.c:293:do_fd_cleanup] 0-video-server: fd cleanup on /xx [2017-10-17 03:15:13.529400] I [MSGID: 101055] [client_t.c:415:gf_client_unref] 0-video-server: Shutting down connection node-dc4-01-6174-2017/07/13-10:46:48:503667-video-client-4-7-597 [2017-10-17 03:15:41.764247] I [MSGID: 115029] [server-handshake.c:692:server_setvolume] 0-video-server: accepted client from
[Gluster-users] gluster usecase CDN
hi everybody, as " http://www.itzgeek.com/how-tos/linux/centos-how-tos/install-and-configure-glusterfs-on-centos-7-rhel-7.html; says: *GlusterFS *is an open-source, scalable network filesystem *suitable *for high data-intensive workloads such as media streaming, cloud storage, and* CDN (Content Delivery Network).* I have googled alot, but couldn't find any architecture for CDN as a use case of glusterfs. Can anybody help me in this case? Thx alot gluster-users;) ___ Gluster-users mailing list Gluster-users@gluster.org http://lists.gluster.org/mailman/listinfo/gluster-users
Re: [Gluster-users] Distribute rebalance issues
On 17 October 2017 at 14:48, Stephen Remdewrote: > Hi, > > > I have a rebalance that has failed on one peer twice now. Rebalance logs > below (directories anonomised and some irrelevant log lines cut). It looks > like it loses connection to the brick, but immediately stops the rebalance on > that peer instead of waiting for reconnection - which happens a second or so > later. > Is this normal behaviour? So far it has been the same server and the same > (remote) brick. > > > The brick shows a high number of disconnects compared to the other bricks on > the same server > > > ./export-md0-brick.log.1 2 > ./export-md1-brick.log.1 2 > ./export-md2-brick.log.1181 > ./export-md3-brick.log.1 2 > > > Any clues? What could be causing this because there is nothing in the log to > indicate cause. > > The rebalance process requires that all DHT child subvols be up during the operation as it needs to reapply the directory layouts (which requires all child subvols to be up). As this is a pure distribute volume, even a single brick getting disconnected is enough to cause the process to stop. You would need to figure out why that brick is disconnecting so often. The brick logs might help with that. Regards, Nithya > > Steve > > > gluster volume info video > > Volume Name: video > Type: Distribute > Volume ID: ccdac37f-9b0e-415f-b62e-9071d8168199 > Status: Started > Snapshot Count: 0 > Number of Bricks: 9 > Transport-type: tcp > Bricks: > Brick1: 10.0.0.31:/export/md0/brick > Brick2: 10.0.0.32:/export/md0/brick > Brick3: 10.0.0.31:/export/md1/brick > Brick4: 10.0.0.32:/export/md1/brick > Brick5: 10.0.0.31:/export/md2/brick > Brick6: 10.0.0.32:/export/md2/brick > Brick7: 10.0.0.31:/export/md3/brick > Brick8: 10.0.0.32:/export/md3/brick > Brick9: 10.0.0.33:/export/md0/brick > Options Reconfigured: > network.ping-timeout: 10 > cluster.min-free-disk: 1% > transport.address-family: inet > performance.readdir-ahead: on > nfs.disable: on > cluster.rebal-throttle: lazy > > [2017-10-12 23:00:55.099153] W [socket.c:590:__socket_rwv] 0-video-client-4: > readv on 10.0.0.31:49164 failed (Connection reset by peer) > [2017-10-12 23:00:55.099709] I [MSGID: 114018] > [client.c:2280:client_rpc_notify] 0-video-client-4: disconnected from > video-client-4. Client process will keep trying to connect to glusterd until > brick's port is available > [2017-10-12 23:00:55.099741] W [MSGID: 109073] [dht-common.c:8839:dht_notify] > 0-video-dht: Received CHILD_DOWN. Exiting > [2017-10-12 23:00:55.099752] I [MSGID: 109029] > [dht-rebalance.c:4195:gf_defrag_stop] 0-: Received stop command on rebalance > [2017-10-12 23:01:05.478462] I [rpc-clnt.c:1947:rpc_clnt_reconfig] > 0-video-client-4: changing port to 49164 (from 0) > [2017-10-12 23:01:05.481180] I [MSGID: 114057] > [client-handshake.c:1446:select_server_supported_programs] 0-video-client-4: > Using Program GlusterFS 3.3, Num (1298437), Version (330) > [2017-10-12 23:01:05.482630] I [MSGID: 114046] > [client-handshake.c:1222:client_setvolume_cbk] 0-video-client-4: Connected to > video-client-4, attached to remote volume '/export/md2/brick'. > [2017-10-12 23:01:05.482659] I [MSGID: 114047] > [client-handshake.c:1233:client_setvolume_cbk] 0-video-client-4: Server and > Client lk-version numbers are not same, reopening the fds > [2017-10-12 23:01:05.483365] I [MSGID: 114035] > [client-handshake.c:201:client_set_lk_version_cbk] 0-video-client-4: Server > lk version = 1 > [2017-10-12 23:01:30.310089] I [dht-rebalance.c:2819:gf_defrag_process_dir] > 0-DHT: Found critical error from gf_defrag_get_entry > [2017-10-12 23:01:30.310166] E [MSGID: 109111] > [dht-rebalance.c:3090:gf_defrag_fix_layout] 0-video-dht: > gf_defrag_process_dir failed for directory: /y/y/y/y/y > [2017-10-12 23:01:30.380574] E [MSGID: 109016] > [dht-rebalance.c:3267:gf_defrag_fix_layout] 0-video-dht: Fix layout failed > for /y/y/y/y/y > [2017-10-12 23:01:30.380756] E [MSGID: 109016] > [dht-rebalance.c:3267:gf_defrag_fix_layout] 0-video-dht: Fix layout failed > for /y/y/y/y > [2017-10-12 23:01:30.380879] E [MSGID: 109016] > [dht-rebalance.c:3267:gf_defrag_fix_layout] 0-video-dht: Fix layout failed > for /y/y/y > [2017-10-12 23:01:30.380965] E [MSGID: 109016] > [dht-rebalance.c:3267:gf_defrag_fix_layout] 0-video-dht: Fix layout failed > for /y/y > [2017-10-12 23:03:09.285157] W [glusterfsd.c:1327:cleanup_and_exit] > (-->/lib/x86_64-linux-gnu/libpthread.so.0(+0x76ba) [0x7f112b6d16ba] > -->/usr/sbin/glusterfs(glusterfs_sigwaiter+0xe5) [0x55b325019545] > -->/usr/sbin/glusterfs(cleanup_and_exit+0x54) [0x55b3250193b4] ) 0-: received > signum (15), shutting down > > [2017-10-17 03:20:28.921512] W [socket.c:590:__socket_rwv] 0-video-client-4: > readv on 10.0.0.31:49164 failed (Connection reset by peer) > [2017-10-17 03:20:28.921554] I [MSGID: 114018] > [client.c:2280:client_rpc_notify] 0-video-client-4: disconnected from > video-client-4. Client process
[Gluster-users] Distribute rebalance issues
Hi, I have a rebalance that has failed on one peer twice now. Rebalance logs below (directories anonomised and some irrelevant log lines cut). It looks like it loses connection to the brick, but immediately stops the rebalance on that peer instead of waiting for reconnection - which happens a second or so later. Is this normal behaviour? So far it has been the same server and the same (remote) brick. The brick shows a high number of disconnects compared to the other bricks on the same server ./export-md0-brick.log.1 2 ./export-md1-brick.log.1 2 ./export-md2-brick.log.1181 ./export-md3-brick.log.1 2 Any clues? What could be causing this because there is nothing in the log to indicate cause. Steve gluster volume info video Volume Name: video Type: Distribute Volume ID: ccdac37f-9b0e-415f-b62e-9071d8168199 Status: Started Snapshot Count: 0 Number of Bricks: 9 Transport-type: tcp Bricks: Brick1: 10.0.0.31:/export/md0/brick Brick2: 10.0.0.32:/export/md0/brick Brick3: 10.0.0.31:/export/md1/brick Brick4: 10.0.0.32:/export/md1/brick Brick5: 10.0.0.31:/export/md2/brick Brick6: 10.0.0.32:/export/md2/brick Brick7: 10.0.0.31:/export/md3/brick Brick8: 10.0.0.32:/export/md3/brick Brick9: 10.0.0.33:/export/md0/brick Options Reconfigured: network.ping-timeout: 10 cluster.min-free-disk: 1% transport.address-family: inet performance.readdir-ahead: on nfs.disable: on cluster.rebal-throttle: lazy [2017-10-12 23:00:55.099153] W [socket.c:590:__socket_rwv] 0-video-client-4: readv on 10.0.0.31:49164 failed (Connection reset by peer) [2017-10-12 23:00:55.099709] I [MSGID: 114018] [client.c:2280:client_rpc_notify] 0-video-client-4: disconnected from video-client-4. Client process will keep trying to connect to glusterd until brick's port is available [2017-10-12 23:00:55.099741] W [MSGID: 109073] [dht-common.c:8839:dht_notify] 0-video-dht: Received CHILD_DOWN. Exiting [2017-10-12 23:00:55.099752] I [MSGID: 109029] [dht-rebalance.c:4195:gf_defrag_stop] 0-: Received stop command on rebalance [2017-10-12 23:01:05.478462] I [rpc-clnt.c:1947:rpc_clnt_reconfig] 0-video-client-4: changing port to 49164 (from 0) [2017-10-12 23:01:05.481180] I [MSGID: 114057] [client-handshake.c:1446:select_server_supported_programs] 0-video-client-4: Using Program GlusterFS 3.3, Num (1298437), Version (330) [2017-10-12 23:01:05.482630] I [MSGID: 114046] [client-handshake.c:1222:client_setvolume_cbk] 0-video-client-4: Connected to video-client-4, attached to remote volume '/export/md2/brick'. [2017-10-12 23:01:05.482659] I [MSGID: 114047] [client-handshake.c:1233:client_setvolume_cbk] 0-video-client-4: Server and Client lk-version numbers are not same, reopening the fds [2017-10-12 23:01:05.483365] I [MSGID: 114035] [client-handshake.c:201:client_set_lk_version_cbk] 0-video-client-4: Server lk version = 1 [2017-10-12 23:01:30.310089] I [dht-rebalance.c:2819:gf_defrag_process_dir] 0-DHT: Found critical error from gf_defrag_get_entry [2017-10-12 23:01:30.310166] E [MSGID: 109111] [dht-rebalance.c:3090:gf_defrag_fix_layout] 0-video-dht: gf_defrag_process_dir failed for directory: /y/y/y/y/y [2017-10-12 23:01:30.380574] E [MSGID: 109016] [dht-rebalance.c:3267:gf_defrag_fix_layout] 0-video-dht: Fix layout failed for /y/y/y/y/y [2017-10-12 23:01:30.380756] E [MSGID: 109016] [dht-rebalance.c:3267:gf_defrag_fix_layout] 0-video-dht: Fix layout failed for /y/y/y/y [2017-10-12 23:01:30.380879] E [MSGID: 109016] [dht-rebalance.c:3267:gf_defrag_fix_layout] 0-video-dht: Fix layout failed for /y/y/y [2017-10-12 23:01:30.380965] E [MSGID: 109016] [dht-rebalance.c:3267:gf_defrag_fix_layout] 0-video-dht: Fix layout failed for /y/y [2017-10-12 23:03:09.285157] W [glusterfsd.c:1327:cleanup_and_exit] (-->/lib/x86_64-linux-gnu/libpthread.so.0(+0x76ba) [0x7f112b6d16ba] -->/usr/sbin/glusterfs(glusterfs_sigwaiter+0xe5) [0x55b325019545] -->/usr/sbin/glusterfs(cleanup_and_exit+0x54) [0x55b3250193b4] ) 0-: received signum (15), shutting down [2017-10-17 03:20:28.921512] W [socket.c:590:__socket_rwv] 0-video-client-4: readv on 10.0.0.31:49164 failed (Connection reset by peer) [2017-10-17 03:20:28.921554] I [MSGID: 114018] [client.c:2280:client_rpc_notify] 0-video-client-4: disconnected from video-client-4. Client process will keep trying to connect to glusterd until brick's port is available [2017-10-17 03:20:28.921570] W [MSGID: 109073] [dht-common.c:8839:dht_notify] 0-video-dht: Received CHILD_DOWN. Exiting [2017-10-17 03:20:28.921578] I [MSGID: 109029] [dht-rebalance.c:4195:gf_defrag_stop] 0-: Received stop command on rebalance [2017-10-17 03:20:39.344417] I [rpc-clnt.c:1947:rpc_clnt_reconfig] 0-video-client-4: changing port to 49164 (from 0) [2017-10-17 03:20:39.347440] I [MSGID: 114057] [client-handshake.c:1446:select_server_supported_programs] 0-video-client-4: Using Program GlusterFS 3.3, Num (1298437), Version (330) [2017-10-17 03:20:39.349244] I [MSGID: 114046] [client-handshake.c:1222:client_setvolume_cbk] 0-video-client-4: Connected to
Re: [Gluster-users] Gluster CLI reference
Hi, I think the cli-main should contain all, even most obscure options. Additionally, I'm not sure whether this is feasible, but maybe the cli-main could include links to all places where an example usage of a option is given. The goal would be: I go to the cli-main page and if there is something I want to do I have links with the examples of the usage. Marcin On Tue, Oct 17, 2017 at 7:55 AM, Nithya Balachandranwrote: > Hi, > > As part of our initiative to improve our docs, we have made a few changes > over the past few weeks. One of this is a CLI reference [1]. This is still > a WIP so not all commands have been documented. > > Is this something you would find useful? > Would you like to see more information captured as part of this? If yes, > what else should we include? > > Thanks, > Nithya > > [1]: http://docs.gluster.org/en/latest/CLI-Reference/cli-main/ > > ___ > Gluster-users mailing list > Gluster-users@gluster.org > http://lists.gluster.org/mailman/listinfo/gluster-users > ___ Gluster-users mailing list Gluster-users@gluster.org http://lists.gluster.org/mailman/listinfo/gluster-users
Re: [Gluster-users] gfid entries in volume heal info that do not heal
Hi Matt, Run these commands on all the bricks of the replica pair to get the attrs set on the backend. On the bricks of first replica set: getfattr -d -e hex -m . /.glusterfs/10/86/ 108694db-c039-4b7c-bd3d-ad6a15d811a2 On the fourth replica set: getfattr -d -e hex -m . /.glusterfs/ e0/c5/e0c56bf7-8bfe-46ca-bde1-e46b92d33df3 Also run the "gluster volume heal " once and send the shd log. And the output of "gluster volume heal info split-brain" Regards, Karthik On Mon, Oct 16, 2017 at 9:51 PM, Matt Waymackwrote: > OK, so here’s my output of the volume info and the heal info. I have not > yet tracked down physical location of these files, any tips to finding them > would be appreciated, but I’m definitely just wanting them gone. I forgot > to mention earlier that the cluster is running 3.12 and was upgraded from > 3.10; these files were likely stuck like this when it was on 3.10. > > > > [root@tpc-cent-glus1-081017 ~]# gluster volume info gv0 > > > > Volume Name: gv0 > > Type: Distributed-Replicate > > Volume ID: 8f07894d-e3ab-4a65-bda1-9d9dd46db007 > > Status: Started > > Snapshot Count: 0 > > Number of Bricks: 4 x (2 + 1) = 12 > > Transport-type: tcp > > Bricks: > > Brick1: tpc-cent-glus1-081017:/exp/b1/gv0 > > Brick2: tpc-cent-glus2-081017:/exp/b1/gv0 > > Brick3: tpc-arbiter1-100617:/exp/b1/gv0 (arbiter) > > Brick4: tpc-cent-glus1-081017:/exp/b2/gv0 > > Brick5: tpc-cent-glus2-081017:/exp/b2/gv0 > > Brick6: tpc-arbiter1-100617:/exp/b2/gv0 (arbiter) > > Brick7: tpc-cent-glus1-081017:/exp/b3/gv0 > > Brick8: tpc-cent-glus2-081017:/exp/b3/gv0 > > Brick9: tpc-arbiter1-100617:/exp/b3/gv0 (arbiter) > > Brick10: tpc-cent-glus1-081017:/exp/b4/gv0 > > Brick11: tpc-cent-glus2-081017:/exp/b4/gv0 > > Brick12: tpc-arbiter1-100617:/exp/b4/gv0 (arbiter) > > Options Reconfigured: > > nfs.disable: on > > transport.address-family: inet > > > > [root@tpc-cent-glus1-081017 ~]# gluster volume heal gv0 info > > Brick tpc-cent-glus1-081017:/exp/b1/gv0 > > > > > > > > > > > > > > > > > > Status: Connected > > Number of entries: 118 > > > > Brick tpc-cent-glus2-081017:/exp/b1/gv0 > > > > > > > > > > > > > > > > > > Status: Connected > > Number of entries: 118 > > > > Brick tpc-arbiter1-100617:/exp/b1/gv0 > > Status: Connected > > Number of entries: 0 > > > > Brick tpc-cent-glus1-081017:/exp/b2/gv0 > > Status: Connected > > Number of entries: 0 > > > > Brick tpc-cent-glus2-081017:/exp/b2/gv0 > > Status: Connected > > Number of entries: 0 > > > > Brick tpc-arbiter1-100617:/exp/b2/gv0 > > Status: Connected > > Number of entries: 0 > > > > Brick tpc-cent-glus1-081017:/exp/b3/gv0 > > Status: Connected > > Number of entries: 0 > > > > Brick tpc-cent-glus2-081017:/exp/b3/gv0 > > Status: Connected > > Number of entries: 0 > > > > Brick tpc-arbiter1-100617:/exp/b3/gv0 > > Status: Connected > > Number of entries: 0 > > > > Brick tpc-cent-glus1-081017:/exp/b4/gv0 > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > Status: Connected > > Number of entries: 24 > > > > Brick tpc-cent-glus2-081017:/exp/b4/gv0 > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > Status: Connected > > Number of entries: 24 > > > > Brick tpc-arbiter1-100617:/exp/b4/gv0 > > Status: Connected > > Number of entries: 0 > > > > Thank you for your help! > > > > *From:* Karthik Subrahmanya [mailto:ksubr...@redhat.com] > *Sent:* Monday, October 16, 2017 10:27 AM > *To:* Matt Waymack > *Cc:* gluster-users > *Subject:* Re: [Gluster-users] gfid entries in volume heal info that do > not heal > > > > Hi Matt, > > > > The files might be in split brain. Could you please send the outputs of > these? > > gluster volume info > > gluster volume heal info > > And also the getfattr output of the files which are in the heal info > output from all the bricks of that replica pair. > > getfattr -d -e hex -m . > > > > Thanks & Regards > > Karthik > > > > On 16-Oct-2017 8:16 PM, "Matt Waymack" wrote: > > Hi all, > > > > I have a volume where the output of volume heal info shows several gfid > entries to be healed, but they’ve been there for weeks and have not > healed. Any normal file that shows up on the heal info does get healed as > expected, but these gfid entries do not. Is there any way to remove these > orphaned entries from the volume so they are no longer stuck in the heal > process? > > > > Thank you! > > > ___ > Gluster-users mailing list > Gluster-users@gluster.org > http://lists.gluster.org/mailman/listinfo/gluster-users > > > ___ Gluster-users mailing list Gluster-users@gluster.org http://lists.gluster.org/mailman/listinfo/gluster-users