On Mon, May 22, 2017 at 7:51 PM, Atin Mukherjee <amukh...@redhat.com> wrote:
> Sorry Pawan, I did miss the other part of the attachments. So looking from > the glusterd.info file from all the hosts, it looks like host2 and host3 > do not have the correct op-version. Can you please set the op-version as > "operating-version=30702" in host2 and host3 and restart glusterd instance > one by one on all the nodes? > Please ensure that all the hosts are upgraded to the same bits before doing this change. > > Apparently it looks like there is a bug which you have uncovered, during > peer handshaking if one of the glusterd instance is running with old bits > then during validating the handshake request there is a possibility that > uuid received will be blank and the same was ignored however there was a > patch http://review.gluster.org/13519 which had some additional changes > which was always looking at this field and doing some extra checks which > was causing the handshake to fail. For now, the above workaround should > suffice. I'll be sending a patch pretty soon. > Posted a patch https://review.gluster.org/#/c/17358 . > > > > On Mon, May 22, 2017 at 11:35 AM, Pawan Alwandi <pa...@platform.sh> wrote: > >> Hello Atin, >> >> The tar's have the content of `/var/lib/glusterd` too for all 3 nodes, >> please check again. >> >> Thanks >> >> On Mon, May 22, 2017 at 11:32 AM, Atin Mukherjee <amukh...@redhat.com> >> wrote: >> >>> Pawan, >>> >>> I see you have provided the log files from the nodes, however it'd be >>> really helpful if you can provide me the content of /var/lib/glusterd from >>> all the nodes to get to the root cause of this issue. >>> >>> On Fri, May 19, 2017 at 12:09 PM, Pawan Alwandi <pa...@platform.sh> >>> wrote: >>> >>>> Hello Atin, >>>> >>>> Thanks for continued support. I've attached requested files from all 3 >>>> nodes. >>>> >>>> (I think we already verified the UUIDs to be correct, anyway let us >>>> know if you find any more info in the logs) >>>> >>>> Pawan >>>> >>>> On Thu, May 18, 2017 at 11:45 PM, Atin Mukherjee <amukh...@redhat.com> >>>> wrote: >>>> >>>>> >>>>> On Thu, 18 May 2017 at 23:40, Atin Mukherjee <amukh...@redhat.com> >>>>> wrote: >>>>> >>>>>> On Wed, 17 May 2017 at 12:47, Pawan Alwandi <pa...@platform.sh> >>>>>> wrote: >>>>>> >>>>>>> Hello Atin, >>>>>>> >>>>>>> I realized that these http://gluster.readthedocs.io/ >>>>>>> en/latest/Upgrade-Guide/upgrade_to_3.10/ instructions only work for >>>>>>> upgrades from 3.7, while we are running 3.6.2. Are there >>>>>>> instructions/suggestion you have for us to upgrade from 3.6 version? >>>>>>> >>>>>>> I believe upgrade from 3.6 to 3.7 and then to 3.10 would work, but I >>>>>>> see similar errors reported when I upgraded to 3.7 too. >>>>>>> >>>>>>> For what its worth, I was able to set the op-version (gluster v set >>>>>>> all cluster.op-version 30702) but that doesn't seem to help. >>>>>>> >>>>>>> [2017-05-17 06:48:33.700014] I [MSGID: 100030] >>>>>>> [glusterfsd.c:2338:main] 0-/usr/sbin/glusterd: Started running >>>>>>> /usr/sbin/glusterd version 3.7.20 (args: /usr/sbin/glusterd -p >>>>>>> /var/run/glusterd.pid) >>>>>>> [2017-05-17 06:48:33.703808] I [MSGID: 106478] >>>>>>> [glusterd.c:1383:init] 0-management: Maximum allowed open file >>>>>>> descriptors >>>>>>> set to 65536 >>>>>>> [2017-05-17 06:48:33.703836] I [MSGID: 106479] >>>>>>> [glusterd.c:1432:init] 0-management: Using /var/lib/glusterd as working >>>>>>> directory >>>>>>> [2017-05-17 06:48:33.708866] W [MSGID: 103071] >>>>>>> [rdma.c:4594:__gf_rdma_ctx_create] 0-rpc-transport/rdma: rdma_cm >>>>>>> event channel creation failed [No such device] >>>>>>> [2017-05-17 06:48:33.709011] W [MSGID: 103055] [rdma.c:4901:init] >>>>>>> 0-rdma.management: Failed to initialize IB Device >>>>>>> [2017-05-17 06:48:33.709033] W [rpc-transport.c:359:rpc_transport_load] >>>>>>> 0-rpc-transport: 'rdma' initialization failed >>>>>>> [2017-05-17 06:48:33.709088] W [rpcsvc.c:1642:rpcsvc_create_listener] >>>>>>> 0-rpc-service: cannot create listener, initing the transport failed >>>>>>> [2017-05-17 06:48:33.709105] E [MSGID: 106243] >>>>>>> [glusterd.c:1656:init] 0-management: creation of 1 listeners failed, >>>>>>> continuing with succeeded transport >>>>>>> [2017-05-17 06:48:35.480043] I [MSGID: 106513] >>>>>>> [glusterd-store.c:2068:glusterd_restore_op_version] 0-glusterd: >>>>>>> retrieved op-version: 30600 >>>>>>> [2017-05-17 06:48:35.605779] I [MSGID: 106498] >>>>>>> [glusterd-handler.c:3640:glusterd_friend_add_from_peerinfo] >>>>>>> 0-management: connect returned 0 >>>>>>> [2017-05-17 06:48:35.607059] I >>>>>>> [rpc-clnt.c:1046:rpc_clnt_connection_init] >>>>>>> 0-management: setting frame-timeout to 600 >>>>>>> [2017-05-17 06:48:35.607670] I >>>>>>> [rpc-clnt.c:1046:rpc_clnt_connection_init] >>>>>>> 0-management: setting frame-timeout to 600 >>>>>>> [2017-05-17 06:48:35.607025] I [MSGID: 106498] >>>>>>> [glusterd-handler.c:3640:glusterd_friend_add_from_peerinfo] >>>>>>> 0-management: connect returned 0 >>>>>>> [2017-05-17 06:48:35.608125] I [MSGID: 106544] >>>>>>> [glusterd.c:159:glusterd_uuid_init] 0-management: retrieved UUID: >>>>>>> 7f2a6e11-2a53-4ab4-9ceb-8be6a9f2d073 >>>>>>> >>>>>> >>>>>>> Final graph: >>>>>>> +----------------------------------------------------------- >>>>>>> -------------------+ >>>>>>> 1: volume management >>>>>>> 2: type mgmt/glusterd >>>>>>> 3: option rpc-auth.auth-glusterfs on >>>>>>> 4: option rpc-auth.auth-unix on >>>>>>> 5: option rpc-auth.auth-null on >>>>>>> 6: option rpc-auth-allow-insecure on >>>>>>> 7: option transport.socket.listen-backlog 128 >>>>>>> 8: option event-threads 1 >>>>>>> 9: option ping-timeout 0 >>>>>>> 10: option transport.socket.read-fail-log off >>>>>>> 11: option transport.socket.keepalive-interval 2 >>>>>>> 12: option transport.socket.keepalive-time 10 >>>>>>> 13: option transport-type rdma >>>>>>> 14: option working-directory /var/lib/glusterd >>>>>>> 15: end-volume >>>>>>> 16: >>>>>>> +----------------------------------------------------------- >>>>>>> -------------------+ >>>>>>> [2017-05-17 06:48:35.609868] I [MSGID: 101190] >>>>>>> [event-epoll.c:632:event_dispatch_epoll_worker] 0-epoll: Started >>>>>>> thread with index 1 >>>>>>> [2017-05-17 06:48:35.610839] W [socket.c:596:__socket_rwv] >>>>>>> 0-management: readv on 192.168.0.7:24007 failed (No data available) >>>>>>> [2017-05-17 06:48:35.611907] E [rpc-clnt.c:370:saved_frames_unwind] >>>>>>> (--> >>>>>>> /usr/lib/x86_64-linux-gnu/libglusterfs.so.0(_gf_log_callingfn+0x1a3)[0x7fd6c2d70bb3] >>>>>>> (--> >>>>>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(saved_frames_unwind+0x1cf)[0x7fd6c2b3a2df] >>>>>>> (--> >>>>>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(saved_frames_destroy+0xe)[0x7fd6c2b3a3fe] >>>>>>> (--> >>>>>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(rpc_clnt_connection_cleanup+0x89)[0x7fd6c2b3ba39] >>>>>>> (--> >>>>>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(rpc_clnt_notify+0x160)[0x7fd6c2b3c380] >>>>>>> ))))) 0-management: forced unwinding frame type(GLUSTERD-DUMP) >>>>>>> op(DUMP(1)) >>>>>>> called at 2017-05-17 06:48:35.609965 (xid=0x1) >>>>>>> [2017-05-17 06:48:35.611928] E [MSGID: 106167] >>>>>>> [glusterd-handshake.c:2091:__glusterd_peer_dump_version_cbk] >>>>>>> 0-management: Error through RPC layer, retry again later >>>>>>> [2017-05-17 06:48:35.611944] I [MSGID: 106004] >>>>>>> [glusterd-handler.c:5201:__glusterd_peer_rpc_notify] 0-management: >>>>>>> Peer <192.168.0.7> (<5ec54b4f-f60c-48c6-9e55-95f2bb58f633>), in >>>>>>> state <Peer in Cluster>, has disconnected from glusterd. >>>>>>> [2017-05-17 06:48:35.612024] W >>>>>>> [glusterd-locks.c:681:glusterd_mgmt_v3_unlock] >>>>>>> (-->/usr/lib/x86_64-linux-gnu/glusterfs/3.7.20/xlator/mgmt/g >>>>>>> lusterd.so(glusterd_big_locked_notify+0x4b) [0x7fd6bdc4912b] >>>>>>> -->/usr/lib/x86_64-linux-gnu/glusterfs/3.7.20/xlator/mgmt/gl >>>>>>> usterd.so(__glusterd_peer_rpc_notify+0x160) [0x7fd6bdc52dd0] >>>>>>> -->/usr/lib/x86_64-linux-gnu/glusterfs/3.7.20/xlator/mgmt/gl >>>>>>> usterd.so(glusterd_mgmt_v3_unlock+0x4c3) [0x7fd6bdcef1b3] ) >>>>>>> 0-management: Lock for vol shared not held >>>>>>> [2017-05-17 06:48:35.612039] W [MSGID: 106118] >>>>>>> [glusterd-handler.c:5223:__glusterd_peer_rpc_notify] 0-management: >>>>>>> Lock not released for shared >>>>>>> [2017-05-17 06:48:35.612079] W [socket.c:596:__socket_rwv] >>>>>>> 0-management: readv on 192.168.0.6:24007 failed (No data available) >>>>>>> [2017-05-17 06:48:35.612179] E [rpc-clnt.c:370:saved_frames_unwind] >>>>>>> (--> >>>>>>> /usr/lib/x86_64-linux-gnu/libglusterfs.so.0(_gf_log_callingfn+0x1a3)[0x7fd6c2d70bb3] >>>>>>> (--> >>>>>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(saved_frames_unwind+0x1cf)[0x7fd6c2b3a2df] >>>>>>> (--> >>>>>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(saved_frames_destroy+0xe)[0x7fd6c2b3a3fe] >>>>>>> (--> >>>>>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(rpc_clnt_connection_cleanup+0x89)[0x7fd6c2b3ba39] >>>>>>> (--> >>>>>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(rpc_clnt_notify+0x160)[0x7fd6c2b3c380] >>>>>>> ))))) 0-management: forced unwinding frame type(GLUSTERD-DUMP) >>>>>>> op(DUMP(1)) >>>>>>> called at 2017-05-17 06:48:35.610007 (xid=0x1) >>>>>>> [2017-05-17 06:48:35.612197] E [MSGID: 106167] >>>>>>> [glusterd-handshake.c:2091:__glusterd_peer_dump_version_cbk] >>>>>>> 0-management: Error through RPC layer, retry again later >>>>>>> [2017-05-17 06:48:35.612211] I [MSGID: 106004] >>>>>>> [glusterd-handler.c:5201:__glusterd_peer_rpc_notify] 0-management: >>>>>>> Peer <192.168.0.6> (<83e9a0b9-6bd5-483b-8516-d8928805ed95>), in >>>>>>> state <Peer in Cluster>, has disconnected from glusterd. >>>>>>> [2017-05-17 06:48:35.612292] W >>>>>>> [glusterd-locks.c:681:glusterd_mgmt_v3_unlock] >>>>>>> (-->/usr/lib/x86_64-linux-gnu/glusterfs/3.7.20/xlator/mgmt/g >>>>>>> lusterd.so(glusterd_big_locked_notify+0x4b) [0x7fd6bdc4912b] >>>>>>> -->/usr/lib/x86_64-linux-gnu/glusterfs/3.7.20/xlator/mgmt/gl >>>>>>> usterd.so(__glusterd_peer_rpc_notify+0x160) [0x7fd6bdc52dd0] >>>>>>> -->/usr/lib/x86_64-linux-gnu/glusterfs/3.7.20/xlator/mgmt/gl >>>>>>> usterd.so(glusterd_mgmt_v3_unlock+0x4c3) [0x7fd6bdcef1b3] ) >>>>>>> 0-management: Lock for vol shared not held >>>>>>> [2017-05-17 06:48:35.613432] W [MSGID: 106118] >>>>>>> [glusterd-handler.c:5223:__glusterd_peer_rpc_notify] 0-management: >>>>>>> Lock not released for shared >>>>>>> [2017-05-17 06:48:35.614317] E [MSGID: 106170] >>>>>>> [glusterd-handshake.c:1051:gd_validate_mgmt_hndsk_req] >>>>>>> 0-management: Request from peer 192.168.0.6:991 has an entry in >>>>>>> peerinfo, but uuid does not match >>>>>>> >>>>>> >>>>>> Apologies for delay. My initial suspect was correct. You have an >>>>>> incorrect UUID in the peer file which is causing this. Can you please >>>>>> provide me the >>>>>> >>>>> >>>>> Clicked the send button accidentally! >>>>> >>>>> Can you please send me the content of /var/lib/glusterd & glusterd log >>>>> from all the nodes? >>>>> >>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> On Mon, May 15, 2017 at 10:31 PM, Atin Mukherjee < >>>>>>> amukh...@redhat.com> wrote: >>>>>>> >>>>>>>> >>>>>>>> On Mon, 15 May 2017 at 11:58, Pawan Alwandi <pa...@platform.sh> >>>>>>>> wrote: >>>>>>>> >>>>>>>>> Hi Atin, >>>>>>>>> >>>>>>>>> I see below error. Do I require gluster to be upgraded on all 3 >>>>>>>>> hosts for this to work? Right now I have host 1 running 3.10.1 and >>>>>>>>> host 2 >>>>>>>>> & 3 running 3.6.2 >>>>>>>>> >>>>>>>>> # gluster v set all cluster.op-version 31001 >>>>>>>>> volume set: failed: Required op_version (31001) is not supported >>>>>>>>> >>>>>>>> >>>>>>>> Yes you should given 3.6 version is EOLed. >>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> On Mon, May 15, 2017 at 3:32 AM, Atin Mukherjee < >>>>>>>>> amukh...@redhat.com> wrote: >>>>>>>>> >>>>>>>>>> On Sun, 14 May 2017 at 21:43, Atin Mukherjee <amukh...@redhat.com> >>>>>>>>>> wrote: >>>>>>>>>> >>>>>>>>>>> Allright, I see that you haven't bumped up the op-version. Can >>>>>>>>>>> you please execute: >>>>>>>>>>> >>>>>>>>>>> gluster v set all cluster.op-version 30101 and then restart >>>>>>>>>>> glusterd on all the nodes and check the brick status? >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> s/30101/31001 >>>>>>>>>> >>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> On Sun, May 14, 2017 at 8:55 PM, Pawan Alwandi < >>>>>>>>>>> pa...@platform.sh> wrote: >>>>>>>>>>> >>>>>>>>>>>> Hello Atin, >>>>>>>>>>>> >>>>>>>>>>>> Thanks for looking at this. Below is the output you requested >>>>>>>>>>>> for. >>>>>>>>>>>> >>>>>>>>>>>> Again, I'm seeing those errors after upgrading gluster on host >>>>>>>>>>>> 1. >>>>>>>>>>>> >>>>>>>>>>>> Host 1 >>>>>>>>>>>> >>>>>>>>>>>> # cat /var/lib/glusterd/glusterd.info >>>>>>>>>>>> UUID=7f2a6e11-2a53-4ab4-9ceb-8be6a9f2d073 >>>>>>>>>>>> operating-version=30600 >>>>>>>>>>>> >>>>>>>>>>>> # cat /var/lib/glusterd/peers/* >>>>>>>>>>>> uuid=5ec54b4f-f60c-48c6-9e55-95f2bb58f633 >>>>>>>>>>>> state=3 >>>>>>>>>>>> hostname1=192.168.0.7 >>>>>>>>>>>> uuid=83e9a0b9-6bd5-483b-8516-d8928805ed95 >>>>>>>>>>>> state=3 >>>>>>>>>>>> hostname1=192.168.0.6 >>>>>>>>>>>> >>>>>>>>>>>> # gluster --version >>>>>>>>>>>> glusterfs 3.10.1 >>>>>>>>>>>> >>>>>>>>>>>> Host 2 >>>>>>>>>>>> >>>>>>>>>>>> # cat /var/lib/glusterd/glusterd.info >>>>>>>>>>>> UUID=83e9a0b9-6bd5-483b-8516-d8928805ed95 >>>>>>>>>>>> operating-version=30600 >>>>>>>>>>>> >>>>>>>>>>>> # cat /var/lib/glusterd/peers/* >>>>>>>>>>>> uuid=5ec54b4f-f60c-48c6-9e55-95f2bb58f633 >>>>>>>>>>>> state=3 >>>>>>>>>>>> hostname1=192.168.0.7 >>>>>>>>>>>> uuid=7f2a6e11-2a53-4ab4-9ceb-8be6a9f2d073 >>>>>>>>>>>> state=3 >>>>>>>>>>>> hostname1=192.168.0.5 >>>>>>>>>>>> >>>>>>>>>>>> # gluster --version >>>>>>>>>>>> glusterfs 3.6.2 built on Jan 21 2015 14:23:44 >>>>>>>>>>>> >>>>>>>>>>>> Host 3 >>>>>>>>>>>> >>>>>>>>>>>> # cat /var/lib/glusterd/glusterd.info >>>>>>>>>>>> UUID=5ec54b4f-f60c-48c6-9e55-95f2bb58f633 >>>>>>>>>>>> operating-version=30600 >>>>>>>>>>>> >>>>>>>>>>>> # cat /var/lib/glusterd/peers/* >>>>>>>>>>>> uuid=7f2a6e11-2a53-4ab4-9ceb-8be6a9f2d073 >>>>>>>>>>>> state=3 >>>>>>>>>>>> hostname1=192.168.0.5 >>>>>>>>>>>> uuid=83e9a0b9-6bd5-483b-8516-d8928805ed95 >>>>>>>>>>>> state=3 >>>>>>>>>>>> hostname1=192.168.0.6 >>>>>>>>>>>> >>>>>>>>>>>> # gluster --version >>>>>>>>>>>> glusterfs 3.6.2 built on Jan 21 2015 14:23:44 >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> On Sat, May 13, 2017 at 6:28 PM, Atin Mukherjee < >>>>>>>>>>>> amukh...@redhat.com> wrote: >>>>>>>>>>>> >>>>>>>>>>>>> I have already asked for the following earlier: >>>>>>>>>>>>> >>>>>>>>>>>>> Can you please provide output of following from all the nodes: >>>>>>>>>>>>> >>>>>>>>>>>>> cat /var/lib/glusterd/glusterd.info >>>>>>>>>>>>> cat /var/lib/glusterd/peers/* >>>>>>>>>>>>> >>>>>>>>>>>>> On Sat, 13 May 2017 at 12:22, Pawan Alwandi <pa...@platform.sh> >>>>>>>>>>>>> wrote: >>>>>>>>>>>>> >>>>>>>>>>>>>> Hello folks, >>>>>>>>>>>>>> >>>>>>>>>>>>>> Does anyone have any idea whats going on here? >>>>>>>>>>>>>> >>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>> Pawan >>>>>>>>>>>>>> >>>>>>>>>>>>>> On Wed, May 10, 2017 at 5:02 PM, Pawan Alwandi < >>>>>>>>>>>>>> pa...@platform.sh> wrote: >>>>>>>>>>>>>> >>>>>>>>>>>>>>> Hello, >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> I'm trying to upgrade gluster from 3.6.2 to 3.10.1 but don't >>>>>>>>>>>>>>> see the glusterfsd and glusterfs processes coming up. >>>>>>>>>>>>>>> http://gluster.readthedocs.io/en/latest/Upgrade-Guide/upgrad >>>>>>>>>>>>>>> e_to_3.10/ is the process that I'm trying to follow. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> This is a 3 node server setup with a replicated volume >>>>>>>>>>>>>>> having replica count of 3. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Logs below: >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> [2017-05-10 09:07:03.507959] I [MSGID: 100030] >>>>>>>>>>>>>>> [glusterfsd.c:2460:main] 0-/usr/sbin/glusterd: Started running >>>>>>>>>>>>>>> /usr/sbin/glusterd version 3.10.1 (args: /usr/sbin/glusterd -p >>>>>>>>>>>>>>> /var/run/glusterd.pid) >>>>>>>>>>>>>>> [2017-05-10 09:07:03.512827] I [MSGID: 106478] >>>>>>>>>>>>>>> [glusterd.c:1449:init] 0-management: Maximum allowed open file >>>>>>>>>>>>>>> descriptors >>>>>>>>>>>>>>> set to 65536 >>>>>>>>>>>>>>> [2017-05-10 09:07:03.512855] I [MSGID: 106479] >>>>>>>>>>>>>>> [glusterd.c:1496:init] 0-management: Using /var/lib/glusterd as >>>>>>>>>>>>>>> working >>>>>>>>>>>>>>> directory >>>>>>>>>>>>>>> [2017-05-10 09:07:03.520426] W [MSGID: 103071] >>>>>>>>>>>>>>> [rdma.c:4590:__gf_rdma_ctx_create] 0-rpc-transport/rdma: >>>>>>>>>>>>>>> rdma_cm event channel creation failed [No such device] >>>>>>>>>>>>>>> [2017-05-10 09:07:03.520452] W [MSGID: 103055] >>>>>>>>>>>>>>> [rdma.c:4897:init] 0-rdma.management: Failed to initialize IB >>>>>>>>>>>>>>> Device >>>>>>>>>>>>>>> [2017-05-10 09:07:03.520465] W >>>>>>>>>>>>>>> [rpc-transport.c:350:rpc_transport_load] >>>>>>>>>>>>>>> 0-rpc-transport: 'rdma' initialization failed >>>>>>>>>>>>>>> [2017-05-10 09:07:03.520518] W >>>>>>>>>>>>>>> [rpcsvc.c:1661:rpcsvc_create_listener] >>>>>>>>>>>>>>> 0-rpc-service: cannot create listener, initing the transport >>>>>>>>>>>>>>> failed >>>>>>>>>>>>>>> [2017-05-10 09:07:03.520534] E [MSGID: 106243] >>>>>>>>>>>>>>> [glusterd.c:1720:init] 0-management: creation of 1 listeners >>>>>>>>>>>>>>> failed, >>>>>>>>>>>>>>> continuing with succeeded transport >>>>>>>>>>>>>>> [2017-05-10 09:07:04.931764] I [MSGID: 106513] >>>>>>>>>>>>>>> [glusterd-store.c:2197:glusterd_restore_op_version] >>>>>>>>>>>>>>> 0-glusterd: retrieved op-version: 30600 >>>>>>>>>>>>>>> [2017-05-10 09:07:04.964354] I [MSGID: 106544] >>>>>>>>>>>>>>> [glusterd.c:158:glusterd_uuid_init] 0-management: retrieved >>>>>>>>>>>>>>> UUID: 7f2a6e11-2a53-4ab4-9ceb-8be6a9f2d073 >>>>>>>>>>>>>>> [2017-05-10 09:07:04.993944] I [MSGID: 106498] >>>>>>>>>>>>>>> [glusterd-handler.c:3669:glusterd_friend_add_from_peerinfo] >>>>>>>>>>>>>>> 0-management: connect returned 0 >>>>>>>>>>>>>>> [2017-05-10 09:07:04.995864] I [MSGID: 106498] >>>>>>>>>>>>>>> [glusterd-handler.c:3669:glusterd_friend_add_from_peerinfo] >>>>>>>>>>>>>>> 0-management: connect returned 0 >>>>>>>>>>>>>>> [2017-05-10 09:07:04.995879] W [MSGID: 106062] >>>>>>>>>>>>>>> [glusterd-handler.c:3466:glusterd_transport_inet_options_build] >>>>>>>>>>>>>>> 0-glusterd: Failed to get tcp-user-timeout >>>>>>>>>>>>>>> [2017-05-10 09:07:04.995903] I >>>>>>>>>>>>>>> [rpc-clnt.c:1059:rpc_clnt_connection_init] >>>>>>>>>>>>>>> 0-management: setting frame-timeout to 600 >>>>>>>>>>>>>>> [2017-05-10 09:07:04.996325] I >>>>>>>>>>>>>>> [rpc-clnt.c:1059:rpc_clnt_connection_init] >>>>>>>>>>>>>>> 0-management: setting frame-timeout to 600 >>>>>>>>>>>>>>> Final graph: >>>>>>>>>>>>>>> +----------------------------------------------------------- >>>>>>>>>>>>>>> -------------------+ >>>>>>>>>>>>>>> 1: volume management >>>>>>>>>>>>>>> 2: type mgmt/glusterd >>>>>>>>>>>>>>> 3: option rpc-auth.auth-glusterfs on >>>>>>>>>>>>>>> 4: option rpc-auth.auth-unix on >>>>>>>>>>>>>>> 5: option rpc-auth.auth-null on >>>>>>>>>>>>>>> 6: option rpc-auth-allow-insecure on >>>>>>>>>>>>>>> 7: option transport.socket.listen-backlog 128 >>>>>>>>>>>>>>> 8: option event-threads 1 >>>>>>>>>>>>>>> 9: option ping-timeout 0 >>>>>>>>>>>>>>> 10: option transport.socket.read-fail-log off >>>>>>>>>>>>>>> 11: option transport.socket.keepalive-interval 2 >>>>>>>>>>>>>>> 12: option transport.socket.keepalive-time 10 >>>>>>>>>>>>>>> 13: option transport-type rdma >>>>>>>>>>>>>>> 14: option working-directory /var/lib/glusterd >>>>>>>>>>>>>>> 15: end-volume >>>>>>>>>>>>>>> 16: >>>>>>>>>>>>>>> +----------------------------------------------------------- >>>>>>>>>>>>>>> -------------------+ >>>>>>>>>>>>>>> [2017-05-10 09:07:04.996310] W [MSGID: 106062] >>>>>>>>>>>>>>> [glusterd-handler.c:3466:glusterd_transport_inet_options_build] >>>>>>>>>>>>>>> 0-glusterd: Failed to get tcp-user-timeout >>>>>>>>>>>>>>> [2017-05-10 09:07:05.000461] I [MSGID: 101190] >>>>>>>>>>>>>>> [event-epoll.c:629:event_dispatch_epoll_worker] 0-epoll: >>>>>>>>>>>>>>> Started thread with index 1 >>>>>>>>>>>>>>> [2017-05-10 09:07:05.001493] W [socket.c:593:__socket_rwv] >>>>>>>>>>>>>>> 0-management: readv on 192.168.0.7:24007 failed (No data >>>>>>>>>>>>>>> available) >>>>>>>>>>>>>>> [2017-05-10 09:07:05.001513] I [MSGID: 106004] >>>>>>>>>>>>>>> [glusterd-handler.c:5882:__glusterd_peer_rpc_notify] >>>>>>>>>>>>>>> 0-management: Peer <192.168.0.7> >>>>>>>>>>>>>>> (<5ec54b4f-f60c-48c6-9e55-95f2bb58f633>), >>>>>>>>>>>>>>> in state <Peer in Cluster>, h >>>>>>>>>>>>>>> as disconnected from glusterd. >>>>>>>>>>>>>>> [2017-05-10 09:07:05.001677] W >>>>>>>>>>>>>>> [glusterd-locks.c:675:glusterd_mgmt_v3_unlock] >>>>>>>>>>>>>>> (-->/usr/lib/x86_64-linux-gnu/glusterfs/3.10.1/xlator/mgmt/glusterd.so(+0x20559) >>>>>>>>>>>>>>> [0x7f0bf9d74559] -->/usr/lib/x86_64-linux-gnu >>>>>>>>>>>>>>> /glusterfs/3.10.1/xlator/mgmt/glusterd.so(+0x29cf0) >>>>>>>>>>>>>>> [0x7f0bf9d7dcf0] -->/usr/lib/x86_64-linux-gnu/g >>>>>>>>>>>>>>> lusterfs/3.10.1/xlator/mgmt/glusterd.so(+0xd5ba3) >>>>>>>>>>>>>>> [0x7f0bf9e29ba3] ) 0-management: Lock for vol shared no >>>>>>>>>>>>>>> t held >>>>>>>>>>>>>>> [2017-05-10 09:07:05.001696] W [MSGID: 106118] >>>>>>>>>>>>>>> [glusterd-handler.c:5907:__glusterd_peer_rpc_notify] >>>>>>>>>>>>>>> 0-management: Lock not released for shared >>>>>>>>>>>>>>> [2017-05-10 09:07:05.003099] E >>>>>>>>>>>>>>> [rpc-clnt.c:365:saved_frames_unwind] >>>>>>>>>>>>>>> (--> /usr/lib/x86_64-linux-gnu/libg >>>>>>>>>>>>>>> lusterfs.so.0(_gf_log_callingfn+0x13c)[0x7f0bfeeca73c] (--> >>>>>>>>>>>>>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(s >>>>>>>>>>>>>>> aved_frames_unwind+0x1cf)[0x7f0bfec904bf] (--> >>>>>>>>>>>>>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(saved_frames_destroy+0xe)[0x7f0bfec905de] >>>>>>>>>>>>>>> (--> /usr/lib/x86_64-linux-gnu/libg >>>>>>>>>>>>>>> frpc.so.0(rpc_clnt_connection_cleanup+0x >>>>>>>>>>>>>>> 91)[0x7f0bfec91c21] (--> /usr/lib/x86_64-linux-gnu/libg >>>>>>>>>>>>>>> frpc.so.0(rpc_clnt_notify+0x290)[0x7f0bfec92710] ))))) >>>>>>>>>>>>>>> 0-management: forced unwinding frame type(GLUSTERD-DUMP) >>>>>>>>>>>>>>> op(DUMP(1)) called >>>>>>>>>>>>>>> at 2017-05-10 09:0 >>>>>>>>>>>>>>> 7:05.000627 (xid=0x1) >>>>>>>>>>>>>>> [2017-05-10 09:07:05.003129] E [MSGID: 106167] >>>>>>>>>>>>>>> [glusterd-handshake.c:2181:__glusterd_peer_dump_version_cbk] >>>>>>>>>>>>>>> 0-management: Error through RPC layer, retry again later >>>>>>>>>>>>>>> [2017-05-10 09:07:05.003251] W [socket.c:593:__socket_rwv] >>>>>>>>>>>>>>> 0-management: readv on 192.168.0.6:24007 failed (No data >>>>>>>>>>>>>>> available) >>>>>>>>>>>>>>> [2017-05-10 09:07:05.003267] I [MSGID: 106004] >>>>>>>>>>>>>>> [glusterd-handler.c:5882:__glusterd_peer_rpc_notify] >>>>>>>>>>>>>>> 0-management: Peer <192.168.0.6> >>>>>>>>>>>>>>> (<83e9a0b9-6bd5-483b-8516-d8928805ed95>), >>>>>>>>>>>>>>> in state <Peer in Cluster>, h >>>>>>>>>>>>>>> as disconnected from glusterd. >>>>>>>>>>>>>>> [2017-05-10 09:07:05.003318] W >>>>>>>>>>>>>>> [glusterd-locks.c:675:glusterd_mgmt_v3_unlock] >>>>>>>>>>>>>>> (-->/usr/lib/x86_64-linux-gnu/glusterfs/3.10.1/xlator/mgmt/glusterd.so(+0x20559) >>>>>>>>>>>>>>> [0x7f0bf9d74559] -->/usr/lib/x86_64-linux-gnu >>>>>>>>>>>>>>> /glusterfs/3.10.1/xlator/mgmt/glusterd.so(+0x29cf0) >>>>>>>>>>>>>>> [0x7f0bf9d7dcf0] -->/usr/lib/x86_64-linux-gnu/g >>>>>>>>>>>>>>> lusterfs/3.10.1/xlator/mgmt/glusterd.so(+0xd5ba3) >>>>>>>>>>>>>>> [0x7f0bf9e29ba3] ) 0-management: Lock for vol shared no >>>>>>>>>>>>>>> t held >>>>>>>>>>>>>>> [2017-05-10 09:07:05.003329] W [MSGID: 106118] >>>>>>>>>>>>>>> [glusterd-handler.c:5907:__glusterd_peer_rpc_notify] >>>>>>>>>>>>>>> 0-management: Lock not released for shared >>>>>>>>>>>>>>> [2017-05-10 09:07:05.003457] E >>>>>>>>>>>>>>> [rpc-clnt.c:365:saved_frames_unwind] >>>>>>>>>>>>>>> (--> /usr/lib/x86_64-linux-gnu/libg >>>>>>>>>>>>>>> lusterfs.so.0(_gf_log_callingfn+0x13c)[0x7f0bfeeca73c] (--> >>>>>>>>>>>>>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(s >>>>>>>>>>>>>>> aved_frames_unwind+0x1cf)[0x7f0bfec904bf] (--> >>>>>>>>>>>>>>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(saved_frames_destroy+0xe)[0x7f0bfec905de] >>>>>>>>>>>>>>> (--> /usr/lib/x86_64-linux-gnu/libg >>>>>>>>>>>>>>> frpc.so.0(rpc_clnt_connection_cleanup+0x >>>>>>>>>>>>>>> 91)[0x7f0bfec91c21] (--> /usr/lib/x86_64-linux-gnu/libg >>>>>>>>>>>>>>> frpc.so.0(rpc_clnt_notify+0x290)[0x7f0bfec92710] ))))) >>>>>>>>>>>>>>> 0-management: forced unwinding frame type(GLUSTERD-DUMP) >>>>>>>>>>>>>>> op(DUMP(1)) called >>>>>>>>>>>>>>> at 2017-05-10 09:0 >>>>>>>>>>>>>>> 7:05.001407 (xid=0x1) >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> There are a bunch of errors reported but I'm not sure which >>>>>>>>>>>>>>> is signal and which ones are noise. Does anyone have any idea >>>>>>>>>>>>>>> whats going >>>>>>>>>>>>>>> on here? >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>>> Pawan >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>> Gluster-users mailing list >>>>>>>>>>>>>> Gluster-users@gluster.org >>>>>>>>>>>>>> http://lists.gluster.org/mailman/listinfo/gluster-users >>>>>>>>>>>>> >>>>>>>>>>>>> -- >>>>>>>>>>>>> - Atin (atinm) >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> -- >>>>>>>>>> - Atin (atinm) >>>>>>>>>> >>>>>>>>> >>>>>>>>> -- >>>>>>>> - Atin (atinm) >>>>>>>> >>>>>>> >>>>>>> -- >>>>>> - Atin (atinm) >>>>>> >>>>> -- >>>>> - Atin (atinm) >>>>> >>>> >>>> >>> >> >
_______________________________________________ Gluster-users mailing list Gluster-users@gluster.org http://lists.gluster.org/mailman/listinfo/gluster-users