I got that too so upgraded to gluster6-rc0 nit still, this morning one
engine brick is down :

[2019-03-04 01:33:22.492206] E [MSGID: 101191]
[event-epoll.c:765:event_dispatch_epoll_worker] 0-epoll: Failed to dispatch
handler
[2019-03-04 01:38:34.601381] I [addr.c:54:compare_addr_and_update]
0-/gluster_bricks/engine/engine: allowed = "*", received addr =
"10.199.211.5"
[2019-03-04 01:38:34.601410] I [login.c:110:gf_auth] 0-auth/login: allowed
user names: 9e360b5b-34d3-4076-bc7e-ed78e4e0dc01
[2019-03-04 01:38:34.601421] I [MSGID: 115029]
[server-handshake.c:550:server_setvolume] 0-engine-server: accepted client
from
CTX_ID:f7603ec6-9914-408b-85e6-e64e9844e326-GRAPH_ID:0-PID:300490-HOST:ps-inf-int-kvm-fr-305-210.hostics.fr-PC_NAME:engine-client-0-RECON_NO:-0
(version: 6.0rc0) with subvol /gluster_bricks/engine/engine
[2019-03-04 01:38:34.610400] I [MSGID: 115036]
[server.c:498:server_rpc_notify] 0-engine-server: disconnecting connection
from
CTX_ID:f7603ec6-9914-408b-85e6-e64e9844e326-GRAPH_ID:0-PID:300490-HOST:ps-inf-int-kvm-fr-305-210.hostics.fr-PC_NAME:engine-client-0-RECON_NO:-0
[2019-03-04 01:38:34.610531] I [MSGID: 101055]
[client_t.c:436:gf_client_unref] 0-engine-server: Shutting down connection
CTX_ID:f7603ec6-9914-408b-85e6-e64e9844e326-GRAPH_ID:0-PID:300490-HOST:ps-inf-int-kvm-fr-305-210.hostics.fr-PC_NAME:engine-client-0-RECON_NO:-0
[2019-03-04 01:38:34.610574] E [MSGID: 101191]
[event-epoll.c:765:event_dispatch_epoll_worker] 0-epoll: Failed to dispatch
handler
[2019-03-04 01:39:18.520347] I [addr.c:54:compare_addr_and_update]
0-/gluster_bricks/engine/engine: allowed = "*", received addr =
"10.199.211.5"
[2019-03-04 01:39:18.520373] I [login.c:110:gf_auth] 0-auth/login: allowed
user names: 9e360b5b-34d3-4076-bc7e-ed78e4e0dc01
[2019-03-04 01:39:18.520383] I [MSGID: 115029]
[server-handshake.c:550:server_setvolume] 0-engine-server: accepted client
from
CTX_ID:f3be82ea-6340-4bd4-afb3-aa9db432f779-GRAPH_ID:0-PID:300885-HOST:ps-inf-int-kvm-fr-305-210.hostics.fr-PC_NAME:engine-client-0-RECON_NO:-0
(version: 6.0rc0) with subvol /gluster_bricks/engine/engine
[2019-03-04 01:39:19.711947] I [MSGID: 115036]
[server.c:498:server_rpc_notify] 0-engine-server: disconnecting connection
from
CTX_ID:f3be82ea-6340-4bd4-afb3-aa9db432f779-GRAPH_ID:0-PID:300885-HOST:ps-inf-int-kvm-fr-305-210.hostics.fr-PC_NAME:engine-client-0-RECON_NO:-0
[2019-03-04 01:39:19.712431] I [MSGID: 101055]
[client_t.c:436:gf_client_unref] 0-engine-server: Shutting down connection
CTX_ID:f3be82ea-6340-4bd4-afb3-aa9db432f779-GRAPH_ID:0-PID:300885-HOST:ps-inf-int-kvm-fr-305-210.hostics.fr-PC_NAME:engine-client-0-RECON_NO:-0
[2019-03-04 01:39:19.712484] E [MSGID: 101191]
[event-epoll.c:765:event_dispatch_epoll_worker] 0-epoll: Failed to dispatch
handler
(END)


Guillaume Pavese
Ingénieur Système et Réseau
Interactiv-Group


On Mon, Mar 4, 2019 at 3:56 AM Endre Karlson <endre.karl...@gmail.com>
wrote:

> I have tried bumping to 5.4 now and still getting alot of "Failed
> Eventhandler" errors in the logs, any ideas guys?
>
> Den søn. 3. mar. 2019 kl. 09:03 skrev Guillaume Pavese <
> guillaume.pav...@interactiv-group.com>:
>
>> Gluster 5.4 is released but not yet in official repository
>> If like me you can not wait the official release of Gluster 5.4 with the
>> instability bugfixes (planned for around March 12 hopefully), you can use
>> the following repository :
>>
>> For Gluster 5.4-1 :
>>
>> #/etc/yum.repos.d/Gluster5-Testing.repo
>> [Gluster5-Testing]
>> name=Gluster5-Testing $basearch
>> baseurl=
>> https://cbs.centos.org/repos/storage7-gluster-5-testing/os/$basearch/
>> enabled=1
>> #metadata_expire=60m
>> gpgcheck=0
>>
>>
>> If adventurous ;)  Gluster 6-rc0 :
>>
>> #/etc/yum.repos.d/Gluster6-Testing.repo
>> [Gluster6-Testing]
>> name=Gluster6-Testing $basearch
>> baseurl=
>> https://cbs.centos.org/repos/storage7-gluster-6-testing/os/$basearch/
>> enabled=1
>> #metadata_expire=60m
>> gpgcheck=0
>>
>>
>> GLHF
>>
>> Guillaume Pavese
>> Ingénieur Système et Réseau
>> Interactiv-Group
>>
>>
>> On Sun, Mar 3, 2019 at 6:16 AM Endre Karlson <endre.karl...@gmail.com>
>> wrote:
>>
>>> Hi, should we downgrade / reinstall our cluster? we have a 4 node
>>> cluster that's breakin apart daily due to the issues with GlusterFS after
>>> upgrading from 4.2.8 that was rock solid. I am wondering why 4.3 was
>>> released as a stable version at all?? **FRUSTRATION**
>>>
>>> Endre
>>> _______________________________________________
>>> Users mailing list -- users@ovirt.org
>>> To unsubscribe send an email to users-le...@ovirt.org
>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>>> oVirt Code of Conduct:
>>> https://www.ovirt.org/community/about/community-guidelines/
>>> List Archives:
>>> https://lists.ovirt.org/archives/list/users@ovirt.org/message/3TJKJGGWCANXWZED2WF5ZHTSRS2DVHR2/
>>>
>>
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/53PH4H7HNDVQOTJSYYUO77KPFUH2TOPT/

Reply via email to