[+users] Can you provide the engine.log to see why the monitoring is not working here. thanks!
On Wed, May 16, 2018 at 2:08 PM, Demeter Tibor <tdeme...@itsmart.hu> wrote: > Hi, > > Meanwhile, I did the upgrade engine, but the gluster state is same on my > first node. > I've attached some screenshot of my problem. > > Thanks > > Tibor > > > > ----- 2018. máj.. 16., 10:16, Demeter Tibor <tdeme...@itsmart.hu> írtaHi, > > > If 4.3.4 will release, i just have to remove the nightly repo and update > to stable? > > I'm sorry for my terrible English, I try to explain what was my problem > with update. > I'm upgraded from 4.1.8. > > I followed up the official hosted-engine update documentation, that was > not clear me, because it has referenced to a lot of old thing (i think). > https://www.ovirt.org/documentation/upgrade-guide/upgrade-guide/ > https://www.ovirt.org/documentation/how-to/hosted- > engine/#upgrade-hosted-engine > > Maybe it need to update, because I had a lot of question under upgrade and > I was not sure in all of necessary steps. For example, If I need to > installing the new, 4.2 repo on the hosts, then need to remove the old repo > from that? > Why I need to do a" yum update -y" on hosts, meanwhile there is an > "Updatehost" menu in the GUI? So, maybe it outdated. > Since upgrade hosted engine, and the first node, I have problems with > gluster. It seems to working fine if you check it from console "gluster > volume status, etc" but not on the Gui, because now it yellow, and the > brick reds in the first node. > > Previously I did a mistake with glusterfs, my gluster config was wrong. I > have corrected them, but it did not helped to me,gluster bricks are reds on > my first node yet.... > > > Now I try to upgrade to nightly, but I'm affraid, because it a living, > productive system, and I don't have downtime. I hope it will help me. > > Thanks for all, > > Regards, > Tibor Demeter > > > > ----- 2018. máj.. 16., 9:58, Sahina Bose <sab...@redhat.com> írta: > > > > On Wed, May 16, 2018 at 1:19 PM, Demeter Tibor <tdeme...@itsmart.hu> > wrote: > >> Hi, >> >> is it a different, unstable repo? I have a productive cluster, how is >> safe that? >> I don't have any experience with nightly build. How can I use this? It >> have to install to the engine VM or all of my hosts? >> Thanks in advance for help me.. >> > > Only on the engine VM. > > Regarding stability - it passes CI so relatively stable, beyond that there > are no guarantees. > > What's the specific problem you're facing with update? Can you elaborate? > > >> Regards, >> >> Tibor >> >> ----- 2018. máj.. 15., 9:58, Demeter Tibor <tdeme...@itsmart.hu> írta: >> >> Hi, >> >> Could you explain how can I use this patch? >> >> R, >> Tibor >> >> >> ----- 2018. máj.. 14., 11:18, Demeter Tibor <tdeme...@itsmart.hu> írta: >> >> Hi, >> >> Sorry for my question, but can you tell me please how can I use this >> patch? >> >> Thanks, >> Regards, >> Tibor >> ----- 2018. máj.. 14., 10:47, Sahina Bose <sab...@redhat.com> írta: >> >> >> >> On Sat, May 12, 2018 at 1:14 PM, Demeter Tibor <tdeme...@itsmart.hu> >> wrote: >> >>> Hi, >>> >>> Could someone help me please ? I can't finish my upgrade process. >>> >> >> https://gerrit.ovirt.org/91164 should fix the error you're facing. >> >> Can you elaborate why this is affecting the upgrade process? >> >> >>> Thanks >>> R >>> Tibor >>> >>> >>> >>> ----- 2018. máj.. 10., 12:51, Demeter Tibor <tdeme...@itsmart.hu> írta: >>> >>> Hi, >>> >>> I've attached the vdsm and supervdsm logs. But I don't have engine.log >>> here, because that is on hosted engine vm. Should I send that ? >>> >>> Thank you >>> >>> Regards, >>> >>> Tibor >>> ----- 2018. máj.. 10., 12:30, Sahina Bose <sab...@redhat.com> írta: >>> >>> There's a bug here. Can you log one attaching this engine.log and also >>> vdsm.log & supervdsm.log from n3.itsmart.cloud >>> >>> On Thu, May 10, 2018 at 3:35 PM, Demeter Tibor <tdeme...@itsmart.hu> >>> wrote: >>> >>>> Hi, >>>> >>>> I found this: >>>> >>>> >>>> 2018-05-10 03:24:19,096+02 INFO [org.ovirt.engine.core. >>>> vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] >>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, >>>> GetGlusterVolumeAdvancedDetailsVDSCommand, >>>> return: org.ovirt.engine.core.common.businessentities.gluster. >>>> GlusterVolumeAdvancedDetails@ca97448e, log id: 347435ae >>>> 2018-05-10 03:24:19,097+02 ERROR >>>> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob] >>>> (DefaultQuartzScheduler7) [43f4eaec] Error while refreshing brick statuses >>>> for volume 'volume2' of cluster 'C6220': null >>>> 2018-05-10 03:24:19,097+02 INFO >>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>> (DefaultQuartzScheduler8) [7715ceda] Failed to acquire lock and wait lock >>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>> sharedLocks=''}' >>>> 2018-05-10 03:24:19,104+02 INFO [org.ovirt.engine.core. >>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>> (DefaultQuartzScheduler7) [43f4eaec] START, >>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName >>>> = n4.itsmart.cloud, VdsIdVDSCommandParametersBase: >>>> {hostId='3ddef95f-158d-407c-a7d8-49641e012755'}), log id: 6908121d >>>> 2018-05-10 03:24:19,106+02 ERROR [org.ovirt.engine.core. >>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>> (DefaultQuartzScheduler7) [43f4eaec] Command ' >>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>> n4.itsmart.cloud, VdsIdVDSCommandParametersBase: >>>> {hostId='3ddef95f-158d-407c-a7d8-49641e012755'})' execution failed: >>>> null >>>> 2018-05-10 03:24:19,106+02 INFO [org.ovirt.engine.core. >>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, >>>> GetGlusterLocalLogicalVolumeListVDSCommand, >>>> log id: 6908121d >>>> 2018-05-10 03:24:19,107+02 INFO [org.ovirt.engine.core. >>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>> (DefaultQuartzScheduler7) [43f4eaec] START, >>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName >>>> = n1.itsmart.cloud, VdsIdVDSCommandParametersBase: >>>> {hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'}), log id: 735c6a5f >>>> 2018-05-10 03:24:19,109+02 ERROR [org.ovirt.engine.core. >>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>> (DefaultQuartzScheduler7) [43f4eaec] Command ' >>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>> n1.itsmart.cloud, VdsIdVDSCommandParametersBase: >>>> {hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'})' execution failed: >>>> null >>>> 2018-05-10 03:24:19,109+02 INFO [org.ovirt.engine.core. >>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, >>>> GetGlusterLocalLogicalVolumeListVDSCommand, >>>> log id: 735c6a5f >>>> 2018-05-10 03:24:19,110+02 INFO [org.ovirt.engine.core. >>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>> (DefaultQuartzScheduler7) [43f4eaec] START, >>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName >>>> = n2.itsmart.cloud, VdsIdVDSCommandParametersBase: >>>> {hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 6f9e9f58 >>>> 2018-05-10 03:24:19,112+02 ERROR [org.ovirt.engine.core. >>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>> (DefaultQuartzScheduler7) [43f4eaec] Command ' >>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>> n2.itsmart.cloud, VdsIdVDSCommandParametersBase: >>>> {hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'})' execution failed: >>>> null >>>> 2018-05-10 03:24:19,112+02 INFO [org.ovirt.engine.core. >>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, >>>> GetGlusterLocalLogicalVolumeListVDSCommand, >>>> log id: 6f9e9f58 >>>> 2018-05-10 03:24:19,113+02 INFO [org.ovirt.engine.core. >>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>> (DefaultQuartzScheduler7) [43f4eaec] START, >>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName >>>> = n3.itsmart.cloud, VdsIdVDSCommandParametersBase: >>>> {hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'}), log id: 2ee46967 >>>> 2018-05-10 03:24:19,115+02 ERROR [org.ovirt.engine.core. >>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>> (DefaultQuartzScheduler7) [43f4eaec] Command ' >>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>> n3.itsmart.cloud, VdsIdVDSCommandParametersBase: >>>> {hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'})' execution failed: >>>> null >>>> 2018-05-10 03:24:19,116+02 INFO [org.ovirt.engine.core. >>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, >>>> GetGlusterLocalLogicalVolumeListVDSCommand, >>>> log id: 2ee46967 >>>> 2018-05-10 03:24:19,117+02 INFO [org.ovirt.engine.core. >>>> vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] >>>> (DefaultQuartzScheduler7) [43f4eaec] START, >>>> GetGlusterVolumeAdvancedDetailsVDSCommand(HostName >>>> = n1.itsmart.cloud, GlusterVolumeAdvancedDetailsVD >>>> SParameters:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57', >>>> volumeName='volume1'}), log id: 7550e5c >>>> 2018-05-10 03:24:20,748+02 INFO [org.ovirt.engine.core. >>>> vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] >>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, >>>> GetGlusterVolumeAdvancedDetailsVDSCommand, >>>> return: org.ovirt.engine.core.common.businessentities.gluster. >>>> GlusterVolumeAdvancedDetails@4a46066f, log id: 7550e5c >>>> 2018-05-10 03:24:20,749+02 ERROR >>>> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob] >>>> (DefaultQuartzScheduler7) [43f4eaec] Error while refreshing brick statuses >>>> for volume 'volume1' of cluster 'C6220': null >>>> 2018-05-10 03:24:20,750+02 INFO [org.ovirt.engine.core. >>>> vdsbroker.gluster.GlusterServersListVDSCommand] >>>> (DefaultQuartzScheduler8) [7715ceda] START, >>>> GlusterServersListVDSCommand(HostName >>>> = n2.itsmart.cloud, VdsIdVDSCommandParametersBase: >>>> {hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 120cc68d >>>> 2018-05-10 03:24:20,930+02 INFO [org.ovirt.engine.core. >>>> vdsbroker.gluster.GlusterServersListVDSCommand] >>>> (DefaultQuartzScheduler8) [7715ceda] FINISH, GlusterServersListVDSCommand, >>>> return: [10.101.0.2/24:CONNECTED, n1.cloudata.local:CONNECTED, >>>> 10.104.0.3:CONNECTED, 10.104.0.4:CONNECTED], log id: 120cc68d >>>> 2018-05-10 03:24:20,949+02 INFO [org.ovirt.engine.core. >>>> vdsbroker.gluster.GlusterVolumesListVDSCommand] >>>> (DefaultQuartzScheduler8) [7715ceda] START, >>>> GlusterVolumesListVDSCommand(HostName >>>> = n2.itsmart.cloud, GlusterVolumesListVDSParameter >>>> s:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 118aa264 >>>> 2018-05-10 03:24:21,048+02 WARN [org.ovirt.engine.core. >>>> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler8) >>>> [7715ceda] Could not associate brick '10.104.0.1:/gluster/brick/brick1' >>>> of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network >>>> as no gluster network found in cluster '59c10db3-0324-0320-0120- >>>> 000000000339' >>>> 2018-05-10 03:24:21,055+02 WARN [org.ovirt.engine.core. >>>> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler8) >>>> [7715ceda] Could not associate brick '10.104.0.1:/gluster/brick/brick2' >>>> of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network >>>> as no gluster network found in cluster '59c10db3-0324-0320-0120- >>>> 000000000339' >>>> 2018-05-10 03:24:21,061+02 WARN [org.ovirt.engine.core. >>>> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler8) >>>> [7715ceda] Could not associate brick '10.104.0.1:/gluster/brick/brick3' >>>> of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network >>>> as no gluster network found in cluster '59c10db3-0324-0320-0120- >>>> 000000000339' >>>> 2018-05-10 03:24:21,067+02 WARN [org.ovirt.engine.core. >>>> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler8) >>>> [7715ceda] Could not associate brick '10.104.0.1:/gluster2/brick/brick1' >>>> of volume '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network >>>> as no gluster network found in cluster '59c10db3-0324-0320-0120- >>>> 000000000339' >>>> 2018-05-10 03:24:21,074+02 WARN [org.ovirt.engine.core. >>>> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler8) >>>> [7715ceda] Could not associate brick '10.104.0.1:/gluster2/brick/brick2' >>>> of volume '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network >>>> as no gluster network found in cluster '59c10db3-0324-0320-0120- >>>> 000000000339' >>>> 2018-05-10 03:24:21,080+02 WARN [org.ovirt.engine.core. >>>> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler8) >>>> [7715ceda] Could not associate brick '10.104.0.1:/gluster2/brick/brick3' >>>> of volume '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network >>>> as no gluster network found in cluster '59c10db3-0324-0320-0120- >>>> 000000000339' >>>> 2018-05-10 03:24:21,081+02 INFO [org.ovirt.engine.core. >>>> vdsbroker.gluster.GlusterVolumesListVDSCommand] >>>> (DefaultQuartzScheduler8) [7715ceda] FINISH, GlusterVolumesListVDSCommand, >>>> return: {68cfb061-1320-4042-abcd-9228da23c0c8=org.ovirt.engine. >>>> core.common.businessentities.gluster.GlusterVolumeEntity@7a6720d, >>>> e0f568fa-987c-4f5c-b853-01bce718ee27=org.ovirt.engine. >>>> core.common.businessentities.g >>>> luster.GlusterVolumeEntity@f88c521b}, log id: 118aa264 >>>> >>>> >>>> >>>> 2018-05-10 11:59:26,047+02 ERROR [org.ovirt.engine.core. >>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>> (DefaultQuartzScheduler4) [400fa486] Command ' >>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>> n1.itsmart.cloud, VdsIdVDSCommandParametersBase: >>>> {hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'})' execution failed: >>>> null >>>> 2018-05-10 11:59:26,047+02 INFO [org.ovirt.engine.core. >>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>> (DefaultQuartzScheduler4) [400fa486] FINISH, >>>> GetGlusterLocalLogicalVolumeListVDSCommand, >>>> log id: 14a71ef0 >>>> 2018-05-10 11:59:26,048+02 INFO [org.ovirt.engine.core. >>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>> (DefaultQuartzScheduler4) [400fa486] START, >>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName >>>> = n4.itsmart.cloud, VdsIdVDSCommandParametersBase: >>>> {hostId='3ddef95f-158d-407c-a7d8-49641e012755'}), log id: 28d9e255 >>>> 2018-05-10 11:59:26,051+02 ERROR [org.ovirt.engine.core. >>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>> (DefaultQuartzScheduler4) [400fa486] Command ' >>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>> n4.itsmart.cloud, VdsIdVDSCommandParametersBase: >>>> {hostId='3ddef95f-158d-407c-a7d8-49641e012755'})' execution failed: >>>> null >>>> 2018-05-10 11:59:26,051+02 INFO [org.ovirt.engine.core. >>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>> (DefaultQuartzScheduler4) [400fa486] FINISH, >>>> GetGlusterLocalLogicalVolumeListVDSCommand, >>>> log id: 28d9e255 >>>> 2018-05-10 11:59:26,052+02 INFO [org.ovirt.engine.core. >>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>> (DefaultQuartzScheduler4) [400fa486] START, >>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName >>>> = n2.itsmart.cloud, VdsIdVDSCommandParametersBase: >>>> {hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 4a7b280e >>>> 2018-05-10 11:59:26,054+02 ERROR [org.ovirt.engine.core. >>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>> (DefaultQuartzScheduler4) [400fa486] Command ' >>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>> n2.itsmart.cloud, VdsIdVDSCommandParametersBase: >>>> {hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'})' execution failed: >>>> null >>>> 2018-05-10 11:59:26,054+02 INFO [org.ovirt.engine.core. >>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>> (DefaultQuartzScheduler4) [400fa486] FINISH, >>>> GetGlusterLocalLogicalVolumeListVDSCommand, >>>> log id: 4a7b280e >>>> 2018-05-10 11:59:26,055+02 INFO [org.ovirt.engine.core. >>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>> (DefaultQuartzScheduler4) [400fa486] START, >>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName >>>> = n3.itsmart.cloud, VdsIdVDSCommandParametersBase: >>>> {hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'}), log id: 18adc534 >>>> 2018-05-10 11:59:26,057+02 ERROR [org.ovirt.engine.core. >>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>> (DefaultQuartzScheduler4) [400fa486] Command ' >>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>> n3.itsmart.cloud, VdsIdVDSCommandParametersBase: >>>> {hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'})' execution failed: >>>> null >>>> 2018-05-10 11:59:26,057+02 INFO [org.ovirt.engine.core. >>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>> (DefaultQuartzScheduler4) [400fa486] FINISH, >>>> GetGlusterLocalLogicalVolumeListVDSCommand, >>>> log id: 18adc534 >>>> 2018-05-10 11:59:26,058+02 INFO [org.ovirt.engine.core. >>>> vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] >>>> (DefaultQuartzScheduler4) [400fa486] START, >>>> GetGlusterVolumeAdvancedDetailsVDSCommand(HostName >>>> = n3.itsmart.cloud, GlusterVolumeAdvancedDetailsVD >>>> SParameters:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec', >>>> volumeName='volume1'}), log id: 3451084f >>>> 2018-05-10 11:59:28,050+02 INFO >>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>> (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock >>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>> sharedLocks=''}' >>>> 2018-05-10 11:59:28,060+02 INFO >>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>> (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock >>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>> sharedLocks=''}' >>>> 2018-05-10 11:59:28,062+02 INFO >>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>> (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock >>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>> sharedLocks=''}' >>>> 2018-05-10 11:59:31,054+02 INFO >>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>> (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock >>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>> sharedLocks=''}' >>>> 2018-05-10 11:59:31,054+02 INFO >>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>> (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock >>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>> sharedLocks=''}' >>>> 2018-05-10 11:59:31,062+02 INFO >>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>> (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock >>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>> sharedLocks=''}' >>>> 2018-05-10 11:59:31,064+02 INFO >>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>> (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock >>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>> sharedLocks=''}' >>>> 2018-05-10 11:59:31,465+02 INFO [org.ovirt.engine.core. >>>> vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] >>>> (DefaultQuartzScheduler4) [400fa486] FINISH, >>>> GetGlusterVolumeAdvancedDetailsVDSCommand, >>>> return: org.ovirt.engine.core.common.businessentities.gluster. >>>> GlusterVolumeAdvancedDetails@3f1b7f43, log id: 3451084f >>>> 2018-05-10 11:59:31,466+02 ERROR >>>> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob] >>>> (DefaultQuartzScheduler4) [400fa486] Error while refreshing brick statuses >>>> for volume 'volume1' of cluster 'C6220': null >>>> >>>> >>>> R >>>> Tibor >>>> >>>> ----- 2018. máj.. 10., 11:43, Sahina Bose <sab...@redhat.com> írta: >>>> >>>> This doesn't affect the monitoring of state. >>>> Any errors in vdsm.log? >>>> Or errors in engine.log of the form "Error while refreshing brick >>>> statuses for volume" >>>> >>>> On Thu, May 10, 2018 at 2:33 PM, Demeter Tibor <tdeme...@itsmart.hu> >>>> wrote: >>>> >>>>> Hi, >>>>> >>>>> Thank you for your fast reply :) >>>>> >>>>> >>>>> 2018-05-10 11:01:51,574+02 INFO [org.ovirt.engine.core. >>>>> vdsbroker.gluster.GlusterServersListVDSCommand] >>>>> (DefaultQuartzScheduler6) [7f01fc2d] START, >>>>> GlusterServersListVDSCommand(HostName >>>>> = n2.itsmart.cloud, VdsIdVDSCommandParametersBase: >>>>> {hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 39adbbb8 >>>>> 2018-05-10 11:01:51,768+02 INFO [org.ovirt.engine.core. >>>>> vdsbroker.gluster.GlusterServersListVDSCommand] >>>>> (DefaultQuartzScheduler6) [7f01fc2d] FINISH, GlusterServersListVDSCommand, >>>>> return: [10.101.0.2/24:CONNECTED, n1.cloudata.local:CONNECTED, >>>>> 10.104.0.3:CONNECTED, 10.104.0.4:CONNECTED], log id: 39adbbb8 >>>>> 2018-05-10 11:01:51,788+02 INFO [org.ovirt.engine.core. >>>>> vdsbroker.gluster.GlusterVolumesListVDSCommand] >>>>> (DefaultQuartzScheduler6) [7f01fc2d] START, >>>>> GlusterVolumesListVDSCommand(HostName >>>>> = n2.itsmart.cloud, GlusterVolumesListVDSParameter >>>>> s:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 738a7261 >>>>> 2018-05-10 11:01:51,892+02 WARN [org.ovirt.engine.core. >>>>> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler6) >>>>> [7f01fc2d] Could not associate brick '10.104.0.1:/gluster/brick/brick1' >>>>> of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network >>>>> as no gluster network found in cluster '59c10db3-0324-0320-0120- >>>>> 000000000339' >>>>> 2018-05-10 11:01:51,898+02 WARN [org.ovirt.engine.core. >>>>> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler6) >>>>> [7f01fc2d] Could not associate brick '10.104.0.1:/gluster/brick/brick2' >>>>> of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network >>>>> as no gluster network found in cluster '59c10db3-0324-0320-0120- >>>>> 000000000339' >>>>> 2018-05-10 11:01:51,905+02 WARN [org.ovirt.engine.core. >>>>> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler6) >>>>> [7f01fc2d] Could not associate brick '10.104.0.1:/gluster/brick/brick3' >>>>> of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network >>>>> as no gluster network found in cluster '59c10db3-0324-0320-0120- >>>>> 000000000339' >>>>> 2018-05-10 11:01:51,911+02 WARN [org.ovirt.engine.core. >>>>> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler6) >>>>> [7f01fc2d] Could not associate brick '10.104.0.1:/gluster2/brick/brick1' >>>>> of volume '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network >>>>> as no gluster network found in cluster '59c10db3-0324-0320-0120- >>>>> 000000000339' >>>>> 2018-05-10 11:01:51,917+02 WARN [org.ovirt.engine.core. >>>>> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler6) >>>>> [7f01fc2d] Could not associate brick '10.104.0.1:/gluster2/brick/brick2' >>>>> of volume '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network >>>>> as no gluster network found in cluster '59c10db3-0324-0320-0120- >>>>> 000000000339' >>>>> 2018-05-10 11:01:51,924+02 WARN [org.ovirt.engine.core. >>>>> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler6) >>>>> [7f01fc2d] Could not associate brick '10.104.0.1:/gluster2/brick/brick3' >>>>> of volume '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network >>>>> as no gluster network found in cluster '59c10db3-0324-0320-0120- >>>>> 000000000339' >>>>> 2018-05-10 11:01:51,925+02 INFO [org.ovirt.engine.core. >>>>> vdsbroker.gluster.GlusterVolumesListVDSCommand] >>>>> (DefaultQuartzScheduler6) [7f01fc2d] FINISH, GlusterVolumesListVDSCommand, >>>>> return: {68cfb061-1320-4042-abcd-9228da23c0c8=org.ovirt.engine. >>>>> core.common.businessentities.gluster.GlusterVolumeEntity@7a6720d, >>>>> e0f568fa-987c-4f5c-b853-01bce718ee27=org.ovirt.engine. >>>>> core.common.businessentities.gluster.GlusterVolumeEntity@f88c521b}, >>>>> log id: 738a7261 >>>>> >>>>> >>>>> This happening continuously. >>>>> >>>>> Thanks! >>>>> Tibor >>>>> >>>>> >>>>> >>>>> ----- 2018. máj.. 10., 10:56, Sahina Bose <sab...@redhat.com> írta: >>>>> >>>>> Could you check the engine.log if there are errors related to getting >>>>> GlusterVolumeAdvancedDetails ? >>>>> >>>>> On Thu, May 10, 2018 at 2:02 PM, Demeter Tibor <tdeme...@itsmart.hu> >>>>> wrote: >>>>> >>>>>> Dear Ovirt Users, >>>>>> I've followed up the self-hosted-engine upgrade documentation, I >>>>>> upgraded my 4.1 system to 4.2.3. >>>>>> I upgaded the first node with yum upgrade, it seems working now fine. >>>>>> But since upgrade, the gluster informations seems to displayed incorrect >>>>>> on >>>>>> the admin panel. The volume yellow, and there are red bricks from that >>>>>> node. >>>>>> I've checked in console, I think my gluster is not degraded: >>>>>> >>>>>> root@n1 ~]# gluster volume list >>>>>> volume1 >>>>>> volume2 >>>>>> [root@n1 ~]# gluster volume info >>>>>> >>>>>> Volume Name: volume1 >>>>>> Type: Distributed-Replicate >>>>>> Volume ID: e0f568fa-987c-4f5c-b853-01bce718ee27 >>>>>> Status: Started >>>>>> Snapshot Count: 0 >>>>>> Number of Bricks: 3 x 3 = 9 >>>>>> Transport-type: tcp >>>>>> Bricks: >>>>>> Brick1: 10.104.0.1:/gluster/brick/brick1 >>>>>> Brick2: 10.104.0.2:/gluster/brick/brick1 >>>>>> Brick3: 10.104.0.3:/gluster/brick/brick1 >>>>>> Brick4: 10.104.0.1:/gluster/brick/brick2 >>>>>> Brick5: 10.104.0.2:/gluster/brick/brick2 >>>>>> Brick6: 10.104.0.3:/gluster/brick/brick2 >>>>>> Brick7: 10.104.0.1:/gluster/brick/brick3 >>>>>> Brick8: 10.104.0.2:/gluster/brick/brick3 >>>>>> Brick9: 10.104.0.3:/gluster/brick/brick3 >>>>>> Options Reconfigured: >>>>>> transport.address-family: inet >>>>>> performance.readdir-ahead: on >>>>>> nfs.disable: on >>>>>> storage.owner-uid: 36 >>>>>> storage.owner-gid: 36 >>>>>> performance.quick-read: off >>>>>> performance.read-ahead: off >>>>>> performance.io-cache: off >>>>>> performance.stat-prefetch: off >>>>>> performance.low-prio-threads: 32 >>>>>> network.remote-dio: enable >>>>>> cluster.eager-lock: enable >>>>>> cluster.quorum-type: auto >>>>>> cluster.server-quorum-type: server >>>>>> cluster.data-self-heal-algorithm: full >>>>>> cluster.locking-scheme: granular >>>>>> cluster.shd-max-threads: 8 >>>>>> cluster.shd-wait-qlength: 10000 >>>>>> features.shard: on >>>>>> user.cifs: off >>>>>> server.allow-insecure: on >>>>>> >>>>>> Volume Name: volume2 >>>>>> Type: Distributed-Replicate >>>>>> Volume ID: 68cfb061-1320-4042-abcd-9228da23c0c8 >>>>>> Status: Started >>>>>> Snapshot Count: 0 >>>>>> Number of Bricks: 3 x 3 = 9 >>>>>> Transport-type: tcp >>>>>> Bricks: >>>>>> Brick1: 10.104.0.1:/gluster2/brick/brick1 >>>>>> Brick2: 10.104.0.2:/gluster2/brick/brick1 >>>>>> Brick3: 10.104.0.3:/gluster2/brick/brick1 >>>>>> Brick4: 10.104.0.1:/gluster2/brick/brick2 >>>>>> Brick5: 10.104.0.2:/gluster2/brick/brick2 >>>>>> Brick6: 10.104.0.3:/gluster2/brick/brick2 >>>>>> Brick7: 10.104.0.1:/gluster2/brick/brick3 >>>>>> Brick8: 10.104.0.2:/gluster2/brick/brick3 >>>>>> Brick9: 10.104.0.3:/gluster2/brick/brick3 >>>>>> Options Reconfigured: >>>>>> nfs.disable: on >>>>>> performance.readdir-ahead: on >>>>>> transport.address-family: inet >>>>>> cluster.quorum-type: auto >>>>>> network.ping-timeout: 10 >>>>>> auth.allow: * >>>>>> performance.quick-read: off >>>>>> performance.read-ahead: off >>>>>> performance.io-cache: off >>>>>> performance.stat-prefetch: off >>>>>> performance.low-prio-threads: 32 >>>>>> network.remote-dio: enable >>>>>> cluster.eager-lock: enable >>>>>> cluster.server-quorum-type: server >>>>>> cluster.data-self-heal-algorithm: full >>>>>> cluster.locking-scheme: granular >>>>>> cluster.shd-max-threads: 8 >>>>>> cluster.shd-wait-qlength: 10000 >>>>>> features.shard: on >>>>>> user.cifs: off >>>>>> storage.owner-uid: 36 >>>>>> storage.owner-gid: 36 >>>>>> server.allow-insecure: on >>>>>> [root@n1 ~]# gluster volume status >>>>>> Status of volume: volume1 >>>>>> Gluster process TCP Port RDMA Port >>>>>> Online Pid >>>>>> ------------------------------------------------------------ >>>>>> ------------------ >>>>>> Brick 10.104.0.1:/gluster/brick/brick1 49152 0 Y >>>>>> 3464 >>>>>> Brick 10.104.0.2:/gluster/brick/brick1 49152 0 Y >>>>>> 68937 >>>>>> Brick 10.104.0.3:/gluster/brick/brick1 49161 0 Y >>>>>> 94506 >>>>>> Brick 10.104.0.1:/gluster/brick/brick2 49153 0 Y >>>>>> 3457 >>>>>> Brick 10.104.0.2:/gluster/brick/brick2 49153 0 Y >>>>>> 68943 >>>>>> Brick 10.104.0.3:/gluster/brick/brick2 49162 0 Y >>>>>> 94514 >>>>>> Brick 10.104.0.1:/gluster/brick/brick3 49154 0 Y >>>>>> 3465 >>>>>> Brick 10.104.0.2:/gluster/brick/brick3 49154 0 Y >>>>>> 68949 >>>>>> Brick 10.104.0.3:/gluster/brick/brick3 49163 0 Y >>>>>> 94520 >>>>>> Self-heal Daemon on localhost N/A N/A Y >>>>>> 54356 >>>>>> Self-heal Daemon on 10.104.0.2 N/A N/A Y >>>>>> 962 >>>>>> Self-heal Daemon on 10.104.0.3 N/A N/A Y >>>>>> 108977 >>>>>> Self-heal Daemon on 10.104.0.4 N/A N/A Y >>>>>> 61603 >>>>>> >>>>>> Task Status of Volume volume1 >>>>>> ------------------------------------------------------------ >>>>>> ------------------ >>>>>> There are no active volume tasks >>>>>> >>>>>> Status of volume: volume2 >>>>>> Gluster process TCP Port RDMA Port >>>>>> Online Pid >>>>>> ------------------------------------------------------------ >>>>>> ------------------ >>>>>> Brick 10.104.0.1:/gluster2/brick/brick1 49155 0 Y >>>>>> 3852 >>>>>> Brick 10.104.0.2:/gluster2/brick/brick1 49158 0 Y >>>>>> 68955 >>>>>> Brick 10.104.0.3:/gluster2/brick/brick1 49164 0 Y >>>>>> 94527 >>>>>> Brick 10.104.0.1:/gluster2/brick/brick2 49156 0 Y >>>>>> 3851 >>>>>> Brick 10.104.0.2:/gluster2/brick/brick2 49159 0 Y >>>>>> 68961 >>>>>> Brick 10.104.0.3:/gluster2/brick/brick2 49165 0 Y >>>>>> 94533 >>>>>> Brick 10.104.0.1:/gluster2/brick/brick3 49157 0 Y >>>>>> 3883 >>>>>> Brick 10.104.0.2:/gluster2/brick/brick3 49160 0 Y >>>>>> 68968 >>>>>> Brick 10.104.0.3:/gluster2/brick/brick3 49166 0 Y >>>>>> 94541 >>>>>> Self-heal Daemon on localhost N/A N/A Y >>>>>> 54356 >>>>>> Self-heal Daemon on 10.104.0.2 N/A N/A Y >>>>>> 962 >>>>>> Self-heal Daemon on 10.104.0.3 N/A N/A Y >>>>>> 108977 >>>>>> Self-heal Daemon on 10.104.0.4 N/A N/A Y >>>>>> 61603 >>>>>> >>>>>> Task Status of Volume volume2 >>>>>> ------------------------------------------------------------ >>>>>> ------------------ >>>>>> There are no active volume tasks >>>>>> >>>>>> I think ovirt can't read valid informations about gluster. >>>>>> I can't contiune upgrade of other hosts until this problem exist. >>>>>> >>>>>> Please help me:) >>>>>> >>>>>> >>>>>> Thanks >>>>>> >>>>>> Regards, >>>>>> >>>>>> Tibor >>>>>> >>>>>> >>>>>> _______________________________________________ >>>>>> Users mailing list -- users@ovirt.org >>>>>> To unsubscribe send an email to users-le...@ovirt.org >>>>>> >>>>>> >>>>> >>>> >>> >>> _______________________________________________ >>> Users mailing list -- users@ovirt.org >>> To unsubscribe send an email to users-le...@ovirt.org >>> >>> >> >> _______________________________________________ >> Users mailing list -- users@ovirt.org >> To unsubscribe send an email to users-le...@ovirt.org >> >> >> _______________________________________________ >> Users mailing list -- users@ovirt.org >> To unsubscribe send an email to users-le...@ovirt.org >> oVirt Code of Conduct: https://www.ovirt.org/community/about/community- >> guidelines/ >> List Archives: >> >> >
_______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org