[+users]

Can you provide the engine.log to see why the monitoring is not working
here. thanks!

On Wed, May 16, 2018 at 2:08 PM, Demeter Tibor <tdeme...@itsmart.hu> wrote:

> Hi,
>
> Meanwhile, I did the upgrade engine, but the gluster state is same on my
> first node.
> I've attached some screenshot of my problem.
>
> Thanks
>
> Tibor
>
>
>
> ----- 2018. máj.. 16., 10:16, Demeter Tibor <tdeme...@itsmart.hu> írtaHi,
>
>
> If 4.3.4 will release, i just have to remove the nightly repo and update
> to stable?
>
> I'm sorry for my terrible English, I try to explain what was my problem
> with update.
> I'm upgraded from 4.1.8.
>
> I followed up the official hosted-engine update documentation, that was
> not clear me, because it has referenced to a lot of old thing (i think).
> https://www.ovirt.org/documentation/upgrade-guide/upgrade-guide/
> https://www.ovirt.org/documentation/how-to/hosted-
> engine/#upgrade-hosted-engine
>
> Maybe it need to update, because I had a lot of question under upgrade and
> I was not sure in all of necessary steps. For example, If I need to
> installing the new, 4.2 repo on the hosts, then need to remove the old repo
> from that?
> Why I need to do a" yum update -y" on hosts, meanwhile there is an
> "Updatehost" menu in the GUI? So, maybe it outdated.
> Since upgrade hosted engine, and the first node, I have problems with
> gluster. It seems to working fine if you check it from console "gluster
> volume status, etc" but not on the Gui, because now it yellow, and the
> brick reds in the first node.
>
> Previously I did a mistake with glusterfs, my gluster config was wrong. I
> have corrected them, but it did not helped to me,gluster bricks are reds on
> my first node yet....
>
>
> Now I try to upgrade to nightly, but I'm affraid, because it a living,
> productive system, and I don't have downtime. I hope it will help me.
>
> Thanks for all,
>
> Regards,
> Tibor Demeter
>
>
>
> ----- 2018. máj.. 16., 9:58, Sahina Bose <sab...@redhat.com> írta:
>
>
>
> On Wed, May 16, 2018 at 1:19 PM, Demeter Tibor <tdeme...@itsmart.hu>
> wrote:
>
>> Hi,
>>
>> is it a different, unstable repo? I have a productive cluster, how is
>> safe that?
>> I don't have any experience with nightly build. How can I use this? It
>> have to install to the engine VM or all of my hosts?
>> Thanks in advance for help me..
>>
>
> Only on the engine VM.
>
> Regarding stability - it passes CI so relatively stable, beyond that there
> are no guarantees.
>
> What's the specific problem you're facing with update? Can you elaborate?
>
>
>> Regards,
>>
>> Tibor
>>
>> ----- 2018. máj.. 15., 9:58, Demeter Tibor <tdeme...@itsmart.hu> írta:
>>
>> Hi,
>>
>> Could you explain how can I use this patch?
>>
>> R,
>> Tibor
>>
>>
>> ----- 2018. máj.. 14., 11:18, Demeter Tibor <tdeme...@itsmart.hu> írta:
>>
>> Hi,
>>
>> Sorry for my question, but can you tell me please how can I use this
>> patch?
>>
>> Thanks,
>> Regards,
>> Tibor
>> ----- 2018. máj.. 14., 10:47, Sahina Bose <sab...@redhat.com> írta:
>>
>>
>>
>> On Sat, May 12, 2018 at 1:14 PM, Demeter Tibor <tdeme...@itsmart.hu>
>> wrote:
>>
>>> Hi,
>>>
>>> Could someone help me please ? I can't finish my upgrade process.
>>>
>>
>> https://gerrit.ovirt.org/91164 should fix the error you're facing.
>>
>> Can you elaborate why this is affecting the upgrade process?
>>
>>
>>> Thanks
>>> R
>>> Tibor
>>>
>>>
>>>
>>> ----- 2018. máj.. 10., 12:51, Demeter Tibor <tdeme...@itsmart.hu> írta:
>>>
>>> Hi,
>>>
>>> I've attached the vdsm and supervdsm logs. But I don't have engine.log
>>> here, because that is on hosted engine vm. Should I send that ?
>>>
>>> Thank you
>>>
>>> Regards,
>>>
>>> Tibor
>>> ----- 2018. máj.. 10., 12:30, Sahina Bose <sab...@redhat.com> írta:
>>>
>>> There's a bug here. Can you log one attaching this engine.log and also
>>> vdsm.log & supervdsm.log from n3.itsmart.cloud
>>>
>>> On Thu, May 10, 2018 at 3:35 PM, Demeter Tibor <tdeme...@itsmart.hu>
>>> wrote:
>>>
>>>> Hi,
>>>>
>>>> I found this:
>>>>
>>>>
>>>> 2018-05-10 03:24:19,096+02 INFO  [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, 
>>>> GetGlusterVolumeAdvancedDetailsVDSCommand,
>>>> return: org.ovirt.engine.core.common.businessentities.gluster.
>>>> GlusterVolumeAdvancedDetails@ca97448e, log id: 347435ae
>>>> 2018-05-10 03:24:19,097+02 ERROR 
>>>> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob]
>>>> (DefaultQuartzScheduler7) [43f4eaec] Error while refreshing brick statuses
>>>> for volume 'volume2' of cluster 'C6220': null
>>>> 2018-05-10 03:24:19,097+02 INFO  
>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
>>>> (DefaultQuartzScheduler8) [7715ceda] Failed to acquire lock and wait lock
>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>> sharedLocks=''}'
>>>> 2018-05-10 03:24:19,104+02 INFO  [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler7) [43f4eaec] START, 
>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName
>>>> = n4.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>>> {hostId='3ddef95f-158d-407c-a7d8-49641e012755'}), log id: 6908121d
>>>> 2018-05-10 03:24:19,106+02 ERROR [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler7) [43f4eaec] Command '
>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName =
>>>> n4.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>>> {hostId='3ddef95f-158d-407c-a7d8-49641e012755'})' execution failed:
>>>> null
>>>> 2018-05-10 03:24:19,106+02 INFO  [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, 
>>>> GetGlusterLocalLogicalVolumeListVDSCommand,
>>>> log id: 6908121d
>>>> 2018-05-10 03:24:19,107+02 INFO  [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler7) [43f4eaec] START, 
>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName
>>>> = n1.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>>> {hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'}), log id: 735c6a5f
>>>> 2018-05-10 03:24:19,109+02 ERROR [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler7) [43f4eaec] Command '
>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName =
>>>> n1.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>>> {hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'})' execution failed:
>>>> null
>>>> 2018-05-10 03:24:19,109+02 INFO  [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, 
>>>> GetGlusterLocalLogicalVolumeListVDSCommand,
>>>> log id: 735c6a5f
>>>> 2018-05-10 03:24:19,110+02 INFO  [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler7) [43f4eaec] START, 
>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName
>>>> = n2.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>>> {hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 6f9e9f58
>>>> 2018-05-10 03:24:19,112+02 ERROR [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler7) [43f4eaec] Command '
>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName =
>>>> n2.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>>> {hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'})' execution failed:
>>>> null
>>>> 2018-05-10 03:24:19,112+02 INFO  [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, 
>>>> GetGlusterLocalLogicalVolumeListVDSCommand,
>>>> log id: 6f9e9f58
>>>> 2018-05-10 03:24:19,113+02 INFO  [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler7) [43f4eaec] START, 
>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName
>>>> = n3.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>>> {hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'}), log id: 2ee46967
>>>> 2018-05-10 03:24:19,115+02 ERROR [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler7) [43f4eaec] Command '
>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName =
>>>> n3.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>>> {hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'})' execution failed:
>>>> null
>>>> 2018-05-10 03:24:19,116+02 INFO  [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, 
>>>> GetGlusterLocalLogicalVolumeListVDSCommand,
>>>> log id: 2ee46967
>>>> 2018-05-10 03:24:19,117+02 INFO  [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
>>>> (DefaultQuartzScheduler7) [43f4eaec] START, 
>>>> GetGlusterVolumeAdvancedDetailsVDSCommand(HostName
>>>> = n1.itsmart.cloud, GlusterVolumeAdvancedDetailsVD
>>>> SParameters:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57',
>>>> volumeName='volume1'}), log id: 7550e5c
>>>> 2018-05-10 03:24:20,748+02 INFO  [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, 
>>>> GetGlusterVolumeAdvancedDetailsVDSCommand,
>>>> return: org.ovirt.engine.core.common.businessentities.gluster.
>>>> GlusterVolumeAdvancedDetails@4a46066f, log id: 7550e5c
>>>> 2018-05-10 03:24:20,749+02 ERROR 
>>>> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob]
>>>> (DefaultQuartzScheduler7) [43f4eaec] Error while refreshing brick statuses
>>>> for volume 'volume1' of cluster 'C6220': null
>>>> 2018-05-10 03:24:20,750+02 INFO  [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GlusterServersListVDSCommand]
>>>> (DefaultQuartzScheduler8) [7715ceda] START, 
>>>> GlusterServersListVDSCommand(HostName
>>>> = n2.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>>> {hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 120cc68d
>>>> 2018-05-10 03:24:20,930+02 INFO  [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GlusterServersListVDSCommand]
>>>> (DefaultQuartzScheduler8) [7715ceda] FINISH, GlusterServersListVDSCommand,
>>>> return: [10.101.0.2/24:CONNECTED, n1.cloudata.local:CONNECTED,
>>>> 10.104.0.3:CONNECTED, 10.104.0.4:CONNECTED], log id: 120cc68d
>>>> 2018-05-10 03:24:20,949+02 INFO  [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>>> (DefaultQuartzScheduler8) [7715ceda] START, 
>>>> GlusterVolumesListVDSCommand(HostName
>>>> = n2.itsmart.cloud, GlusterVolumesListVDSParameter
>>>> s:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 118aa264
>>>> 2018-05-10 03:24:21,048+02 WARN  [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler8)
>>>> [7715ceda] Could not associate brick '10.104.0.1:/gluster/brick/brick1'
>>>> of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network
>>>> as no gluster network found in cluster '59c10db3-0324-0320-0120-
>>>> 000000000339'
>>>> 2018-05-10 03:24:21,055+02 WARN  [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler8)
>>>> [7715ceda] Could not associate brick '10.104.0.1:/gluster/brick/brick2'
>>>> of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network
>>>> as no gluster network found in cluster '59c10db3-0324-0320-0120-
>>>> 000000000339'
>>>> 2018-05-10 03:24:21,061+02 WARN  [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler8)
>>>> [7715ceda] Could not associate brick '10.104.0.1:/gluster/brick/brick3'
>>>> of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network
>>>> as no gluster network found in cluster '59c10db3-0324-0320-0120-
>>>> 000000000339'
>>>> 2018-05-10 03:24:21,067+02 WARN  [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler8)
>>>> [7715ceda] Could not associate brick '10.104.0.1:/gluster2/brick/brick1'
>>>> of volume '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network
>>>> as no gluster network found in cluster '59c10db3-0324-0320-0120-
>>>> 000000000339'
>>>> 2018-05-10 03:24:21,074+02 WARN  [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler8)
>>>> [7715ceda] Could not associate brick '10.104.0.1:/gluster2/brick/brick2'
>>>> of volume '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network
>>>> as no gluster network found in cluster '59c10db3-0324-0320-0120-
>>>> 000000000339'
>>>> 2018-05-10 03:24:21,080+02 WARN  [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler8)
>>>> [7715ceda] Could not associate brick '10.104.0.1:/gluster2/brick/brick3'
>>>> of volume '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network
>>>> as no gluster network found in cluster '59c10db3-0324-0320-0120-
>>>> 000000000339'
>>>> 2018-05-10 03:24:21,081+02 INFO  [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>>> (DefaultQuartzScheduler8) [7715ceda] FINISH, GlusterVolumesListVDSCommand,
>>>> return: {68cfb061-1320-4042-abcd-9228da23c0c8=org.ovirt.engine.
>>>> core.common.businessentities.gluster.GlusterVolumeEntity@7a6720d,
>>>> e0f568fa-987c-4f5c-b853-01bce718ee27=org.ovirt.engine.
>>>> core.common.businessentities.g
>>>> luster.GlusterVolumeEntity@f88c521b}, log id: 118aa264
>>>>
>>>>
>>>>
>>>> 2018-05-10 11:59:26,047+02 ERROR [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler4) [400fa486] Command '
>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName =
>>>> n1.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>>> {hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'})' execution failed:
>>>> null
>>>> 2018-05-10 11:59:26,047+02 INFO  [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler4) [400fa486] FINISH, 
>>>> GetGlusterLocalLogicalVolumeListVDSCommand,
>>>> log id: 14a71ef0
>>>> 2018-05-10 11:59:26,048+02 INFO  [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler4) [400fa486] START, 
>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName
>>>> = n4.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>>> {hostId='3ddef95f-158d-407c-a7d8-49641e012755'}), log id: 28d9e255
>>>> 2018-05-10 11:59:26,051+02 ERROR [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler4) [400fa486] Command '
>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName =
>>>> n4.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>>> {hostId='3ddef95f-158d-407c-a7d8-49641e012755'})' execution failed:
>>>> null
>>>> 2018-05-10 11:59:26,051+02 INFO  [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler4) [400fa486] FINISH, 
>>>> GetGlusterLocalLogicalVolumeListVDSCommand,
>>>> log id: 28d9e255
>>>> 2018-05-10 11:59:26,052+02 INFO  [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler4) [400fa486] START, 
>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName
>>>> = n2.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>>> {hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 4a7b280e
>>>> 2018-05-10 11:59:26,054+02 ERROR [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler4) [400fa486] Command '
>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName =
>>>> n2.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>>> {hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'})' execution failed:
>>>> null
>>>> 2018-05-10 11:59:26,054+02 INFO  [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler4) [400fa486] FINISH, 
>>>> GetGlusterLocalLogicalVolumeListVDSCommand,
>>>> log id: 4a7b280e
>>>> 2018-05-10 11:59:26,055+02 INFO  [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler4) [400fa486] START, 
>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName
>>>> = n3.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>>> {hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'}), log id: 18adc534
>>>> 2018-05-10 11:59:26,057+02 ERROR [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler4) [400fa486] Command '
>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName =
>>>> n3.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>>> {hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'})' execution failed:
>>>> null
>>>> 2018-05-10 11:59:26,057+02 INFO  [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler4) [400fa486] FINISH, 
>>>> GetGlusterLocalLogicalVolumeListVDSCommand,
>>>> log id: 18adc534
>>>> 2018-05-10 11:59:26,058+02 INFO  [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
>>>> (DefaultQuartzScheduler4) [400fa486] START, 
>>>> GetGlusterVolumeAdvancedDetailsVDSCommand(HostName
>>>> = n3.itsmart.cloud, GlusterVolumeAdvancedDetailsVD
>>>> SParameters:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec',
>>>> volumeName='volume1'}), log id: 3451084f
>>>> 2018-05-10 11:59:28,050+02 INFO  
>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
>>>> (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock
>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>> sharedLocks=''}'
>>>> 2018-05-10 11:59:28,060+02 INFO  
>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
>>>> (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock
>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>> sharedLocks=''}'
>>>> 2018-05-10 11:59:28,062+02 INFO  
>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
>>>> (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock
>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>> sharedLocks=''}'
>>>> 2018-05-10 11:59:31,054+02 INFO  
>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
>>>> (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock
>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>> sharedLocks=''}'
>>>> 2018-05-10 11:59:31,054+02 INFO  
>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
>>>> (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock
>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>> sharedLocks=''}'
>>>> 2018-05-10 11:59:31,062+02 INFO  
>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
>>>> (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock
>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>> sharedLocks=''}'
>>>> 2018-05-10 11:59:31,064+02 INFO  
>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
>>>> (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock
>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>> sharedLocks=''}'
>>>> 2018-05-10 11:59:31,465+02 INFO  [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
>>>> (DefaultQuartzScheduler4) [400fa486] FINISH, 
>>>> GetGlusterVolumeAdvancedDetailsVDSCommand,
>>>> return: org.ovirt.engine.core.common.businessentities.gluster.
>>>> GlusterVolumeAdvancedDetails@3f1b7f43, log id: 3451084f
>>>> 2018-05-10 11:59:31,466+02 ERROR 
>>>> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob]
>>>> (DefaultQuartzScheduler4) [400fa486] Error while refreshing brick statuses
>>>> for volume 'volume1' of cluster 'C6220': null
>>>>
>>>>
>>>> R
>>>> Tibor
>>>>
>>>> ----- 2018. máj.. 10., 11:43, Sahina Bose <sab...@redhat.com> írta:
>>>>
>>>> This doesn't affect the monitoring of state.
>>>> Any errors in vdsm.log?
>>>> Or errors in engine.log of the form "Error while refreshing brick
>>>> statuses for volume"
>>>>
>>>> On Thu, May 10, 2018 at 2:33 PM, Demeter Tibor <tdeme...@itsmart.hu>
>>>> wrote:
>>>>
>>>>> Hi,
>>>>>
>>>>> Thank you for your fast reply :)
>>>>>
>>>>>
>>>>> 2018-05-10 11:01:51,574+02 INFO  [org.ovirt.engine.core.
>>>>> vdsbroker.gluster.GlusterServersListVDSCommand]
>>>>> (DefaultQuartzScheduler6) [7f01fc2d] START, 
>>>>> GlusterServersListVDSCommand(HostName
>>>>> = n2.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>>>> {hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 39adbbb8
>>>>> 2018-05-10 11:01:51,768+02 INFO  [org.ovirt.engine.core.
>>>>> vdsbroker.gluster.GlusterServersListVDSCommand]
>>>>> (DefaultQuartzScheduler6) [7f01fc2d] FINISH, GlusterServersListVDSCommand,
>>>>> return: [10.101.0.2/24:CONNECTED, n1.cloudata.local:CONNECTED,
>>>>> 10.104.0.3:CONNECTED, 10.104.0.4:CONNECTED], log id: 39adbbb8
>>>>> 2018-05-10 11:01:51,788+02 INFO  [org.ovirt.engine.core.
>>>>> vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>>>> (DefaultQuartzScheduler6) [7f01fc2d] START, 
>>>>> GlusterVolumesListVDSCommand(HostName
>>>>> = n2.itsmart.cloud, GlusterVolumesListVDSParameter
>>>>> s:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 738a7261
>>>>> 2018-05-10 11:01:51,892+02 WARN  [org.ovirt.engine.core.
>>>>> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler6)
>>>>> [7f01fc2d] Could not associate brick '10.104.0.1:/gluster/brick/brick1'
>>>>> of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network
>>>>> as no gluster network found in cluster '59c10db3-0324-0320-0120-
>>>>> 000000000339'
>>>>> 2018-05-10 11:01:51,898+02 WARN  [org.ovirt.engine.core.
>>>>> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler6)
>>>>> [7f01fc2d] Could not associate brick '10.104.0.1:/gluster/brick/brick2'
>>>>> of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network
>>>>> as no gluster network found in cluster '59c10db3-0324-0320-0120-
>>>>> 000000000339'
>>>>> 2018-05-10 11:01:51,905+02 WARN  [org.ovirt.engine.core.
>>>>> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler6)
>>>>> [7f01fc2d] Could not associate brick '10.104.0.1:/gluster/brick/brick3'
>>>>> of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network
>>>>> as no gluster network found in cluster '59c10db3-0324-0320-0120-
>>>>> 000000000339'
>>>>> 2018-05-10 11:01:51,911+02 WARN  [org.ovirt.engine.core.
>>>>> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler6)
>>>>> [7f01fc2d] Could not associate brick '10.104.0.1:/gluster2/brick/brick1'
>>>>> of volume '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network
>>>>> as no gluster network found in cluster '59c10db3-0324-0320-0120-
>>>>> 000000000339'
>>>>> 2018-05-10 11:01:51,917+02 WARN  [org.ovirt.engine.core.
>>>>> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler6)
>>>>> [7f01fc2d] Could not associate brick '10.104.0.1:/gluster2/brick/brick2'
>>>>> of volume '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network
>>>>> as no gluster network found in cluster '59c10db3-0324-0320-0120-
>>>>> 000000000339'
>>>>> 2018-05-10 11:01:51,924+02 WARN  [org.ovirt.engine.core.
>>>>> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler6)
>>>>> [7f01fc2d] Could not associate brick '10.104.0.1:/gluster2/brick/brick3'
>>>>> of volume '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network
>>>>> as no gluster network found in cluster '59c10db3-0324-0320-0120-
>>>>> 000000000339'
>>>>> 2018-05-10 11:01:51,925+02 INFO  [org.ovirt.engine.core.
>>>>> vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>>>> (DefaultQuartzScheduler6) [7f01fc2d] FINISH, GlusterVolumesListVDSCommand,
>>>>> return: {68cfb061-1320-4042-abcd-9228da23c0c8=org.ovirt.engine.
>>>>> core.common.businessentities.gluster.GlusterVolumeEntity@7a6720d,
>>>>> e0f568fa-987c-4f5c-b853-01bce718ee27=org.ovirt.engine.
>>>>> core.common.businessentities.gluster.GlusterVolumeEntity@f88c521b},
>>>>> log id: 738a7261
>>>>>
>>>>>
>>>>> This happening continuously.
>>>>>
>>>>> Thanks!
>>>>> Tibor
>>>>>
>>>>>
>>>>>
>>>>> ----- 2018. máj.. 10., 10:56, Sahina Bose <sab...@redhat.com> írta:
>>>>>
>>>>> Could you check the engine.log if there are errors related to getting
>>>>> GlusterVolumeAdvancedDetails ?
>>>>>
>>>>> On Thu, May 10, 2018 at 2:02 PM, Demeter Tibor <tdeme...@itsmart.hu>
>>>>> wrote:
>>>>>
>>>>>> Dear Ovirt Users,
>>>>>> I've followed up the self-hosted-engine upgrade documentation, I
>>>>>> upgraded my 4.1 system to 4.2.3.
>>>>>> I upgaded the first node with yum upgrade, it seems working now fine.
>>>>>> But since upgrade, the gluster informations seems to displayed incorrect 
>>>>>> on
>>>>>> the admin panel. The volume yellow, and there are red bricks from that 
>>>>>> node.
>>>>>> I've checked in console, I think my gluster is not degraded:
>>>>>>
>>>>>> root@n1 ~]# gluster volume list
>>>>>> volume1
>>>>>> volume2
>>>>>> [root@n1 ~]# gluster volume info
>>>>>>
>>>>>> Volume Name: volume1
>>>>>> Type: Distributed-Replicate
>>>>>> Volume ID: e0f568fa-987c-4f5c-b853-01bce718ee27
>>>>>> Status: Started
>>>>>> Snapshot Count: 0
>>>>>> Number of Bricks: 3 x 3 = 9
>>>>>> Transport-type: tcp
>>>>>> Bricks:
>>>>>> Brick1: 10.104.0.1:/gluster/brick/brick1
>>>>>> Brick2: 10.104.0.2:/gluster/brick/brick1
>>>>>> Brick3: 10.104.0.3:/gluster/brick/brick1
>>>>>> Brick4: 10.104.0.1:/gluster/brick/brick2
>>>>>> Brick5: 10.104.0.2:/gluster/brick/brick2
>>>>>> Brick6: 10.104.0.3:/gluster/brick/brick2
>>>>>> Brick7: 10.104.0.1:/gluster/brick/brick3
>>>>>> Brick8: 10.104.0.2:/gluster/brick/brick3
>>>>>> Brick9: 10.104.0.3:/gluster/brick/brick3
>>>>>> Options Reconfigured:
>>>>>> transport.address-family: inet
>>>>>> performance.readdir-ahead: on
>>>>>> nfs.disable: on
>>>>>> storage.owner-uid: 36
>>>>>> storage.owner-gid: 36
>>>>>> performance.quick-read: off
>>>>>> performance.read-ahead: off
>>>>>> performance.io-cache: off
>>>>>> performance.stat-prefetch: off
>>>>>> performance.low-prio-threads: 32
>>>>>> network.remote-dio: enable
>>>>>> cluster.eager-lock: enable
>>>>>> cluster.quorum-type: auto
>>>>>> cluster.server-quorum-type: server
>>>>>> cluster.data-self-heal-algorithm: full
>>>>>> cluster.locking-scheme: granular
>>>>>> cluster.shd-max-threads: 8
>>>>>> cluster.shd-wait-qlength: 10000
>>>>>> features.shard: on
>>>>>> user.cifs: off
>>>>>> server.allow-insecure: on
>>>>>>
>>>>>> Volume Name: volume2
>>>>>> Type: Distributed-Replicate
>>>>>> Volume ID: 68cfb061-1320-4042-abcd-9228da23c0c8
>>>>>> Status: Started
>>>>>> Snapshot Count: 0
>>>>>> Number of Bricks: 3 x 3 = 9
>>>>>> Transport-type: tcp
>>>>>> Bricks:
>>>>>> Brick1: 10.104.0.1:/gluster2/brick/brick1
>>>>>> Brick2: 10.104.0.2:/gluster2/brick/brick1
>>>>>> Brick3: 10.104.0.3:/gluster2/brick/brick1
>>>>>> Brick4: 10.104.0.1:/gluster2/brick/brick2
>>>>>> Brick5: 10.104.0.2:/gluster2/brick/brick2
>>>>>> Brick6: 10.104.0.3:/gluster2/brick/brick2
>>>>>> Brick7: 10.104.0.1:/gluster2/brick/brick3
>>>>>> Brick8: 10.104.0.2:/gluster2/brick/brick3
>>>>>> Brick9: 10.104.0.3:/gluster2/brick/brick3
>>>>>> Options Reconfigured:
>>>>>> nfs.disable: on
>>>>>> performance.readdir-ahead: on
>>>>>> transport.address-family: inet
>>>>>> cluster.quorum-type: auto
>>>>>> network.ping-timeout: 10
>>>>>> auth.allow: *
>>>>>> performance.quick-read: off
>>>>>> performance.read-ahead: off
>>>>>> performance.io-cache: off
>>>>>> performance.stat-prefetch: off
>>>>>> performance.low-prio-threads: 32
>>>>>> network.remote-dio: enable
>>>>>> cluster.eager-lock: enable
>>>>>> cluster.server-quorum-type: server
>>>>>> cluster.data-self-heal-algorithm: full
>>>>>> cluster.locking-scheme: granular
>>>>>> cluster.shd-max-threads: 8
>>>>>> cluster.shd-wait-qlength: 10000
>>>>>> features.shard: on
>>>>>> user.cifs: off
>>>>>> storage.owner-uid: 36
>>>>>> storage.owner-gid: 36
>>>>>> server.allow-insecure: on
>>>>>> [root@n1 ~]# gluster volume status
>>>>>> Status of volume: volume1
>>>>>> Gluster process                             TCP Port  RDMA Port
>>>>>> Online  Pid
>>>>>> ------------------------------------------------------------
>>>>>> ------------------
>>>>>> Brick 10.104.0.1:/gluster/brick/brick1      49152     0          Y
>>>>>>      3464
>>>>>> Brick 10.104.0.2:/gluster/brick/brick1      49152     0          Y
>>>>>>      68937
>>>>>> Brick 10.104.0.3:/gluster/brick/brick1      49161     0          Y
>>>>>>      94506
>>>>>> Brick 10.104.0.1:/gluster/brick/brick2      49153     0          Y
>>>>>>      3457
>>>>>> Brick 10.104.0.2:/gluster/brick/brick2      49153     0          Y
>>>>>>      68943
>>>>>> Brick 10.104.0.3:/gluster/brick/brick2      49162     0          Y
>>>>>>      94514
>>>>>> Brick 10.104.0.1:/gluster/brick/brick3      49154     0          Y
>>>>>>      3465
>>>>>> Brick 10.104.0.2:/gluster/brick/brick3      49154     0          Y
>>>>>>      68949
>>>>>> Brick 10.104.0.3:/gluster/brick/brick3      49163     0          Y
>>>>>>      94520
>>>>>> Self-heal Daemon on localhost               N/A       N/A        Y
>>>>>>    54356
>>>>>> Self-heal Daemon on 10.104.0.2              N/A       N/A        Y
>>>>>>    962
>>>>>> Self-heal Daemon on 10.104.0.3              N/A       N/A        Y
>>>>>>    108977
>>>>>> Self-heal Daemon on 10.104.0.4              N/A       N/A        Y
>>>>>>    61603
>>>>>>
>>>>>> Task Status of Volume volume1
>>>>>> ------------------------------------------------------------
>>>>>> ------------------
>>>>>> There are no active volume tasks
>>>>>>
>>>>>> Status of volume: volume2
>>>>>> Gluster process                             TCP Port  RDMA Port
>>>>>> Online  Pid
>>>>>> ------------------------------------------------------------
>>>>>> ------------------
>>>>>> Brick 10.104.0.1:/gluster2/brick/brick1     49155     0          Y
>>>>>>      3852
>>>>>> Brick 10.104.0.2:/gluster2/brick/brick1     49158     0          Y
>>>>>>      68955
>>>>>> Brick 10.104.0.3:/gluster2/brick/brick1     49164     0          Y
>>>>>>      94527
>>>>>> Brick 10.104.0.1:/gluster2/brick/brick2     49156     0          Y
>>>>>>      3851
>>>>>> Brick 10.104.0.2:/gluster2/brick/brick2     49159     0          Y
>>>>>>      68961
>>>>>> Brick 10.104.0.3:/gluster2/brick/brick2     49165     0          Y
>>>>>>      94533
>>>>>> Brick 10.104.0.1:/gluster2/brick/brick3     49157     0          Y
>>>>>>      3883
>>>>>> Brick 10.104.0.2:/gluster2/brick/brick3     49160     0          Y
>>>>>>      68968
>>>>>> Brick 10.104.0.3:/gluster2/brick/brick3     49166     0          Y
>>>>>>      94541
>>>>>> Self-heal Daemon on localhost               N/A       N/A        Y
>>>>>>    54356
>>>>>> Self-heal Daemon on 10.104.0.2              N/A       N/A        Y
>>>>>>    962
>>>>>> Self-heal Daemon on 10.104.0.3              N/A       N/A        Y
>>>>>>    108977
>>>>>> Self-heal Daemon on 10.104.0.4              N/A       N/A        Y
>>>>>>    61603
>>>>>>
>>>>>> Task Status of Volume volume2
>>>>>> ------------------------------------------------------------
>>>>>> ------------------
>>>>>> There are no active volume tasks
>>>>>>
>>>>>> I think ovirt can't read valid informations about gluster.
>>>>>> I can't contiune upgrade of other hosts until this problem exist.
>>>>>>
>>>>>> Please help me:)
>>>>>>
>>>>>>
>>>>>> Thanks
>>>>>>
>>>>>> Regards,
>>>>>>
>>>>>> Tibor
>>>>>>
>>>>>>
>>>>>> _______________________________________________
>>>>>> Users mailing list -- users@ovirt.org
>>>>>> To unsubscribe send an email to users-le...@ovirt.org
>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>> _______________________________________________
>>> Users mailing list -- users@ovirt.org
>>> To unsubscribe send an email to users-le...@ovirt.org
>>>
>>>
>>
>> _______________________________________________
>> Users mailing list -- users@ovirt.org
>> To unsubscribe send an email to users-le...@ovirt.org
>>
>>
>> _______________________________________________
>> Users mailing list -- users@ovirt.org
>> To unsubscribe send an email to users-le...@ovirt.org
>> oVirt Code of Conduct: https://www.ovirt.org/community/about/community-
>> guidelines/
>> List Archives:
>>
>>
>
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org

Reply via email to