Hi, 

4.2.4-0.0.master.20180515183442.git00e1340.el7.centos 

Firstly, I did a yum update "ovirt-*-setup*" 
second, I have ran engine-setup to upgrade. 

I didn't remove the old repos, just installed the nightly repo. 

Thank you again, 

Regards, 

Tibor 

----- 2018. máj.. 17., 15:02, Sahina Bose <sab...@redhat.com> írta: 

> It doesn't look like the patch was applied. Still see the same error in
> engine.log
> "Error while refreshing brick statuses for volume 'volume1' of cluster 
> 'C6220':
> null"\

> Did you use engine-setup to upgrade? What's the version of ovirt-engine
> currently installed?

> On Thu, May 17, 2018 at 5:10 PM, Demeter Tibor < [ mailto:tdeme...@itsmart.hu 
> |
> tdeme...@itsmart.hu ] > wrote:

>> Hi,

>> sure,

>> Thank you for your time!

>> R
>> Tibor

>> ----- 2018. máj.. 17., 12:19, Sahina Bose < [ mailto:sab...@redhat.com |
>> sab...@redhat.com ] > írta:

>>> [+users]

>>> Can you provide the engine.log to see why the monitoring is not working 
>>> here.
>>> thanks!

>>> On Wed, May 16, 2018 at 2:08 PM, Demeter Tibor < [ 
>>> mailto:tdeme...@itsmart.hu |
>>> tdeme...@itsmart.hu ] > wrote:

>>>> Hi,

>>>> Meanwhile, I did the upgrade engine, but the gluster state is same on my 
>>>> first
>>>> node.
>>>> I've attached some screenshot of my problem.

>>>> Thanks

>>>> Tibor

>>>> ----- 2018. máj.. 16., 10:16, Demeter Tibor < [ mailto:tdeme...@itsmart.hu 
>>>> |
>>>> tdeme...@itsmart.hu ] > írta Hi,

>>>>> If 4.3.4 will release, i just have to remove the nightly repo and update 
>>>>> to
>>>>> stable?

>>>>> I'm sorry for my terrible English, I try to explain what was my problem 
>>>>> with
>>>>> update.
>>>>> I'm upgraded from 4.1.8.

>>>>> I followed up the official hosted-engine update documentation, that was 
>>>>> not
>>>>> clear me, because it has referenced to a lot of old thing (i think).
>>>>> [ https://www.ovirt.org/documentation/upgrade-guide/upgrade-guide/ |
>>>>> https://www.ovirt.org/documentation/upgrade-guide/upgrade-guide/ ]
>>>>> [
>>>>> https://www.ovirt.org/documentation/how-to/hosted-engine/#upgrade-hosted-engine
>>>>> |
>>>>> https://www.ovirt.org/documentation/how-to/hosted-engine/#upgrade-hosted-engine
>>>>> ]

>>>>> Maybe it need to update, because I had a lot of question under upgrade 
>>>>> and I was
>>>>> not sure in all of necessary steps. For example, If I need to installing 
>>>>> the
>>>>> new, 4.2 repo on the hosts, then need to remove the old repo from that?
>>>>> Why I need to do a" yum update -y" on hosts, meanwhile there is an 
>>>>> "Updatehost"
>>>>> menu in the GUI? So, maybe it outdated.
>>>>> Since upgrade hosted engine, and the first node, I have problems with 
>>>>> gluster.
>>>>> It seems to working fine if you check it from console "gluster volume 
>>>>> status,
>>>>> etc" but not on the Gui, because now it yellow, and the brick reds in the 
>>>>> first
>>>>> node.

>>>>> Previously I did a mistake with glusterfs, my gluster config was wrong. I 
>>>>> have
>>>>> corrected them, but it did not helped to me,gluster bricks are reds on my 
>>>>> first
>>>>> node yet....

>>>>> Now I try to upgrade to nightly, but I'm affraid, because it a living,
>>>>> productive system, and I don't have downtime. I hope it will help me.

>>>>> Thanks for all,

>>>>> Regards,
>>>>> Tibor Demeter

>>>>> ----- 2018. máj.. 16., 9:58, Sahina Bose < [ mailto:sab...@redhat.com |
>>>>> sab...@redhat.com ] > írta:

>>>>>> On Wed, May 16, 2018 at 1:19 PM, Demeter Tibor < [ 
>>>>>> mailto:tdeme...@itsmart.hu |
>>>>>> tdeme...@itsmart.hu ] > wrote:

>>>>>>> Hi,

>>>>>>> is it a different, unstable repo? I have a productive cluster, how is 
>>>>>>> safe that?
>>>>>>> I don't have any experience with nightly build. How can I use this? It 
>>>>>>> have to
>>>>>>> install to the engine VM or all of my hosts?
>>>>>>> Thanks in advance for help me..

>>>>>> Only on the engine VM.

>>>>>> Regarding stability - it passes CI so relatively stable, beyond that 
>>>>>> there are
>>>>>> no guarantees.

>>>>>> What's the specific problem you're facing with update? Can you elaborate?

>>>>>>> Regards,

>>>>>>> Tibor

>>>>>>> ----- 2018. máj.. 15., 9:58, Demeter Tibor < [ 
>>>>>>> mailto:tdeme...@itsmart.hu |
>>>>>>> tdeme...@itsmart.hu ] > írta:

>>>>>>>> Hi,

>>>>>>>> Could you explain how can I use this patch?

>>>>>>>> R,
>>>>>>>> Tibor

>>>>>>>> ----- 2018. máj.. 14., 11:18, Demeter Tibor < [ 
>>>>>>>> mailto:tdeme...@itsmart.hu |
>>>>>>>> tdeme...@itsmart.hu ] > írta:

>>>>>>>>> Hi,

>>>>>>>>> Sorry for my question, but can you tell me please how can I use this 
>>>>>>>>> patch?

>>>>>>>>> Thanks,
>>>>>>>>> Regards,
>>>>>>>>> Tibor
>>>>>>>>> ----- 2018. máj.. 14., 10:47, Sahina Bose < [ 
>>>>>>>>> mailto:sab...@redhat.com |
>>>>>>>>> sab...@redhat.com ] > írta:

>>>>>>>>>> On Sat, May 12, 2018 at 1:14 PM, Demeter Tibor < [ 
>>>>>>>>>> mailto:tdeme...@itsmart.hu |
>>>>>>>>>> tdeme...@itsmart.hu ] > wrote:

>>>>>>>>>>> Hi,

>>>>>>>>>>> Could someone help me please ? I can't finish my upgrade process.

>>>>>>>>>> [ https://gerrit.ovirt.org/91164 | https://gerrit.ovirt.org/91164 ] 
>>>>>>>>>> should fix
>>>>>>>>>> the error you're facing.

>>>>>>>>>> Can you elaborate why this is affecting the upgrade process?

>>>>>>>>>>> Thanks
>>>>>>>>>>> R
>>>>>>>>>>> Tibor

>>>>>>>>>>> ----- 2018. máj.. 10., 12:51, Demeter Tibor < [ 
>>>>>>>>>>> mailto:tdeme...@itsmart.hu |
>>>>>>>>>>> tdeme...@itsmart.hu ] > írta:

>>>>>>>>>>>> Hi,

>>>>>>>>>>>> I've attached the vdsm and supervdsm logs. But I don't have 
>>>>>>>>>>>> engine.log here,
>>>>>>>>>>>> because that is on hosted engine vm. Should I send that ?

>>>>>>>>>>>> Thank you

>>>>>>>>>>>> Regards,

>>>>>>>>>>>> Tibor
>>>>>>>>>>>> ----- 2018. máj.. 10., 12:30, Sahina Bose < [ 
>>>>>>>>>>>> mailto:sab...@redhat.com |
>>>>>>>>>>>> sab...@redhat.com ] > írta:

>>>>>>>>>>>>> There's a bug here. Can you log one attaching this engine.log and 
>>>>>>>>>>>>> also vdsm.log
>>>>>>>>>>>>> & supervdsm.log from n3.itsmart.cloud

>>>>>>>>>>>>> On Thu, May 10, 2018 at 3:35 PM, Demeter Tibor < [ 
>>>>>>>>>>>>> mailto:tdeme...@itsmart.hu |
>>>>>>>>>>>>> tdeme...@itsmart.hu ] > wrote:

>>>>>>>>>>>>>> Hi,

>>>>>>>>>>>>>> I found this:

>>>>>>>>>>>>>> 2018-05-10 03:24:19,096+02 INFO
>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH,
>>>>>>>>>>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand, return:
>>>>>>>>>>>>>> org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeAdvancedDetails@ca97448e,
>>>>>>>>>>>>>> log id: 347435ae
>>>>>>>>>>>>>> 2018-05-10 03:24:19,097+02 ERROR
>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob] 
>>>>>>>>>>>>>> (DefaultQuartzScheduler7)
>>>>>>>>>>>>>> [43f4eaec] Error while refreshing brick statuses for volume 
>>>>>>>>>>>>>> 'volume2' of
>>>>>>>>>>>>>> cluster 'C6220': null
>>>>>>>>>>>>>> 2018-05-10 03:24:19,097+02 INFO
>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] 
>>>>>>>>>>>>>> (DefaultQuartzScheduler8)
>>>>>>>>>>>>>> [7715ceda] Failed to acquire lock and wait lock
>>>>>>>>>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>>>>>>>>>>>> sharedLocks=''}'
>>>>>>>>>>>>>> 2018-05-10 03:24:19,104+02 INFO
>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] START,
>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = 
>>>>>>>>>>>>>> n4.itsmart.cloud,
>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'}),
>>>>>>>>>>>>>> log id: 6908121d
>>>>>>>>>>>>>> 2018-05-10 03:24:19,106+02 ERROR
>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] Command
>>>>>>>>>>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = 
>>>>>>>>>>>>>> n4.itsmart.cloud,
>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'})'
>>>>>>>>>>>>>> execution failed: null
>>>>>>>>>>>>>> 2018-05-10 03:24:19,106+02 INFO
>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH,
>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 6908121d
>>>>>>>>>>>>>> 2018-05-10 03:24:19,107+02 INFO
>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] START,
>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = 
>>>>>>>>>>>>>> n1.itsmart.cloud,
>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'}),
>>>>>>>>>>>>>> log id: 735c6a5f
>>>>>>>>>>>>>> 2018-05-10 03:24:19,109+02 ERROR
>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] Command
>>>>>>>>>>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = 
>>>>>>>>>>>>>> n1.itsmart.cloud,
>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'})'
>>>>>>>>>>>>>> execution failed: null
>>>>>>>>>>>>>> 2018-05-10 03:24:19,109+02 INFO
>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH,
>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 735c6a5f
>>>>>>>>>>>>>> 2018-05-10 03:24:19,110+02 INFO
>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] START,
>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = 
>>>>>>>>>>>>>> n2.itsmart.cloud,
>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}),
>>>>>>>>>>>>>> log id: 6f9e9f58
>>>>>>>>>>>>>> 2018-05-10 03:24:19,112+02 ERROR
>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] Command
>>>>>>>>>>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = 
>>>>>>>>>>>>>> n2.itsmart.cloud,
>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'})'
>>>>>>>>>>>>>> execution failed: null
>>>>>>>>>>>>>> 2018-05-10 03:24:19,112+02 INFO
>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH,
>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 6f9e9f58
>>>>>>>>>>>>>> 2018-05-10 03:24:19,113+02 INFO
>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] START,
>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = 
>>>>>>>>>>>>>> n3.itsmart.cloud,
>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'}),
>>>>>>>>>>>>>> log id: 2ee46967
>>>>>>>>>>>>>> 2018-05-10 03:24:19,115+02 ERROR
>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] Command
>>>>>>>>>>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = 
>>>>>>>>>>>>>> n3.itsmart.cloud,
>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'})'
>>>>>>>>>>>>>> execution failed: null
>>>>>>>>>>>>>> 2018-05-10 03:24:19,116+02 INFO
>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH,
>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 2ee46967
>>>>>>>>>>>>>> 2018-05-10 03:24:19,117+02 INFO
>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] START,
>>>>>>>>>>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand(HostName = 
>>>>>>>>>>>>>> n1.itsmart.cloud,
>>>>>>>>>>>>>> GlusterVolumeAdvancedDetailsVDSParameters:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57',
>>>>>>>>>>>>>> volumeName='volume1'}), log id: 7550e5c
>>>>>>>>>>>>>> 2018-05-10 03:24:20,748+02 INFO
>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH,
>>>>>>>>>>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand, return:
>>>>>>>>>>>>>> org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeAdvancedDetails@4a46066f,
>>>>>>>>>>>>>> log id: 7550e5c
>>>>>>>>>>>>>> 2018-05-10 03:24:20,749+02 ERROR
>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob] 
>>>>>>>>>>>>>> (DefaultQuartzScheduler7)
>>>>>>>>>>>>>> [43f4eaec] Error while refreshing brick statuses for volume 
>>>>>>>>>>>>>> 'volume1' of
>>>>>>>>>>>>>> cluster 'C6220': null
>>>>>>>>>>>>>> 2018-05-10 03:24:20,750+02 INFO
>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
>>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] START,
>>>>>>>>>>>>>> GlusterServersListVDSCommand(HostName = n2.itsmart.cloud,
>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}),
>>>>>>>>>>>>>> log id: 120cc68d
>>>>>>>>>>>>>> 2018-05-10 03:24:20,930+02 INFO
>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
>>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] FINISH, 
>>>>>>>>>>>>>> GlusterServersListVDSCommand,
>>>>>>>>>>>>>> return: [ [ http://10.101.0.2/24:CONNECTED | 
>>>>>>>>>>>>>> 10.101.0.2/24:CONNECTED ] ,
>>>>>>>>>>>>>> n1.cloudata.local:CONNECTED, 10.104.0.3:CONNECTED, 
>>>>>>>>>>>>>> 10.104.0.4:CONNECTED], log
>>>>>>>>>>>>>> id: 120cc68d
>>>>>>>>>>>>>> 2018-05-10 03:24:20,949+02 INFO
>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] START,
>>>>>>>>>>>>>> GlusterVolumesListVDSCommand(HostName = n2.itsmart.cloud,
>>>>>>>>>>>>>> GlusterVolumesListVDSParameters:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}),
>>>>>>>>>>>>>> log id: 118aa264
>>>>>>>>>>>>>> 2018-05-10 03:24:21,048+02 WARN
>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick
>>>>>>>>>>>>>> '10.104.0.1:/gluster/brick/brick1' of volume
>>>>>>>>>>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as 
>>>>>>>>>>>>>> no gluster
>>>>>>>>>>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339'
>>>>>>>>>>>>>> 2018-05-10 03:24:21,055+02 WARN
>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick
>>>>>>>>>>>>>> '10.104.0.1:/gluster/brick/brick2' of volume
>>>>>>>>>>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as 
>>>>>>>>>>>>>> no gluster
>>>>>>>>>>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339'
>>>>>>>>>>>>>> 2018-05-10 03:24:21,061+02 WARN
>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick
>>>>>>>>>>>>>> '10.104.0.1:/gluster/brick/brick3' of volume
>>>>>>>>>>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as 
>>>>>>>>>>>>>> no gluster
>>>>>>>>>>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339'
>>>>>>>>>>>>>> 2018-05-10 03:24:21,067+02 WARN
>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick
>>>>>>>>>>>>>> '10.104.0.1:/gluster2/brick/brick1' of volume
>>>>>>>>>>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as 
>>>>>>>>>>>>>> no gluster
>>>>>>>>>>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339'
>>>>>>>>>>>>>> 2018-05-10 03:24:21,074+02 WARN
>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick
>>>>>>>>>>>>>> '10.104.0.1:/gluster2/brick/brick2' of volume
>>>>>>>>>>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as 
>>>>>>>>>>>>>> no gluster
>>>>>>>>>>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339'
>>>>>>>>>>>>>> 2018-05-10 03:24:21,080+02 WARN
>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick
>>>>>>>>>>>>>> '10.104.0.1:/gluster2/brick/brick3' of volume
>>>>>>>>>>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as 
>>>>>>>>>>>>>> no gluster
>>>>>>>>>>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339'
>>>>>>>>>>>>>> 2018-05-10 03:24:21,081+02 INFO
>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] FINISH, 
>>>>>>>>>>>>>> GlusterVolumesListVDSCommand,
>>>>>>>>>>>>>> return:
>>>>>>>>>>>>>> {68cfb061-1320-4042-abcd-9228da23c0c8=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7a6720d,
>>>>>>>>>>>>>> e0f568fa-987c-4f5c-b853-01bce718ee27=org.ovirt.engine.core.common.businessentities.g
>>>>>>>>>>>>>> luster.GlusterVolumeEntity@f88c521b}, log id: 118aa264

>>>>>>>>>>>>>> 2018-05-10 11:59:26,047+02 ERROR
>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] Command
>>>>>>>>>>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = 
>>>>>>>>>>>>>> n1.itsmart.cloud,
>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'})'
>>>>>>>>>>>>>> execution failed: null
>>>>>>>>>>>>>> 2018-05-10 11:59:26,047+02 INFO
>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH,
>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 14a71ef0
>>>>>>>>>>>>>> 2018-05-10 11:59:26,048+02 INFO
>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] START,
>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = 
>>>>>>>>>>>>>> n4.itsmart.cloud,
>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'}),
>>>>>>>>>>>>>> log id: 28d9e255
>>>>>>>>>>>>>> 2018-05-10 11:59:26,051+02 ERROR
>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] Command
>>>>>>>>>>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = 
>>>>>>>>>>>>>> n4.itsmart.cloud,
>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'})'
>>>>>>>>>>>>>> execution failed: null
>>>>>>>>>>>>>> 2018-05-10 11:59:26,051+02 INFO
>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH,
>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 28d9e255
>>>>>>>>>>>>>> 2018-05-10 11:59:26,052+02 INFO
>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] START,
>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = 
>>>>>>>>>>>>>> n2.itsmart.cloud,
>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}),
>>>>>>>>>>>>>> log id: 4a7b280e
>>>>>>>>>>>>>> 2018-05-10 11:59:26,054+02 ERROR
>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] Command
>>>>>>>>>>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = 
>>>>>>>>>>>>>> n2.itsmart.cloud,
>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'})'
>>>>>>>>>>>>>> execution failed: null
>>>>>>>>>>>>>> 2018-05-10 11:59:26,054+02 INFO
>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH,
>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 4a7b280e
>>>>>>>>>>>>>> 2018-05-10 11:59:26,055+02 INFO
>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] START,
>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = 
>>>>>>>>>>>>>> n3.itsmart.cloud,
>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'}),
>>>>>>>>>>>>>> log id: 18adc534
>>>>>>>>>>>>>> 2018-05-10 11:59:26,057+02 ERROR
>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] Command
>>>>>>>>>>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = 
>>>>>>>>>>>>>> n3.itsmart.cloud,
>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'})'
>>>>>>>>>>>>>> execution failed: null
>>>>>>>>>>>>>> 2018-05-10 11:59:26,057+02 INFO
>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH,
>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 18adc534
>>>>>>>>>>>>>> 2018-05-10 11:59:26,058+02 INFO
>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] START,
>>>>>>>>>>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand(HostName = 
>>>>>>>>>>>>>> n3.itsmart.cloud,
>>>>>>>>>>>>>> GlusterVolumeAdvancedDetailsVDSParameters:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec',
>>>>>>>>>>>>>> volumeName='volume1'}), log id: 3451084f
>>>>>>>>>>>>>> 2018-05-10 11:59:28,050+02 INFO
>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] 
>>>>>>>>>>>>>> (DefaultQuartzScheduler3)
>>>>>>>>>>>>>> [2eb1c389] Failed to acquire lock and wait lock
>>>>>>>>>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>>>>>>>>>>>> sharedLocks=''}'
>>>>>>>>>>>>>> 2018-05-10 11:59:28,060+02 INFO
>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] 
>>>>>>>>>>>>>> (DefaultQuartzScheduler3)
>>>>>>>>>>>>>> [2eb1c389] Failed to acquire lock and wait lock
>>>>>>>>>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>>>>>>>>>>>> sharedLocks=''}'
>>>>>>>>>>>>>> 2018-05-10 11:59:28,062+02 INFO
>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] 
>>>>>>>>>>>>>> (DefaultQuartzScheduler3)
>>>>>>>>>>>>>> [2eb1c389] Failed to acquire lock and wait lock
>>>>>>>>>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>>>>>>>>>>>> sharedLocks=''}'
>>>>>>>>>>>>>> 2018-05-10 11:59:31,054+02 INFO
>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] 
>>>>>>>>>>>>>> (DefaultQuartzScheduler3)
>>>>>>>>>>>>>> [2eb1c389] Failed to acquire lock and wait lock
>>>>>>>>>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>>>>>>>>>>>> sharedLocks=''}'
>>>>>>>>>>>>>> 2018-05-10 11:59:31,054+02 INFO
>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] 
>>>>>>>>>>>>>> (DefaultQuartzScheduler3)
>>>>>>>>>>>>>> [2eb1c389] Failed to acquire lock and wait lock
>>>>>>>>>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>>>>>>>>>>>> sharedLocks=''}'
>>>>>>>>>>>>>> 2018-05-10 11:59:31,062+02 INFO
>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] 
>>>>>>>>>>>>>> (DefaultQuartzScheduler3)
>>>>>>>>>>>>>> [2eb1c389] Failed to acquire lock and wait lock
>>>>>>>>>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>>>>>>>>>>>> sharedLocks=''}'
>>>>>>>>>>>>>> 2018-05-10 11:59:31,064+02 INFO
>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] 
>>>>>>>>>>>>>> (DefaultQuartzScheduler3)
>>>>>>>>>>>>>> [2eb1c389] Failed to acquire lock and wait lock
>>>>>>>>>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>>>>>>>>>>>> sharedLocks=''}'
>>>>>>>>>>>>>> 2018-05-10 11:59:31,465+02 INFO
>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH,
>>>>>>>>>>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand, return:
>>>>>>>>>>>>>> org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeAdvancedDetails@3f1b7f43,
>>>>>>>>>>>>>> log id: 3451084f
>>>>>>>>>>>>>> 2018-05-10 11:59:31,466+02 ERROR
>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob] 
>>>>>>>>>>>>>> (DefaultQuartzScheduler4)
>>>>>>>>>>>>>> [400fa486] Error while refreshing brick statuses for volume 
>>>>>>>>>>>>>> 'volume1' of
>>>>>>>>>>>>>> cluster 'C6220': null

>>>>>>>>>>>>>> R
>>>>>>>>>>>>>> Tibor

>>>>>>>>>>>>>> ----- 2018. máj.. 10., 11:43, Sahina Bose < [ 
>>>>>>>>>>>>>> mailto:sab...@redhat.com |
>>>>>>>>>>>>>> sab...@redhat.com ] > írta:

>>>>>>>>>>>>>>> This doesn't affect the monitoring of state.
>>>>>>>>>>>>>>> Any errors in vdsm.log?
>>>>>>>>>>>>>>> Or errors in engine.log of the form "Error while refreshing 
>>>>>>>>>>>>>>> brick statuses for
>>>>>>>>>>>>>>> volume"

>>>>>>>>>>>>>>> On Thu, May 10, 2018 at 2:33 PM, Demeter Tibor < [ 
>>>>>>>>>>>>>>> mailto:tdeme...@itsmart.hu |
>>>>>>>>>>>>>>> tdeme...@itsmart.hu ] > wrote:

>>>>>>>>>>>>>>>> Hi,

>>>>>>>>>>>>>>>> Thank you for your fast reply :)

>>>>>>>>>>>>>>>> 2018-05-10 11:01:51,574+02 INFO
>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
>>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] START,
>>>>>>>>>>>>>>>> GlusterServersListVDSCommand(HostName = n2.itsmart.cloud,
>>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}),
>>>>>>>>>>>>>>>> log id: 39adbbb8
>>>>>>>>>>>>>>>> 2018-05-10 11:01:51,768+02 INFO
>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
>>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] FINISH, 
>>>>>>>>>>>>>>>> GlusterServersListVDSCommand,
>>>>>>>>>>>>>>>> return: [ [ http://10.101.0.2/24:CONNECTED | 
>>>>>>>>>>>>>>>> 10.101.0.2/24:CONNECTED ] ,
>>>>>>>>>>>>>>>> n1.cloudata.local:CONNECTED, 10.104.0.3:CONNECTED, 
>>>>>>>>>>>>>>>> 10.104.0.4:CONNECTED], log
>>>>>>>>>>>>>>>> id: 39adbbb8
>>>>>>>>>>>>>>>> 2018-05-10 11:01:51,788+02 INFO
>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] START,
>>>>>>>>>>>>>>>> GlusterVolumesListVDSCommand(HostName = n2.itsmart.cloud,
>>>>>>>>>>>>>>>> GlusterVolumesListVDSParameters:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}),
>>>>>>>>>>>>>>>> log id: 738a7261
>>>>>>>>>>>>>>>> 2018-05-10 11:01:51,892+02 WARN
>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick
>>>>>>>>>>>>>>>> '10.104.0.1:/gluster/brick/brick1' of volume
>>>>>>>>>>>>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as 
>>>>>>>>>>>>>>>> no gluster
>>>>>>>>>>>>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339'
>>>>>>>>>>>>>>>> 2018-05-10 11:01:51,898+02 WARN
>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick
>>>>>>>>>>>>>>>> '10.104.0.1:/gluster/brick/brick2' of volume
>>>>>>>>>>>>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as 
>>>>>>>>>>>>>>>> no gluster
>>>>>>>>>>>>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339'
>>>>>>>>>>>>>>>> 2018-05-10 11:01:51,905+02 WARN
>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick
>>>>>>>>>>>>>>>> '10.104.0.1:/gluster/brick/brick3' of volume
>>>>>>>>>>>>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as 
>>>>>>>>>>>>>>>> no gluster
>>>>>>>>>>>>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339'
>>>>>>>>>>>>>>>> 2018-05-10 11:01:51,911+02 WARN
>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick
>>>>>>>>>>>>>>>> '10.104.0.1:/gluster2/brick/brick1' of volume
>>>>>>>>>>>>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as 
>>>>>>>>>>>>>>>> no gluster
>>>>>>>>>>>>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339'
>>>>>>>>>>>>>>>> 2018-05-10 11:01:51,917+02 WARN
>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick
>>>>>>>>>>>>>>>> '10.104.0.1:/gluster2/brick/brick2' of volume
>>>>>>>>>>>>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as 
>>>>>>>>>>>>>>>> no gluster
>>>>>>>>>>>>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339'
>>>>>>>>>>>>>>>> 2018-05-10 11:01:51,924+02 WARN
>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick
>>>>>>>>>>>>>>>> '10.104.0.1:/gluster2/brick/brick3' of volume
>>>>>>>>>>>>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as 
>>>>>>>>>>>>>>>> no gluster
>>>>>>>>>>>>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339'
>>>>>>>>>>>>>>>> 2018-05-10 11:01:51,925+02 INFO
>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] FINISH, 
>>>>>>>>>>>>>>>> GlusterVolumesListVDSCommand,
>>>>>>>>>>>>>>>> return:
>>>>>>>>>>>>>>>> {68cfb061-1320-4042-abcd-9228da23c0c8=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7a6720d,
>>>>>>>>>>>>>>>> e0f568fa-987c-4f5c-b853-01bce718ee27=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@f88c521b},
>>>>>>>>>>>>>>>> log id: 738a7261

>>>>>>>>>>>>>>>> This happening continuously.

>>>>>>>>>>>>>>>> Thanks!
>>>>>>>>>>>>>>>> Tibor

>>>>>>>>>>>>>>>> ----- 2018. máj.. 10., 10:56, Sahina Bose < [ 
>>>>>>>>>>>>>>>> mailto:sab...@redhat.com |
>>>>>>>>>>>>>>>> sab...@redhat.com ] > írta:

>>>>>>>>>>>>>>>>> Could you check the engine.log if there are errors related to 
>>>>>>>>>>>>>>>>> getting
>>>>>>>>>>>>>>>>> GlusterVolumeAdvancedDetails ?

>>>>>>>>>>>>>>>>> On Thu, May 10, 2018 at 2:02 PM, Demeter Tibor < [ 
>>>>>>>>>>>>>>>>> mailto:tdeme...@itsmart.hu |
>>>>>>>>>>>>>>>>> tdeme...@itsmart.hu ] > wrote:

>>>>>>>>>>>>>>>>>> Dear Ovirt Users,
>>>>>>>>>>>>>>>>>> I've followed up the self-hosted-engine upgrade 
>>>>>>>>>>>>>>>>>> documentation, I upgraded my 4.1
>>>>>>>>>>>>>>>>>> system to 4.2.3.
>>>>>>>>>>>>>>>>>> I upgaded the first node with yum upgrade, it seems working 
>>>>>>>>>>>>>>>>>> now fine. But since
>>>>>>>>>>>>>>>>>> upgrade, the gluster informations seems to displayed 
>>>>>>>>>>>>>>>>>> incorrect on the admin
>>>>>>>>>>>>>>>>>> panel. The volume yellow, and there are red bricks from that 
>>>>>>>>>>>>>>>>>> node.
>>>>>>>>>>>>>>>>>> I've checked in console, I think my gluster is not degraded:

>>>>>>>>>>>>>>>>>> root@n1 ~]# gluster volume list
>>>>>>>>>>>>>>>>>> volume1
>>>>>>>>>>>>>>>>>> volume2
>>>>>>>>>>>>>>>>>> [root@n1 ~]# gluster volume info
>>>>>>>>>>>>>>>>>> Volume Name: volume1
>>>>>>>>>>>>>>>>>> Type: Distributed-Replicate
>>>>>>>>>>>>>>>>>> Volume ID: e0f568fa-987c-4f5c-b853-01bce718ee27
>>>>>>>>>>>>>>>>>> Status: Started
>>>>>>>>>>>>>>>>>> Snapshot Count: 0
>>>>>>>>>>>>>>>>>> Number of Bricks: 3 x 3 = 9
>>>>>>>>>>>>>>>>>> Transport-type: tcp
>>>>>>>>>>>>>>>>>> Bricks:
>>>>>>>>>>>>>>>>>> Brick1: 10.104.0.1:/gluster/brick/brick1
>>>>>>>>>>>>>>>>>> Brick2: 10.104.0.2:/gluster/brick/brick1
>>>>>>>>>>>>>>>>>> Brick3: 10.104.0.3:/gluster/brick/brick1
>>>>>>>>>>>>>>>>>> Brick4: 10.104.0.1:/gluster/brick/brick2
>>>>>>>>>>>>>>>>>> Brick5: 10.104.0.2:/gluster/brick/brick2
>>>>>>>>>>>>>>>>>> Brick6: 10.104.0.3:/gluster/brick/brick2
>>>>>>>>>>>>>>>>>> Brick7: 10.104.0.1:/gluster/brick/brick3
>>>>>>>>>>>>>>>>>> Brick8: 10.104.0.2:/gluster/brick/brick3
>>>>>>>>>>>>>>>>>> Brick9: 10.104.0.3:/gluster/brick/brick3
>>>>>>>>>>>>>>>>>> Options Reconfigured:
>>>>>>>>>>>>>>>>>> transport.address-family: inet
>>>>>>>>>>>>>>>>>> performance.readdir-ahead: on
>>>>>>>>>>>>>>>>>> nfs.disable: on
>>>>>>>>>>>>>>>>>> storage.owner-uid: 36
>>>>>>>>>>>>>>>>>> storage.owner-gid: 36
>>>>>>>>>>>>>>>>>> performance.quick-read: off
>>>>>>>>>>>>>>>>>> performance.read-ahead: off
>>>>>>>>>>>>>>>>>> performance.io-cache: off
>>>>>>>>>>>>>>>>>> performance.stat-prefetch: off
>>>>>>>>>>>>>>>>>> performance.low-prio-threads: 32
>>>>>>>>>>>>>>>>>> network.remote-dio: enable
>>>>>>>>>>>>>>>>>> cluster.eager-lock: enable
>>>>>>>>>>>>>>>>>> cluster.quorum-type: auto
>>>>>>>>>>>>>>>>>> cluster.server-quorum-type: server
>>>>>>>>>>>>>>>>>> cluster.data-self-heal-algorithm: full
>>>>>>>>>>>>>>>>>> cluster.locking-scheme: granular
>>>>>>>>>>>>>>>>>> cluster.shd-max-threads: 8
>>>>>>>>>>>>>>>>>> cluster.shd-wait-qlength: 10000
>>>>>>>>>>>>>>>>>> features.shard: on
>>>>>>>>>>>>>>>>>> user.cifs: off
>>>>>>>>>>>>>>>>>> server.allow-insecure: on
>>>>>>>>>>>>>>>>>> Volume Name: volume2
>>>>>>>>>>>>>>>>>> Type: Distributed-Replicate
>>>>>>>>>>>>>>>>>> Volume ID: 68cfb061-1320-4042-abcd-9228da23c0c8
>>>>>>>>>>>>>>>>>> Status: Started
>>>>>>>>>>>>>>>>>> Snapshot Count: 0
>>>>>>>>>>>>>>>>>> Number of Bricks: 3 x 3 = 9
>>>>>>>>>>>>>>>>>> Transport-type: tcp
>>>>>>>>>>>>>>>>>> Bricks:
>>>>>>>>>>>>>>>>>> Brick1: 10.104.0.1:/gluster2/brick/brick1
>>>>>>>>>>>>>>>>>> Brick2: 10.104.0.2:/gluster2/brick/brick1
>>>>>>>>>>>>>>>>>> Brick3: 10.104.0.3:/gluster2/brick/brick1
>>>>>>>>>>>>>>>>>> Brick4: 10.104.0.1:/gluster2/brick/brick2
>>>>>>>>>>>>>>>>>> Brick5: 10.104.0.2:/gluster2/brick/brick2
>>>>>>>>>>>>>>>>>> Brick6: 10.104.0.3:/gluster2/brick/brick2
>>>>>>>>>>>>>>>>>> Brick7: 10.104.0.1:/gluster2/brick/brick3
>>>>>>>>>>>>>>>>>> Brick8: 10.104.0.2:/gluster2/brick/brick3
>>>>>>>>>>>>>>>>>> Brick9: 10.104.0.3:/gluster2/brick/brick3
>>>>>>>>>>>>>>>>>> Options Reconfigured:
>>>>>>>>>>>>>>>>>> nfs.disable: on
>>>>>>>>>>>>>>>>>> performance.readdir-ahead: on
>>>>>>>>>>>>>>>>>> transport.address-family: inet
>>>>>>>>>>>>>>>>>> cluster.quorum-type: auto
>>>>>>>>>>>>>>>>>> network.ping-timeout: 10
>>>>>>>>>>>>>>>>>> auth.allow: *
>>>>>>>>>>>>>>>>>> performance.quick-read: off
>>>>>>>>>>>>>>>>>> performance.read-ahead: off
>>>>>>>>>>>>>>>>>> performance.io-cache: off
>>>>>>>>>>>>>>>>>> performance.stat-prefetch: off
>>>>>>>>>>>>>>>>>> performance.low-prio-threads: 32
>>>>>>>>>>>>>>>>>> network.remote-dio: enable
>>>>>>>>>>>>>>>>>> cluster.eager-lock: enable
>>>>>>>>>>>>>>>>>> cluster.server-quorum-type: server
>>>>>>>>>>>>>>>>>> cluster.data-self-heal-algorithm: full
>>>>>>>>>>>>>>>>>> cluster.locking-scheme: granular
>>>>>>>>>>>>>>>>>> cluster.shd-max-threads: 8
>>>>>>>>>>>>>>>>>> cluster.shd-wait-qlength: 10000
>>>>>>>>>>>>>>>>>> features.shard: on
>>>>>>>>>>>>>>>>>> user.cifs: off
>>>>>>>>>>>>>>>>>> storage.owner-uid: 36
>>>>>>>>>>>>>>>>>> storage.owner-gid: 36
>>>>>>>>>>>>>>>>>> server.allow-insecure: on
>>>>>>>>>>>>>>>>>> [root@n1 ~]# gluster volume status
>>>>>>>>>>>>>>>>>> Status of volume: volume1
>>>>>>>>>>>>>>>>>> Gluster process TCP Port RDMA Port Online Pid
>>>>>>>>>>>>>>>>>> ------------------------------------------------------------------------------
>>>>>>>>>>>>>>>>>> Brick 10.104.0.1:/gluster/brick/brick1 49152 0 Y 3464
>>>>>>>>>>>>>>>>>> Brick 10.104.0.2:/gluster/brick/brick1 49152 0 Y 68937
>>>>>>>>>>>>>>>>>> Brick 10.104.0.3:/gluster/brick/brick1 49161 0 Y 94506
>>>>>>>>>>>>>>>>>> Brick 10.104.0.1:/gluster/brick/brick2 49153 0 Y 3457
>>>>>>>>>>>>>>>>>> Brick 10.104.0.2:/gluster/brick/brick2 49153 0 Y 68943
>>>>>>>>>>>>>>>>>> Brick 10.104.0.3:/gluster/brick/brick2 49162 0 Y 94514
>>>>>>>>>>>>>>>>>> Brick 10.104.0.1:/gluster/brick/brick3 49154 0 Y 3465
>>>>>>>>>>>>>>>>>> Brick 10.104.0.2:/gluster/brick/brick3 49154 0 Y 68949
>>>>>>>>>>>>>>>>>> Brick 10.104.0.3:/gluster/brick/brick3 49163 0 Y 94520
>>>>>>>>>>>>>>>>>> Self-heal Daemon on localhost N/A N/A Y 54356
>>>>>>>>>>>>>>>>>> Self-heal Daemon on 10.104.0.2 N/A N/A Y 962
>>>>>>>>>>>>>>>>>> Self-heal Daemon on 10.104.0.3 N/A N/A Y 108977
>>>>>>>>>>>>>>>>>> Self-heal Daemon on 10.104.0.4 N/A N/A Y 61603
>>>>>>>>>>>>>>>>>> Task Status of Volume volume1
>>>>>>>>>>>>>>>>>> ------------------------------------------------------------------------------
>>>>>>>>>>>>>>>>>> There are no active volume tasks
>>>>>>>>>>>>>>>>>> Status of volume: volume2
>>>>>>>>>>>>>>>>>> Gluster process TCP Port RDMA Port Online Pid
>>>>>>>>>>>>>>>>>> ------------------------------------------------------------------------------
>>>>>>>>>>>>>>>>>> Brick 10.104.0.1:/gluster2/brick/brick1 49155 0 Y 3852
>>>>>>>>>>>>>>>>>> Brick 10.104.0.2:/gluster2/brick/brick1 49158 0 Y 68955
>>>>>>>>>>>>>>>>>> Brick 10.104.0.3:/gluster2/brick/brick1 49164 0 Y 94527
>>>>>>>>>>>>>>>>>> Brick 10.104.0.1:/gluster2/brick/brick2 49156 0 Y 3851
>>>>>>>>>>>>>>>>>> Brick 10.104.0.2:/gluster2/brick/brick2 49159 0 Y 68961
>>>>>>>>>>>>>>>>>> Brick 10.104.0.3:/gluster2/brick/brick2 49165 0 Y 94533
>>>>>>>>>>>>>>>>>> Brick 10.104.0.1:/gluster2/brick/brick3 49157 0 Y 3883
>>>>>>>>>>>>>>>>>> Brick 10.104.0.2:/gluster2/brick/brick3 49160 0 Y 68968
>>>>>>>>>>>>>>>>>> Brick 10.104.0.3:/gluster2/brick/brick3 49166 0 Y 94541
>>>>>>>>>>>>>>>>>> Self-heal Daemon on localhost N/A N/A Y 54356
>>>>>>>>>>>>>>>>>> Self-heal Daemon on 10.104.0.2 N/A N/A Y 962
>>>>>>>>>>>>>>>>>> Self-heal Daemon on 10.104.0.3 N/A N/A Y 108977
>>>>>>>>>>>>>>>>>> Self-heal Daemon on 10.104.0.4 N/A N/A Y 61603
>>>>>>>>>>>>>>>>>> Task Status of Volume volume2
>>>>>>>>>>>>>>>>>> ------------------------------------------------------------------------------
>>>>>>>>>>>>>>>>>> There are no active volume tasks
>>>>>>>>>>>>>>>>>> I think ovirt can't read valid informations about gluster.
>>>>>>>>>>>>>>>>>> I can't contiune upgrade of other hosts until this problem 
>>>>>>>>>>>>>>>>>> exist.

>>>>>>>>>>>>>>>>>> Please help me:)

>>>>>>>>>>>>>>>>>> Thanks

>>>>>>>>>>>>>>>>>> Regards,

>>>>>>>>>>>>>>>>>> Tibor

>>>>>>>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>>>>>>>> Users mailing list -- [ mailto:users@ovirt.org | 
>>>>>>>>>>>>>>>>>> users@ovirt.org ]
>>>>>>>>>>>>>>>>>> To unsubscribe send an email to [ 
>>>>>>>>>>>>>>>>>> mailto:users-le...@ovirt.org |
>>>>>>>>>>>>>>>>>> users-le...@ovirt.org ]

>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>> Users mailing list -- [ mailto:users@ovirt.org | users@ovirt.org ]
>>>>>>>>>>>> To unsubscribe send an email to [ mailto:users-le...@ovirt.org |
>>>>>>>>>>>> users-le...@ovirt.org ]

>>>>>>>>> _______________________________________________
>>>>>>>>> Users mailing list -- [ mailto:users@ovirt.org | users@ovirt.org ]
>>>>>>>>> To unsubscribe send an email to [ mailto:users-le...@ovirt.org |
>>>>>>>>> users-le...@ovirt.org ]

>>>>>>>> _______________________________________________
>>>>>>>> Users mailing list -- [ mailto:users@ovirt.org | users@ovirt.org ]
>>>>>>>> To unsubscribe send an email to [ mailto:users-le...@ovirt.org |
>>>>>>>> users-le...@ovirt.org ]
>>>>>>>> oVirt Code of Conduct: [
>>>>>>>> https://www.ovirt.org/community/about/community-guidelines/ |
>>>>>>>> https://www.ovirt.org/community/about/community-guidelines/ ]
>>>>>>>> List Archives:
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org

Reply via email to