Hi, 4.2.4-0.0.master.20180515183442.git00e1340.el7.centos
Firstly, I did a yum update "ovirt-*-setup*" second, I have ran engine-setup to upgrade. I didn't remove the old repos, just installed the nightly repo. Thank you again, Regards, Tibor ----- 2018. máj.. 17., 15:02, Sahina Bose <sab...@redhat.com> írta: > It doesn't look like the patch was applied. Still see the same error in > engine.log > "Error while refreshing brick statuses for volume 'volume1' of cluster > 'C6220': > null"\ > Did you use engine-setup to upgrade? What's the version of ovirt-engine > currently installed? > On Thu, May 17, 2018 at 5:10 PM, Demeter Tibor < [ mailto:tdeme...@itsmart.hu > | > tdeme...@itsmart.hu ] > wrote: >> Hi, >> sure, >> Thank you for your time! >> R >> Tibor >> ----- 2018. máj.. 17., 12:19, Sahina Bose < [ mailto:sab...@redhat.com | >> sab...@redhat.com ] > írta: >>> [+users] >>> Can you provide the engine.log to see why the monitoring is not working >>> here. >>> thanks! >>> On Wed, May 16, 2018 at 2:08 PM, Demeter Tibor < [ >>> mailto:tdeme...@itsmart.hu | >>> tdeme...@itsmart.hu ] > wrote: >>>> Hi, >>>> Meanwhile, I did the upgrade engine, but the gluster state is same on my >>>> first >>>> node. >>>> I've attached some screenshot of my problem. >>>> Thanks >>>> Tibor >>>> ----- 2018. máj.. 16., 10:16, Demeter Tibor < [ mailto:tdeme...@itsmart.hu >>>> | >>>> tdeme...@itsmart.hu ] > írta Hi, >>>>> If 4.3.4 will release, i just have to remove the nightly repo and update >>>>> to >>>>> stable? >>>>> I'm sorry for my terrible English, I try to explain what was my problem >>>>> with >>>>> update. >>>>> I'm upgraded from 4.1.8. >>>>> I followed up the official hosted-engine update documentation, that was >>>>> not >>>>> clear me, because it has referenced to a lot of old thing (i think). >>>>> [ https://www.ovirt.org/documentation/upgrade-guide/upgrade-guide/ | >>>>> https://www.ovirt.org/documentation/upgrade-guide/upgrade-guide/ ] >>>>> [ >>>>> https://www.ovirt.org/documentation/how-to/hosted-engine/#upgrade-hosted-engine >>>>> | >>>>> https://www.ovirt.org/documentation/how-to/hosted-engine/#upgrade-hosted-engine >>>>> ] >>>>> Maybe it need to update, because I had a lot of question under upgrade >>>>> and I was >>>>> not sure in all of necessary steps. For example, If I need to installing >>>>> the >>>>> new, 4.2 repo on the hosts, then need to remove the old repo from that? >>>>> Why I need to do a" yum update -y" on hosts, meanwhile there is an >>>>> "Updatehost" >>>>> menu in the GUI? So, maybe it outdated. >>>>> Since upgrade hosted engine, and the first node, I have problems with >>>>> gluster. >>>>> It seems to working fine if you check it from console "gluster volume >>>>> status, >>>>> etc" but not on the Gui, because now it yellow, and the brick reds in the >>>>> first >>>>> node. >>>>> Previously I did a mistake with glusterfs, my gluster config was wrong. I >>>>> have >>>>> corrected them, but it did not helped to me,gluster bricks are reds on my >>>>> first >>>>> node yet.... >>>>> Now I try to upgrade to nightly, but I'm affraid, because it a living, >>>>> productive system, and I don't have downtime. I hope it will help me. >>>>> Thanks for all, >>>>> Regards, >>>>> Tibor Demeter >>>>> ----- 2018. máj.. 16., 9:58, Sahina Bose < [ mailto:sab...@redhat.com | >>>>> sab...@redhat.com ] > írta: >>>>>> On Wed, May 16, 2018 at 1:19 PM, Demeter Tibor < [ >>>>>> mailto:tdeme...@itsmart.hu | >>>>>> tdeme...@itsmart.hu ] > wrote: >>>>>>> Hi, >>>>>>> is it a different, unstable repo? I have a productive cluster, how is >>>>>>> safe that? >>>>>>> I don't have any experience with nightly build. How can I use this? It >>>>>>> have to >>>>>>> install to the engine VM or all of my hosts? >>>>>>> Thanks in advance for help me.. >>>>>> Only on the engine VM. >>>>>> Regarding stability - it passes CI so relatively stable, beyond that >>>>>> there are >>>>>> no guarantees. >>>>>> What's the specific problem you're facing with update? Can you elaborate? >>>>>>> Regards, >>>>>>> Tibor >>>>>>> ----- 2018. máj.. 15., 9:58, Demeter Tibor < [ >>>>>>> mailto:tdeme...@itsmart.hu | >>>>>>> tdeme...@itsmart.hu ] > írta: >>>>>>>> Hi, >>>>>>>> Could you explain how can I use this patch? >>>>>>>> R, >>>>>>>> Tibor >>>>>>>> ----- 2018. máj.. 14., 11:18, Demeter Tibor < [ >>>>>>>> mailto:tdeme...@itsmart.hu | >>>>>>>> tdeme...@itsmart.hu ] > írta: >>>>>>>>> Hi, >>>>>>>>> Sorry for my question, but can you tell me please how can I use this >>>>>>>>> patch? >>>>>>>>> Thanks, >>>>>>>>> Regards, >>>>>>>>> Tibor >>>>>>>>> ----- 2018. máj.. 14., 10:47, Sahina Bose < [ >>>>>>>>> mailto:sab...@redhat.com | >>>>>>>>> sab...@redhat.com ] > írta: >>>>>>>>>> On Sat, May 12, 2018 at 1:14 PM, Demeter Tibor < [ >>>>>>>>>> mailto:tdeme...@itsmart.hu | >>>>>>>>>> tdeme...@itsmart.hu ] > wrote: >>>>>>>>>>> Hi, >>>>>>>>>>> Could someone help me please ? I can't finish my upgrade process. >>>>>>>>>> [ https://gerrit.ovirt.org/91164 | https://gerrit.ovirt.org/91164 ] >>>>>>>>>> should fix >>>>>>>>>> the error you're facing. >>>>>>>>>> Can you elaborate why this is affecting the upgrade process? >>>>>>>>>>> Thanks >>>>>>>>>>> R >>>>>>>>>>> Tibor >>>>>>>>>>> ----- 2018. máj.. 10., 12:51, Demeter Tibor < [ >>>>>>>>>>> mailto:tdeme...@itsmart.hu | >>>>>>>>>>> tdeme...@itsmart.hu ] > írta: >>>>>>>>>>>> Hi, >>>>>>>>>>>> I've attached the vdsm and supervdsm logs. But I don't have >>>>>>>>>>>> engine.log here, >>>>>>>>>>>> because that is on hosted engine vm. Should I send that ? >>>>>>>>>>>> Thank you >>>>>>>>>>>> Regards, >>>>>>>>>>>> Tibor >>>>>>>>>>>> ----- 2018. máj.. 10., 12:30, Sahina Bose < [ >>>>>>>>>>>> mailto:sab...@redhat.com | >>>>>>>>>>>> sab...@redhat.com ] > írta: >>>>>>>>>>>>> There's a bug here. Can you log one attaching this engine.log and >>>>>>>>>>>>> also vdsm.log >>>>>>>>>>>>> & supervdsm.log from n3.itsmart.cloud >>>>>>>>>>>>> On Thu, May 10, 2018 at 3:35 PM, Demeter Tibor < [ >>>>>>>>>>>>> mailto:tdeme...@itsmart.hu | >>>>>>>>>>>>> tdeme...@itsmart.hu ] > wrote: >>>>>>>>>>>>>> Hi, >>>>>>>>>>>>>> I found this: >>>>>>>>>>>>>> 2018-05-10 03:24:19,096+02 INFO >>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] >>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, >>>>>>>>>>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand, return: >>>>>>>>>>>>>> org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeAdvancedDetails@ca97448e, >>>>>>>>>>>>>> log id: 347435ae >>>>>>>>>>>>>> 2018-05-10 03:24:19,097+02 ERROR >>>>>>>>>>>>>> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob] >>>>>>>>>>>>>> (DefaultQuartzScheduler7) >>>>>>>>>>>>>> [43f4eaec] Error while refreshing brick statuses for volume >>>>>>>>>>>>>> 'volume2' of >>>>>>>>>>>>>> cluster 'C6220': null >>>>>>>>>>>>>> 2018-05-10 03:24:19,097+02 INFO >>>>>>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>>>>>>>>>> (DefaultQuartzScheduler8) >>>>>>>>>>>>>> [7715ceda] Failed to acquire lock and wait lock >>>>>>>>>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>>>>>>>>>>>> sharedLocks=''}' >>>>>>>>>>>>>> 2018-05-10 03:24:19,104+02 INFO >>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] START, >>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>> n4.itsmart.cloud, >>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'}), >>>>>>>>>>>>>> log id: 6908121d >>>>>>>>>>>>>> 2018-05-10 03:24:19,106+02 ERROR >>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] Command >>>>>>>>>>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>> n4.itsmart.cloud, >>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'})' >>>>>>>>>>>>>> execution failed: null >>>>>>>>>>>>>> 2018-05-10 03:24:19,106+02 INFO >>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, >>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 6908121d >>>>>>>>>>>>>> 2018-05-10 03:24:19,107+02 INFO >>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] START, >>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>> n1.itsmart.cloud, >>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'}), >>>>>>>>>>>>>> log id: 735c6a5f >>>>>>>>>>>>>> 2018-05-10 03:24:19,109+02 ERROR >>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] Command >>>>>>>>>>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>> n1.itsmart.cloud, >>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'})' >>>>>>>>>>>>>> execution failed: null >>>>>>>>>>>>>> 2018-05-10 03:24:19,109+02 INFO >>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, >>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 735c6a5f >>>>>>>>>>>>>> 2018-05-10 03:24:19,110+02 INFO >>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] START, >>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>> n2.itsmart.cloud, >>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), >>>>>>>>>>>>>> log id: 6f9e9f58 >>>>>>>>>>>>>> 2018-05-10 03:24:19,112+02 ERROR >>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] Command >>>>>>>>>>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>> n2.itsmart.cloud, >>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'})' >>>>>>>>>>>>>> execution failed: null >>>>>>>>>>>>>> 2018-05-10 03:24:19,112+02 INFO >>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, >>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 6f9e9f58 >>>>>>>>>>>>>> 2018-05-10 03:24:19,113+02 INFO >>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] START, >>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>> n3.itsmart.cloud, >>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'}), >>>>>>>>>>>>>> log id: 2ee46967 >>>>>>>>>>>>>> 2018-05-10 03:24:19,115+02 ERROR >>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] Command >>>>>>>>>>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>> n3.itsmart.cloud, >>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'})' >>>>>>>>>>>>>> execution failed: null >>>>>>>>>>>>>> 2018-05-10 03:24:19,116+02 INFO >>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, >>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 2ee46967 >>>>>>>>>>>>>> 2018-05-10 03:24:19,117+02 INFO >>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] >>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] START, >>>>>>>>>>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand(HostName = >>>>>>>>>>>>>> n1.itsmart.cloud, >>>>>>>>>>>>>> GlusterVolumeAdvancedDetailsVDSParameters:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57', >>>>>>>>>>>>>> volumeName='volume1'}), log id: 7550e5c >>>>>>>>>>>>>> 2018-05-10 03:24:20,748+02 INFO >>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] >>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, >>>>>>>>>>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand, return: >>>>>>>>>>>>>> org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeAdvancedDetails@4a46066f, >>>>>>>>>>>>>> log id: 7550e5c >>>>>>>>>>>>>> 2018-05-10 03:24:20,749+02 ERROR >>>>>>>>>>>>>> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob] >>>>>>>>>>>>>> (DefaultQuartzScheduler7) >>>>>>>>>>>>>> [43f4eaec] Error while refreshing brick statuses for volume >>>>>>>>>>>>>> 'volume1' of >>>>>>>>>>>>>> cluster 'C6220': null >>>>>>>>>>>>>> 2018-05-10 03:24:20,750+02 INFO >>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] >>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] START, >>>>>>>>>>>>>> GlusterServersListVDSCommand(HostName = n2.itsmart.cloud, >>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), >>>>>>>>>>>>>> log id: 120cc68d >>>>>>>>>>>>>> 2018-05-10 03:24:20,930+02 INFO >>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] >>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] FINISH, >>>>>>>>>>>>>> GlusterServersListVDSCommand, >>>>>>>>>>>>>> return: [ [ http://10.101.0.2/24:CONNECTED | >>>>>>>>>>>>>> 10.101.0.2/24:CONNECTED ] , >>>>>>>>>>>>>> n1.cloudata.local:CONNECTED, 10.104.0.3:CONNECTED, >>>>>>>>>>>>>> 10.104.0.4:CONNECTED], log >>>>>>>>>>>>>> id: 120cc68d >>>>>>>>>>>>>> 2018-05-10 03:24:20,949+02 INFO >>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] >>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] START, >>>>>>>>>>>>>> GlusterVolumesListVDSCommand(HostName = n2.itsmart.cloud, >>>>>>>>>>>>>> GlusterVolumesListVDSParameters:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), >>>>>>>>>>>>>> log id: 118aa264 >>>>>>>>>>>>>> 2018-05-10 03:24:21,048+02 WARN >>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick >>>>>>>>>>>>>> '10.104.0.1:/gluster/brick/brick1' of volume >>>>>>>>>>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as >>>>>>>>>>>>>> no gluster >>>>>>>>>>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339' >>>>>>>>>>>>>> 2018-05-10 03:24:21,055+02 WARN >>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick >>>>>>>>>>>>>> '10.104.0.1:/gluster/brick/brick2' of volume >>>>>>>>>>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as >>>>>>>>>>>>>> no gluster >>>>>>>>>>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339' >>>>>>>>>>>>>> 2018-05-10 03:24:21,061+02 WARN >>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick >>>>>>>>>>>>>> '10.104.0.1:/gluster/brick/brick3' of volume >>>>>>>>>>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as >>>>>>>>>>>>>> no gluster >>>>>>>>>>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339' >>>>>>>>>>>>>> 2018-05-10 03:24:21,067+02 WARN >>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick >>>>>>>>>>>>>> '10.104.0.1:/gluster2/brick/brick1' of volume >>>>>>>>>>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as >>>>>>>>>>>>>> no gluster >>>>>>>>>>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339' >>>>>>>>>>>>>> 2018-05-10 03:24:21,074+02 WARN >>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick >>>>>>>>>>>>>> '10.104.0.1:/gluster2/brick/brick2' of volume >>>>>>>>>>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as >>>>>>>>>>>>>> no gluster >>>>>>>>>>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339' >>>>>>>>>>>>>> 2018-05-10 03:24:21,080+02 WARN >>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick >>>>>>>>>>>>>> '10.104.0.1:/gluster2/brick/brick3' of volume >>>>>>>>>>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as >>>>>>>>>>>>>> no gluster >>>>>>>>>>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339' >>>>>>>>>>>>>> 2018-05-10 03:24:21,081+02 INFO >>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] >>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] FINISH, >>>>>>>>>>>>>> GlusterVolumesListVDSCommand, >>>>>>>>>>>>>> return: >>>>>>>>>>>>>> {68cfb061-1320-4042-abcd-9228da23c0c8=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7a6720d, >>>>>>>>>>>>>> e0f568fa-987c-4f5c-b853-01bce718ee27=org.ovirt.engine.core.common.businessentities.g >>>>>>>>>>>>>> luster.GlusterVolumeEntity@f88c521b}, log id: 118aa264 >>>>>>>>>>>>>> 2018-05-10 11:59:26,047+02 ERROR >>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] Command >>>>>>>>>>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>> n1.itsmart.cloud, >>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'})' >>>>>>>>>>>>>> execution failed: null >>>>>>>>>>>>>> 2018-05-10 11:59:26,047+02 INFO >>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH, >>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 14a71ef0 >>>>>>>>>>>>>> 2018-05-10 11:59:26,048+02 INFO >>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] START, >>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>> n4.itsmart.cloud, >>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'}), >>>>>>>>>>>>>> log id: 28d9e255 >>>>>>>>>>>>>> 2018-05-10 11:59:26,051+02 ERROR >>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] Command >>>>>>>>>>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>> n4.itsmart.cloud, >>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'})' >>>>>>>>>>>>>> execution failed: null >>>>>>>>>>>>>> 2018-05-10 11:59:26,051+02 INFO >>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH, >>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 28d9e255 >>>>>>>>>>>>>> 2018-05-10 11:59:26,052+02 INFO >>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] START, >>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>> n2.itsmart.cloud, >>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), >>>>>>>>>>>>>> log id: 4a7b280e >>>>>>>>>>>>>> 2018-05-10 11:59:26,054+02 ERROR >>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] Command >>>>>>>>>>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>> n2.itsmart.cloud, >>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'})' >>>>>>>>>>>>>> execution failed: null >>>>>>>>>>>>>> 2018-05-10 11:59:26,054+02 INFO >>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH, >>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 4a7b280e >>>>>>>>>>>>>> 2018-05-10 11:59:26,055+02 INFO >>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] START, >>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>> n3.itsmart.cloud, >>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'}), >>>>>>>>>>>>>> log id: 18adc534 >>>>>>>>>>>>>> 2018-05-10 11:59:26,057+02 ERROR >>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] Command >>>>>>>>>>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>> n3.itsmart.cloud, >>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'})' >>>>>>>>>>>>>> execution failed: null >>>>>>>>>>>>>> 2018-05-10 11:59:26,057+02 INFO >>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH, >>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 18adc534 >>>>>>>>>>>>>> 2018-05-10 11:59:26,058+02 INFO >>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] >>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] START, >>>>>>>>>>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand(HostName = >>>>>>>>>>>>>> n3.itsmart.cloud, >>>>>>>>>>>>>> GlusterVolumeAdvancedDetailsVDSParameters:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec', >>>>>>>>>>>>>> volumeName='volume1'}), log id: 3451084f >>>>>>>>>>>>>> 2018-05-10 11:59:28,050+02 INFO >>>>>>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>>>>>>>>>> (DefaultQuartzScheduler3) >>>>>>>>>>>>>> [2eb1c389] Failed to acquire lock and wait lock >>>>>>>>>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>>>>>>>>>>>> sharedLocks=''}' >>>>>>>>>>>>>> 2018-05-10 11:59:28,060+02 INFO >>>>>>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>>>>>>>>>> (DefaultQuartzScheduler3) >>>>>>>>>>>>>> [2eb1c389] Failed to acquire lock and wait lock >>>>>>>>>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>>>>>>>>>>>> sharedLocks=''}' >>>>>>>>>>>>>> 2018-05-10 11:59:28,062+02 INFO >>>>>>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>>>>>>>>>> (DefaultQuartzScheduler3) >>>>>>>>>>>>>> [2eb1c389] Failed to acquire lock and wait lock >>>>>>>>>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>>>>>>>>>>>> sharedLocks=''}' >>>>>>>>>>>>>> 2018-05-10 11:59:31,054+02 INFO >>>>>>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>>>>>>>>>> (DefaultQuartzScheduler3) >>>>>>>>>>>>>> [2eb1c389] Failed to acquire lock and wait lock >>>>>>>>>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>>>>>>>>>>>> sharedLocks=''}' >>>>>>>>>>>>>> 2018-05-10 11:59:31,054+02 INFO >>>>>>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>>>>>>>>>> (DefaultQuartzScheduler3) >>>>>>>>>>>>>> [2eb1c389] Failed to acquire lock and wait lock >>>>>>>>>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>>>>>>>>>>>> sharedLocks=''}' >>>>>>>>>>>>>> 2018-05-10 11:59:31,062+02 INFO >>>>>>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>>>>>>>>>> (DefaultQuartzScheduler3) >>>>>>>>>>>>>> [2eb1c389] Failed to acquire lock and wait lock >>>>>>>>>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>>>>>>>>>>>> sharedLocks=''}' >>>>>>>>>>>>>> 2018-05-10 11:59:31,064+02 INFO >>>>>>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>>>>>>>>>> (DefaultQuartzScheduler3) >>>>>>>>>>>>>> [2eb1c389] Failed to acquire lock and wait lock >>>>>>>>>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>>>>>>>>>>>> sharedLocks=''}' >>>>>>>>>>>>>> 2018-05-10 11:59:31,465+02 INFO >>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] >>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH, >>>>>>>>>>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand, return: >>>>>>>>>>>>>> org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeAdvancedDetails@3f1b7f43, >>>>>>>>>>>>>> log id: 3451084f >>>>>>>>>>>>>> 2018-05-10 11:59:31,466+02 ERROR >>>>>>>>>>>>>> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob] >>>>>>>>>>>>>> (DefaultQuartzScheduler4) >>>>>>>>>>>>>> [400fa486] Error while refreshing brick statuses for volume >>>>>>>>>>>>>> 'volume1' of >>>>>>>>>>>>>> cluster 'C6220': null >>>>>>>>>>>>>> R >>>>>>>>>>>>>> Tibor >>>>>>>>>>>>>> ----- 2018. máj.. 10., 11:43, Sahina Bose < [ >>>>>>>>>>>>>> mailto:sab...@redhat.com | >>>>>>>>>>>>>> sab...@redhat.com ] > írta: >>>>>>>>>>>>>>> This doesn't affect the monitoring of state. >>>>>>>>>>>>>>> Any errors in vdsm.log? >>>>>>>>>>>>>>> Or errors in engine.log of the form "Error while refreshing >>>>>>>>>>>>>>> brick statuses for >>>>>>>>>>>>>>> volume" >>>>>>>>>>>>>>> On Thu, May 10, 2018 at 2:33 PM, Demeter Tibor < [ >>>>>>>>>>>>>>> mailto:tdeme...@itsmart.hu | >>>>>>>>>>>>>>> tdeme...@itsmart.hu ] > wrote: >>>>>>>>>>>>>>>> Hi, >>>>>>>>>>>>>>>> Thank you for your fast reply :) >>>>>>>>>>>>>>>> 2018-05-10 11:01:51,574+02 INFO >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] START, >>>>>>>>>>>>>>>> GlusterServersListVDSCommand(HostName = n2.itsmart.cloud, >>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), >>>>>>>>>>>>>>>> log id: 39adbbb8 >>>>>>>>>>>>>>>> 2018-05-10 11:01:51,768+02 INFO >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] FINISH, >>>>>>>>>>>>>>>> GlusterServersListVDSCommand, >>>>>>>>>>>>>>>> return: [ [ http://10.101.0.2/24:CONNECTED | >>>>>>>>>>>>>>>> 10.101.0.2/24:CONNECTED ] , >>>>>>>>>>>>>>>> n1.cloudata.local:CONNECTED, 10.104.0.3:CONNECTED, >>>>>>>>>>>>>>>> 10.104.0.4:CONNECTED], log >>>>>>>>>>>>>>>> id: 39adbbb8 >>>>>>>>>>>>>>>> 2018-05-10 11:01:51,788+02 INFO >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] START, >>>>>>>>>>>>>>>> GlusterVolumesListVDSCommand(HostName = n2.itsmart.cloud, >>>>>>>>>>>>>>>> GlusterVolumesListVDSParameters:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), >>>>>>>>>>>>>>>> log id: 738a7261 >>>>>>>>>>>>>>>> 2018-05-10 11:01:51,892+02 WARN >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick >>>>>>>>>>>>>>>> '10.104.0.1:/gluster/brick/brick1' of volume >>>>>>>>>>>>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as >>>>>>>>>>>>>>>> no gluster >>>>>>>>>>>>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339' >>>>>>>>>>>>>>>> 2018-05-10 11:01:51,898+02 WARN >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick >>>>>>>>>>>>>>>> '10.104.0.1:/gluster/brick/brick2' of volume >>>>>>>>>>>>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as >>>>>>>>>>>>>>>> no gluster >>>>>>>>>>>>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339' >>>>>>>>>>>>>>>> 2018-05-10 11:01:51,905+02 WARN >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick >>>>>>>>>>>>>>>> '10.104.0.1:/gluster/brick/brick3' of volume >>>>>>>>>>>>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as >>>>>>>>>>>>>>>> no gluster >>>>>>>>>>>>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339' >>>>>>>>>>>>>>>> 2018-05-10 11:01:51,911+02 WARN >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick >>>>>>>>>>>>>>>> '10.104.0.1:/gluster2/brick/brick1' of volume >>>>>>>>>>>>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as >>>>>>>>>>>>>>>> no gluster >>>>>>>>>>>>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339' >>>>>>>>>>>>>>>> 2018-05-10 11:01:51,917+02 WARN >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick >>>>>>>>>>>>>>>> '10.104.0.1:/gluster2/brick/brick2' of volume >>>>>>>>>>>>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as >>>>>>>>>>>>>>>> no gluster >>>>>>>>>>>>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339' >>>>>>>>>>>>>>>> 2018-05-10 11:01:51,924+02 WARN >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick >>>>>>>>>>>>>>>> '10.104.0.1:/gluster2/brick/brick3' of volume >>>>>>>>>>>>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as >>>>>>>>>>>>>>>> no gluster >>>>>>>>>>>>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339' >>>>>>>>>>>>>>>> 2018-05-10 11:01:51,925+02 INFO >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] FINISH, >>>>>>>>>>>>>>>> GlusterVolumesListVDSCommand, >>>>>>>>>>>>>>>> return: >>>>>>>>>>>>>>>> {68cfb061-1320-4042-abcd-9228da23c0c8=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7a6720d, >>>>>>>>>>>>>>>> e0f568fa-987c-4f5c-b853-01bce718ee27=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@f88c521b}, >>>>>>>>>>>>>>>> log id: 738a7261 >>>>>>>>>>>>>>>> This happening continuously. >>>>>>>>>>>>>>>> Thanks! >>>>>>>>>>>>>>>> Tibor >>>>>>>>>>>>>>>> ----- 2018. máj.. 10., 10:56, Sahina Bose < [ >>>>>>>>>>>>>>>> mailto:sab...@redhat.com | >>>>>>>>>>>>>>>> sab...@redhat.com ] > írta: >>>>>>>>>>>>>>>>> Could you check the engine.log if there are errors related to >>>>>>>>>>>>>>>>> getting >>>>>>>>>>>>>>>>> GlusterVolumeAdvancedDetails ? >>>>>>>>>>>>>>>>> On Thu, May 10, 2018 at 2:02 PM, Demeter Tibor < [ >>>>>>>>>>>>>>>>> mailto:tdeme...@itsmart.hu | >>>>>>>>>>>>>>>>> tdeme...@itsmart.hu ] > wrote: >>>>>>>>>>>>>>>>>> Dear Ovirt Users, >>>>>>>>>>>>>>>>>> I've followed up the self-hosted-engine upgrade >>>>>>>>>>>>>>>>>> documentation, I upgraded my 4.1 >>>>>>>>>>>>>>>>>> system to 4.2.3. >>>>>>>>>>>>>>>>>> I upgaded the first node with yum upgrade, it seems working >>>>>>>>>>>>>>>>>> now fine. But since >>>>>>>>>>>>>>>>>> upgrade, the gluster informations seems to displayed >>>>>>>>>>>>>>>>>> incorrect on the admin >>>>>>>>>>>>>>>>>> panel. The volume yellow, and there are red bricks from that >>>>>>>>>>>>>>>>>> node. >>>>>>>>>>>>>>>>>> I've checked in console, I think my gluster is not degraded: >>>>>>>>>>>>>>>>>> root@n1 ~]# gluster volume list >>>>>>>>>>>>>>>>>> volume1 >>>>>>>>>>>>>>>>>> volume2 >>>>>>>>>>>>>>>>>> [root@n1 ~]# gluster volume info >>>>>>>>>>>>>>>>>> Volume Name: volume1 >>>>>>>>>>>>>>>>>> Type: Distributed-Replicate >>>>>>>>>>>>>>>>>> Volume ID: e0f568fa-987c-4f5c-b853-01bce718ee27 >>>>>>>>>>>>>>>>>> Status: Started >>>>>>>>>>>>>>>>>> Snapshot Count: 0 >>>>>>>>>>>>>>>>>> Number of Bricks: 3 x 3 = 9 >>>>>>>>>>>>>>>>>> Transport-type: tcp >>>>>>>>>>>>>>>>>> Bricks: >>>>>>>>>>>>>>>>>> Brick1: 10.104.0.1:/gluster/brick/brick1 >>>>>>>>>>>>>>>>>> Brick2: 10.104.0.2:/gluster/brick/brick1 >>>>>>>>>>>>>>>>>> Brick3: 10.104.0.3:/gluster/brick/brick1 >>>>>>>>>>>>>>>>>> Brick4: 10.104.0.1:/gluster/brick/brick2 >>>>>>>>>>>>>>>>>> Brick5: 10.104.0.2:/gluster/brick/brick2 >>>>>>>>>>>>>>>>>> Brick6: 10.104.0.3:/gluster/brick/brick2 >>>>>>>>>>>>>>>>>> Brick7: 10.104.0.1:/gluster/brick/brick3 >>>>>>>>>>>>>>>>>> Brick8: 10.104.0.2:/gluster/brick/brick3 >>>>>>>>>>>>>>>>>> Brick9: 10.104.0.3:/gluster/brick/brick3 >>>>>>>>>>>>>>>>>> Options Reconfigured: >>>>>>>>>>>>>>>>>> transport.address-family: inet >>>>>>>>>>>>>>>>>> performance.readdir-ahead: on >>>>>>>>>>>>>>>>>> nfs.disable: on >>>>>>>>>>>>>>>>>> storage.owner-uid: 36 >>>>>>>>>>>>>>>>>> storage.owner-gid: 36 >>>>>>>>>>>>>>>>>> performance.quick-read: off >>>>>>>>>>>>>>>>>> performance.read-ahead: off >>>>>>>>>>>>>>>>>> performance.io-cache: off >>>>>>>>>>>>>>>>>> performance.stat-prefetch: off >>>>>>>>>>>>>>>>>> performance.low-prio-threads: 32 >>>>>>>>>>>>>>>>>> network.remote-dio: enable >>>>>>>>>>>>>>>>>> cluster.eager-lock: enable >>>>>>>>>>>>>>>>>> cluster.quorum-type: auto >>>>>>>>>>>>>>>>>> cluster.server-quorum-type: server >>>>>>>>>>>>>>>>>> cluster.data-self-heal-algorithm: full >>>>>>>>>>>>>>>>>> cluster.locking-scheme: granular >>>>>>>>>>>>>>>>>> cluster.shd-max-threads: 8 >>>>>>>>>>>>>>>>>> cluster.shd-wait-qlength: 10000 >>>>>>>>>>>>>>>>>> features.shard: on >>>>>>>>>>>>>>>>>> user.cifs: off >>>>>>>>>>>>>>>>>> server.allow-insecure: on >>>>>>>>>>>>>>>>>> Volume Name: volume2 >>>>>>>>>>>>>>>>>> Type: Distributed-Replicate >>>>>>>>>>>>>>>>>> Volume ID: 68cfb061-1320-4042-abcd-9228da23c0c8 >>>>>>>>>>>>>>>>>> Status: Started >>>>>>>>>>>>>>>>>> Snapshot Count: 0 >>>>>>>>>>>>>>>>>> Number of Bricks: 3 x 3 = 9 >>>>>>>>>>>>>>>>>> Transport-type: tcp >>>>>>>>>>>>>>>>>> Bricks: >>>>>>>>>>>>>>>>>> Brick1: 10.104.0.1:/gluster2/brick/brick1 >>>>>>>>>>>>>>>>>> Brick2: 10.104.0.2:/gluster2/brick/brick1 >>>>>>>>>>>>>>>>>> Brick3: 10.104.0.3:/gluster2/brick/brick1 >>>>>>>>>>>>>>>>>> Brick4: 10.104.0.1:/gluster2/brick/brick2 >>>>>>>>>>>>>>>>>> Brick5: 10.104.0.2:/gluster2/brick/brick2 >>>>>>>>>>>>>>>>>> Brick6: 10.104.0.3:/gluster2/brick/brick2 >>>>>>>>>>>>>>>>>> Brick7: 10.104.0.1:/gluster2/brick/brick3 >>>>>>>>>>>>>>>>>> Brick8: 10.104.0.2:/gluster2/brick/brick3 >>>>>>>>>>>>>>>>>> Brick9: 10.104.0.3:/gluster2/brick/brick3 >>>>>>>>>>>>>>>>>> Options Reconfigured: >>>>>>>>>>>>>>>>>> nfs.disable: on >>>>>>>>>>>>>>>>>> performance.readdir-ahead: on >>>>>>>>>>>>>>>>>> transport.address-family: inet >>>>>>>>>>>>>>>>>> cluster.quorum-type: auto >>>>>>>>>>>>>>>>>> network.ping-timeout: 10 >>>>>>>>>>>>>>>>>> auth.allow: * >>>>>>>>>>>>>>>>>> performance.quick-read: off >>>>>>>>>>>>>>>>>> performance.read-ahead: off >>>>>>>>>>>>>>>>>> performance.io-cache: off >>>>>>>>>>>>>>>>>> performance.stat-prefetch: off >>>>>>>>>>>>>>>>>> performance.low-prio-threads: 32 >>>>>>>>>>>>>>>>>> network.remote-dio: enable >>>>>>>>>>>>>>>>>> cluster.eager-lock: enable >>>>>>>>>>>>>>>>>> cluster.server-quorum-type: server >>>>>>>>>>>>>>>>>> cluster.data-self-heal-algorithm: full >>>>>>>>>>>>>>>>>> cluster.locking-scheme: granular >>>>>>>>>>>>>>>>>> cluster.shd-max-threads: 8 >>>>>>>>>>>>>>>>>> cluster.shd-wait-qlength: 10000 >>>>>>>>>>>>>>>>>> features.shard: on >>>>>>>>>>>>>>>>>> user.cifs: off >>>>>>>>>>>>>>>>>> storage.owner-uid: 36 >>>>>>>>>>>>>>>>>> storage.owner-gid: 36 >>>>>>>>>>>>>>>>>> server.allow-insecure: on >>>>>>>>>>>>>>>>>> [root@n1 ~]# gluster volume status >>>>>>>>>>>>>>>>>> Status of volume: volume1 >>>>>>>>>>>>>>>>>> Gluster process TCP Port RDMA Port Online Pid >>>>>>>>>>>>>>>>>> ------------------------------------------------------------------------------ >>>>>>>>>>>>>>>>>> Brick 10.104.0.1:/gluster/brick/brick1 49152 0 Y 3464 >>>>>>>>>>>>>>>>>> Brick 10.104.0.2:/gluster/brick/brick1 49152 0 Y 68937 >>>>>>>>>>>>>>>>>> Brick 10.104.0.3:/gluster/brick/brick1 49161 0 Y 94506 >>>>>>>>>>>>>>>>>> Brick 10.104.0.1:/gluster/brick/brick2 49153 0 Y 3457 >>>>>>>>>>>>>>>>>> Brick 10.104.0.2:/gluster/brick/brick2 49153 0 Y 68943 >>>>>>>>>>>>>>>>>> Brick 10.104.0.3:/gluster/brick/brick2 49162 0 Y 94514 >>>>>>>>>>>>>>>>>> Brick 10.104.0.1:/gluster/brick/brick3 49154 0 Y 3465 >>>>>>>>>>>>>>>>>> Brick 10.104.0.2:/gluster/brick/brick3 49154 0 Y 68949 >>>>>>>>>>>>>>>>>> Brick 10.104.0.3:/gluster/brick/brick3 49163 0 Y 94520 >>>>>>>>>>>>>>>>>> Self-heal Daemon on localhost N/A N/A Y 54356 >>>>>>>>>>>>>>>>>> Self-heal Daemon on 10.104.0.2 N/A N/A Y 962 >>>>>>>>>>>>>>>>>> Self-heal Daemon on 10.104.0.3 N/A N/A Y 108977 >>>>>>>>>>>>>>>>>> Self-heal Daemon on 10.104.0.4 N/A N/A Y 61603 >>>>>>>>>>>>>>>>>> Task Status of Volume volume1 >>>>>>>>>>>>>>>>>> ------------------------------------------------------------------------------ >>>>>>>>>>>>>>>>>> There are no active volume tasks >>>>>>>>>>>>>>>>>> Status of volume: volume2 >>>>>>>>>>>>>>>>>> Gluster process TCP Port RDMA Port Online Pid >>>>>>>>>>>>>>>>>> ------------------------------------------------------------------------------ >>>>>>>>>>>>>>>>>> Brick 10.104.0.1:/gluster2/brick/brick1 49155 0 Y 3852 >>>>>>>>>>>>>>>>>> Brick 10.104.0.2:/gluster2/brick/brick1 49158 0 Y 68955 >>>>>>>>>>>>>>>>>> Brick 10.104.0.3:/gluster2/brick/brick1 49164 0 Y 94527 >>>>>>>>>>>>>>>>>> Brick 10.104.0.1:/gluster2/brick/brick2 49156 0 Y 3851 >>>>>>>>>>>>>>>>>> Brick 10.104.0.2:/gluster2/brick/brick2 49159 0 Y 68961 >>>>>>>>>>>>>>>>>> Brick 10.104.0.3:/gluster2/brick/brick2 49165 0 Y 94533 >>>>>>>>>>>>>>>>>> Brick 10.104.0.1:/gluster2/brick/brick3 49157 0 Y 3883 >>>>>>>>>>>>>>>>>> Brick 10.104.0.2:/gluster2/brick/brick3 49160 0 Y 68968 >>>>>>>>>>>>>>>>>> Brick 10.104.0.3:/gluster2/brick/brick3 49166 0 Y 94541 >>>>>>>>>>>>>>>>>> Self-heal Daemon on localhost N/A N/A Y 54356 >>>>>>>>>>>>>>>>>> Self-heal Daemon on 10.104.0.2 N/A N/A Y 962 >>>>>>>>>>>>>>>>>> Self-heal Daemon on 10.104.0.3 N/A N/A Y 108977 >>>>>>>>>>>>>>>>>> Self-heal Daemon on 10.104.0.4 N/A N/A Y 61603 >>>>>>>>>>>>>>>>>> Task Status of Volume volume2 >>>>>>>>>>>>>>>>>> ------------------------------------------------------------------------------ >>>>>>>>>>>>>>>>>> There are no active volume tasks >>>>>>>>>>>>>>>>>> I think ovirt can't read valid informations about gluster. >>>>>>>>>>>>>>>>>> I can't contiune upgrade of other hosts until this problem >>>>>>>>>>>>>>>>>> exist. >>>>>>>>>>>>>>>>>> Please help me:) >>>>>>>>>>>>>>>>>> Thanks >>>>>>>>>>>>>>>>>> Regards, >>>>>>>>>>>>>>>>>> Tibor >>>>>>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>>>>>> Users mailing list -- [ mailto:users@ovirt.org | >>>>>>>>>>>>>>>>>> users@ovirt.org ] >>>>>>>>>>>>>>>>>> To unsubscribe send an email to [ >>>>>>>>>>>>>>>>>> mailto:users-le...@ovirt.org | >>>>>>>>>>>>>>>>>> users-le...@ovirt.org ] >>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>> Users mailing list -- [ mailto:users@ovirt.org | users@ovirt.org ] >>>>>>>>>>>> To unsubscribe send an email to [ mailto:users-le...@ovirt.org | >>>>>>>>>>>> users-le...@ovirt.org ] >>>>>>>>> _______________________________________________ >>>>>>>>> Users mailing list -- [ mailto:users@ovirt.org | users@ovirt.org ] >>>>>>>>> To unsubscribe send an email to [ mailto:users-le...@ovirt.org | >>>>>>>>> users-le...@ovirt.org ] >>>>>>>> _______________________________________________ >>>>>>>> Users mailing list -- [ mailto:users@ovirt.org | users@ovirt.org ] >>>>>>>> To unsubscribe send an email to [ mailto:users-le...@ovirt.org | >>>>>>>> users-le...@ovirt.org ] >>>>>>>> oVirt Code of Conduct: [ >>>>>>>> https://www.ovirt.org/community/about/community-guidelines/ | >>>>>>>> https://www.ovirt.org/community/about/community-guidelines/ ] >>>>>>>> List Archives:
_______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org