Hi Adrian,
 Looks like issue is with lvm (
https://bugzilla.redhat.com/show_bug.cgi?id=502671  and
https://bugzilla.redhat.com/show_bug.cgi?id=1664461 )
And it's fixed in lvm2-2.02.46-3.el5
Can you please check your lvm2 version?

On Mon, Mar 11, 2019 at 1:57 PM Sahina Bose <sab...@redhat.com> wrote:

> +Gobinda Das +Dhanjal Parth
>
> On Mon, Mar 11, 2019 at 1:42 AM <adrianquint...@gmail.com> wrote:
> >
> > Hello I am trying to run a Hyperconverged setup "COnfigure gluster
> storage and ovirt hosted engine", however  I get the following error
> >
> >
> __________________________________________________________________________________________________
> >  PLAY [gluster_servers]
> *********************************************************
> >
> > TASK [Create LVs with specified size for the VGs]
> ******************************
> > failed: [ovirt01.grupokino.com] (item={u'lv': u'gluster_thinpool_sdb',
> u'size': u'45GB', u'extent': u'100%FREE', u'vg': u'gluster_vg_sdb'}) =>
> {"changed": false, "item": {"extent": "100%FREE", "lv":
> "gluster_thinpool_sdb", "size": "45GB", "vg": "gluster_vg_sdb"}, "msg":
> "lvcreate: metadata/pv_map.c:198: consume_pv_area: Assertion `to_go <=
> pva->count' failed.\n", "rc": -6}
> >         to retry, use: --limit @/tmp/tmpwo4SNB/lvcreate.retry
> >
> > PLAY RECAP
> *********************************************************************
> > ovirt01.grupokino.com      : ok=0    changed=0    unreachable=0
> failed=1
> >
> __________________________________________________________________________________________________
> >
> > I know that oVirt Hosted Engine Setup GUI for "gluster wizard (gluster
> deployment) does not populate the geodeployConfig.conf file properly
> (Generated Gdeploy configuration :
> /var/lib/ovirt-hosted-engine-setup/gdeploy/gdeployConfig.conf) so I have
> tried to modify it to fit our needs but keep getting the above error
> everytime.
> >
> > Any ideas or comments are welcome...Thanks!
> >
> >
> >
> >
> > My servers are setup with 4x50GB disks, 1 for the OS and the rest for
> Gluster Hyperconverged setup.
> >
> __________________________________________________________________________________________________
> > my gdeployConfig.conf file:
> >
> __________________________________________________________________________________________________
> > #gdeploy configuration generated by cockpit-gluster plugin
> > [hosts]
> > ovirt01.mydomain.com
> > ovirt02.mydomain.com
> > ovirt03.mydomain.com
> >
> > [script1:ovirt01.mydomain.com]
> > action=execute
> > ignore_script_errors=no
> > file=/usr/share/gdeploy/scripts/grafton-sanity-check.sh -d sdb,sdc,sdd
> -h ovirt01.mydomain.com, ovirt02.mydomain.com, ovirt03.mydomain.com
> >
> > [script1:ovirt02.mydomain.com]
> > action=execute
> > ignore_script_errors=no
> > file=/usr/share/gdeploy/scripts/grafton-sanity-check.sh -d sdb,sdc,sdd
> -h ovirt01.mydomain.com, ovirt02.mydomain.com, ovirt03.mydomain.com
> >
> > [script1:ovirt03.mydomain.com]
> > action=execute
> > ignore_script_errors=no
> > file=/usr/share/gdeploy/scripts/grafton-sanity-check.sh -d sdb,sdc,sdd
> -h ovirt01.mydomain.com, ovirt02.mydomain.com, ovirt03.mydomain.com
> >
> > [disktype]
> > jbod
> >
> > [diskcount]
> > 3
> >
> > [stripesize]
> > 256
> >
> > [service1]
> > action=enable
> > service=chronyd
> >
> > [service2]
> > action=restart
> > service=chronyd
> >
> > [shell2]
> > action=execute
> > command=vdsm-tool configure --force
> >
> > [script3]
> > action=execute
> > file=/usr/share/gdeploy/scripts/blacklist_all_disks.sh
> > ignore_script_errors=no
> >
> > [pv1:ovirt01.mydomain.com]
> > action=create
> > devices=sdb
> > ignore_pv_errors=no
> >
> > [pv1:ovirt02.mydomain.com]
> > action=create
> > devices=sdb
> > ignore_pv_errors=no
> >
> > [pv1:ovirt03.mydomain.com]
> > action=create
> > devices=sdb
> > ignore_pv_errors=no
> >
> > [pv2:ovirt01.mydomain.com]
> > action=create
> > devices=sdc
> > ignore_pv_errors=no
> >
> > [pv2:ovirt02.mydomain.com]
> > action=create
> > devices=sdc
> > ignore_pv_errors=no
> >
> > [pv2:ovirt03.mydomain.com]
> > action=create
> > devices=sdc
> > ignore_pv_errors=no
> >
> > [pv3:ovirt01.mydomain.com]
> > action=create
> > devices=sdd
> > ignore_pv_errors=no
> >
> > [pv3:ovirt02.mydomain.com]
> > action=create
> > devices=sdd
> > ignore_pv_errors=no
> >
> > [pv3:ovirt03.mydomain.com]
> > action=create
> > devices=sdd
> > ignore_pv_errors=no
> >
> > [vg1:ovirt01.mydomain.com]
> > action=create
> > vgname=gluster_vg_sdb
> > pvname=sdb
> > ignore_vg_errors=no
> >
> > [vg1:ovirt02.mydomain.com]
> > action=create
> > vgname=gluster_vg_sdb
> > pvname=sdb
> > ignore_vg_errors=no
> >
> > [vg1:ovirt03.mydomain.com]
> > action=create
> > vgname=gluster_vg_sdb
> > pvname=sdb
> > ignore_vg_errors=no
> >
> > [vg2:ovirt01.mydomain.com]
> > action=create
> > vgname=gluster_vg_sdc
> > pvname=sdc
> > ignore_vg_errors=no
> >
> > [vg2:ovirt02.mydomain.com]
> > action=create
> > vgname=gluster_vg_sdc
> > pvname=sdc
> > ignore_vg_errors=no
> >
> > [vg2:ovirt03.mydomain.com]
> > action=create
> > vgname=gluster_vg_sdc
> > pvname=sdc
> > ignore_vg_errors=no
> >
> > [vg3:ovirt01.mydomain.com]
> > action=create
> > vgname=gluster_vg_sdd
> > pvname=sdd
> > ignore_vg_errors=no
> >
> > [vg3:ovirt02.mydomain.com]
> > action=create
> > vgname=gluster_vg_sdd
> > pvname=sdd
> > ignore_vg_errors=no
> >
> > [vg3:ovirt03.mydomain.com]
> > action=create
> > vgname=gluster_vg_sdd
> > pvname=sdd
> > ignore_vg_errors=no
> >
> > [lv1:ovirt01.mydomain.com]
> > action=create
> > poolname=gluster_thinpool_sdb
> > ignore_lv_errors=no
> > vgname=gluster_vg_sdb
> > lvtype=thinpool
> > size=45GB
> > poolmetadatasize=3GB
> >
> > [lv2:ovirt02.mydomain.com]
> > action=create
> > poolname=gluster_thinpool_sdc
> > ignore_lv_errors=no
> > vgname=gluster_vg_sdc
> > lvtype=thinpool
> > size=45GB
> > poolmetadatasize=3GB
> >
> > [lv3:ovirt03.mydomain.com]
> > action=create
> > poolname=gluster_thinpool_sdd
> > ignore_lv_errors=no
> > vgname=gluster_vg_sdd
> > lvtype=thinpool
> > size=45GB
> > poolmetadatasize=3GB
> >
> > [lv4:ovirt01.mydomain.com]
> > action=create
> > lvname=gluster_lv_engine
> > ignore_lv_errors=no
> > vgname=gluster_vg_sdb
> > mount=/gluster_bricks/engine
> > size=42GB
> > lvtype=thick
> >
> > [lv5:ovirt01.mydomain.com]
> > action=create
> > lvname=gluster_lv_data
> > ignore_lv_errors=no
> > vgname=gluster_vg_sdc
> > mount=/gluster_bricks/data
> > lvtype=thinlv
> > poolname=gluster_thinpool_sdc
> > virtualsize=42GB
> >
> > [lv6:ovirt01.mydomain.com]
> > action=create
> > lvname=gluster_lv_vmstore
> > ignore_lv_errors=no
> > vgname=gluster_vg_sdd
> > mount=/gluster_bricks/vmstore
> > lvtype=thinlv
> > poolname=gluster_thinpool_sdd
> > virtualsize=42GB
> >
> > [lv7:ovirt02.mydomain.com]
> > action=create
> > lvname=gluster_lv_engine
> > ignore_lv_errors=no
> > vgname=gluster_vg_sdb
> > mount=/gluster_bricks/engine
> > size=42GB
> > lvtype=thick
> >
> > [lv8:ovirt02.mydomain.com]
> > action=create
> > lvname=gluster_lv_data
> > ignore_lv_errors=no
> > vgname=gluster_vg_sdc
> > mount=/gluster_bricks/data
> > lvtype=thinlv
> > poolname=gluster_thinpool_sdc
> > virtualsize=42GB
> >
> > [lv9:ovirt02.mydomain.com]
> > action=create
> > lvname=gluster_lv_vmstore
> > ignore_lv_errors=no
> > vgname=gluster_vg_sdd
> > mount=/gluster_bricks/vmstore
> > lvtype=thinlv
> > poolname=gluster_thinpool_sdd
> > virtualsize=42GB
> >
> > [lv10:ovirt03.mydomain.com]
> > action=create
> > lvname=gluster_lv_engine
> > ignore_lv_errors=no
> > vgname=gluster_vg_sdb
> > mount=/gluster_bricks/engine
> > size=42GB
> > lvtype=thick
> >
> > [lv11:ovirt03.mydomain.com]
> > action=create
> > lvname=gluster_lv_data
> > ignore_lv_errors=no
> > vgname=gluster_vg_sdc
> > mount=/gluster_bricks/data
> > lvtype=thinlv
> > poolname=gluster_thinpool_sdc
> > virtualsize=42GB
> >
> > [lv12:ovirt03.mydomain.com]
> > action=create
> > lvname=gluster_lv_vmstore
> > ignore_lv_errors=no
> > vgname=gluster_vg_sdd
> > mount=/gluster_bricks/vmstore
> > lvtype=thinlv
> > poolname=gluster_thinpool_sdd
> > virtualsize=42GB
> >
> > [selinux]
> > yes
> >
> > [service3]
> > action=restart
> > service=glusterd
> > slice_setup=yes
> >
> > [firewalld]
> > action=add
> >
> ports=111/tcp,2049/tcp,54321/tcp,5900/tcp,5900-6923/tcp,5666/tcp,16514/tcp,54322/tcp
> > services=glusterfs
> >
> > [script2]
> > action=execute
> > file=/usr/share/gdeploy/scripts/disable-gluster-hooks.sh
> >
> > [shell3]
> > action=execute
> > command=usermod -a -G gluster qemu
> >
> > [volume1]
> > action=create
> > volname=engine
> > transport=tcp
> > replica=yes
> > replica_count=3
> >
> key=group,storage.owner-uid,storage.owner-gid,network.ping-timeout,performance.strict-o-direct,network.remote-dio,cluster.granular-entry-heal
> > value=virt,36,36,30,on,off,enable
> > brick_dirs=ovirt01.mydomain.com:
> /gluster_bricks/engine/engine,ovirt02.mydomain.com:
> /gluster_bricks/engine/engine,ovirt03.mydomain.com:
> /gluster_bricks/engine/engine
> > ignore_volume_errors=no
> >
> > [volume2]
> > action=create
> > volname=data
> > transport=tcp
> > replica=yes
> > replica_count=3
> >
> key=group,storage.owner-uid,storage.owner-gid,network.ping-timeout,performance.strict-o-direct,network.remote-dio,cluster.granular-entry-heal
> > value=virt,36,36,30,on,off,enable
> > brick_dirs=ovirt01.mydomain.com:
> /gluster_bricks/data/data,ovirt02.mydomain.com:
> /gluster_bricks/data/data,ovirt03.mydomain.com:/gluster_bricks/data/data
> > ignore_volume_errors=no
> >
> > [volume3]
> > action=create
> > volname=vmstore
> > transport=tcp
> > replica=yes
> > replica_count=3
> >
> key=group,storage.owner-uid,storage.owner-gid,network.ping-timeout,performance.strict-o-direct,network.remote-dio,cluster.granular-entry-heal
> > value=virt,36,36,30,on,off,enable
> > brick_dirs=ovirt01.mydomain.com:
> /gluster_bricks/vmstore/vmstore,ovirt02.mydomain.com:
> /gluster_bricks/vmstore/vmstore,ovirt03.mydomain.com:
> /gluster_bricks/vmstore/vmstore
> > ignore_volume_errors=no
> >
> ---------------------------------------------------------------------------------------------------
> > _______________________________________________
> > Users mailing list -- users@ovirt.org
> > To unsubscribe send an email to users-le...@ovirt.org
> > Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> > oVirt Code of Conduct:
> https://www.ovirt.org/community/about/community-guidelines/
> > List Archives:
> https://lists.ovirt.org/archives/list/users@ovirt.org/message/4DCUGOSWOK6KNT26ROWYOITD5JQMQ3NE/
>


-- 


Thanks,
Gobinda
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/WCUHKVEWW3XGKSEEI7ZCUV7CJIKKBGJL/

Reply via email to