Hi Adrian, Looks like issue is with lvm ( https://bugzilla.redhat.com/show_bug.cgi?id=502671 and https://bugzilla.redhat.com/show_bug.cgi?id=1664461 ) And it's fixed in lvm2-2.02.46-3.el5 Can you please check your lvm2 version?
On Mon, Mar 11, 2019 at 1:57 PM Sahina Bose <sab...@redhat.com> wrote: > +Gobinda Das +Dhanjal Parth > > On Mon, Mar 11, 2019 at 1:42 AM <adrianquint...@gmail.com> wrote: > > > > Hello I am trying to run a Hyperconverged setup "COnfigure gluster > storage and ovirt hosted engine", however I get the following error > > > > > __________________________________________________________________________________________________ > > PLAY [gluster_servers] > ********************************************************* > > > > TASK [Create LVs with specified size for the VGs] > ****************************** > > failed: [ovirt01.grupokino.com] (item={u'lv': u'gluster_thinpool_sdb', > u'size': u'45GB', u'extent': u'100%FREE', u'vg': u'gluster_vg_sdb'}) => > {"changed": false, "item": {"extent": "100%FREE", "lv": > "gluster_thinpool_sdb", "size": "45GB", "vg": "gluster_vg_sdb"}, "msg": > "lvcreate: metadata/pv_map.c:198: consume_pv_area: Assertion `to_go <= > pva->count' failed.\n", "rc": -6} > > to retry, use: --limit @/tmp/tmpwo4SNB/lvcreate.retry > > > > PLAY RECAP > ********************************************************************* > > ovirt01.grupokino.com : ok=0 changed=0 unreachable=0 > failed=1 > > > __________________________________________________________________________________________________ > > > > I know that oVirt Hosted Engine Setup GUI for "gluster wizard (gluster > deployment) does not populate the geodeployConfig.conf file properly > (Generated Gdeploy configuration : > /var/lib/ovirt-hosted-engine-setup/gdeploy/gdeployConfig.conf) so I have > tried to modify it to fit our needs but keep getting the above error > everytime. > > > > Any ideas or comments are welcome...Thanks! > > > > > > > > > > My servers are setup with 4x50GB disks, 1 for the OS and the rest for > Gluster Hyperconverged setup. > > > __________________________________________________________________________________________________ > > my gdeployConfig.conf file: > > > __________________________________________________________________________________________________ > > #gdeploy configuration generated by cockpit-gluster plugin > > [hosts] > > ovirt01.mydomain.com > > ovirt02.mydomain.com > > ovirt03.mydomain.com > > > > [script1:ovirt01.mydomain.com] > > action=execute > > ignore_script_errors=no > > file=/usr/share/gdeploy/scripts/grafton-sanity-check.sh -d sdb,sdc,sdd > -h ovirt01.mydomain.com, ovirt02.mydomain.com, ovirt03.mydomain.com > > > > [script1:ovirt02.mydomain.com] > > action=execute > > ignore_script_errors=no > > file=/usr/share/gdeploy/scripts/grafton-sanity-check.sh -d sdb,sdc,sdd > -h ovirt01.mydomain.com, ovirt02.mydomain.com, ovirt03.mydomain.com > > > > [script1:ovirt03.mydomain.com] > > action=execute > > ignore_script_errors=no > > file=/usr/share/gdeploy/scripts/grafton-sanity-check.sh -d sdb,sdc,sdd > -h ovirt01.mydomain.com, ovirt02.mydomain.com, ovirt03.mydomain.com > > > > [disktype] > > jbod > > > > [diskcount] > > 3 > > > > [stripesize] > > 256 > > > > [service1] > > action=enable > > service=chronyd > > > > [service2] > > action=restart > > service=chronyd > > > > [shell2] > > action=execute > > command=vdsm-tool configure --force > > > > [script3] > > action=execute > > file=/usr/share/gdeploy/scripts/blacklist_all_disks.sh > > ignore_script_errors=no > > > > [pv1:ovirt01.mydomain.com] > > action=create > > devices=sdb > > ignore_pv_errors=no > > > > [pv1:ovirt02.mydomain.com] > > action=create > > devices=sdb > > ignore_pv_errors=no > > > > [pv1:ovirt03.mydomain.com] > > action=create > > devices=sdb > > ignore_pv_errors=no > > > > [pv2:ovirt01.mydomain.com] > > action=create > > devices=sdc > > ignore_pv_errors=no > > > > [pv2:ovirt02.mydomain.com] > > action=create > > devices=sdc > > ignore_pv_errors=no > > > > [pv2:ovirt03.mydomain.com] > > action=create > > devices=sdc > > ignore_pv_errors=no > > > > [pv3:ovirt01.mydomain.com] > > action=create > > devices=sdd > > ignore_pv_errors=no > > > > [pv3:ovirt02.mydomain.com] > > action=create > > devices=sdd > > ignore_pv_errors=no > > > > [pv3:ovirt03.mydomain.com] > > action=create > > devices=sdd > > ignore_pv_errors=no > > > > [vg1:ovirt01.mydomain.com] > > action=create > > vgname=gluster_vg_sdb > > pvname=sdb > > ignore_vg_errors=no > > > > [vg1:ovirt02.mydomain.com] > > action=create > > vgname=gluster_vg_sdb > > pvname=sdb > > ignore_vg_errors=no > > > > [vg1:ovirt03.mydomain.com] > > action=create > > vgname=gluster_vg_sdb > > pvname=sdb > > ignore_vg_errors=no > > > > [vg2:ovirt01.mydomain.com] > > action=create > > vgname=gluster_vg_sdc > > pvname=sdc > > ignore_vg_errors=no > > > > [vg2:ovirt02.mydomain.com] > > action=create > > vgname=gluster_vg_sdc > > pvname=sdc > > ignore_vg_errors=no > > > > [vg2:ovirt03.mydomain.com] > > action=create > > vgname=gluster_vg_sdc > > pvname=sdc > > ignore_vg_errors=no > > > > [vg3:ovirt01.mydomain.com] > > action=create > > vgname=gluster_vg_sdd > > pvname=sdd > > ignore_vg_errors=no > > > > [vg3:ovirt02.mydomain.com] > > action=create > > vgname=gluster_vg_sdd > > pvname=sdd > > ignore_vg_errors=no > > > > [vg3:ovirt03.mydomain.com] > > action=create > > vgname=gluster_vg_sdd > > pvname=sdd > > ignore_vg_errors=no > > > > [lv1:ovirt01.mydomain.com] > > action=create > > poolname=gluster_thinpool_sdb > > ignore_lv_errors=no > > vgname=gluster_vg_sdb > > lvtype=thinpool > > size=45GB > > poolmetadatasize=3GB > > > > [lv2:ovirt02.mydomain.com] > > action=create > > poolname=gluster_thinpool_sdc > > ignore_lv_errors=no > > vgname=gluster_vg_sdc > > lvtype=thinpool > > size=45GB > > poolmetadatasize=3GB > > > > [lv3:ovirt03.mydomain.com] > > action=create > > poolname=gluster_thinpool_sdd > > ignore_lv_errors=no > > vgname=gluster_vg_sdd > > lvtype=thinpool > > size=45GB > > poolmetadatasize=3GB > > > > [lv4:ovirt01.mydomain.com] > > action=create > > lvname=gluster_lv_engine > > ignore_lv_errors=no > > vgname=gluster_vg_sdb > > mount=/gluster_bricks/engine > > size=42GB > > lvtype=thick > > > > [lv5:ovirt01.mydomain.com] > > action=create > > lvname=gluster_lv_data > > ignore_lv_errors=no > > vgname=gluster_vg_sdc > > mount=/gluster_bricks/data > > lvtype=thinlv > > poolname=gluster_thinpool_sdc > > virtualsize=42GB > > > > [lv6:ovirt01.mydomain.com] > > action=create > > lvname=gluster_lv_vmstore > > ignore_lv_errors=no > > vgname=gluster_vg_sdd > > mount=/gluster_bricks/vmstore > > lvtype=thinlv > > poolname=gluster_thinpool_sdd > > virtualsize=42GB > > > > [lv7:ovirt02.mydomain.com] > > action=create > > lvname=gluster_lv_engine > > ignore_lv_errors=no > > vgname=gluster_vg_sdb > > mount=/gluster_bricks/engine > > size=42GB > > lvtype=thick > > > > [lv8:ovirt02.mydomain.com] > > action=create > > lvname=gluster_lv_data > > ignore_lv_errors=no > > vgname=gluster_vg_sdc > > mount=/gluster_bricks/data > > lvtype=thinlv > > poolname=gluster_thinpool_sdc > > virtualsize=42GB > > > > [lv9:ovirt02.mydomain.com] > > action=create > > lvname=gluster_lv_vmstore > > ignore_lv_errors=no > > vgname=gluster_vg_sdd > > mount=/gluster_bricks/vmstore > > lvtype=thinlv > > poolname=gluster_thinpool_sdd > > virtualsize=42GB > > > > [lv10:ovirt03.mydomain.com] > > action=create > > lvname=gluster_lv_engine > > ignore_lv_errors=no > > vgname=gluster_vg_sdb > > mount=/gluster_bricks/engine > > size=42GB > > lvtype=thick > > > > [lv11:ovirt03.mydomain.com] > > action=create > > lvname=gluster_lv_data > > ignore_lv_errors=no > > vgname=gluster_vg_sdc > > mount=/gluster_bricks/data > > lvtype=thinlv > > poolname=gluster_thinpool_sdc > > virtualsize=42GB > > > > [lv12:ovirt03.mydomain.com] > > action=create > > lvname=gluster_lv_vmstore > > ignore_lv_errors=no > > vgname=gluster_vg_sdd > > mount=/gluster_bricks/vmstore > > lvtype=thinlv > > poolname=gluster_thinpool_sdd > > virtualsize=42GB > > > > [selinux] > > yes > > > > [service3] > > action=restart > > service=glusterd > > slice_setup=yes > > > > [firewalld] > > action=add > > > ports=111/tcp,2049/tcp,54321/tcp,5900/tcp,5900-6923/tcp,5666/tcp,16514/tcp,54322/tcp > > services=glusterfs > > > > [script2] > > action=execute > > file=/usr/share/gdeploy/scripts/disable-gluster-hooks.sh > > > > [shell3] > > action=execute > > command=usermod -a -G gluster qemu > > > > [volume1] > > action=create > > volname=engine > > transport=tcp > > replica=yes > > replica_count=3 > > > key=group,storage.owner-uid,storage.owner-gid,network.ping-timeout,performance.strict-o-direct,network.remote-dio,cluster.granular-entry-heal > > value=virt,36,36,30,on,off,enable > > brick_dirs=ovirt01.mydomain.com: > /gluster_bricks/engine/engine,ovirt02.mydomain.com: > /gluster_bricks/engine/engine,ovirt03.mydomain.com: > /gluster_bricks/engine/engine > > ignore_volume_errors=no > > > > [volume2] > > action=create > > volname=data > > transport=tcp > > replica=yes > > replica_count=3 > > > key=group,storage.owner-uid,storage.owner-gid,network.ping-timeout,performance.strict-o-direct,network.remote-dio,cluster.granular-entry-heal > > value=virt,36,36,30,on,off,enable > > brick_dirs=ovirt01.mydomain.com: > /gluster_bricks/data/data,ovirt02.mydomain.com: > /gluster_bricks/data/data,ovirt03.mydomain.com:/gluster_bricks/data/data > > ignore_volume_errors=no > > > > [volume3] > > action=create > > volname=vmstore > > transport=tcp > > replica=yes > > replica_count=3 > > > key=group,storage.owner-uid,storage.owner-gid,network.ping-timeout,performance.strict-o-direct,network.remote-dio,cluster.granular-entry-heal > > value=virt,36,36,30,on,off,enable > > brick_dirs=ovirt01.mydomain.com: > /gluster_bricks/vmstore/vmstore,ovirt02.mydomain.com: > /gluster_bricks/vmstore/vmstore,ovirt03.mydomain.com: > /gluster_bricks/vmstore/vmstore > > ignore_volume_errors=no > > > --------------------------------------------------------------------------------------------------- > > _______________________________________________ > > Users mailing list -- users@ovirt.org > > To unsubscribe send an email to users-le...@ovirt.org > > Privacy Statement: https://www.ovirt.org/site/privacy-policy/ > > oVirt Code of Conduct: > https://www.ovirt.org/community/about/community-guidelines/ > > List Archives: > https://lists.ovirt.org/archives/list/users@ovirt.org/message/4DCUGOSWOK6KNT26ROWYOITD5JQMQ3NE/ > -- Thanks, Gobinda
_______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/WCUHKVEWW3XGKSEEI7ZCUV7CJIKKBGJL/