Now that you have identified the problem, should I run the following commands and send you another SOS?
ovirt-hosted-engine-cleanup vdsm-tool configure --force systemctl restart libvirtd systemctl restart vdsm Or is there a different plan in mind? Thank you, Brendan On Tue, May 30, 2017 at 11:42 PM, Sandro Bonazzola <sbona...@redhat.com> wrote: > > > On Wed, May 31, 2017 at 4:45 AM, Brendan Hartzell <mrr...@gmail.com> > wrote: > >> Can you please elaborate about the failure you see here and how are you >> trying to manually partition the host? >> >> Sure, I will start from the beginning. >> - Using: ovirt-node-ng-installer-ovirt-4.1-2017052604 <(201)%20705-2604> >> .iso >> - During installation I setup one of the two interfaces and check the box >> to automatically use the connection. >> - I'm currently providing a host name of node-1.test.net until I have a >> successful process. >> - I configure date and time for my timezone and to use an internal NTP >> server. >> - On Installation Destination, I pick my 128GB USB3.0 SanDisk flash >> drive, check the box that I would like to make additional space, and click >> done. In the reclaim disk space window, I click delete all, and then >> reclaim space. I go back into the Installation Destination, select that I >> will configure partitioning, and click done. The Manual Partitioning >> window opens, I use the option to automatically create mount points. >> > > In this screen, please change partitioning scheme from LVM to LVM Thin > Provisioning: it should solve your following error. > > > > >> At this point, /boot is 1024MB, /var is 15GB, / is 88.11 GB, and swap >> is 11.57GB. I then change / to 23.11 GB, update settings, change /var to >> 80GB, update settings again, and click done. I accept the changes and >> begin installation. >> >> I tried these changes based on this article: http://www.ovirt.org/ >> documentation/self-hosted/chap-Deploying_Self-Hosted_Engine/ >> >> The article does say that you can specify a different directory than >> /var/tmp, but I don't recall seeing that option. >> > > If the setup detects not enough space in /var/tmp for extracting the > appliance it will ask about a different directory. > > > >> >> After some time, I get the following error: >> There was an error running the kickstart script at line 7. This is a >> fatal error and installation will be aborted. The details of this error >> are: >> >> [INFO] Trying to create a manageable base from '/' >> [ERROR] LVM Thin Provisioning partitioning scheme is required. For >> autoinstall via Kickstart with LVM Thin Provisioning check options >> --thinpool and --grow. Please consult documentation for details. >> > > > ^^ this one should be solved by the LVM Thin Provisioning scheme mentioned > above.. > > > >> >> Traceback (most recent call last): >> File "/usr/lib64/python2.7/runpy.py", line 162, in _run_module_as_main >> "__main__", fname, loader, pkg_name) >> File "/usr/lib64/python2.7runpy.py", line 72, in _run_code exec code in >> run_globals >> File "/usr/lib/python2.7/site-packages/imgbased/__main__.py", line 51, >> in <module> CliApplication() >> File "/usr/lib/python2.7/site-packages/imgbased/__init__.py", line 82, >> in CliApplication() >> File "/usr/lib/python2.7/site-packages/imgbased/hooks.py", line 120, in >> emit cb(self.context, *args) >> File "/usr/lib/python2.7/site-packages/imgbased/plugins/core.py", line >> 169, in post_argparse layout.initialize(args.source, args.init_nvr) >> File "/usr/lib/python2.7/site-packages/imgbased/plugins/core.py", line >> 216, in initialize self.app.imgbase.init_layout_from(source, init_nvr) >> File "/usr/lib/python2.7/site-packages/imgbased/imgbase.py", line 271, >> in init_layout_from self.init_tags_on(existing_lv) >> File "/usr/lib/python2.7/site-packages/imgbased/imgbase.py", line 243, >> in init_tags_on pool = lv.thinpool() >> File "/usr/lib/python2.7/site-packages/imgbased/lvm.py", line 250, in >> thinpool raise MissingLvmThinPool() imgbased.lvm.MissingLvmThinPool >> >> At this point, the only option is to exit the installer. >> >> **************************** >> >> Being this a new install, please use 4.1. oVirt 4.0 is not supported >> anymore. >> >> Not a problem. >> >> **************************** >> >> Can you please provide hosted engine setup logs or better a full sos >> report? (sosreport -a) >> >> Again, the process I'm following: >> - Using: ovirt-node-ng-installer-ovirt-4.1-2017052604 <(201)%20705-2604> >> .iso >> - During installation I setup one of the two interfaces and check the box >> to automatically use the connection. >> - I'm currently providing a host name of node-1.test.net until I have a >> successful process. >> - I configure date and time for my timezone and to use an internal NTP >> server. >> - On Installation Destination, I pick my 128GB USB3.0 SanDisk flash >> drive, check the box that I would like to make additional space, and click >> done. In the reclaim disk space window, I click delete all, and then >> reclaim space. >> - Begin Installation and set a root password. >> - Perform a yum update - no packages marked for update (as expected) >> - Use vi to update /etc/hosts with a reference for node-1.test.net and >> engine.test.net >> - First attempt at hosted-engine from web-UI >> - Setup downloads and installs ovirt-engine-applianc >> e-4.1-20170523.1.el7.centos.noarch.rpm >> *Failed to execute stage 'Environment setup': Failed to reconfigure >> libvirt for VDSM >> *Hosted Engine deployment failed >> - Attached SOS report >> The checksum is: aa56097edc0b63c49caaf1a1fde021bc >> >> At this point, I would run ovirt-hosted-engine-cleanup and I would get >> further along in the install process. However, because this is a fresh >> install, I'm going to leave things here for now so you can review the SOS. >> > > Thanks for the SOS report! > Hosted Engine setup fails on: > > 2017-05-30 19:24:39 DEBUG otopi.plugins.gr_he_setup.system.vdsmenv > plugin.execute:921 execute-output: ('/bin/vdsm-tool', 'configure', > '--force') stdout: > > Checking configuration status... > > Current revision of multipath.conf detected, preserving > lvm is configured for vdsm > libvirt is already configured for vdsm > SUCCESS: ssl configured to true. No conflicts > > Running configure... > Reconfiguration of libvirt is done. > > 2017-05-30 19:24:39 DEBUG otopi.plugins.gr_he_setup.system.vdsmenv > plugin.execute:926 execute-output: ('/bin/vdsm-tool', 'configure', > '--force') stderr: > Error: ServiceOperationError: _systemctlStart failed > Job for libvirtd.service failed because the control process exited with > error code. See "systemctl status libvirtd.service" and "journalctl -xe" > for details. > > At the same time journalctl shows: > > May 30 19:24:39 node-1.test.net libvirtd[20954]: libvirt version: 2.0.0, > package: 10.el7_3.5 (CentOS BuildSystem <http://bugs.centos.org>, > 2017-03-03-02:09:45, c1bm.rdu2.centos.org) > May 30 19:24:39 node-1.test.net libvirtd[20954]: hostname: node-1.test.net > May 30 19:24:39 node-1.test.net libvirtd[20954]: The server certificate > /etc/pki/vdsm/certs/vdsmcert.pem is not yet active > May 30 19:24:39 node-1.test.net systemd[1]: libvirtd.service: main > process exited, code=exited, status=6/NOTCONFIGURED > May 30 19:24:39 node-1.test.net systemd[1]: Failed to start > Virtualization daemon. > May 30 19:24:39 node-1.test.net systemd[1]: Unit libvirtd.service entered > failed state. > May 30 19:24:39 node-1.test.net systemd[1]: libvirtd.service failed. > May 30 19:24:39 node-1.test.net systemd[1]: libvirtd.service holdoff time > over, scheduling restart. > > > > > > > > > >> >> **************************** >> >> I'd like to understand the issues you faced before suggesting to restart >> from scratch. >> >> Too late... I did two re-installs to get a more accurate account of my >> install process for above. >> >> **************************** >> >> Thank you for your help! >> >> Brendan >> >> On Tue, May 30, 2017 at 12:17 AM, Sandro Bonazzola <sbona...@redhat.com> >> wrote: >> >>> >>> >>> On Tue, May 30, 2017 at 6:49 AM, Brendan Hartzell <mrr...@gmail.com> >>> wrote: >>> >>>> oVirt users list, >>>> >>>> Long story short, I've been spending weeks on this project for my home >>>> lab with no success. >>>> >>>> I would like to successfully install two nodes that host a highly >>>> available engine with an iSCSI storage back-end. >>>> >>>> I have read through most, if not all, of the guides on ovirt.org with >>>> no substantial help. >>>> >>>> Successfully, I have done the following: >>>> Install oVirt Engine on a bare metal system, added a node, and started >>>> exploring - not desired. >>>> Install oVirt Node 4.0.6 on a bare metal system - fails if partitions >>>> are not done automatically. >>>> Install oVirt Node 4.1.2 on a bare metal system - fails if partitions >>>> are not done automatically. >>>> >>> >>> Can you please elaborate about the failure you see here and how are you >>> trying to manually partition the host? >>> >>> >>> >>> >>> >>>> >>>> My process after installing a Node: >>>> Run a yum update - just to be sure, but I am using latest iso images >>>> from downloads section. >>>> Edit /etc/hosts for local name resolution - the goal is to host DNS as >>>> a virtual machine, eventually. >>>> On 4.1 if I install ovirt-engine-appliance from yum, it does simplify >>>> one step in the hosted engine setup. If I do this on 4.0 it discards the >>>> image and uses the default. >>>> >>> >>> Being this a new install, please use 4.1. oVirt 4.0 is not supported >>> anymore. >>> >>> >>> >>>> On 4.1 the hosted engine setup fails immediately unless I run the >>>> hosted engine cleanup from the shell. >>>> >>> >>> Can you please provide hosted engine setup logs or better a full sos >>> report? (sosreport -a) >>> >>> >>> >>> >>>> If I do this, I can typically get to the point of installing. >>>> >>>> When I do get to the installation phase, I get to a point just after >>>> extracting the OVA that I get a message on the shell saying something about >>>> the watchdog running the whole time and then the node reboots. >>>> >>>> I found one email thread that sounded like my issue and suggested the >>>> following commands: >>>> vdsm-tool configure --force >>>> systemctl restart libvirtd >>>> systemctl restart vdsmd >>>> >>>> Unfortunately, these commands did not help my situation like the other >>>> individual. >>>> >>>> What log file would everyone like to see first? Given that I still >>>> consider myself relatively new to Linux, please identify the path for the >>>> log file requested. >>>> >>> >>> See above >>> >>> >>>> >>>> Also, because I plan on performing a clean install for this thread >>>> using my process from above (I'm not expecting my outcome to be any >>>> different), are there any tips and tricks that might result in a success? >>>> >>> >>> I'd like to understand the issues you faced before suggesting to restart >>> from scratch. >>> Adding some people who may help as well. >>> >>> >>> >>>> >>>> Thank you for any and all help, >>>> Brendan >>>> >>>> _______________________________________________ >>>> Users mailing list >>>> Users@ovirt.org >>>> http://lists.ovirt.org/mailman/listinfo/users >>>> >>>> >>> >>> >>> -- >>> >>> SANDRO BONAZZOLA >>> >>> ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG VIRTUALIZATION R&D >>> >>> Red Hat EMEA <https://www.redhat.com/> >>> <https://red.ht/sig> >>> TRIED. TESTED. TRUSTED. <https://redhat.com/trusted> >>> >> >> > > > -- > > SANDRO BONAZZOLA > > ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG VIRTUALIZATION R&D > > Red Hat EMEA <https://www.redhat.com/> > <https://red.ht/sig> > TRIED. TESTED. TRUSTED. <https://redhat.com/trusted> >
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users