As requested, The output of ovirt-hosted-engine-cleanup
[root@node-1 ~]# ovirt-hosted-engine-cleanup This will de-configure the host to run ovirt-hosted-engine-setup from scratch. Caution, this operation should be used with care. Are you sure you want to proceed? [y/n] y -=== Destroy hosted-engine VM ===- You must run deploy first -=== Stop HA services ===- -=== Shutdown sanlock ===- shutdown force 1 wait 0 shutdown done 0 -=== Disconnecting the hosted-engine storage domain ===- You must run deploy first -=== De-configure VDSM networks ===- -=== Stop other services ===- -=== De-configure external daemons ===- -=== Removing configuration files ===- ? /etc/init/libvirtd.conf already missing - removing /etc/libvirt/nwfilter/vdsm-no-mac-spoofing.xml ? /etc/ovirt-hosted-engine/answers.conf already missing ? /etc/ovirt-hosted-engine/hosted-engine.conf already missing - removing /etc/vdsm/vdsm.conf - removing /etc/pki/vdsm/certs/cacert.pem - removing /etc/pki/vdsm/certs/vdsmcert.pem - removing /etc/pki/vdsm/keys/vdsmkey.pem - removing /etc/pki/vdsm/libvirt-spice/ca-cert.pem - removing /etc/pki/vdsm/libvirt-spice/ca-key.pem - removing /etc/pki/vdsm/libvirt-spice/server-cert.pem - removing /etc/pki/vdsm/libvirt-spice/server-key.pem ? /etc/pki/CA/cacert.pem already missing ? /etc/pki/libvirt/*.pem already missing ? /etc/pki/libvirt/private/*.pem already missing ? /etc/pki/ovirt-vmconsole/*.pem already missing - removing /var/cache/libvirt/qemu ? /var/run/ovirt-hosted-engine-ha/* already missing [root@node-1 ~]# Output of sanlock client status: [root@node-1 ~]# sanlock client status [root@node-1 ~]# Thank you for your help! On Mon, Jun 5, 2017 at 7:25 AM, Simone Tiraboschi <stira...@redhat.com> wrote: > > > On Mon, Jun 5, 2017 at 3:57 PM, Brendan Hartzell <mrr...@gmail.com> wrote: > >> After letting this sit for a few days, does anyone have any ideas as to >> how to deal with my situation? Would anyone like me to send the SOS report >> directly to them? It's a 9MB file. >> >> If nothing comes up, I'm going to try and sift through the SOS report >> tonight, but I won't know what I'm trying to find. >> >> Thank you for any and all help. >> >> On Thu, Jun 1, 2017 at 1:15 AM, Sandro Bonazzola <sbona...@redhat.com> >> wrote: >> >>> >>> >>> On Thu, Jun 1, 2017 at 6:36 AM, Brendan Hartzell <mrr...@gmail.com> >>> wrote: >>> >>>> Ran the 4 commands listed above, no errors on the screen. >>>> >>>> Started the hosted-engine standard setup from the web-UI. >>>> >>>> Using iSCSI for the storage. >>>> >>>> Using mostly default options, I got these errors in the web-UI. >>>> >>>> Error creating Volume Group: Failed to initialize physical device: >>>> ("[u'/dev/mapper/36589cfc000000de7482638fcfcebbbb4']",) >>>> Failed to execute stage 'Misc configuration': Failed to initialize >>>> physical device: ("[u'/dev/mapper/36589cfc00000 >>>> 0de7482638fcfcebbbb4']",) >>>> Hosted Engine deployment failed: this system is not reliable, please >>>> check the issue,fix and redeploy >>>> >>>> I rebuilt my iSCSI (I don't think I cleaned it up from a previous >>>> install). >>>> Re-ran the above 4 commands. >>>> Restarted hosted engine standard setup from web-UI. >>>> Install moved past "Connecting Storage Pool" so I believe the above was >>>> my fault. >>>> >>>> These are the last messages displayed on the web-UI. >>>> Creating Storage Pool >>>> Connecting Storage Pool >>>> Verifying sanlock lockspace initialization >>>> Creating Image for 'hosted-engine.lockspace' ... >>>> Image for 'hosted-engine.lockspace' created successfully >>>> Creating Image for 'hosted-engine.metadata' ... >>>> Image for 'hosted-engine.metadata' created successfully >>>> Creating VM Image >>>> Extracting disk image from OVF archive (could take a few minutes >>>> depending on archive size) >>>> Validating pre-allocated volume size >>>> Uploading volume to data domain (could take a few minutes depending on >>>> archive size) >>>> >>>> At the host terminal, I got the error "watchdog watchdog0: watchdog did >>>> not stop!" >>>> Then the host restarted. >>>> >>> >>> Simone, can you help here? >>> >>> > Ok, sorry for the delay. > The second installation attempt seams fine but it seams that > ovirt-hosted-engine-cleanup failed stopping sanlock and so the > watchdog kick in rebooting your system in the middle of deployment attempt. > > could you please post the output of > ovirt-hosted-engine-cleanup > sanlock client status > ? > > > >> >>> >>> >>>> >>>> This is as far as I've gotten in previous attempts. >>>> >>>> Attaching the hosted-engine-setup log. >>>> >>>> The SOS report is 9MB and the ovirt users group will drop the email. >>>> >>>> On Wed, May 31, 2017 at 6:59 AM, Sandro Bonazzola <sbona...@redhat.com> >>>> wrote: >>>> >>>>> >>>>> >>>>> On Wed, May 31, 2017 at 3:10 PM, Brendan Hartzell <mrr...@gmail.com> >>>>> wrote: >>>>> >>>>>> Now that you have identified the problem, should I run the following >>>>>> commands and send you another SOS? >>>>>> >>>>>> ovirt-hosted-engine-cleanup >>>>>> vdsm-tool configure --force >>>>>> systemctl restart libvirtd >>>>>> systemctl restart vdsm >>>>>> >>>>>> Or is there a different plan in mind? >>>>>> >>>>> >>>>> I would have expected someone from virt team to follow up for further >>>>> investigations :-) >>>>> above commands should work. >>>>> >>>>> >>>>> >>>>>> >>>>>> Thank you, >>>>>> >>>>>> Brendan >>>>>> >>>>>> On Tue, May 30, 2017 at 11:42 PM, Sandro Bonazzola < >>>>>> sbona...@redhat.com> wrote: >>>>>> >>>>>>> >>>>>>> >>>>>>> On Wed, May 31, 2017 at 4:45 AM, Brendan Hartzell <mrr...@gmail.com> >>>>>>> wrote: >>>>>>> >>>>>>>> Can you please elaborate about the failure you see here and how are >>>>>>>> you trying to manually partition the host? >>>>>>>> >>>>>>>> Sure, I will start from the beginning. >>>>>>>> - Using: ovirt-node-ng-installer-ovirt-4.1-2017052604 >>>>>>>> <(201)%20705-2604>.iso >>>>>>>> - During installation I setup one of the two interfaces and check >>>>>>>> the box to automatically use the connection. >>>>>>>> - I'm currently providing a host name of node-1.test.net until I >>>>>>>> have a successful process. >>>>>>>> - I configure date and time for my timezone and to use an internal >>>>>>>> NTP server. >>>>>>>> - On Installation Destination, I pick my 128GB USB3.0 SanDisk flash >>>>>>>> drive, check the box that I would like to make additional space, and >>>>>>>> click >>>>>>>> done. In the reclaim disk space window, I click delete all, and then >>>>>>>> reclaim space. I go back into the Installation Destination, select >>>>>>>> that I >>>>>>>> will configure partitioning, and click done. The Manual Partitioning >>>>>>>> window opens, I use the option to automatically create mount points. >>>>>>>> >>>>>>> >>>>>>> In this screen, please change partitioning scheme from LVM to LVM >>>>>>> Thin Provisioning: it should solve your following error. >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>>> At this point, /boot is 1024MB, /var is 15GB, / is 88.11 GB, and >>>>>>>> swap is 11.57GB. I then change / to 23.11 GB, update settings, change >>>>>>>> /var >>>>>>>> to 80GB, update settings again, and click done. I accept the changes >>>>>>>> and >>>>>>>> begin installation. >>>>>>>> >>>>>>>> I tried these changes based on this article: http://www.ovirt.org/ >>>>>>>> documentation/self-hosted/chap-Deploying_Self-Hosted_Engine/ >>>>>>>> >>>>>>>> The article does say that you can specify a different directory >>>>>>>> than /var/tmp, but I don't recall seeing that option. >>>>>>>> >>>>>>> >>>>>>> If the setup detects not enough space in /var/tmp for extracting the >>>>>>> appliance it will ask about a different directory. >>>>>>> >>>>>>> >>>>>>> >>>>>>>> >>>>>>>> After some time, I get the following error: >>>>>>>> There was an error running the kickstart script at line 7. This is >>>>>>>> a fatal error and installation will be aborted. The details of this >>>>>>>> error >>>>>>>> are: >>>>>>>> >>>>>>>> [INFO] Trying to create a manageable base from '/' >>>>>>>> [ERROR] LVM Thin Provisioning partitioning scheme is required. For >>>>>>>> autoinstall via Kickstart with LVM Thin Provisioning check options >>>>>>>> --thinpool and --grow. Please consult documentation for details. >>>>>>>> >>>>>>> >>>>>>> >>>>>>> ^^ this one should be solved by the LVM Thin Provisioning scheme >>>>>>> mentioned above.. >>>>>>> >>>>>>> >>>>>>> >>>>>>>> >>>>>>>> Traceback (most recent call last): >>>>>>>> File "/usr/lib64/python2.7/runpy.py", line 162, in >>>>>>>> _run_module_as_main "__main__", fname, loader, pkg_name) >>>>>>>> File "/usr/lib64/python2.7runpy.py", line 72, in _run_code exec >>>>>>>> code in run_globals >>>>>>>> File "/usr/lib/python2.7/site-packages/imgbased/__main__.py", line >>>>>>>> 51, in <module> CliApplication() >>>>>>>> File "/usr/lib/python2.7/site-packages/imgbased/__init__.py", line >>>>>>>> 82, in CliApplication() >>>>>>>> File "/usr/lib/python2.7/site-packages/imgbased/hooks.py", line >>>>>>>> 120, in emit cb(self.context, *args) >>>>>>>> File "/usr/lib/python2.7/site-packages/imgbased/plugins/core.py", >>>>>>>> line 169, in post_argparse layout.initialize(args.source, >>>>>>>> args.init_nvr) >>>>>>>> File "/usr/lib/python2.7/site-packages/imgbased/plugins/core.py", >>>>>>>> line 216, in initialize self.app.imgbase.init_layout_from(source, >>>>>>>> init_nvr) >>>>>>>> File "/usr/lib/python2.7/site-packages/imgbased/imgbase.py", line >>>>>>>> 271, in init_layout_from self.init_tags_on(existing_lv) >>>>>>>> File "/usr/lib/python2.7/site-packages/imgbased/imgbase.py", line >>>>>>>> 243, in init_tags_on pool = lv.thinpool() >>>>>>>> File "/usr/lib/python2.7/site-packages/imgbased/lvm.py", line 250, >>>>>>>> in thinpool raise MissingLvmThinPool() imgbased.lvm.MissingLvmThinPoo >>>>>>>> l >>>>>>>> >>>>>>>> At this point, the only option is to exit the installer. >>>>>>>> >>>>>>>> **************************** >>>>>>>> >>>>>>>> Being this a new install, please use 4.1. oVirt 4.0 is not >>>>>>>> supported anymore. >>>>>>>> >>>>>>>> Not a problem. >>>>>>>> >>>>>>>> **************************** >>>>>>>> >>>>>>>> Can you please provide hosted engine setup logs or better a full >>>>>>>> sos report? (sosreport -a) >>>>>>>> >>>>>>>> Again, the process I'm following: >>>>>>>> - Using: ovirt-node-ng-installer-ovirt-4.1-2017052604 >>>>>>>> <(201)%20705-2604>.iso >>>>>>>> - During installation I setup one of the two interfaces and check >>>>>>>> the box to automatically use the connection. >>>>>>>> - I'm currently providing a host name of node-1.test.net until I >>>>>>>> have a successful process. >>>>>>>> - I configure date and time for my timezone and to use an internal >>>>>>>> NTP server. >>>>>>>> - On Installation Destination, I pick my 128GB USB3.0 SanDisk flash >>>>>>>> drive, check the box that I would like to make additional space, and >>>>>>>> click >>>>>>>> done. In the reclaim disk space window, I click delete all, and then >>>>>>>> reclaim space. >>>>>>>> - Begin Installation and set a root password. >>>>>>>> - Perform a yum update - no packages marked for update (as expected) >>>>>>>> - Use vi to update /etc/hosts with a reference for node-1.test.net >>>>>>>> and engine.test.net >>>>>>>> - First attempt at hosted-engine from web-UI >>>>>>>> - Setup downloads and installs ovirt-engine-applianc >>>>>>>> e-4.1-20170523.1.el7.centos.noarch.rpm >>>>>>>> *Failed to execute stage 'Environment setup': Failed to reconfigure >>>>>>>> libvirt for VDSM >>>>>>>> *Hosted Engine deployment failed >>>>>>>> - Attached SOS report >>>>>>>> The checksum is: aa56097edc0b63c49caaf1a1fde021bc >>>>>>>> >>>>>>>> At this point, I would run ovirt-hosted-engine-cleanup and I would >>>>>>>> get further along in the install process. However, because this is a >>>>>>>> fresh >>>>>>>> install, I'm going to leave things here for now so you can review the >>>>>>>> SOS. >>>>>>>> >>>>>>> >>>>>>> Thanks for the SOS report! >>>>>>> Hosted Engine setup fails on: >>>>>>> >>>>>>> 2017-05-30 19:24:39 DEBUG otopi.plugins.gr_he_setup.system.vdsmenv >>>>>>> plugin.execute:921 execute-output: ('/bin/vdsm-tool', 'configure', >>>>>>> '--force') stdout: >>>>>>> >>>>>>> Checking configuration status... >>>>>>> >>>>>>> Current revision of multipath.conf detected, preserving >>>>>>> lvm is configured for vdsm >>>>>>> libvirt is already configured for vdsm >>>>>>> SUCCESS: ssl configured to true. No conflicts >>>>>>> >>>>>>> Running configure... >>>>>>> Reconfiguration of libvirt is done. >>>>>>> >>>>>>> 2017-05-30 19:24:39 DEBUG otopi.plugins.gr_he_setup.system.vdsmenv >>>>>>> plugin.execute:926 execute-output: ('/bin/vdsm-tool', 'configure', >>>>>>> '--force') stderr: >>>>>>> Error: ServiceOperationError: _systemctlStart failed >>>>>>> Job for libvirtd.service failed because the control process exited >>>>>>> with error code. See "systemctl status libvirtd.service" and "journalctl >>>>>>> -xe" for details. >>>>>>> >>>>>>> At the same time journalctl shows: >>>>>>> >>>>>>> May 30 19:24:39 node-1.test.net libvirtd[20954]: libvirt version: >>>>>>> 2.0.0, package: 10.el7_3.5 (CentOS BuildSystem < >>>>>>> http://bugs.centos.org>, 2017-03-03-02:09:45, c1bm.rdu2.centos.org) >>>>>>> May 30 19:24:39 node-1.test.net libvirtd[20954]: hostname: >>>>>>> node-1.test.net >>>>>>> May 30 19:24:39 node-1.test.net libvirtd[20954]: The server >>>>>>> certificate /etc/pki/vdsm/certs/vdsmcert.pem is not yet active >>>>>>> May 30 19:24:39 node-1.test.net systemd[1]: libvirtd.service: main >>>>>>> process exited, code=exited, status=6/NOTCONFIGURED >>>>>>> May 30 19:24:39 node-1.test.net systemd[1]: Failed to start >>>>>>> Virtualization daemon. >>>>>>> May 30 19:24:39 node-1.test.net systemd[1]: Unit libvirtd.service >>>>>>> entered failed state. >>>>>>> May 30 19:24:39 node-1.test.net systemd[1]: libvirtd.service failed. >>>>>>> May 30 19:24:39 node-1.test.net systemd[1]: libvirtd.service >>>>>>> holdoff time over, scheduling restart. >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>>> >>>>>>>> **************************** >>>>>>>> >>>>>>>> I'd like to understand the issues you faced before suggesting to >>>>>>>> restart from scratch. >>>>>>>> >>>>>>>> Too late... I did two re-installs to get a more accurate account of >>>>>>>> my install process for above. >>>>>>>> >>>>>>>> **************************** >>>>>>>> >>>>>>>> Thank you for your help! >>>>>>>> >>>>>>>> Brendan >>>>>>>> >>>>>>>> On Tue, May 30, 2017 at 12:17 AM, Sandro Bonazzola < >>>>>>>> sbona...@redhat.com> wrote: >>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> On Tue, May 30, 2017 at 6:49 AM, Brendan Hartzell < >>>>>>>>> mrr...@gmail.com> wrote: >>>>>>>>> >>>>>>>>>> oVirt users list, >>>>>>>>>> >>>>>>>>>> Long story short, I've been spending weeks on this project for my >>>>>>>>>> home lab with no success. >>>>>>>>>> >>>>>>>>>> I would like to successfully install two nodes that host a highly >>>>>>>>>> available engine with an iSCSI storage back-end. >>>>>>>>>> >>>>>>>>>> I have read through most, if not all, of the guides on ovirt.org >>>>>>>>>> with no substantial help. >>>>>>>>>> >>>>>>>>>> Successfully, I have done the following: >>>>>>>>>> Install oVirt Engine on a bare metal system, added a node, and >>>>>>>>>> started exploring - not desired. >>>>>>>>>> Install oVirt Node 4.0.6 on a bare metal system - fails if >>>>>>>>>> partitions are not done automatically. >>>>>>>>>> Install oVirt Node 4.1.2 on a bare metal system - fails if >>>>>>>>>> partitions are not done automatically. >>>>>>>>>> >>>>>>>>> >>>>>>>>> Can you please elaborate about the failure you see here and how >>>>>>>>> are you trying to manually partition the host? >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>>> >>>>>>>>>> My process after installing a Node: >>>>>>>>>> Run a yum update - just to be sure, but I am using latest iso >>>>>>>>>> images from downloads section. >>>>>>>>>> Edit /etc/hosts for local name resolution - the goal is to host >>>>>>>>>> DNS as a virtual machine, eventually. >>>>>>>>>> On 4.1 if I install ovirt-engine-appliance from yum, it does >>>>>>>>>> simplify one step in the hosted engine setup. If I do this on 4.0 it >>>>>>>>>> discards the image and uses the default. >>>>>>>>>> >>>>>>>>> >>>>>>>>> Being this a new install, please use 4.1. oVirt 4.0 is not >>>>>>>>> supported anymore. >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>>> On 4.1 the hosted engine setup fails immediately unless I run the >>>>>>>>>> hosted engine cleanup from the shell. >>>>>>>>>> >>>>>>>>> >>>>>>>>> Can you please provide hosted engine setup logs or better a full >>>>>>>>> sos report? (sosreport -a) >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>>> If I do this, I can typically get to the point of installing. >>>>>>>>>> >>>>>>>>>> When I do get to the installation phase, I get to a point just >>>>>>>>>> after extracting the OVA that I get a message on the shell saying >>>>>>>>>> something >>>>>>>>>> about the watchdog running the whole time and then the node reboots. >>>>>>>>>> >>>>>>>>>> I found one email thread that sounded like my issue and suggested >>>>>>>>>> the following commands: >>>>>>>>>> vdsm-tool configure --force >>>>>>>>>> systemctl restart libvirtd >>>>>>>>>> systemctl restart vdsmd >>>>>>>>>> >>>>>>>>>> Unfortunately, these commands did not help my situation like the >>>>>>>>>> other individual. >>>>>>>>>> >>>>>>>>>> What log file would everyone like to see first? Given that I >>>>>>>>>> still consider myself relatively new to Linux, please identify the >>>>>>>>>> path for >>>>>>>>>> the log file requested. >>>>>>>>>> >>>>>>>>> >>>>>>>>> See above >>>>>>>>> >>>>>>>>> >>>>>>>>>> >>>>>>>>>> Also, because I plan on performing a clean install for this >>>>>>>>>> thread using my process from above (I'm not expecting my outcome to >>>>>>>>>> be any >>>>>>>>>> different), are there any tips and tricks that might result in a >>>>>>>>>> success? >>>>>>>>>> >>>>>>>>> >>>>>>>>> I'd like to understand the issues you faced before suggesting to >>>>>>>>> restart from scratch. >>>>>>>>> Adding some people who may help as well. >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>>> >>>>>>>>>> Thank you for any and all help, >>>>>>>>>> Brendan >>>>>>>>>> >>>>>>>>>> _______________________________________________ >>>>>>>>>> Users mailing list >>>>>>>>>> Users@ovirt.org >>>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users >>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> -- >>>>>>>>> >>>>>>>>> SANDRO BONAZZOLA >>>>>>>>> >>>>>>>>> ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG VIRTUALIZATION >>>>>>>>> R&D >>>>>>>>> >>>>>>>>> Red Hat EMEA <https://www.redhat.com/> >>>>>>>>> <https://red.ht/sig> >>>>>>>>> TRIED. TESTED. TRUSTED. <https://redhat.com/trusted> >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>> >>>>>>> >>>>>>> -- >>>>>>> >>>>>>> SANDRO BONAZZOLA >>>>>>> >>>>>>> ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG VIRTUALIZATION R&D >>>>>>> >>>>>>> Red Hat EMEA <https://www.redhat.com/> >>>>>>> <https://red.ht/sig> >>>>>>> TRIED. TESTED. TRUSTED. <https://redhat.com/trusted> >>>>>>> >>>>>> >>>>>> >>>>> >>>>> >>>>> -- >>>>> >>>>> SANDRO BONAZZOLA >>>>> >>>>> ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG VIRTUALIZATION R&D >>>>> >>>>> Red Hat EMEA <https://www.redhat.com/> >>>>> <https://red.ht/sig> >>>>> TRIED. TESTED. TRUSTED. <https://redhat.com/trusted> >>>>> >>>> >>>> >>> >>> >>> -- >>> >>> SANDRO BONAZZOLA >>> >>> ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG VIRTUALIZATION R&D >>> >>> Red Hat EMEA <https://www.redhat.com/> >>> <https://red.ht/sig> >>> TRIED. TESTED. TRUSTED. <https://redhat.com/trusted> >>> >> >> >
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users