HI Brendan, Can you please send the output for systemctl status vdsmd and journalctl -u vdsmd.service ?
Thanks, On Wed, Jun 7, 2017 at 9:32 AM, Sandro Bonazzola <sbona...@redhat.com> wrote: > > > On Tue, Jun 6, 2017 at 2:56 PM, Brendan Hartzell <mrr...@gmail.com> wrote: > >> Upon login to the server, to watch terminal output, I noticed that the >> node status is degraded. >> >> [root@node-1 ~]# nodectl check >> Status: WARN >> Bootloader ... OK >> Layer boot entries ... OK >> Valid boot entries ... OK >> Mount points ... OK >> Separate /var ... OK >> Discard is used ... OK >> Basic storage ... OK >> Initialized VG ... OK >> Initialized Thin Pool ... OK >> Initialized LVs ... OK >> Thin storage ... OK >> Checking available space in thinpool ... OK >> Checking thinpool auto-extend ... OK >> vdsmd ... BAD >> > > Yuval, can you help here? > > > >> >> >> Pressing forward with the retry using the web-UI. >> >> After resetting my iSCSI storage (on the storage server side), Install >> started. >> >> Status in the web-UI: >> Creating Storage Domain >> Creating Storage Pool >> Connecting Storage Pool >> Verifying sanlock lockspace initialization >> Creating Image for 'hosted-engine.lockspace' ... >> Image for 'hosted-engine.lockspace' created successfully >> Creating Image for 'hosted-engine.metadata' ... >> Image for 'hosted-engine.metadata' created successfully >> Creating VM Image >> Extracting disk image from OVF archive (could take a few minutes >> depending on archive size) >> Validating pre-allocated volume size >> >> Output from the terminal: >> [45863.076979]watchdog watchdog0: watchdog did not stop! >> >> System restarted. >> >> Attaching ovirt-hosted-engine-setup log. >> >> I'm running an SOS report, but it's too big for the users list. I can >> email it directly to you upon request. >> >> On Tue, Jun 6, 2017 at 12:12 AM, Simone Tiraboschi <stira...@redhat.com> >> wrote: >> >>> >>> >>> On Tue, Jun 6, 2017 at 2:10 AM, Brendan Hartzell <mrr...@gmail.com> >>> wrote: >>> >>>> As requested, >>>> >>> >>> It seams fine, there are no pending locks now. >>> Could you please retry? >>> >>> >>>> >>>> The output of ovirt-hosted-engine-cleanup >>>> >>>> [root@node-1 ~]# ovirt-hosted-engine-cleanup >>>> This will de-configure the host to run ovirt-hosted-engine-setup from >>>> scratch. >>>> Caution, this operation should be used with care. >>>> >>>> Are you sure you want to proceed? [y/n] >>>> y >>>> -=== Destroy hosted-engine VM ===- >>>> You must run deploy first >>>> -=== Stop HA services ===- >>>> -=== Shutdown sanlock ===- >>>> shutdown force 1 wait 0 >>>> shutdown done 0 >>>> -=== Disconnecting the hosted-engine storage domain ===- >>>> You must run deploy first >>>> -=== De-configure VDSM networks ===- >>>> -=== Stop other services ===- >>>> -=== De-configure external daemons ===- >>>> -=== Removing configuration files ===- >>>> ? /etc/init/libvirtd.conf already missing >>>> - removing /etc/libvirt/nwfilter/vdsm-no-mac-spoofing.xml >>>> ? /etc/ovirt-hosted-engine/answers.conf already missing >>>> ? /etc/ovirt-hosted-engine/hosted-engine.conf already missing >>>> - removing /etc/vdsm/vdsm.conf >>>> - removing /etc/pki/vdsm/certs/cacert.pem >>>> - removing /etc/pki/vdsm/certs/vdsmcert.pem >>>> - removing /etc/pki/vdsm/keys/vdsmkey.pem >>>> - removing /etc/pki/vdsm/libvirt-spice/ca-cert.pem >>>> - removing /etc/pki/vdsm/libvirt-spice/ca-key.pem >>>> - removing /etc/pki/vdsm/libvirt-spice/server-cert.pem >>>> - removing /etc/pki/vdsm/libvirt-spice/server-key.pem >>>> ? /etc/pki/CA/cacert.pem already missing >>>> ? /etc/pki/libvirt/*.pem already missing >>>> ? /etc/pki/libvirt/private/*.pem already missing >>>> ? /etc/pki/ovirt-vmconsole/*.pem already missing >>>> - removing /var/cache/libvirt/qemu >>>> ? /var/run/ovirt-hosted-engine-ha/* already missing >>>> [root@node-1 ~]# >>>> >>>> Output of sanlock client status: >>>> [root@node-1 ~]# sanlock client status >>>> [root@node-1 ~]# >>>> >>>> Thank you for your help! >>>> >>>> On Mon, Jun 5, 2017 at 7:25 AM, Simone Tiraboschi <stira...@redhat.com> >>>> wrote: >>>> >>>>> >>>>> >>>>> On Mon, Jun 5, 2017 at 3:57 PM, Brendan Hartzell <mrr...@gmail.com> >>>>> wrote: >>>>> >>>>>> After letting this sit for a few days, does anyone have any ideas as >>>>>> to how to deal with my situation? Would anyone like me to send the SOS >>>>>> report directly to them? It's a 9MB file. >>>>>> >>>>>> If nothing comes up, I'm going to try and sift through the SOS report >>>>>> tonight, but I won't know what I'm trying to find. >>>>>> >>>>>> Thank you for any and all help. >>>>>> >>>>>> On Thu, Jun 1, 2017 at 1:15 AM, Sandro Bonazzola <sbona...@redhat.com >>>>>> > wrote: >>>>>> >>>>>>> >>>>>>> >>>>>>> On Thu, Jun 1, 2017 at 6:36 AM, Brendan Hartzell <mrr...@gmail.com> >>>>>>> wrote: >>>>>>> >>>>>>>> Ran the 4 commands listed above, no errors on the screen. >>>>>>>> >>>>>>>> Started the hosted-engine standard setup from the web-UI. >>>>>>>> >>>>>>>> Using iSCSI for the storage. >>>>>>>> >>>>>>>> Using mostly default options, I got these errors in the web-UI. >>>>>>>> >>>>>>>> Error creating Volume Group: Failed to initialize physical device: >>>>>>>> ("[u'/dev/mapper/36589cfc000000de7482638fcfcebbbb4']",) >>>>>>>> Failed to execute stage 'Misc configuration': Failed to initialize >>>>>>>> physical device: ("[u'/dev/mapper/36589cfc00000 >>>>>>>> 0de7482638fcfcebbbb4']",) >>>>>>>> Hosted Engine deployment failed: this system is not reliable, >>>>>>>> please check the issue,fix and redeploy >>>>>>>> >>>>>>>> I rebuilt my iSCSI (I don't think I cleaned it up from a previous >>>>>>>> install). >>>>>>>> Re-ran the above 4 commands. >>>>>>>> Restarted hosted engine standard setup from web-UI. >>>>>>>> Install moved past "Connecting Storage Pool" so I believe the above >>>>>>>> was my fault. >>>>>>>> >>>>>>>> These are the last messages displayed on the web-UI. >>>>>>>> Creating Storage Pool >>>>>>>> Connecting Storage Pool >>>>>>>> Verifying sanlock lockspace initialization >>>>>>>> Creating Image for 'hosted-engine.lockspace' ... >>>>>>>> Image for 'hosted-engine.lockspace' created successfully >>>>>>>> Creating Image for 'hosted-engine.metadata' ... >>>>>>>> Image for 'hosted-engine.metadata' created successfully >>>>>>>> Creating VM Image >>>>>>>> Extracting disk image from OVF archive (could take a few minutes >>>>>>>> depending on archive size) >>>>>>>> Validating pre-allocated volume size >>>>>>>> Uploading volume to data domain (could take a few minutes depending >>>>>>>> on archive size) >>>>>>>> >>>>>>>> At the host terminal, I got the error "watchdog watchdog0: watchdog >>>>>>>> did not stop!" >>>>>>>> Then the host restarted. >>>>>>>> >>>>>>> >>>>>>> Simone, can you help here? >>>>>>> >>>>>>> >>>>> Ok, sorry for the delay. >>>>> The second installation attempt seams fine but it seams that >>>>> ovirt-hosted-engine-cleanup failed stopping sanlock and so the >>>>> watchdog kick in rebooting your system in the middle of deployment >>>>> attempt. >>>>> >>>>> could you please post the output of >>>>> ovirt-hosted-engine-cleanup >>>>> sanlock client status >>>>> ? >>>>> >>>>> >>>>> >>>>>> >>>>>>> >>>>>>> >>>>>>>> >>>>>>>> This is as far as I've gotten in previous attempts. >>>>>>>> >>>>>>>> Attaching the hosted-engine-setup log. >>>>>>>> >>>>>>>> The SOS report is 9MB and the ovirt users group will drop the email. >>>>>>>> >>>>>>>> On Wed, May 31, 2017 at 6:59 AM, Sandro Bonazzola < >>>>>>>> sbona...@redhat.com> wrote: >>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> On Wed, May 31, 2017 at 3:10 PM, Brendan Hartzell < >>>>>>>>> mrr...@gmail.com> wrote: >>>>>>>>> >>>>>>>>>> Now that you have identified the problem, should I run the >>>>>>>>>> following commands and send you another SOS? >>>>>>>>>> >>>>>>>>>> ovirt-hosted-engine-cleanup >>>>>>>>>> vdsm-tool configure --force >>>>>>>>>> systemctl restart libvirtd >>>>>>>>>> systemctl restart vdsm >>>>>>>>>> >>>>>>>>>> Or is there a different plan in mind? >>>>>>>>>> >>>>>>>>> >>>>>>>>> I would have expected someone from virt team to follow up for >>>>>>>>> further investigations :-) >>>>>>>>> above commands should work. >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>>> >>>>>>>>>> Thank you, >>>>>>>>>> >>>>>>>>>> Brendan >>>>>>>>>> >>>>>>>>>> On Tue, May 30, 2017 at 11:42 PM, Sandro Bonazzola < >>>>>>>>>> sbona...@redhat.com> wrote: >>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> On Wed, May 31, 2017 at 4:45 AM, Brendan Hartzell < >>>>>>>>>>> mrr...@gmail.com> wrote: >>>>>>>>>>> >>>>>>>>>>>> Can you please elaborate about the failure you see here and how >>>>>>>>>>>> are you trying to manually partition the host? >>>>>>>>>>>> >>>>>>>>>>>> Sure, I will start from the beginning. >>>>>>>>>>>> - Using: ovirt-node-ng-installer-ovirt-4.1-2017052604 >>>>>>>>>>>> <(201)%20705-2604>.iso >>>>>>>>>>>> - During installation I setup one of the two interfaces and >>>>>>>>>>>> check the box to automatically use the connection. >>>>>>>>>>>> - I'm currently providing a host name of node-1.test.net until >>>>>>>>>>>> I have a successful process. >>>>>>>>>>>> - I configure date and time for my timezone and to use an >>>>>>>>>>>> internal NTP server. >>>>>>>>>>>> - On Installation Destination, I pick my 128GB USB3.0 SanDisk >>>>>>>>>>>> flash drive, check the box that I would like to make additional >>>>>>>>>>>> space, and >>>>>>>>>>>> click done. In the reclaim disk space window, I click delete all, >>>>>>>>>>>> and then >>>>>>>>>>>> reclaim space. I go back into the Installation Destination, >>>>>>>>>>>> select that I >>>>>>>>>>>> will configure partitioning, and click done. The Manual >>>>>>>>>>>> Partitioning >>>>>>>>>>>> window opens, I use the option to automatically create mount >>>>>>>>>>>> points. >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> In this screen, please change partitioning scheme from LVM to >>>>>>>>>>> LVM Thin Provisioning: it should solve your following error. >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>>> At this point, /boot is 1024MB, /var is 15GB, / is 88.11 GB, >>>>>>>>>>>> and swap is 11.57GB. I then change / to 23.11 GB, update >>>>>>>>>>>> settings, change >>>>>>>>>>>> /var to 80GB, update settings again, and click done. I accept the >>>>>>>>>>>> changes >>>>>>>>>>>> and begin installation. >>>>>>>>>>>> >>>>>>>>>>>> I tried these changes based on this article: >>>>>>>>>>>> http://www.ovirt.org/documentation/self-hosted/chap >>>>>>>>>>>> -Deploying_Self-Hosted_Engine/ >>>>>>>>>>>> >>>>>>>>>>>> The article does say that you can specify a different directory >>>>>>>>>>>> than /var/tmp, but I don't recall seeing that option. >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> If the setup detects not enough space in /var/tmp for extracting >>>>>>>>>>> the appliance it will ask about a different directory. >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> After some time, I get the following error: >>>>>>>>>>>> There was an error running the kickstart script at line 7. >>>>>>>>>>>> This is a fatal error and installation will be aborted. The >>>>>>>>>>>> details of >>>>>>>>>>>> this error are: >>>>>>>>>>>> >>>>>>>>>>>> [INFO] Trying to create a manageable base from '/' >>>>>>>>>>>> [ERROR] LVM Thin Provisioning partitioning scheme is required. >>>>>>>>>>>> For autoinstall via Kickstart with LVM Thin Provisioning check >>>>>>>>>>>> options >>>>>>>>>>>> --thinpool and --grow. Please consult documentation for details. >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> ^^ this one should be solved by the LVM Thin Provisioning scheme >>>>>>>>>>> mentioned above.. >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> Traceback (most recent call last): >>>>>>>>>>>> File "/usr/lib64/python2.7/runpy.py", line 162, in >>>>>>>>>>>> _run_module_as_main "__main__", fname, loader, pkg_name) >>>>>>>>>>>> File "/usr/lib64/python2.7runpy.py", line 72, in _run_code >>>>>>>>>>>> exec code in run_globals >>>>>>>>>>>> File "/usr/lib/python2.7/site-packages/imgbased/__main__.py", >>>>>>>>>>>> line 51, in <module> CliApplication() >>>>>>>>>>>> File "/usr/lib/python2.7/site-packages/imgbased/__init__.py", >>>>>>>>>>>> line 82, in CliApplication() >>>>>>>>>>>> File "/usr/lib/python2.7/site-packages/imgbased/hooks.py", >>>>>>>>>>>> line 120, in emit cb(self.context, *args) >>>>>>>>>>>> File "/usr/lib/python2.7/site-packages/imgbased/plugins/core.py", >>>>>>>>>>>> line 169, in post_argparse layout.initialize(args.source, >>>>>>>>>>>> args.init_nvr) >>>>>>>>>>>> File "/usr/lib/python2.7/site-packages/imgbased/plugins/core.py", >>>>>>>>>>>> line 216, in initialize self.app.imgbase.init_layout_from(source, >>>>>>>>>>>> init_nvr) >>>>>>>>>>>> File "/usr/lib/python2.7/site-packages/imgbased/imgbase.py", >>>>>>>>>>>> line 271, in init_layout_from self.init_tags_on(existing_lv) >>>>>>>>>>>> File "/usr/lib/python2.7/site-packages/imgbased/imgbase.py", >>>>>>>>>>>> line 243, in init_tags_on pool = lv.thinpool() >>>>>>>>>>>> File "/usr/lib/python2.7/site-packages/imgbased/lvm.py", line >>>>>>>>>>>> 250, in thinpool raise MissingLvmThinPool() >>>>>>>>>>>> imgbased.lvm.MissingLvmThinPoo >>>>>>>>>>>> l >>>>>>>>>>>> >>>>>>>>>>>> At this point, the only option is to exit the installer. >>>>>>>>>>>> >>>>>>>>>>>> **************************** >>>>>>>>>>>> >>>>>>>>>>>> Being this a new install, please use 4.1. oVirt 4.0 is not >>>>>>>>>>>> supported anymore. >>>>>>>>>>>> >>>>>>>>>>>> Not a problem. >>>>>>>>>>>> >>>>>>>>>>>> **************************** >>>>>>>>>>>> >>>>>>>>>>>> Can you please provide hosted engine setup logs or better a >>>>>>>>>>>> full sos report? (sosreport -a) >>>>>>>>>>>> >>>>>>>>>>>> Again, the process I'm following: >>>>>>>>>>>> - Using: ovirt-node-ng-installer-ovirt-4.1-2017052604 >>>>>>>>>>>> <(201)%20705-2604>.iso >>>>>>>>>>>> - During installation I setup one of the two interfaces and >>>>>>>>>>>> check the box to automatically use the connection. >>>>>>>>>>>> - I'm currently providing a host name of node-1.test.net until >>>>>>>>>>>> I have a successful process. >>>>>>>>>>>> - I configure date and time for my timezone and to use an >>>>>>>>>>>> internal NTP server. >>>>>>>>>>>> - On Installation Destination, I pick my 128GB USB3.0 SanDisk >>>>>>>>>>>> flash drive, check the box that I would like to make additional >>>>>>>>>>>> space, and >>>>>>>>>>>> click done. In the reclaim disk space window, I click delete all, >>>>>>>>>>>> and then >>>>>>>>>>>> reclaim space. >>>>>>>>>>>> - Begin Installation and set a root password. >>>>>>>>>>>> - Perform a yum update - no packages marked for update (as >>>>>>>>>>>> expected) >>>>>>>>>>>> - Use vi to update /etc/hosts with a reference for >>>>>>>>>>>> node-1.test.net and engine.test.net >>>>>>>>>>>> - First attempt at hosted-engine from web-UI >>>>>>>>>>>> - Setup downloads and installs ovirt-engine-applianc >>>>>>>>>>>> e-4.1-20170523.1.el7.centos.noarch.rpm >>>>>>>>>>>> *Failed to execute stage 'Environment setup': Failed to >>>>>>>>>>>> reconfigure libvirt for VDSM >>>>>>>>>>>> *Hosted Engine deployment failed >>>>>>>>>>>> - Attached SOS report >>>>>>>>>>>> The checksum is: aa56097edc0b63c49caaf1a1fde021bc >>>>>>>>>>>> >>>>>>>>>>>> At this point, I would run ovirt-hosted-engine-cleanup and I >>>>>>>>>>>> would get further along in the install process. However, because >>>>>>>>>>>> this is a >>>>>>>>>>>> fresh install, I'm going to leave things here for now so you can >>>>>>>>>>>> review the >>>>>>>>>>>> SOS. >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> Thanks for the SOS report! >>>>>>>>>>> Hosted Engine setup fails on: >>>>>>>>>>> >>>>>>>>>>> 2017-05-30 19:24:39 DEBUG otopi.plugins.gr_he_setup.system.vdsmenv >>>>>>>>>>> plugin.execute:921 execute-output: ('/bin/vdsm-tool', 'configure', >>>>>>>>>>> '--force') stdout: >>>>>>>>>>> >>>>>>>>>>> Checking configuration status... >>>>>>>>>>> >>>>>>>>>>> Current revision of multipath.conf detected, preserving >>>>>>>>>>> lvm is configured for vdsm >>>>>>>>>>> libvirt is already configured for vdsm >>>>>>>>>>> SUCCESS: ssl configured to true. No conflicts >>>>>>>>>>> >>>>>>>>>>> Running configure... >>>>>>>>>>> Reconfiguration of libvirt is done. >>>>>>>>>>> >>>>>>>>>>> 2017-05-30 19:24:39 DEBUG otopi.plugins.gr_he_setup.system.vdsmenv >>>>>>>>>>> plugin.execute:926 execute-output: ('/bin/vdsm-tool', 'configure', >>>>>>>>>>> '--force') stderr: >>>>>>>>>>> Error: ServiceOperationError: _systemctlStart failed >>>>>>>>>>> Job for libvirtd.service failed because the control process >>>>>>>>>>> exited with error code. See "systemctl status libvirtd.service" and >>>>>>>>>>> "journalctl -xe" for details. >>>>>>>>>>> >>>>>>>>>>> At the same time journalctl shows: >>>>>>>>>>> >>>>>>>>>>> May 30 19:24:39 node-1.test.net libvirtd[20954]: libvirt >>>>>>>>>>> version: 2.0.0, package: 10.el7_3.5 (CentOS BuildSystem < >>>>>>>>>>> http://bugs.centos.org>, 2017-03-03-02:09:45, >>>>>>>>>>> c1bm.rdu2.centos.org) >>>>>>>>>>> May 30 19:24:39 node-1.test.net libvirtd[20954]: hostname: >>>>>>>>>>> node-1.test.net >>>>>>>>>>> May 30 19:24:39 node-1.test.net libvirtd[20954]: The server >>>>>>>>>>> certificate /etc/pki/vdsm/certs/vdsmcert.pem is not yet active >>>>>>>>>>> May 30 19:24:39 node-1.test.net systemd[1]: libvirtd.service: >>>>>>>>>>> main process exited, code=exited, status=6/NOTCONFIGURED >>>>>>>>>>> May 30 19:24:39 node-1.test.net systemd[1]: Failed to start >>>>>>>>>>> Virtualization daemon. >>>>>>>>>>> May 30 19:24:39 node-1.test.net systemd[1]: Unit >>>>>>>>>>> libvirtd.service entered failed state. >>>>>>>>>>> May 30 19:24:39 node-1.test.net systemd[1]: libvirtd.service >>>>>>>>>>> failed. >>>>>>>>>>> May 30 19:24:39 node-1.test.net systemd[1]: libvirtd.service >>>>>>>>>>> holdoff time over, scheduling restart. >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> **************************** >>>>>>>>>>>> >>>>>>>>>>>> I'd like to understand the issues you faced before suggesting >>>>>>>>>>>> to restart from scratch. >>>>>>>>>>>> >>>>>>>>>>>> Too late... I did two re-installs to get a more accurate >>>>>>>>>>>> account of my install process for above. >>>>>>>>>>>> >>>>>>>>>>>> **************************** >>>>>>>>>>>> >>>>>>>>>>>> Thank you for your help! >>>>>>>>>>>> >>>>>>>>>>>> Brendan >>>>>>>>>>>> >>>>>>>>>>>> On Tue, May 30, 2017 at 12:17 AM, Sandro Bonazzola < >>>>>>>>>>>> sbona...@redhat.com> wrote: >>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> On Tue, May 30, 2017 at 6:49 AM, Brendan Hartzell < >>>>>>>>>>>>> mrr...@gmail.com> wrote: >>>>>>>>>>>>> >>>>>>>>>>>>>> oVirt users list, >>>>>>>>>>>>>> >>>>>>>>>>>>>> Long story short, I've been spending weeks on this project >>>>>>>>>>>>>> for my home lab with no success. >>>>>>>>>>>>>> >>>>>>>>>>>>>> I would like to successfully install two nodes that host a >>>>>>>>>>>>>> highly available engine with an iSCSI storage back-end. >>>>>>>>>>>>>> >>>>>>>>>>>>>> I have read through most, if not all, of the guides on >>>>>>>>>>>>>> ovirt.org with no substantial help. >>>>>>>>>>>>>> >>>>>>>>>>>>>> Successfully, I have done the following: >>>>>>>>>>>>>> Install oVirt Engine on a bare metal system, added a node, >>>>>>>>>>>>>> and started exploring - not desired. >>>>>>>>>>>>>> Install oVirt Node 4.0.6 on a bare metal system - fails if >>>>>>>>>>>>>> partitions are not done automatically. >>>>>>>>>>>>>> Install oVirt Node 4.1.2 on a bare metal system - fails if >>>>>>>>>>>>>> partitions are not done automatically. >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> Can you please elaborate about the failure you see here and >>>>>>>>>>>>> how are you trying to manually partition the host? >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> My process after installing a Node: >>>>>>>>>>>>>> Run a yum update - just to be sure, but I am using latest iso >>>>>>>>>>>>>> images from downloads section. >>>>>>>>>>>>>> Edit /etc/hosts for local name resolution - the goal is to >>>>>>>>>>>>>> host DNS as a virtual machine, eventually. >>>>>>>>>>>>>> On 4.1 if I install ovirt-engine-appliance from yum, it does >>>>>>>>>>>>>> simplify one step in the hosted engine setup. If I do this on >>>>>>>>>>>>>> 4.0 it >>>>>>>>>>>>>> discards the image and uses the default. >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> Being this a new install, please use 4.1. oVirt 4.0 is not >>>>>>>>>>>>> supported anymore. >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>>> On 4.1 the hosted engine setup fails immediately unless I run >>>>>>>>>>>>>> the hosted engine cleanup from the shell. >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> Can you please provide hosted engine setup logs or better a >>>>>>>>>>>>> full sos report? (sosreport -a) >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>>> If I do this, I can typically get to the point of >>>>>>>>>>>>>> installing. >>>>>>>>>>>>>> >>>>>>>>>>>>>> When I do get to the installation phase, I get to a point >>>>>>>>>>>>>> just after extracting the OVA that I get a message on the shell >>>>>>>>>>>>>> saying >>>>>>>>>>>>>> something about the watchdog running the whole time and then the >>>>>>>>>>>>>> node >>>>>>>>>>>>>> reboots. >>>>>>>>>>>>>> >>>>>>>>>>>>>> I found one email thread that sounded like my issue and >>>>>>>>>>>>>> suggested the following commands: >>>>>>>>>>>>>> vdsm-tool configure --force >>>>>>>>>>>>>> systemctl restart libvirtd >>>>>>>>>>>>>> systemctl restart vdsmd >>>>>>>>>>>>>> >>>>>>>>>>>>>> Unfortunately, these commands did not help my situation like >>>>>>>>>>>>>> the other individual. >>>>>>>>>>>>>> >>>>>>>>>>>>>> What log file would everyone like to see first? Given that I >>>>>>>>>>>>>> still consider myself relatively new to Linux, please identify >>>>>>>>>>>>>> the path for >>>>>>>>>>>>>> the log file requested. >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> See above >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> Also, because I plan on performing a clean install for this >>>>>>>>>>>>>> thread using my process from above (I'm not expecting my outcome >>>>>>>>>>>>>> to be any >>>>>>>>>>>>>> different), are there any tips and tricks that might result in a >>>>>>>>>>>>>> success? >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> I'd like to understand the issues you faced before suggesting >>>>>>>>>>>>> to restart from scratch. >>>>>>>>>>>>> Adding some people who may help as well. >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> Thank you for any and all help, >>>>>>>>>>>>>> Brendan >>>>>>>>>>>>>> >>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>> Users mailing list >>>>>>>>>>>>>> Users@ovirt.org >>>>>>>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> -- >>>>>>>>>>>>> >>>>>>>>>>>>> SANDRO BONAZZOLA >>>>>>>>>>>>> >>>>>>>>>>>>> ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG >>>>>>>>>>>>> VIRTUALIZATION R&D >>>>>>>>>>>>> >>>>>>>>>>>>> Red Hat EMEA <https://www.redhat.com/> >>>>>>>>>>>>> <https://red.ht/sig> >>>>>>>>>>>>> TRIED. TESTED. TRUSTED. <https://redhat.com/trusted> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> -- >>>>>>>>>>> >>>>>>>>>>> SANDRO BONAZZOLA >>>>>>>>>>> >>>>>>>>>>> ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG VIRTUALIZATION >>>>>>>>>>> R&D >>>>>>>>>>> >>>>>>>>>>> Red Hat EMEA <https://www.redhat.com/> >>>>>>>>>>> <https://red.ht/sig> >>>>>>>>>>> TRIED. TESTED. TRUSTED. <https://redhat.com/trusted> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> -- >>>>>>>>> >>>>>>>>> SANDRO BONAZZOLA >>>>>>>>> >>>>>>>>> ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG VIRTUALIZATION >>>>>>>>> R&D >>>>>>>>> >>>>>>>>> Red Hat EMEA <https://www.redhat.com/> >>>>>>>>> <https://red.ht/sig> >>>>>>>>> TRIED. TESTED. TRUSTED. <https://redhat.com/trusted> >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>> >>>>>>> >>>>>>> -- >>>>>>> >>>>>>> SANDRO BONAZZOLA >>>>>>> >>>>>>> ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG VIRTUALIZATION R&D >>>>>>> >>>>>>> Red Hat EMEA <https://www.redhat.com/> >>>>>>> <https://red.ht/sig> >>>>>>> TRIED. TESTED. TRUSTED. <https://redhat.com/trusted> >>>>>>> >>>>>> >>>>>> >>>>> >>>> >>> >> > > > -- > > SANDRO BONAZZOLA > > ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG VIRTUALIZATION R&D > > Red Hat EMEA <https://www.redhat.com/> > <https://red.ht/sig> > TRIED. TESTED. TRUSTED. <https://redhat.com/trusted> >
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users