Can you please check if on your host you have /var/log/ovirt-imageio-daemon and its ownership and permissions (it should be vdsm:kvm,700)? Can you please report which version of ovirt-imageio-daemon you are using? We had a bug there but it has been fixed long time ago.
On Wed, Sep 5, 2018 at 12:04 PM Sakhi Hadebe <sa...@sanren.ac.za> wrote: > Sorry, I mistakenly send the email: > > Below is the output of: > [root@glustermount ~]# systemctl status ovirt-imageio-daemon.service -l > ● ovirt-imageio-daemon.service - oVirt ImageIO Daemon > Loaded: loaded (/usr/lib/systemd/system/ovirt-imageio-daemon.service; > disabled; vendor preset: disabled) > Active: failed (Result: start-limit) since Tue 2018-09-04 16:55:16 > SAST; 19h ago > Condition: start condition failed at Wed 2018-09-05 11:56:46 SAST; 2min 9s > ago > ConditionPathExists=/etc/pki/vdsm/certs/vdsmcert.pem was not met > Process: 11345 ExecStart=/usr/bin/ovirt-imageio-daemon (code=exited, > status=1/FAILURE) > Main PID: 11345 (code=exited, status=1/FAILURE) > > Sep 04 16:55:16 glustermount.goku systemd[1]: Failed to start oVirt > ImageIO Daemon. > Sep 04 16:55:16 glustermount.goku systemd[1]: Unit > ovirt-imageio-daemon.service entered failed state. > Sep 04 16:55:16 glustermount.goku systemd[1]: ovirt-imageio-daemon.service > failed. > Sep 04 16:55:16 glustermount.goku systemd[1]: ovirt-imageio-daemon.service > holdoff time over, scheduling restart. > Sep 04 16:55:16 glustermount.goku systemd[1]: start request repeated too > quickly for ovirt-imageio-daemon.service > Sep 04 16:55:16 glustermount.goku systemd[1]: Failed to start oVirt > ImageIO Daemon. > Sep 04 16:55:16 glustermount.goku systemd[1]: Unit > ovirt-imageio-daemon.service entered failed state. > Sep 04 16:55:16 glustermount.goku systemd[1]: ovirt-imageio-daemon.service > failed. > [root@glustermount ~]# journalctl -xe -u ovirt-imageio-daemon.service > Sep 04 16:55:16 glustermount.goku ovirt-imageio-daemon[11345]: File > "/usr/lib64/python2.7/logging/handlers.py", > Sep 04 16:55:16 glustermount.goku ovirt-imageio-daemon[11345]: > BaseRotatingHandler.__init__(self, filename, mode > Sep 04 16:55:16 glustermount.goku ovirt-imageio-daemon[11345]: File > "/usr/lib64/python2.7/logging/handlers.py", > Sep 04 16:55:16 glustermount.goku ovirt-imageio-daemon[11345]: > logging.FileHandler.__init__(self, filename, mode > Sep 04 16:55:16 glustermount.goku ovirt-imageio-daemon[11345]: File > "/usr/lib64/python2.7/logging/__init__.py", > Sep 04 16:55:16 glustermount.goku ovirt-imageio-daemon[11345]: > StreamHandler.__init__(self, self._open()) > Sep 04 16:55:16 glustermount.goku ovirt-imageio-daemon[11345]: File > "/usr/lib64/python2.7/logging/__init__.py", > Sep 04 16:55:16 glustermount.goku ovirt-imageio-daemon[11345]: stream = > open(self.baseFilename, self.mode) > Sep 04 16:55:16 glustermount.goku ovirt-imageio-daemon[11345]: IOError: > [Errno 2] No such file or directory: '/v > Sep 04 16:55:16 glustermount.goku systemd[1]: > ovirt-imageio-daemon.service: main process exited, code=exited, st > Sep 04 16:55:16 glustermount.goku systemd[1]: Failed to start oVirt > ImageIO Daemon. > -- Subject: Unit ovirt-imageio-daemon.service has failed > -- Defined-By: systemd > -- Support: http://lists.freedesktop.org/mailman/listinfo/systemd-devel > -- > -- Unit ovirt-imageio-daemon.service has failed. > -- > -- The result is failed. > Sep 04 16:55:16 glustermount.goku systemd[1]: Unit > ovirt-imageio-daemon.service entered failed state. > Sep 04 16:55:16 glustermount.goku systemd[1]: ovirt-imageio-daemon.service > failed. > Sep 04 16:55:16 glustermount.goku systemd[1]: ovirt-imageio-daemon.service > holdoff time over, scheduling restart > Sep 04 16:55:16 glustermount.goku systemd[1]: start request repeated too > quickly for ovirt-imageio-daemon.servic > Sep 04 16:55:16 glustermount.goku systemd[1]: Failed to start oVirt > ImageIO Daemon. > -- Subject: Unit ovirt-imageio-daemon.service has failed > -- Defined-By: systemd > -- Support: http://lists.freedesktop.org/mailman/listinfo/systemd-devel > -- > -- Unit ovirt-imageio-daemon.service has failed. > -- > -- The result is failed. > Sep 04 16:55:16 glustermount.goku systemd[1]: Unit > ovirt-imageio-daemon.service entered failed state. > Sep 04 16:55:16 glustermount.goku systemd[1]: ovirt-imageio-daemon.service > failed. > > > On Wed, Sep 5, 2018 at 12:01 PM, Sakhi Hadebe <sa...@sanren.ac.za> wrote: > >> # systemctl status ovirt-imageio-daemon.service >> ● ovirt-imageio-daemon.service - oVirt ImageIO Daemon >> Loaded: loaded (/usr/lib/systemd/system/ovirt-imageio-daemon.service; >> disabled; vendor preset: disabled) >> Active: failed (Result: start-limit) since Tue 2018-09-04 16:55:16 >> SAST; 19h ago >> Condition: start condition failed at Wed 2018-09-05 11:56:46 SAST; 1min >> 58s ago >> ConditionPathExists=/etc/pki/vdsm/certs/vdsmcert.pem was not >> met >> Process: 11345 ExecStart=/usr/bin/ovirt-imageio-daemon (code=exited, >> status=1/FAILURE) >> Main PID: 11345 (code=exited, status=1/FAILURE) >> >> Sep 04 16:55:16 glustermount.goku systemd[1]: Failed to start oVirt >> ImageIO Daemon. >> Sep 04 16:55:16 glustermount.goku systemd[1]: Unit >> ovirt-imageio-daemon.service entered failed state. >> Sep 04 16:55:16 glustermount.goku systemd[1]: >> ovirt-imageio-daemon.service failed. >> Sep 04 16:55:16 glustermount.goku systemd[1]: >> ovirt-imageio-daemon.service holdoff time over, scheduling ...art. >> Sep 04 16:55:16 glustermount.goku systemd[1]: start request repeated too >> quickly for ovirt-imageio-daemon...vice >> Sep 04 16:55:16 glustermount.goku systemd[1]: Failed to start oVirt >> ImageIO Daemon. >> Sep 04 16:55:16 glustermount.goku systemd[1]: Unit >> ovirt-imageio-daemon.service entered failed state. >> Sep 04 16:55:16 glustermount.goku systemd[1]: >> ovirt-imageio-daemon.service failed. >> Hint: Some lines were ellipsized, use -l to show in full. >> [root@glustermount ~]# systemctl status ovirt-imageio-daemon.service -l >> ● ovirt-imageio-daemon.service - oVirt ImageIO Daemon >> Loaded: loaded (/usr/lib/systemd/system/ovirt-imageio-daemon.service; >> disabled; vendor preset: disabled) >> Active: failed (Result: start-limit) since Tue 2018-09-04 16:55:16 >> SAST; 19h ago >> Condition: start condition failed at Wed 2018-09-05 11:56:46 SAST; 2min >> 9s ago >> ConditionPathExists=/etc/pki/vdsm/certs/vdsmcert.pem was not >> met >> Process: 11345 ExecStart=/usr/bin/ovirt-imageio-daemon (code=exited, >> status=1/FAILURE) >> Main PID: 11345 (code=exited, status=1/FAILURE) >> >> Sep 04 16:55:16 glustermount.goku systemd[1]: Failed to start oVirt >> ImageIO Daemon. >> Sep 04 16:55:16 glustermount.goku systemd[1]: Unit >> ovirt-imageio-daemon.service entered failed state. >> Sep 04 16:55:16 glustermount.goku systemd[1]: >> ovirt-imageio-daemon.service failed. >> Sep 04 16:55:16 glustermount.goku systemd[1]: >> ovirt-imageio-daemon.service holdoff time over, scheduling restart. >> Sep 04 16:55:16 glustermount.goku systemd[1]: start request repeated too >> quickly for ovirt-imageio-daemon.service >> Sep 04 16:55:16 glustermount.goku systemd[1]: Failed to start oVirt >> ImageIO Daemon. >> Sep 04 16:55:16 glustermount.goku systemd[1]: Unit >> ovirt-imageio-daemon.service entered failed state. >> Sep 04 16:55:16 glustermount.goku systemd[1]: >> ovirt-imageio-daemon.service failed. >> >> Output of: >> >> On Wed, Sep 5, 2018 at 11:35 AM, Simone Tiraboschi <stira...@redhat.com> >> wrote: >> >>> >>> >>> On Wed, Sep 5, 2018 at 11:10 AM Sakhi Hadebe <sa...@sanren.ac.za> wrote: >>> >>>> Hi All, >>>> >>>> The host deploy logs are showing the below errors: >>>> >>>> [root@garlic engine-logs-2018-09-05T08:48:22Z]# cat >>>> /var/log/ovirt-hosted-engine-setup/engine-logs-2018-09-05T08\:34\:55Z/ovirt-engine/host-deploy/ovirt-host-deploy-20180905103605-garlic.sanren.ac.za-543b536b.log >>>> | grep -i error >>>> 2018-09-05 10:35:46,909+0200 DEBUG otopi.context >>>> context.dumpEnvironment:869 ENV BASE/error=bool:'False' >>>> 2018-09-05 10:35:47,116 [ERROR] __main__.py:8011:MainThread >>>> @identity.py:145 - Reload of consumer identity cert >>>> /etc/pki/consumer/cert.pem raised an exception with msg: [Errno 2] No such >>>> file or directory: '/etc/pki/consumer/key.pem' >>>> 2018-09-05 10:35:47,383+0200 DEBUG otopi.context >>>> context.dumpEnvironment:869 ENV BASE/error=bool:'False' >>>> 2018-09-05 10:35:47,593 [ERROR] __main__.py:8011:MainThread >>>> @identity.py:145 - Reload of consumer identity cert >>>> /etc/pki/consumer/cert.pem raised an exception with msg: [Errno 2] No such >>>> file or directory: '/etc/pki/consumer/key.pem' >>>> 2018-09-05 10:35:48,245+0200 DEBUG otopi.context >>>> context.dumpEnvironment:869 ENV BASE/error=bool:'False' >>>> Job for ovirt-imageio-daemon.service failed because the control process >>>> exited with error code. See "systemctl status ovirt-imageio-daemon.service" >>>> and "journalctl -xe" for details. >>>> RuntimeError: Failed to start service 'ovirt-imageio-daemon' >>>> 2018-09-05 10:36:05,098+0200 ERROR otopi.context >>>> context._executeMethod:152 Failed to execute stage 'Closing up': Failed to >>>> start service 'ovirt-imageio-daemon' >>>> 2018-09-05 10:36:05,099+0200 DEBUG otopi.context >>>> context.dumpEnvironment:869 ENV BASE/error=bool:'True' >>>> 2018-09-05 10:36:05,099+0200 DEBUG otopi.context >>>> context.dumpEnvironment:869 ENV BASE/exceptionInfo=list:'[(<type >>>> 'exceptions.RuntimeError'>, RuntimeError("Failed to start service >>>> 'ovirt-imageio-daemon'",), <traceback object at 0x7f4de25ff320>)]' >>>> 2018-09-05 10:36:05,106+0200 DEBUG otopi.context >>>> context.dumpEnvironment:869 ENV BASE/error=bool:'True' >>>> 2018-09-05 10:36:05,106+0200 DEBUG otopi.context >>>> context.dumpEnvironment:869 ENV BASE/exceptionInfo=list:'[(<type >>>> 'exceptions.RuntimeError'>, RuntimeError("Failed to start service >>>> 'ovirt-imageio-daemon'",), <traceback object at 0x7f4de25ff320>)]' >>>> >>>> I coudn't find anything helpful from the internet. >>>> >>> >>> something relevant int he output of >>> systemctl status ovirt-imageio-daemon.service >>> and >>> journalctl -xe -u ovirt-imageio-daemon.service >>> ? >>> >>> >>> >>>> >>>> On Tue, Sep 4, 2018 at 6:46 PM, Simone Tiraboschi <stira...@redhat.com> >>>> wrote: >>>> >>>>> >>>>> >>>>> On Tue, Sep 4, 2018 at 6:07 PM Sakhi Hadebe <sa...@sanren.ac.za> >>>>> wrote: >>>>> >>>>>> Hi Sahina, >>>>>> >>>>>> I am sorry I can't reproduce the error nor access the logs since I >>>>>> did a fresh installed pn nodes. However now I can't even react that far >>>>>> because the engine deployment fails to start the host up: >>>>>> >>>>>> >>>>>> [ INFO ] TASK [Wait for the host to be up] >>>>>> [ ERROR ] fatal: [localhost]: FAILED! => {"ansible_facts": >>>>>> {"ovirt_hosts": [{"address": "goku.sanren.ac.za", "affinity_labels": >>>>>> [], "auto_numa_status": "unknown", "certificate": {"organization": " >>>>>> sanren.ac.za", "subject": "O=sanren.ac.za,CN=goku.sanren.ac.za"}, >>>>>> "cluster": {"href": "/ovirt-engine/api/clusters/ >>>>>> 1ca368cc-b052-11e8-b7de-00163e008187", "id": >>>>>> "1ca368cc-b052-11e8-b7de-00163e008187"}, "comment": "", "cpu": >>>>>> {"speed": 0.0, "topology": {}}, "device_passthrough": {"enabled": false}, >>>>>> "devices": [], "external_network_provider_configurations": [], >>>>>> "external_status": "ok", "hardware_information": >>>>>> {"supported_rng_sources": >>>>>> []}, "hooks": [], "href": "/ovirt-engine/api/hosts/ >>>>>> 1c575995-70b1-43f7-b348-4a9788e070cd", "id": >>>>>> "1c575995-70b1-43f7-b348-4a9788e070cd", "katello_errata": [], >>>>>> "kdump_status": "unknown", "ksm": {"enabled": false}, >>>>>> "max_scheduling_memory": 0, "memory": 0, "name": "goku.sanren.ac.za", >>>>>> "network_attachments": [], "nics": [], "numa_nodes": [], >>>>>> "numa_supported": >>>>>> false, "os": {"custom_kernel_cmdline": ""}, "permissions": [], "port": >>>>>> 54321, "power_management": {"automatic_pm_enabled": true, "enabled": >>>>>> false, >>>>>> "kdump_detection": true, "pm_proxies": []}, "protocol": "stomp", >>>>>> "se_linux": {}, "spm": {"priority": 5, "status": "none"}, "ssh": >>>>>> {"fingerprint": "SHA256:B3/PDH551EFid93fm6PoRryi6/cXuVE8yNgiiiROh84", >>>>>> "port": 22}, "statistics": [], "status": "install_failed", >>>>>> "storage_connection_extensions": [], "summary": {"total": 0}, >>>>>> "tags": [], "transparent_huge_pages": {"enabled": false}, "type": >>>>>> "ovirt_node", "unmanaged_networks": [], "update_available": false}]}, >>>>>> "attempts": 120, "changed": false} >>>>>> >>>>>> "status": "install_failed" >>>>> >>>>> You have to check host-deploy logs to get a details error message. >>>>> >>>>> >>>>>> >>>>>> Please help. >>>>>> >>>>>> On Mon, Sep 3, 2018 at 1:34 PM, Sahina Bose <sab...@redhat.com> >>>>>> wrote: >>>>>> >>>>>>> >>>>>>> >>>>>>> On Wed, Aug 29, 2018 at 8:39 PM, Sakhi Hadebe <sa...@sanren.ac.za> >>>>>>> wrote: >>>>>>> >>>>>>>> Hi, >>>>>>>> >>>>>>>> I am sorry to bother you again. >>>>>>>> >>>>>>>> I am trying to deploy an oVirt engine for oVirtNode-4.2.5.1. I get >>>>>>>> the same error I encountered before: >>>>>>>> >>>>>>>> [ INFO ] TASK [Add glusterfs storage domain] >>>>>>>> [ ERROR ] Error: Fault reason is "Operation Failed". Fault detail >>>>>>>> is "[Problem while trying to mount target]". HTTP response code is >>>>>>>> 400. >>>>>>>> [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": >>>>>>>> "Fault reason is \"Operation Failed\". Fault detail is \"[Problem >>>>>>>> while trying to mount target]\". HTTP response code is 400."} >>>>>>>> Please specify the storage you would like to use >>>>>>>> (glusterfs, iscsi, fc, nfs)[nfs]: >>>>>>>> >>>>>>>> The glusterd daemon is running. >>>>>>>> >>>>>>> >>>>>>> mounting 172.16.4.18:/engine at >>>>>>> /rhev/data-center/mnt/glusterSD/172.16.4.18:_engine (mount:204) >>>>>>> 2018-08-29 16:47:28,846+0200 ERROR (jsonrpc/3) [storage.HSM] Could >>>>>>> not connect to storageServer (hsm:2398) >>>>>>> >>>>>>> Can you try to see if you are able to mount 172.16.4.18:/engine on >>>>>>> the server you're deploying Hosted Engine using "mount -t glusterfs >>>>>>> 172.16.4.18:/engine /mnt/test" >>>>>>> >>>>>>> >>>>>>>> During the deployment of the engine it sets the engine entry in the >>>>>>>> /etc/hosts file with the IP Address of 192.168.124.* which it gets >>>>>>>> form the >>>>>>>> virbr0 bridge interface. I stopped the bridge and deleted it, but still >>>>>>>> giving the same error. Not sure what causes it to use that interface. >>>>>>>> Please help! >>>>>>>> >>>>>>>> But I give the engine an IP of 192.168.1.10 same subnet as my >>>>>>>> gateway and my ovirtmgmt bridge. Attached is the ifconfig output of my >>>>>>>> Node, engine.log and vdsm.log. >>>>>>>> >>>>>>>> Your assistance is always appreciated >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> On Wed, Jul 11, 2018 at 11:47 AM, Sahina Bose <sab...@redhat.com> >>>>>>>> wrote: >>>>>>>> >>>>>>>>> Is glusterd running on the server: goku.sanren.** >>>>>>>>> There's an error >>>>>>>>> Failed to get volume info: Command execution failed >>>>>>>>> error: Connection failed. Please check if gluster daemon is >>>>>>>>> operational >>>>>>>>> >>>>>>>>> Please check the volume status using "gluster volume status engine" >>>>>>>>> >>>>>>>>> and if all looks ok, attach the mount logs from /var/log/glusterfs >>>>>>>>> >>>>>>>>> On Wed, Jul 11, 2018 at 1:57 PM, Sakhi Hadebe <sa...@sanren.ac.za> >>>>>>>>> wrote: >>>>>>>>> >>>>>>>>>> Hi, >>>>>>>>>> >>>>>>>>>> I have managed to fix the error by enabling the DMA >>>>>>>>>> Virtualisation in BIOS. I am now hit with a new error: It's failing >>>>>>>>>> to add >>>>>>>>>> a glusterfs storage domain: >>>>>>>>>> >>>>>>>>>> [ INFO ] TASK [Add glusterfs storage domain] >>>>>>>>>> [ ERROR ] Error: Fault reason is "Operation Failed". Fault detail >>>>>>>>>> is "[Problem while trying to mount target]". HTTP response code is >>>>>>>>>> 400. >>>>>>>>>> [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, >>>>>>>>>> "msg": "Fault reason is \"Operation Failed\". Fault detail is >>>>>>>>>> \"[Problem >>>>>>>>>> while trying to mount target]\". HTTP response code is 400."} >>>>>>>>>> Please specify the storage you would like to use >>>>>>>>>> (glusterfs, iscsi, fc, nfs)[nfs]: >>>>>>>>>> >>>>>>>>>> Attached are vdsm and engine log files. >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> On Wed, Jul 11, 2018 at 9:57 AM, Sakhi Hadebe <sa...@sanren.ac.za >>>>>>>>>> > wrote: >>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> On Wed, Jul 11, 2018 at 9:33 AM, Sakhi Hadebe < >>>>>>>>>>> sa...@sanren.ac.za> wrote: >>>>>>>>>>> >>>>>>>>>>>> Hi, >>>>>>>>>>>> >>>>>>>>>>>> Below are the versions of packages installed. Please find the >>>>>>>>>>>> logs attached. >>>>>>>>>>>> Qemu: >>>>>>>>>>>> ipxe-roms-qemu-20170123-1.git4e85b27.el7_4.1.noarch >>>>>>>>>>>> libvirt-daemon-driver-qemu-3.9.0-14.el7_5.6.x86_64 >>>>>>>>>>>> qemu-img-ev-2.10.0-21.el7_5.4.1.x86_64 >>>>>>>>>>>> qemu-kvm-ev-2.10.0-21.el7_5.4.1.x86_64 >>>>>>>>>>>> qemu-kvm-common-ev-2.10.0-21.el7_5.4.1.x86_64 >>>>>>>>>>>> >>>>>>>>>>>> Libvirt installed packages: >>>>>>>>>>>> libvirt-daemon-driver-storage-disk-3.9.0-14.el7_5.6.x86_64 >>>>>>>>>>>> libvirt-daemon-config-nwfilter-3.9.0-14.el7_5.6.x86_64 >>>>>>>>>>>> libvirt-daemon-driver-storage-iscsi-3.9.0-14.el7_5.6.x86_64 >>>>>>>>>>>> libvirt-daemon-driver-network-3.9.0-14.el7_5.6.x86_64 >>>>>>>>>>>> libvirt-libs-3.9.0-14.el7_5.6.x86_64 >>>>>>>>>>>> libvirt-daemon-driver-secret-3.9.0-14.el7_5.6.x86_64 >>>>>>>>>>>> libvirt-daemon-driver-storage-core-3.9.0-14.el7_5.6.x86_64 >>>>>>>>>>>> libvirt-daemon-driver-storage-gluster-3.9.0-14.el7_5.6.x86_64 >>>>>>>>>>>> libvirt-daemon-driver-storage-3.9.0-14.el7_5.6.x86_64 >>>>>>>>>>>> libvirt-daemon-driver-qemu-3.9.0-14.el7_5.6.x86_64 >>>>>>>>>>>> libvirt-3.9.0-14.el7_5.6.x86_64 >>>>>>>>>>>> libvirt-python-3.9.0-1.el7.x86_64 >>>>>>>>>>>> libvirt-daemon-driver-nodedev-3.9.0-14.el7_5.6.x86_64 >>>>>>>>>>>> libvirt-daemon-driver-storage-rbd-3.9.0-14.el7_5.6.x86_64 >>>>>>>>>>>> libvirt-daemon-driver-storage-scsi-3.9.0-14.el7_5.6.x86_64 >>>>>>>>>>>> libvirt-daemon-config-network-3.9.0-14.el7_5.6.x86_64 >>>>>>>>>>>> libvirt-client-3.9.0-14.el7_5.6.x86_64 >>>>>>>>>>>> libvirt-daemon-kvm-3.9.0-14.el7_5.6.x86_64 >>>>>>>>>>>> libvirt-daemon-driver-storage-logical-3.9.0-14.el7_5.6.x86_64 >>>>>>>>>>>> libvirt-daemon-3.9.0-14.el7_5.6.x86_64 >>>>>>>>>>>> libvirt-daemon-driver-interface-3.9.0-14.el7_5.6.x86_64 >>>>>>>>>>>> libvirt-lock-sanlock-3.9.0-14.el7_5.6.x86_64 >>>>>>>>>>>> libvirt-daemon-driver-storage-mpath-3.9.0-14.el7_5.6.x86_64 >>>>>>>>>>>> libvirt-daemon-driver-lxc-3.9.0-14.el7_5.6.x86_64 >>>>>>>>>>>> libvirt-daemon-driver-nwfilter-3.9.0-14.el7_5.6.x86_64 >>>>>>>>>>>> >>>>>>>>>>>> Virt-manager: >>>>>>>>>>>> virt-manager-common-1.4.3-3.el7.noarch >>>>>>>>>>>> >>>>>>>>>>>> oVirt: >>>>>>>>>>>> [root@localhost network-scripts]# rpm -qa | grep ovirt >>>>>>>>>>>> ovirt-setup-lib-1.1.4-1.el7.centos.noarch >>>>>>>>>>>> cockpit-ovirt-dashboard-0.11.28-1.el7.noarch >>>>>>>>>>>> ovirt-imageio-common-1.3.1.2-0.el7.centos.noarch >>>>>>>>>>>> ovirt-vmconsole-host-1.0.5-4.el7.centos.noarch >>>>>>>>>>>> ovirt-host-dependencies-4.2.3-1.el7.x86_64 >>>>>>>>>>>> ovirt-engine-sdk-python-3.6.9.1-1.el7.noarch >>>>>>>>>>>> ovirt-imageio-daemon-1.3.1.2-0.el7.centos.noarch >>>>>>>>>>>> ovirt-host-4.2.3-1.el7.x86_64 >>>>>>>>>>>> python-ovirt-engine-sdk4-4.2.7-2.el7.x86_64 >>>>>>>>>>>> ovirt-host-deploy-1.7.4-1.el7.noarch >>>>>>>>>>>> cockpit-machines-ovirt-169-1.el7.noarch >>>>>>>>>>>> ovirt-hosted-engine-ha-2.2.14-1.el7.noarch >>>>>>>>>>>> ovirt-vmconsole-1.0.5-4.el7.centos.noarch >>>>>>>>>>>> ovirt-provider-ovn-driver-1.2.11-1.el7.noarch >>>>>>>>>>>> ovirt-engine-appliance-4.2-20180626.1.el7.noarch >>>>>>>>>>>> ovirt-release42-4.2.4-1.el7.noarch >>>>>>>>>>>> ovirt-hosted-engine-setup-2.2.22.1-1.el7.noarch >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> On Wed, Jul 11, 2018 at 6:48 AM, Yedidyah Bar David < >>>>>>>>>>>> d...@redhat.com> wrote: >>>>>>>>>>>> >>>>>>>>>>>>> On Tue, Jul 10, 2018 at 11:32 PM, Sakhi Hadebe < >>>>>>>>>>>>> sa...@sanren.ac.za> wrote: >>>>>>>>>>>>> >>>>>>>>>>>>>> Hi, >>>>>>>>>>>>>> >>>>>>>>>>>>>> I did not select any CPU architecture. It doenst gove me the >>>>>>>>>>>>>> option to select one. It only states the number of virtual CPUs >>>>>>>>>>>>>> and the >>>>>>>>>>>>>> memory for the engine VM. >>>>>>>>>>>>>> >>>>>>>>>>>>>> Looking at the documentation of installing >>>>>>>>>>>>>> ovirt-release36.rpm....it does allow you to select te CPU, but >>>>>>>>>>>>>> not when >>>>>>>>>>>>>> installing ovirt-release42.rpm >>>>>>>>>>>>>> >>>>>>>>>>>>>> On Tuesday, July 10, 2018, Alastair Neil < >>>>>>>>>>>>>> ajneil.t...@gmail.com> wrote: >>>>>>>>>>>>>> >>>>>>>>>>>>>>> what did you select as your CPU architecture when you >>>>>>>>>>>>>>> created the cluster? It looks like the VM is trying to use a >>>>>>>>>>>>>>> CPU type of >>>>>>>>>>>>>>> "Custom", how many nodes in your cluster? I suggest you >>>>>>>>>>>>>>> specify the lowest >>>>>>>>>>>>>>> common denominator of CPU architecture (e.g. Sandybridge) of >>>>>>>>>>>>>>> the nodes as >>>>>>>>>>>>>>> the CPU architecture of the cluster.. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> On Tue, 10 Jul 2018 at 12:01, Sakhi Hadebe < >>>>>>>>>>>>>>> sa...@sanren.ac.za> wrote: >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Hi, >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> I have just re-installed centOS 7 in 3 servers and have >>>>>>>>>>>>>>>> configured gluster volumes following this documentation: >>>>>>>>>>>>>>>> https://www.ovirt.org/blog/2016/03/up-and-running-with-ovirt-3-6/, >>>>>>>>>>>>>>>> But I have installed >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> http://resources.ovirt.org/pub/yum-repo/ovirt-release42.rpm >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> package. >>>>>>>>>>>>>>>> Hosted-engine --deploy is failing with this error: >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> "rhel7", "--virt-type", "kvm", "--memory", "16384", >>>>>>>>>>>>>>>> "--vcpus", "4", "--network", >>>>>>>>>>>>>>>> "network=default,mac=00:16:3e:09:5e:5d,model=virtio", "--disk", >>>>>>>>>>>>>>>> "/var/tmp/localvm0nnJH9/images/eacac30d-0304-4c77-8753-6965e4b8c2e7/d494577e-027a-4209-895b-6132e6fc6b9a", >>>>>>>>>>>>>>>> "--import", "--disk", >>>>>>>>>>>>>>>> "path=/var/tmp/localvm0nnJH9/seed.iso,device=cdrom", >>>>>>>>>>>>>>>> "--noautoconsole", "--rng", "/dev/random", "--graphics", >>>>>>>>>>>>>>>> "vnc", "--video", >>>>>>>>>>>>>>>> "vga", "--sound", "none", "--controller", "usb,model=none", >>>>>>>>>>>>>>>> "--memballoon", >>>>>>>>>>>>>>>> "none", "--boot", "hd,menu=off", "--clock", >>>>>>>>>>>>>>>> "kvmclock_present=yes"], >>>>>>>>>>>>>>>> "delta": "0:00:00.979003", "end": "2018-07-10 >>>>>>>>>>>>>>>> 17:55:11.308555", "msg": >>>>>>>>>>>>>>>> "non-zero return code", "rc": 1, "start": "2018-07-10 >>>>>>>>>>>>>>>> 17:55:10.329552", >>>>>>>>>>>>>>>> "stderr": "ERROR unsupported configuration: CPU mode >>>>>>>>>>>>>>>> 'custom' for x86_64 >>>>>>>>>>>>>>>> kvm domain on x86_64 host is not supported by >>>>>>>>>>>>>>>> hypervisor\nDomain >>>>>>>>>>>>>>>> installation does not appear to have been successful.\nIf it >>>>>>>>>>>>>>>> was, you can >>>>>>>>>>>>>>>> restart your domain by running:\n virsh --connect >>>>>>>>>>>>>>>> qemu:///system start >>>>>>>>>>>>>>>> HostedEngineLocal\notherwise, please restart your >>>>>>>>>>>>>>>> installation.", >>>>>>>>>>>>>>>> "stderr_lines": ["ERROR unsupported configuration: CPU mode >>>>>>>>>>>>>>>> 'custom' for >>>>>>>>>>>>>>>> x86_64 kvm domain on x86_64 host is not supported by >>>>>>>>>>>>>>>> hypervisor", "Domain >>>>>>>>>>>>>>>> installation does not appear to have been successful.", "If it >>>>>>>>>>>>>>>> was, you can >>>>>>>>>>>>>>>> restart your domain by running:", " virsh --connect >>>>>>>>>>>>>>>> qemu:///system start >>>>>>>>>>>>>>>> HostedEngineLocal", "otherwise, please restart your >>>>>>>>>>>>>>>> installation."], >>>>>>>>>>>>>>>> "stdout": "\nStarting install...", "stdout_lines": ["", >>>>>>>>>>>>>>>> "Starting >>>>>>>>>>>>>>>> install..."]} >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>> This seems to be in the phase where we create a local vm for >>>>>>>>>>>>> the engine. We do this with plain virt-install, nothing fancy. >>>>>>>>>>>>> Searching >>>>>>>>>>>>> the net for "unsupported configuration: CPU mode 'custom'" finds >>>>>>>>>>>>> other >>>>>>>>>>>>> relevant reports, you might want to check them. You can see the >>>>>>>>>>>>> command in >>>>>>>>>>>>> bootstrap_local_vm.yml . >>>>>>>>>>>>> >>>>>>>>>>>>> Please check/share versions of relevant packages (libvirt*, >>>>>>>>>>>>> qemu*, etc) and relevant logs (libvirt). >>>>>>>>>>>>> >>>>>>>>>>>>> Also updating the subject line and adding Simone. >>>>>>>>>>>>> >>>>>>>>>>>>> Best regards, >>>>>>>>>>>>> -- >>>>>>>>>>>>> Didi >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> -- >>>>>>>>>>>> Regards, >>>>>>>>>>>> Sakhi Hadebe >>>>>>>>>>>> >>>>>>>>>>>> Engineer: South African National Research Network >>>>>>>>>>>> (SANReN)Competency Area, Meraka, CSIR >>>>>>>>>>>> >>>>>>>>>>>> Tel: +27 12 841 2308 <+27128414213> >>>>>>>>>>>> Fax: +27 12 841 4223 <+27128414223> >>>>>>>>>>>> Cell: +27 71 331 9622 <+27823034657> >>>>>>>>>>>> Email: sa...@sanren.ac.za <shad...@csir.co.za> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> -- >>>>>>>>>>> Regards, >>>>>>>>>>> Sakhi Hadebe >>>>>>>>>>> >>>>>>>>>>> Engineer: South African National Research Network >>>>>>>>>>> (SANReN)Competency Area, Meraka, CSIR >>>>>>>>>>> >>>>>>>>>>> Tel: +27 12 841 2308 <+27128414213> >>>>>>>>>>> Fax: +27 12 841 4223 <+27128414223> >>>>>>>>>>> Cell: +27 71 331 9622 <+27823034657> >>>>>>>>>>> Email: sa...@sanren.ac.za <shad...@csir.co.za> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> Regards, >>>>>>>>>> Sakhi Hadebe >>>>>>>>>> >>>>>>>>>> Engineer: South African National Research Network (SANReN)Competency >>>>>>>>>> Area, Meraka, CSIR >>>>>>>>>> >>>>>>>>>> Tel: +27 12 841 2308 <+27128414213> >>>>>>>>>> Fax: +27 12 841 4223 <+27128414223> >>>>>>>>>> Cell: +27 71 331 9622 <+27823034657> >>>>>>>>>> Email: sa...@sanren.ac.za <shad...@csir.co.za> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> _______________________________________________ >>>>>>>>>> Users mailing list -- users@ovirt.org >>>>>>>>>> To unsubscribe send an email to users-le...@ovirt.org >>>>>>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>>>>>>>>> oVirt Code of Conduct: >>>>>>>>>> https://www.ovirt.org/community/about/community-guidelines/ >>>>>>>>>> List Archives: >>>>>>>>>> https://lists.ovirt.org/archives/list/users@ovirt.org/message/YHKUKW22QLRVS56XZBXEWOGORFWFEGIA/ >>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> Regards, >>>>>>>> Sakhi Hadebe >>>>>>>> >>>>>>>> Engineer: South African National Research Network (SANReN)Competency >>>>>>>> Area, Meraka, CSIR >>>>>>>> >>>>>>>> Tel: +27 12 841 2308 <+27128414213> >>>>>>>> Fax: +27 12 841 4223 <+27128414223> >>>>>>>> Cell: +27 71 331 9622 <+27823034657> >>>>>>>> Email: sa...@sanren.ac.za <shad...@csir.co.za> >>>>>>>> >>>>>>>> >>>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> Regards, >>>>>> Sakhi Hadebe >>>>>> >>>>>> Engineer: South African National Research Network (SANReN)Competency >>>>>> Area, Meraka, CSIR >>>>>> >>>>>> Tel: +27 12 841 2308 <+27128414213> >>>>>> Fax: +27 12 841 4223 <+27128414223> >>>>>> Cell: +27 71 331 9622 <+27823034657> >>>>>> Email: sa...@sanren.ac.za <shad...@csir.co.za> >>>>>> >>>>>> _______________________________________________ >>>>>> Users mailing list -- users@ovirt.org >>>>>> To unsubscribe send an email to users-le...@ovirt.org >>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>>>>> oVirt Code of Conduct: >>>>>> https://www.ovirt.org/community/about/community-guidelines/ >>>>>> List Archives: >>>>>> https://lists.ovirt.org/archives/list/users@ovirt.org/message/DS7O33ZSGV5ZAHXKA3KT2ABYW3VJJWXQ/ >>>>>> >>>>> >>>> >>>> >>>> -- >>>> Regards, >>>> Sakhi Hadebe >>>> >>>> Engineer: South African National Research Network (SANReN)Competency Area, >>>> Meraka, CSIR >>>> >>>> Tel: +27 12 841 2308 <+27128414213> >>>> Fax: +27 12 841 4223 <+27128414223> >>>> Cell: +27 71 331 9622 <+27823034657> >>>> Email: sa...@sanren.ac.za <shad...@csir.co.za> >>>> >>>> >> >> >> -- >> Regards, >> Sakhi Hadebe >> >> Engineer: South African National Research Network (SANReN)Competency Area, >> Meraka, CSIR >> >> Tel: +27 12 841 2308 <+27128414213> >> Fax: +27 12 841 4223 <+27128414223> >> Cell: +27 71 331 9622 <+27823034657> >> Email: sa...@sanren.ac.za <shad...@csir.co.za> >> >> > > > -- > Regards, > Sakhi Hadebe > > Engineer: South African National Research Network (SANReN)Competency Area, > Meraka, CSIR > > Tel: +27 12 841 2308 <+27128414213> > Fax: +27 12 841 4223 <+27128414223> > Cell: +27 71 331 9622 <+27823034657> > Email: sa...@sanren.ac.za <shad...@csir.co.za> > >
_______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/OD5XLK7DRUHSF33S7QJZFDNRPMREGJ3G/