[ovirt-users] Unable to change engine-config
I wanted to change the default keyboard layout. I ran # engine-config -s VncKeyboardLayout=it Picked up JAVA_TOOL_OPTIONS: -Dcom.redhat.fips=false Cannot set value it to key VncKeyboardLayout. Valid values are [ar,cz,da,de,de-ch,en-gb,en-us,es,et,fi,fo,fr,fr-be,fr-ca,fr-ch,hr,hu,is,it,ja,lt,lv,mk,nl,no,pl,pt,pt-br,ru,sl,sv,th,tr] I have tried with other values but the same result: # engine-config -s ClientModeConsoleDefault=vnc Picked up JAVA_TOOL_OPTIONS: -Dcom.redhat.fips=false Cannot set value vnc to key ClientModeConsoleDefault. Valid values are [vnc,spice] Am I making a mistake with the syntax or should I open a bug? Thanks, James ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/3T4RKHHLW7PWW2K4UDN2OJWCB7VP46FM/
[ovirt-users] Re: Unable to reinstall 2nd and 3rd Host after restore HE backup onto 1st host
Well after many hours of debugging the issue was not with hosts but with the ovirt-engine. We didn't think it was the engine because we had successfully redeployed the engine on the first host. However it was the engine and it was due to the fact it was running the latest version of ansible 2.14.2 which uses python3.11. Unfortunately python3.11 does not have all the module yet compiled for this version so the adding of the host failed due to the netaddr module being missing. So we put python3.6 as the default version on the system because vdsm and other service need this version. We then copied from python3.9 the module netaddr to the python3.11 module folder. At this point we could reinstall the hosts. Somebody else had exactly the same problem at the same time and you can read the issue report here. https://github.com/oVirt/ovirt-ansible-collection/issues/695 ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/JRQEKUJ6J4KPMHIDH4UGIR2ZBD64BW5E/
[ovirt-users] Unable to reinstall 2nd and 3rd Host after restore HE backup onto 1st host
Hi there, We are running ovirt 4.5.4 on 3 RHEL 8.7 host with a on self hosted engine again using RHEL 8.7. We are making some changes to our company network so we followed these instructions here to move the ovirt engine onto a new storage with different IP address. https://access.redhat.com/solutions/6529691 We used a host to redeploy the hosted engine from a backup and we managed to get the hosted engine running on the new storage with the IP address. We then tried to reinstall the 2 other hosts but the reinstallation failed with the message: Host ovirt2.ad.tintolav.com installation failed. Task Restart services failed to execute. Please check logs for more details: /var/log/ovirt-engine/host-deploy/ovirt-host-deploy-ansible-20230323083407-ovirt2.ad.tintolav.com-d950815b-f1b9-4dcf-b609-3ff1866a6b70.log In the ovirt-engie host deploy log we see [ "Traceback (most recent call last):", " File \"/usr/bin/vdsm-tool\", line 18, in ", "import vdsm.tool", "ModuleNotFoundError: No module named 'vdsm'" ] In the vdsm log on the host there are no errors except for a couple of warnings 2023-03-23 07:58:22,335+0100 WARN (periodic/1) [throttled] MOM not available. Error: [Errno 2] No such file or directory (throttledlog:87) 2023-03-23 07:58:22,336+0100 WARN (periodic/1) [throttled] MOM not available, KSM stats will be missing. Error: (throttledlog:87) We have tried a completely new install for ovirt2, but we have finished back at the same point with the same error. [root@ovirt2 ~]# systemctl status vdsmd ● vdsmd.service - Virtual Desktop Server Manager Loaded: loaded (/usr/lib/systemd/system/vdsmd.service; enabled; vendor preset: disabled) Active: inactive (dead) since Thu 2023-03-23 08:35:32 CET; 56min ago Process: 22942 ExecStart=/usr/libexec/vdsm/daemonAdapter -0 /dev/null -1 /dev/null -2 /dev/null /usr/libexec/vdsm/vdsmd (code=exited, status=0/SUCCESS) Process: 22876 ExecStartPre=/usr/libexec/vdsm/vdsmd_init_common.sh --pre-start (code=exited, status=0/SUCCESS) Main PID: 22942 (code=exited, status=0/SUCCESS) Mar 23 07:58:20 ovirt2.ad.tintolav.com vdsmd_init_common.sh[22876]: vdsm: Running test_space Mar 23 07:58:20 ovirt2.ad.tintolav.com vdsmd_init_common.sh[22876]: vdsm: Running test_lo Mar 23 07:58:20 ovirt2.ad.tintolav.com systemd[1]: Started Virtual Desktop Server Manager. Mar 23 07:58:22 ovirt2.ad.tintolav.com vdsm[22942]: WARN MOM not available. Error: [Errno 2] No such file or directory Mar 23 07:58:22 ovirt2.ad.tintolav.com vdsm[22942]: WARN MOM not available, KSM stats will be missing. Error: Mar 23 08:35:32 ovirt2.ad.tintolav.com systemd[1]: Stopping Virtual Desktop Server Manager... Mar 23 08:35:32 ovirt2.ad.tintolav.com systemd[1]: vdsmd.service: Succeeded. Mar 23 08:35:32 ovirt2.ad.tintolav.com systemd[1]: Stopped Virtual Desktop Server Manager. Mar 23 08:35:33 ovirt2.ad.tintolav.com systemd[1]: Dependency failed for Virtual Desktop Server Manager. Mar 23 08:35:33 ovirt2.ad.tintolav.com systemd[1]: vdsmd.service: Job vdsmd.service/start failed with result 'dependency'. Mar 23 08:35:33 ovirt2.ad.tintolav.com systemd[1]: vdsmd.service: Job vdsmd.service/start failed with result 'dependency'. [root@ovirt2 ~]# systemctl start vdsmd A dependency job for vdsmd.service failed. See 'journalctl -xe' for details. [root@ovirt2 ~]# journalctl -xe Mar 23 09:34:35 ovirt2.ad.tintolav.com systemd[1]: Stopped Auxiliary vdsm service for running helper functions as root. -- Subject: Unit supervdsmd.service has finished shutting down -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit supervdsmd.service has finished shutting down. Mar 23 09:34:35 ovirt2.ad.tintolav.com systemd[1]: Started Auxiliary vdsm service for running helper functions as root. -- Subject: Unit supervdsmd.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit supervdsmd.service has finished starting up. -- -- The start-up result is done. Mar 23 09:34:35 ovirt2.ad.tintolav.com daemonAdapter[28677]: Traceback (most recent call last): Mar 23 09:34:35 ovirt2.ad.tintolav.com daemonAdapter[28677]: File "/usr/libexec/vdsm/daemonAdapter", line 16, in Mar 23 09:34:35 ovirt2.ad.tintolav.com daemonAdapter[28677]: from vdsm.config import config Mar 23 09:34:35 ovirt2.ad.tintolav.com daemonAdapter[28677]: ModuleNotFoundError: No module named 'vdsm' Mar 23 09:34:35 ovirt2.ad.tintolav.com systemd[1]: supervdsmd.service: Main process exited, code=exited, status=1/FAILURE Mar 23 09:34:35 ovirt2.ad.tintolav.com systemd[1]: supervdsmd.service: Failed with result 'exit-code'. -- Subject: Unit failed -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit supervdsmd.service has entered the 'failed' state with result 'exit-code'. Mar 23 09:34:35 ovirt2.ad.tintolav.com systemd[1]: supervdsmd.service: Service RestartSec=100ms expired, scheduling restart. Mar 23 09:34:35
[ovirt-users] Re: Ovirt 4.4.10 running on RHEL 8.7
Thanks for your reply. After studying the documentation further I think the safest option will to set the rhel release to 8.5 while we are running 4.4 subscription-manager release --set=8.5 subscription-manager release --show Then update the cluster to 4.4.10. After which we set the release to rhel 8.6 and update to ovit 4.5 following the instruction in the link https://www.ovirt.org/download/install_on_rhel.html What are people's thoughts on this? ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/JDMQPVGCVOXMESVH6AUK3HYXQVRRHTCY/
[ovirt-users] Ovirt 4.4.10 running on RHEL 8.7
Hi there, we currently have ovirt 4.4.9 running on RHEL 8.5. Can we upgrade the cluster to ovirt 4.4.10 on RHEL 8.7? Do we need to following these steps in this link? https://www.ovirt.org/download/install_on_rhel.html Many thanks, James ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/JPUUQEUSEIBYCLL4DQENIXCRCPKHUABE/
[ovirt-users] Re: Cannot to update hosts, nothing provides libvirt-daemon-kvm >= 7.6.0-2 needed by vdsm-4.40.90.4-1.el8.x86_64
We had the same problem when updating to 4.4.9 on RHEL 8.4 Essentially to get past the problem we installed the following repo # /etc/yum/repos.d/ovirt-4.4-stream-advanced-virt.repo [ovirt-centos-stream-advanced-virtualization] name=Advanced Virtualization CentOS Stream testing packages for $basearch baseurl=https://buildlogs.centos.org/centos/8-stream/virt/$basearch/advancedvirt-common/ enabled=1 gpgcheck=0 module_hotfixes=1 The solution can be found in this thread https://www.mail-archive.com/users@ovirt.org/msg68142.html ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/RFQJBVEUGQVS6LARQLWCEF4H7JYPE5GV/
[ovirt-users] Impossible to move disk after a previous disk move failed
This is the log of when it fails 2021-08-23 21:24:10,667+0200 WARN (tasks/0) [storage.LVM] Command with specific filter failed or returned no data, retrying with a wider filter: LVM command failed: 'cmd=[\'/sbin/lvm\', \'lvcreate\', \'--config\', \'devices { preferred_names=["^/dev/mapper/"] ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3 filter=["a|^/dev/mapper/36001405299f83b19569473f9c580660c$|", "r|.*|"] hints="none" obtain_device_list_from_udev=0 } global { locking_type=1 prioritise_write_locks=1 wait_for_locks=1 use_lvmetad=0 use_lvmpolld=1 } backup { retain_min=50 retain_days=0 }\', \'--autobackup\', \'n\', \'--contiguous\', \'n\', \'--size\', \'40960m\', \'--wipesignatures\', \'n\', \'--addtag\', \'OVIRT_VOL_INITIALIZING\', \'--name\', \'432ceb20-efb7-4a40-8431-1b5c825a6168\', \'c23a5bef-48e0-46c7-9d5b-93c97f0240c0\'] rc=5 out=[] err=[\' Logical Volume "432ceb20-efb7-4a40-8431-1b5c825a6168" already exists in volume group "c23a5bef-48e0-46c7-9d5b-93c97f0240c0"\']' (l vm:534) 2021-08-23 21:24:10,859+0200 WARN (tasks/0) [storage.LVM] All 2 tries have failed: LVM command failed: 'cmd=[\'/sbin/lvm\', \'lvcreate\', \'--config\', \'devices { preferred_names=["^/dev/mapper/"] ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3 filter=["a|^/dev/mapper/36001405299f83b19569473f9c580660c$|^/dev/mapper/36001405cdf35411dd040d4121d9326d1$|^/dev/mapper/36001405df393063de6f0d4451d8a61d3$|", "r|.*|"] hints="none" obtain_device_list_from_udev=0 } global { locking_type=1 prioritise_write_locks=1 wait_for_locks=1 use_lvmetad=0 use_lvmpolld=1 } backup { retain_min=50 retain_days=0 }\', \'--autobackup\', \'n\', \'--contiguous\', \'n\', \'--size\', \'40960m\', \'--wipesignatures\', \'n\', \'--addtag\', \'OVIRT_VOL_INITIALIZING\', \'--name\', \'432ceb20-efb7-4a40-8431-1b5c825a6168\', \'c23a5bef-48e0-46c7-9d5b-93c97f0240c0\'] rc=5 out=[] err=[\' Logical Volume "432ceb20-efb7-4a40-8431-1b5c825a6168" already exists in volume group "c23a5bef-4 8e0-46c7-9d5b-93c97f0240c0"\']' (lvm:561) 2021-08-23 21:24:10,859+0200 ERROR (tasks/0) [storage.Volume] Failed to create volume /rhev/data-center/mnt/blockSD/c23a5bef-48e0-46c7-9d5b-93c97f0240c0/images/2172a4ac-6992-4cc2-be1b-6b9290bc9798/432ceb20-efb7-4a40-8431-1b5c825a6168: Cannot create Logical Volume: 'vgname=c23a5bef-48e0-46c7-9d5b-93c97f0240c0 lvname=432ceb20-efb7-4a40-8431-1b5c825a6168 err=[\' Logical Volume "432ceb20-efb7-4a40-8431-1b5c825a6168" already exists in volume group "c23a5bef-48e0-46c7-9d5b-93c97f0240c0"\']' (volume:1257) 2021-08-23 21:24:10,860+0200 ERROR (tasks/0) [storage.Volume] Unexpected error (volume:1293) Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/volume.py", line 1254, in create add_bitmaps=add_bitmaps) File "/usr/lib/python3.6/site-packages/vdsm/storage/blockVolume.py", line 508, in _create initialTags=(sc.TAG_VOL_UNINIT,)) File "/usr/lib/python3.6/site-packages/vdsm/storage/lvm.py", line 1633, in createLV raise se.CannotCreateLogicalVolume(vgName, lvName, err) vdsm.storage.exception.CannotCreateLogicalVolume: Cannot create Logical Volume: 'vgname=c23a5bef-48e0-46c7-9d5b-93c97f0240c0 lvname=432ceb20-efb7-4a40-8431-1b5c825a6168 err=[\' Logical Volume "432ceb20-efb7-4a40-8431-1b5c825a6168" already exists in volume group "c23a5bef-48e0-46c7-9d5b-93c97f0240c0"\']' 2021-08-23 21:24:10,860+0200 ERROR (tasks/0) [storage.TaskManager.Task] (Task='55a4e8dc-9408-4969-b0ba-b9a556bccba1') Unexpected error (task:877) Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/task.py", line 884, in _run return fn(*args, **kargs) File "/usr/lib/python3.6/site-packages/vdsm/storage/task.py", line 350, in run return self.cmd(*self.argslist, **self.argsdict) File "/usr/lib/python3.6/site-packages/vdsm/storage/securable.py", line 79, in wrapper return method(self, *args, **kwargs) File "/usr/lib/python3.6/site-packages/vdsm/storage/sp.py", line 1945, in createVolume initial_size=initialSize, add_bitmaps=addBitmaps) File "/usr/lib/python3.6/site-packages/vdsm/storage/sd.py", line 1216, in createVolume initial_size=initial_size, add_bitmaps=add_bitmaps) File "/usr/lib/python3.6/site-packages/vdsm/storage/volume.py", line 1254, in create add_bitmaps=add_bitmaps) File "/usr/lib/python3.6/site-packages/vdsm/storage/blockVolume.py", line 508, in _create initialTags=(sc.TAG_VOL_UNINIT,)) File "/usr/lib/python3.6/site-packages/vdsm/storage/lvm.py", line 1633, in createLV raise se.CannotCreateLogicalVolume(vgName, lvName, err) vdsm.storage.exception.CannotCreateLogicalVolume: Cannot create Logical Volume: 'vgname=c23a5bef-48e0-46c7-9d5b-93c97f0240c0 lvname=432ceb20-efb7-4a40-8431-1b5c825a6168 err=[\' Logical Volume "432ceb20-efb7-4a40-8431-1b5c825a6168" already exists in volume group