On Tue, Jun 6, 2017 at 2:10 AM, Brendan Hartzell <mrr...@gmail.com> wrote:
> As requested, > It seams fine, there are no pending locks now. Could you please retry? > > The output of ovirt-hosted-engine-cleanup > > [root@node-1 ~]# ovirt-hosted-engine-cleanup > This will de-configure the host to run ovirt-hosted-engine-setup from > scratch. > Caution, this operation should be used with care. > > Are you sure you want to proceed? [y/n] > y > -=== Destroy hosted-engine VM ===- > You must run deploy first > -=== Stop HA services ===- > -=== Shutdown sanlock ===- > shutdown force 1 wait 0 > shutdown done 0 > -=== Disconnecting the hosted-engine storage domain ===- > You must run deploy first > -=== De-configure VDSM networks ===- > -=== Stop other services ===- > -=== De-configure external daemons ===- > -=== Removing configuration files ===- > ? /etc/init/libvirtd.conf already missing > - removing /etc/libvirt/nwfilter/vdsm-no-mac-spoofing.xml > ? /etc/ovirt-hosted-engine/answers.conf already missing > ? /etc/ovirt-hosted-engine/hosted-engine.conf already missing > - removing /etc/vdsm/vdsm.conf > - removing /etc/pki/vdsm/certs/cacert.pem > - removing /etc/pki/vdsm/certs/vdsmcert.pem > - removing /etc/pki/vdsm/keys/vdsmkey.pem > - removing /etc/pki/vdsm/libvirt-spice/ca-cert.pem > - removing /etc/pki/vdsm/libvirt-spice/ca-key.pem > - removing /etc/pki/vdsm/libvirt-spice/server-cert.pem > - removing /etc/pki/vdsm/libvirt-spice/server-key.pem > ? /etc/pki/CA/cacert.pem already missing > ? /etc/pki/libvirt/*.pem already missing > ? /etc/pki/libvirt/private/*.pem already missing > ? /etc/pki/ovirt-vmconsole/*.pem already missing > - removing /var/cache/libvirt/qemu > ? /var/run/ovirt-hosted-engine-ha/* already missing > [root@node-1 ~]# > > Output of sanlock client status: > [root@node-1 ~]# sanlock client status > [root@node-1 ~]# > > Thank you for your help! > > On Mon, Jun 5, 2017 at 7:25 AM, Simone Tiraboschi <stira...@redhat.com> > wrote: > >> >> >> On Mon, Jun 5, 2017 at 3:57 PM, Brendan Hartzell <mrr...@gmail.com> >> wrote: >> >>> After letting this sit for a few days, does anyone have any ideas as to >>> how to deal with my situation? Would anyone like me to send the SOS report >>> directly to them? It's a 9MB file. >>> >>> If nothing comes up, I'm going to try and sift through the SOS report >>> tonight, but I won't know what I'm trying to find. >>> >>> Thank you for any and all help. >>> >>> On Thu, Jun 1, 2017 at 1:15 AM, Sandro Bonazzola <sbona...@redhat.com> >>> wrote: >>> >>>> >>>> >>>> On Thu, Jun 1, 2017 at 6:36 AM, Brendan Hartzell <mrr...@gmail.com> >>>> wrote: >>>> >>>>> Ran the 4 commands listed above, no errors on the screen. >>>>> >>>>> Started the hosted-engine standard setup from the web-UI. >>>>> >>>>> Using iSCSI for the storage. >>>>> >>>>> Using mostly default options, I got these errors in the web-UI. >>>>> >>>>> Error creating Volume Group: Failed to initialize physical device: >>>>> ("[u'/dev/mapper/36589cfc000000de7482638fcfcebbbb4']",) >>>>> Failed to execute stage 'Misc configuration': Failed to initialize >>>>> physical device: ("[u'/dev/mapper/36589cfc00000 >>>>> 0de7482638fcfcebbbb4']",) >>>>> Hosted Engine deployment failed: this system is not reliable, please >>>>> check the issue,fix and redeploy >>>>> >>>>> I rebuilt my iSCSI (I don't think I cleaned it up from a previous >>>>> install). >>>>> Re-ran the above 4 commands. >>>>> Restarted hosted engine standard setup from web-UI. >>>>> Install moved past "Connecting Storage Pool" so I believe the above >>>>> was my fault. >>>>> >>>>> These are the last messages displayed on the web-UI. >>>>> Creating Storage Pool >>>>> Connecting Storage Pool >>>>> Verifying sanlock lockspace initialization >>>>> Creating Image for 'hosted-engine.lockspace' ... >>>>> Image for 'hosted-engine.lockspace' created successfully >>>>> Creating Image for 'hosted-engine.metadata' ... >>>>> Image for 'hosted-engine.metadata' created successfully >>>>> Creating VM Image >>>>> Extracting disk image from OVF archive (could take a few minutes >>>>> depending on archive size) >>>>> Validating pre-allocated volume size >>>>> Uploading volume to data domain (could take a few minutes depending on >>>>> archive size) >>>>> >>>>> At the host terminal, I got the error "watchdog watchdog0: watchdog >>>>> did not stop!" >>>>> Then the host restarted. >>>>> >>>> >>>> Simone, can you help here? >>>> >>>> >> Ok, sorry for the delay. >> The second installation attempt seams fine but it seams that >> ovirt-hosted-engine-cleanup failed stopping sanlock and so the >> watchdog kick in rebooting your system in the middle of deployment attempt. >> >> could you please post the output of >> ovirt-hosted-engine-cleanup >> sanlock client status >> ? >> >> >> >>> >>>> >>>> >>>>> >>>>> This is as far as I've gotten in previous attempts. >>>>> >>>>> Attaching the hosted-engine-setup log. >>>>> >>>>> The SOS report is 9MB and the ovirt users group will drop the email. >>>>> >>>>> On Wed, May 31, 2017 at 6:59 AM, Sandro Bonazzola <sbona...@redhat.com >>>>> > wrote: >>>>> >>>>>> >>>>>> >>>>>> On Wed, May 31, 2017 at 3:10 PM, Brendan Hartzell <mrr...@gmail.com> >>>>>> wrote: >>>>>> >>>>>>> Now that you have identified the problem, should I run the following >>>>>>> commands and send you another SOS? >>>>>>> >>>>>>> ovirt-hosted-engine-cleanup >>>>>>> vdsm-tool configure --force >>>>>>> systemctl restart libvirtd >>>>>>> systemctl restart vdsm >>>>>>> >>>>>>> Or is there a different plan in mind? >>>>>>> >>>>>> >>>>>> I would have expected someone from virt team to follow up for further >>>>>> investigations :-) >>>>>> above commands should work. >>>>>> >>>>>> >>>>>> >>>>>>> >>>>>>> Thank you, >>>>>>> >>>>>>> Brendan >>>>>>> >>>>>>> On Tue, May 30, 2017 at 11:42 PM, Sandro Bonazzola < >>>>>>> sbona...@redhat.com> wrote: >>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> On Wed, May 31, 2017 at 4:45 AM, Brendan Hartzell <mrr...@gmail.com >>>>>>>> > wrote: >>>>>>>> >>>>>>>>> Can you please elaborate about the failure you see here and how >>>>>>>>> are you trying to manually partition the host? >>>>>>>>> >>>>>>>>> Sure, I will start from the beginning. >>>>>>>>> - Using: ovirt-node-ng-installer-ovirt-4.1-2017052604 >>>>>>>>> <(201)%20705-2604>.iso >>>>>>>>> - During installation I setup one of the two interfaces and check >>>>>>>>> the box to automatically use the connection. >>>>>>>>> - I'm currently providing a host name of node-1.test.net until I >>>>>>>>> have a successful process. >>>>>>>>> - I configure date and time for my timezone and to use an internal >>>>>>>>> NTP server. >>>>>>>>> - On Installation Destination, I pick my 128GB USB3.0 SanDisk >>>>>>>>> flash drive, check the box that I would like to make additional >>>>>>>>> space, and >>>>>>>>> click done. In the reclaim disk space window, I click delete all, >>>>>>>>> and then >>>>>>>>> reclaim space. I go back into the Installation Destination, select >>>>>>>>> that I >>>>>>>>> will configure partitioning, and click done. The Manual Partitioning >>>>>>>>> window opens, I use the option to automatically create mount points. >>>>>>>>> >>>>>>>> >>>>>>>> In this screen, please change partitioning scheme from LVM to LVM >>>>>>>> Thin Provisioning: it should solve your following error. >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>>> At this point, /boot is 1024MB, /var is 15GB, / is 88.11 GB, and >>>>>>>>> swap is 11.57GB. I then change / to 23.11 GB, update settings, >>>>>>>>> change /var >>>>>>>>> to 80GB, update settings again, and click done. I accept the changes >>>>>>>>> and >>>>>>>>> begin installation. >>>>>>>>> >>>>>>>>> I tried these changes based on this article: http://www.ovirt.org/ >>>>>>>>> documentation/self-hosted/chap-Deploying_Self-Hosted_Engine/ >>>>>>>>> >>>>>>>>> The article does say that you can specify a different directory >>>>>>>>> than /var/tmp, but I don't recall seeing that option. >>>>>>>>> >>>>>>>> >>>>>>>> If the setup detects not enough space in /var/tmp for extracting >>>>>>>> the appliance it will ask about a different directory. >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>>> >>>>>>>>> After some time, I get the following error: >>>>>>>>> There was an error running the kickstart script at line 7. This >>>>>>>>> is a fatal error and installation will be aborted. The details of >>>>>>>>> this >>>>>>>>> error are: >>>>>>>>> >>>>>>>>> [INFO] Trying to create a manageable base from '/' >>>>>>>>> [ERROR] LVM Thin Provisioning partitioning scheme is required. >>>>>>>>> For autoinstall via Kickstart with LVM Thin Provisioning check options >>>>>>>>> --thinpool and --grow. Please consult documentation for details. >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> ^^ this one should be solved by the LVM Thin Provisioning scheme >>>>>>>> mentioned above.. >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>>> >>>>>>>>> Traceback (most recent call last): >>>>>>>>> File "/usr/lib64/python2.7/runpy.py", line 162, in >>>>>>>>> _run_module_as_main "__main__", fname, loader, pkg_name) >>>>>>>>> File "/usr/lib64/python2.7runpy.py", line 72, in _run_code exec >>>>>>>>> code in run_globals >>>>>>>>> File "/usr/lib/python2.7/site-packages/imgbased/__main__.py", >>>>>>>>> line 51, in <module> CliApplication() >>>>>>>>> File "/usr/lib/python2.7/site-packages/imgbased/__init__.py", >>>>>>>>> line 82, in CliApplication() >>>>>>>>> File "/usr/lib/python2.7/site-packages/imgbased/hooks.py", line >>>>>>>>> 120, in emit cb(self.context, *args) >>>>>>>>> File "/usr/lib/python2.7/site-packages/imgbased/plugins/core.py", >>>>>>>>> line 169, in post_argparse layout.initialize(args.source, >>>>>>>>> args.init_nvr) >>>>>>>>> File "/usr/lib/python2.7/site-packages/imgbased/plugins/core.py", >>>>>>>>> line 216, in initialize self.app.imgbase.init_layout_from(source, >>>>>>>>> init_nvr) >>>>>>>>> File "/usr/lib/python2.7/site-packages/imgbased/imgbase.py", line >>>>>>>>> 271, in init_layout_from self.init_tags_on(existing_lv) >>>>>>>>> File "/usr/lib/python2.7/site-packages/imgbased/imgbase.py", line >>>>>>>>> 243, in init_tags_on pool = lv.thinpool() >>>>>>>>> File "/usr/lib/python2.7/site-packages/imgbased/lvm.py", line >>>>>>>>> 250, in thinpool raise MissingLvmThinPool() >>>>>>>>> imgbased.lvm.MissingLvmThinPoo >>>>>>>>> l >>>>>>>>> >>>>>>>>> At this point, the only option is to exit the installer. >>>>>>>>> >>>>>>>>> **************************** >>>>>>>>> >>>>>>>>> Being this a new install, please use 4.1. oVirt 4.0 is not >>>>>>>>> supported anymore. >>>>>>>>> >>>>>>>>> Not a problem. >>>>>>>>> >>>>>>>>> **************************** >>>>>>>>> >>>>>>>>> Can you please provide hosted engine setup logs or better a full >>>>>>>>> sos report? (sosreport -a) >>>>>>>>> >>>>>>>>> Again, the process I'm following: >>>>>>>>> - Using: ovirt-node-ng-installer-ovirt-4.1-2017052604 >>>>>>>>> <(201)%20705-2604>.iso >>>>>>>>> - During installation I setup one of the two interfaces and check >>>>>>>>> the box to automatically use the connection. >>>>>>>>> - I'm currently providing a host name of node-1.test.net until I >>>>>>>>> have a successful process. >>>>>>>>> - I configure date and time for my timezone and to use an internal >>>>>>>>> NTP server. >>>>>>>>> - On Installation Destination, I pick my 128GB USB3.0 SanDisk >>>>>>>>> flash drive, check the box that I would like to make additional >>>>>>>>> space, and >>>>>>>>> click done. In the reclaim disk space window, I click delete all, and >>>>>>>>> then >>>>>>>>> reclaim space. >>>>>>>>> - Begin Installation and set a root password. >>>>>>>>> - Perform a yum update - no packages marked for update (as >>>>>>>>> expected) >>>>>>>>> - Use vi to update /etc/hosts with a reference for node-1.test.net >>>>>>>>> and engine.test.net >>>>>>>>> - First attempt at hosted-engine from web-UI >>>>>>>>> - Setup downloads and installs ovirt-engine-applianc >>>>>>>>> e-4.1-20170523.1.el7.centos.noarch.rpm >>>>>>>>> *Failed to execute stage 'Environment setup': Failed to >>>>>>>>> reconfigure libvirt for VDSM >>>>>>>>> *Hosted Engine deployment failed >>>>>>>>> - Attached SOS report >>>>>>>>> The checksum is: aa56097edc0b63c49caaf1a1fde021bc >>>>>>>>> >>>>>>>>> At this point, I would run ovirt-hosted-engine-cleanup and I >>>>>>>>> would get further along in the install process. However, because >>>>>>>>> this is a >>>>>>>>> fresh install, I'm going to leave things here for now so you can >>>>>>>>> review the >>>>>>>>> SOS. >>>>>>>>> >>>>>>>> >>>>>>>> Thanks for the SOS report! >>>>>>>> Hosted Engine setup fails on: >>>>>>>> >>>>>>>> 2017-05-30 19:24:39 DEBUG otopi.plugins.gr_he_setup.system.vdsmenv >>>>>>>> plugin.execute:921 execute-output: ('/bin/vdsm-tool', 'configure', >>>>>>>> '--force') stdout: >>>>>>>> >>>>>>>> Checking configuration status... >>>>>>>> >>>>>>>> Current revision of multipath.conf detected, preserving >>>>>>>> lvm is configured for vdsm >>>>>>>> libvirt is already configured for vdsm >>>>>>>> SUCCESS: ssl configured to true. No conflicts >>>>>>>> >>>>>>>> Running configure... >>>>>>>> Reconfiguration of libvirt is done. >>>>>>>> >>>>>>>> 2017-05-30 19:24:39 DEBUG otopi.plugins.gr_he_setup.system.vdsmenv >>>>>>>> plugin.execute:926 execute-output: ('/bin/vdsm-tool', 'configure', >>>>>>>> '--force') stderr: >>>>>>>> Error: ServiceOperationError: _systemctlStart failed >>>>>>>> Job for libvirtd.service failed because the control process exited >>>>>>>> with error code. See "systemctl status libvirtd.service" and >>>>>>>> "journalctl >>>>>>>> -xe" for details. >>>>>>>> >>>>>>>> At the same time journalctl shows: >>>>>>>> >>>>>>>> May 30 19:24:39 node-1.test.net libvirtd[20954]: libvirt version: >>>>>>>> 2.0.0, package: 10.el7_3.5 (CentOS BuildSystem < >>>>>>>> http://bugs.centos.org>, 2017-03-03-02:09:45, c1bm.rdu2.centos.org) >>>>>>>> May 30 19:24:39 node-1.test.net libvirtd[20954]: hostname: >>>>>>>> node-1.test.net >>>>>>>> May 30 19:24:39 node-1.test.net libvirtd[20954]: The server >>>>>>>> certificate /etc/pki/vdsm/certs/vdsmcert.pem is not yet active >>>>>>>> May 30 19:24:39 node-1.test.net systemd[1]: libvirtd.service: main >>>>>>>> process exited, code=exited, status=6/NOTCONFIGURED >>>>>>>> May 30 19:24:39 node-1.test.net systemd[1]: Failed to start >>>>>>>> Virtualization daemon. >>>>>>>> May 30 19:24:39 node-1.test.net systemd[1]: Unit libvirtd.service >>>>>>>> entered failed state. >>>>>>>> May 30 19:24:39 node-1.test.net systemd[1]: libvirtd.service >>>>>>>> failed. >>>>>>>> May 30 19:24:39 node-1.test.net systemd[1]: libvirtd.service >>>>>>>> holdoff time over, scheduling restart. >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>>> >>>>>>>>> **************************** >>>>>>>>> >>>>>>>>> I'd like to understand the issues you faced before suggesting to >>>>>>>>> restart from scratch. >>>>>>>>> >>>>>>>>> Too late... I did two re-installs to get a more accurate account >>>>>>>>> of my install process for above. >>>>>>>>> >>>>>>>>> **************************** >>>>>>>>> >>>>>>>>> Thank you for your help! >>>>>>>>> >>>>>>>>> Brendan >>>>>>>>> >>>>>>>>> On Tue, May 30, 2017 at 12:17 AM, Sandro Bonazzola < >>>>>>>>> sbona...@redhat.com> wrote: >>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> On Tue, May 30, 2017 at 6:49 AM, Brendan Hartzell < >>>>>>>>>> mrr...@gmail.com> wrote: >>>>>>>>>> >>>>>>>>>>> oVirt users list, >>>>>>>>>>> >>>>>>>>>>> Long story short, I've been spending weeks on this project for >>>>>>>>>>> my home lab with no success. >>>>>>>>>>> >>>>>>>>>>> I would like to successfully install two nodes that host a >>>>>>>>>>> highly available engine with an iSCSI storage back-end. >>>>>>>>>>> >>>>>>>>>>> I have read through most, if not all, of the guides on ovirt.org >>>>>>>>>>> with no substantial help. >>>>>>>>>>> >>>>>>>>>>> Successfully, I have done the following: >>>>>>>>>>> Install oVirt Engine on a bare metal system, added a node, and >>>>>>>>>>> started exploring - not desired. >>>>>>>>>>> Install oVirt Node 4.0.6 on a bare metal system - fails if >>>>>>>>>>> partitions are not done automatically. >>>>>>>>>>> Install oVirt Node 4.1.2 on a bare metal system - fails if >>>>>>>>>>> partitions are not done automatically. >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> Can you please elaborate about the failure you see here and how >>>>>>>>>> are you trying to manually partition the host? >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> My process after installing a Node: >>>>>>>>>>> Run a yum update - just to be sure, but I am using latest iso >>>>>>>>>>> images from downloads section. >>>>>>>>>>> Edit /etc/hosts for local name resolution - the goal is to host >>>>>>>>>>> DNS as a virtual machine, eventually. >>>>>>>>>>> On 4.1 if I install ovirt-engine-appliance from yum, it does >>>>>>>>>>> simplify one step in the hosted engine setup. If I do this on 4.0 >>>>>>>>>>> it >>>>>>>>>>> discards the image and uses the default. >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> Being this a new install, please use 4.1. oVirt 4.0 is not >>>>>>>>>> supported anymore. >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>>> On 4.1 the hosted engine setup fails immediately unless I run >>>>>>>>>>> the hosted engine cleanup from the shell. >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> Can you please provide hosted engine setup logs or better a full >>>>>>>>>> sos report? (sosreport -a) >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>>> If I do this, I can typically get to the point of installing. >>>>>>>>>>> >>>>>>>>>>> When I do get to the installation phase, I get to a point just >>>>>>>>>>> after extracting the OVA that I get a message on the shell saying >>>>>>>>>>> something >>>>>>>>>>> about the watchdog running the whole time and then the node reboots. >>>>>>>>>>> >>>>>>>>>>> I found one email thread that sounded like my issue and >>>>>>>>>>> suggested the following commands: >>>>>>>>>>> vdsm-tool configure --force >>>>>>>>>>> systemctl restart libvirtd >>>>>>>>>>> systemctl restart vdsmd >>>>>>>>>>> >>>>>>>>>>> Unfortunately, these commands did not help my situation like the >>>>>>>>>>> other individual. >>>>>>>>>>> >>>>>>>>>>> What log file would everyone like to see first? Given that I >>>>>>>>>>> still consider myself relatively new to Linux, please identify the >>>>>>>>>>> path for >>>>>>>>>>> the log file requested. >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> See above >>>>>>>>>> >>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> Also, because I plan on performing a clean install for this >>>>>>>>>>> thread using my process from above (I'm not expecting my outcome to >>>>>>>>>>> be any >>>>>>>>>>> different), are there any tips and tricks that might result in a >>>>>>>>>>> success? >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> I'd like to understand the issues you faced before suggesting to >>>>>>>>>> restart from scratch. >>>>>>>>>> Adding some people who may help as well. >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> Thank you for any and all help, >>>>>>>>>>> Brendan >>>>>>>>>>> >>>>>>>>>>> _______________________________________________ >>>>>>>>>>> Users mailing list >>>>>>>>>>> Users@ovirt.org >>>>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> >>>>>>>>>> SANDRO BONAZZOLA >>>>>>>>>> >>>>>>>>>> ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG VIRTUALIZATION >>>>>>>>>> R&D >>>>>>>>>> >>>>>>>>>> Red Hat EMEA <https://www.redhat.com/> >>>>>>>>>> <https://red.ht/sig> >>>>>>>>>> TRIED. TESTED. TRUSTED. <https://redhat.com/trusted> >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> >>>>>>>> SANDRO BONAZZOLA >>>>>>>> >>>>>>>> ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG VIRTUALIZATION R&D >>>>>>>> >>>>>>>> Red Hat EMEA <https://www.redhat.com/> >>>>>>>> <https://red.ht/sig> >>>>>>>> TRIED. TESTED. TRUSTED. <https://redhat.com/trusted> >>>>>>>> >>>>>>> >>>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> >>>>>> SANDRO BONAZZOLA >>>>>> >>>>>> ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG VIRTUALIZATION R&D >>>>>> >>>>>> Red Hat EMEA <https://www.redhat.com/> >>>>>> <https://red.ht/sig> >>>>>> TRIED. TESTED. TRUSTED. <https://redhat.com/trusted> >>>>>> >>>>> >>>>> >>>> >>>> >>>> -- >>>> >>>> SANDRO BONAZZOLA >>>> >>>> ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG VIRTUALIZATION R&D >>>> >>>> Red Hat EMEA <https://www.redhat.com/> >>>> <https://red.ht/sig> >>>> TRIED. TESTED. TRUSTED. <https://redhat.com/trusted> >>>> >>> >>> >> >
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users