I have installed KVM in the nested environment in ESXi6.x version is that recommended ? apart from Hosted engine is there any other alternate way to configure Engine HA cluster ?
-Nagaraju On Wed, Dec 2, 2015 at 4:11 PM, Simone Tiraboschi <stira...@redhat.com> wrote: > > > On Wed, Dec 2, 2015 at 11:25 AM, Budur Nagaraju <nbud...@gmail.com> wrote: > >> pls fine the logs from the below mentioned URL, >> >> http://pastebin.com/ZeKyyFbN >> > > OK, the issue is here: > > Thread-88::ERROR::2015-12-02 > 15:06:27,735::vm::2358::vm.Vm::(_startUnderlyingVm) > vmId=`93db4369-285f-48bc-bc68-181d9de41a3c`::The vm start process failed > Traceback (most recent call last): > File "/usr/share/vdsm/virt/vm.py", line 2298, in _startUnderlyingVm > self._run() > File "/usr/share/vdsm/virt/vm.py", line 3363, in _run > self._connection.createXML(domxml, flags), > File "/usr/lib/python2.6/site-packages/vdsm/libvirtconnection.py", line > 119, in wrapper > ret = f(*args, **kwargs) > File "/usr/lib64/python2.6/site-packages/libvirt.py", line 2709, in > createXML > if ret is None:raise libvirtError('virDomainCreateXML() failed', > conn=self) > libvirtError: unsupported configuration: Domain requires KVM, but it is > not available. Check that virtualization is enabled in the host BIOS, and > host configuration is setup to load the kvm modules. > Thread-88::DEBUG::2015-12-02 > 15:06:27,751::vm::2813::vm.Vm::(setDownStatus) > vmId=`93db4369-285f-48bc-bc68-181d9de41a3c`::Changed state to Down: > unsupported configuration: Domain requires KVM, but it is not available. > Check that virtualization is enabled in the host BIOS, and host > configuration is setup to load the kvm modules. (code=1) > > but it's pretty strange cause hosted-engine-setup already explicitly check > for visualization support and just exits with a clear error if not. > Did you played with the kvm module while hosted-engine-setup was running? > > Can you please hosted-engine-setup logs? > > >> >> On Fri, Nov 27, 2015 at 6:39 PM, Simone Tiraboschi <stira...@redhat.com> >> wrote: >> >>> >>> >>> On Fri, Nov 27, 2015 at 12:42 PM, Maxim Kovgan <kovg...@gmail.com> >>> wrote: >>> >>>> Maybe even makes sense to open a bugzilla ticket already. Better safe >>>> than sorry. >>>> >>> >>> We still need at least one log file to understand what happened. >>> >>> >>>> On Nov 27, 2015 11:35 AM, "Simone Tiraboschi" <stira...@redhat.com> >>>> wrote: >>>> >>>>> >>>>> On Fri, Nov 27, 2015 at 10:10 AM, Budur Nagaraju <nbud...@gmail.com> >>>>> wrote: >>>>> >>>>>> I do not know what logs you are expecting ? the logs which I got is >>>>>> pasted in the mail if you require in pastebin let me know I will upload >>>>>> there . >>>>>> >>>>> >>>>> >>>>> Please run sosreport utility and share the resulting archive where you >>>>> prefer. >>>>> You can follow this guide: >>>>> http://www.linuxtechi.com/how-to-create-sosreport-in-linux/ >>>>> >>>>>> >>>>>> >>>>>> On Fri, Nov 27, 2015 at 1:58 PM, Sandro Bonazzola < >>>>>> sbona...@redhat.com> wrote: >>>>>> >>>>>>> >>>>>>> >>>>>>> On Fri, Nov 27, 2015 at 8:34 AM, Budur Nagaraju <nbud...@gmail.com> >>>>>>> wrote: >>>>>>> >>>>>>>> I got only 10lines to in the vdsm logs and are below , >>>>>>>> >>>>>>>> >>>>>>> Can you please provide full sos report? >>>>>>> >>>>>>> >>>>>>> >>>>>>>> >>>>>>>> [root@he /]# tail -f /var/log/vdsm/vdsm.log >>>>>>>> Thread-100::DEBUG::2015-11-27 >>>>>>>> 12:58:57,360::resourceManager::616::Storage.ResourceManager::(releaseResource) >>>>>>>> Trying to release resource 'Storage.HsmDomainMonitorLock' >>>>>>>> Thread-100::DEBUG::2015-11-27 >>>>>>>> 12:58:57,360::resourceManager::635::Storage.ResourceManager::(releaseResource) >>>>>>>> Released resource 'Storage.HsmDomainMonitorLock' (0 active users) >>>>>>>> Thread-100::DEBUG::2015-11-27 >>>>>>>> 12:58:57,360::resourceManager::641::Storage.ResourceManager::(releaseResource) >>>>>>>> Resource 'Storage.HsmDomainMonitorLock' is free, finding out if anyone >>>>>>>> is >>>>>>>> waiting for it. >>>>>>>> Thread-100::DEBUG::2015-11-27 >>>>>>>> 12:58:57,360::resourceManager::649::Storage.ResourceManager::(releaseResource) >>>>>>>> No one is waiting for resource 'Storage.HsmDomainMonitorLock', Clearing >>>>>>>> records. >>>>>>>> Thread-100::INFO::2015-11-27 >>>>>>>> 12:58:57,360::logUtils::47::dispatcher::(wrapper) Run and protect: >>>>>>>> stopMonitoringDomain, Return response: None >>>>>>>> Thread-100::DEBUG::2015-11-27 >>>>>>>> 12:58:57,361::task::1191::Storage.TaskManager.Task::(prepare) >>>>>>>> Task=`0128b179-fdb3-474b-a196-8cc81a72a837`::finished: None >>>>>>>> Thread-100::DEBUG::2015-11-27 >>>>>>>> 12:58:57,361::task::595::Storage.TaskManager.Task::(_updateState) >>>>>>>> Task=`0128b179-fdb3-474b-a196-8cc81a72a837`::moving from state >>>>>>>> preparing -> >>>>>>>> state finished >>>>>>>> Thread-100::DEBUG::2015-11-27 >>>>>>>> 12:58:57,361::resourceManager::940::Storage.ResourceManager.Owner::(releaseAll) >>>>>>>> Owner.releaseAll requests {} resources {} >>>>>>>> Thread-100::DEBUG::2015-11-27 >>>>>>>> 12:58:57,361::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll) >>>>>>>> Owner.cancelAll requests {} >>>>>>>> Thread-100::DEBUG::2015-11-27 >>>>>>>> 12:58:57,361::task::993::Storage.TaskManager.Task::(_decref) >>>>>>>> Task=`0128b179-fdb3-474b-a196-8cc81a72a837`::ref 0 aborting False >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> On Thu, Nov 26, 2015 at 4:20 PM, Simone Tiraboschi < >>>>>>>> stira...@redhat.com> wrote: >>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> On Thu, Nov 26, 2015 at 11:05 AM, Budur Nagaraju < >>>>>>>>> nbud...@gmail.com> wrote: >>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> *Below are the entire logs* >>>>>>>>>> >>>>>>>>>> >>>>>>>>> Sorry, with the entire log I mean if you can attach or share >>>>>>>>> somewhere the whole /var/log/vdsm/vdsm.log cause the latest ten >>>>>>>>> lines are >>>>>>>>> not enough to point out the issue. >>>>>>>>> >>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> *[root@he ~]# tail -f /var/log/vdsm/vdsm.log * >>>>>>>>>> >>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>> 15:16:05,622::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read) >>>>>>>>>> Detected protocol xml from 127.0.0.1:50944 >>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>> 15:16:05,623::BindingXMLRPC::1173::XmlDetector::(handleSocket) xml >>>>>>>>>> over >>>>>>>>>> http detected from ('127.0.0.1', 50944) >>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>> 15:16:05,703::protocoldetector::187::vds.MultiProtocolAcceptor::(_add_connection) >>>>>>>>>> Adding connection from 127.0.0.1:50945 >>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>> 15:16:06,101::protocoldetector::201::vds.MultiProtocolAcceptor::(_remove_connection) >>>>>>>>>> Connection removed from 127.0.0.1:50945 >>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>> 15:16:06,101::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read) >>>>>>>>>> Detected protocol xml from 127.0.0.1:50945 >>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>> 15:16:06,101::BindingXMLRPC::1173::XmlDetector::(handleSocket) xml >>>>>>>>>> over >>>>>>>>>> http detected from ('127.0.0.1', 50945) >>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>> 15:16:06,182::protocoldetector::187::vds.MultiProtocolAcceptor::(_add_connection) >>>>>>>>>> Adding connection from 127.0.0.1:50946 >>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>> 15:16:06,710::protocoldetector::201::vds.MultiProtocolAcceptor::(_remove_connection) >>>>>>>>>> Connection removed from 127.0.0.1:50946 >>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>> 15:16:06,711::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read) >>>>>>>>>> Detected protocol xml from 127.0.0.1:50946 >>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>> 15:16:06,711::BindingXMLRPC::1173::XmlDetector::(handleSocket) xml >>>>>>>>>> over >>>>>>>>>> http detected from ('127.0.0.1', 50946) >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> *[root@he ~]# tail -f /var/log/vdsm/supervdsm.log * >>>>>>>>>> >>>>>>>>>> MainProcess::DEBUG::2015-11-26 >>>>>>>>>> 15:13:30,234::supervdsmServer::102::SuperVdsm.ServerCallback::(wrapper) >>>>>>>>>> call readMultipathConf with () {} >>>>>>>>>> MainProcess::DEBUG::2015-11-26 >>>>>>>>>> 15:13:30,234::supervdsmServer::109::SuperVdsm.ServerCallback::(wrapper) >>>>>>>>>> return readMultipathConf with ['# RHEV REVISION 1.1', '', 'defaults >>>>>>>>>> {', >>>>>>>>>> ' polling_interval 5', ' getuid_callout >>>>>>>>>> "/lib/udev/scsi_id --whitelisted --replace-whitespace >>>>>>>>>> --device=/dev/%n"', >>>>>>>>>> ' no_path_retry fail', ' user_friendly_names >>>>>>>>>> no', ' >>>>>>>>>> flush_on_last_del yes', ' fast_io_fail_tmo 5', ' >>>>>>>>>> dev_loss_tmo 30', ' max_fds 4096', >>>>>>>>>> '}', '', >>>>>>>>>> 'devices {', 'device {', ' vendor "HITACHI"', ' >>>>>>>>>> product "DF.*"', ' getuid_callout >>>>>>>>>> "/lib/udev/scsi_id --whitelisted --replace-whitespace >>>>>>>>>> --device=/dev/%n"', >>>>>>>>>> '}', 'device {', ' vendor "COMPELNT"', ' >>>>>>>>>> product "Compellent Vol"', ' no_path_retry >>>>>>>>>> fail', '}', 'device {', ' # multipath.conf.default', ' >>>>>>>>>> vendor "DGC"', ' product ".*"', ' >>>>>>>>>> product_blacklist "LUNZ"', ' path_grouping_policy >>>>>>>>>> "group_by_prio"', ' path_checker "emc_clariion"', ' >>>>>>>>>> hardware_handler "1 emc"', ' prio >>>>>>>>>> "emc"', ' >>>>>>>>>> failback immediate', ' rr_weight >>>>>>>>>> "uniform"', ' # vdsm required configuration', ' >>>>>>>>>> getuid_callout "/lib/udev/scsi_id --whitelisted >>>>>>>>>> --replace-whitespace --device=/dev/%n"', ' features >>>>>>>>>> "0"', >>>>>>>>>> ' no_path_retry fail', '}', '}'] >>>>>>>>>> MainProcess|Thread-13::DEBUG::2015-11-26 >>>>>>>>>> 15:13:31,365::supervdsmServer::102::SuperVdsm.ServerCallback::(wrapper) >>>>>>>>>> call getHardwareInfo with () {} >>>>>>>>>> MainProcess|Thread-13::DEBUG::2015-11-26 >>>>>>>>>> 15:13:31,397::supervdsmServer::109::SuperVdsm.ServerCallback::(wrapper) >>>>>>>>>> return getHardwareInfo with {'systemProductName': 'KVM', >>>>>>>>>> 'systemUUID': >>>>>>>>>> 'f91632f2-7a17-4ddb-9631-742f82a77480', 'systemFamily': 'Red Hat >>>>>>>>>> Enterprise >>>>>>>>>> Linux', 'systemVersion': 'RHEL 7.0.0 PC (i440FX + PIIX, 1996)', >>>>>>>>>> 'systemManufacturer': 'Red Hat'} >>>>>>>>>> MainProcess|Thread-21::DEBUG::2015-11-26 >>>>>>>>>> 15:13:35,393::supervdsmServer::102::SuperVdsm.ServerCallback::(wrapper) >>>>>>>>>> call validateAccess with ('qemu', ('qemu', 'kvm'), >>>>>>>>>> '/rhev/data-center/mnt/10.204.207.152:_home_vms', 5) {} >>>>>>>>>> MainProcess|Thread-21::DEBUG::2015-11-26 >>>>>>>>>> 15:13:35,395::supervdsmServer::109::SuperVdsm.ServerCallback::(wrapper) >>>>>>>>>> return validateAccess with None >>>>>>>>>> MainProcess|Thread-22::DEBUG::2015-11-26 >>>>>>>>>> 15:13:36,067::supervdsmServer::102::SuperVdsm.ServerCallback::(wrapper) >>>>>>>>>> call validateAccess with ('qemu', ('qemu', 'kvm'), >>>>>>>>>> '/rhev/data-center/mnt/10.204.207.152:_home_vms', 5) {} >>>>>>>>>> MainProcess|Thread-22::DEBUG::2015-11-26 >>>>>>>>>> 15:13:36,069::supervdsmServer::109::SuperVdsm.ServerCallback::(wrapper) >>>>>>>>>> return validateAccess with None >>>>>>>>>> MainProcess|PolicyEngine::DEBUG::2015-11-26 >>>>>>>>>> 15:13:40,619::supervdsmServer::102::SuperVdsm.ServerCallback::(wrapper) >>>>>>>>>> call ksmTune with ({'run': 0},) {} >>>>>>>>>> MainProcess|PolicyEngine::DEBUG::2015-11-26 >>>>>>>>>> 15:13:40,619::supervdsmServer::109::SuperVdsm.ServerCallback::(wrapper) >>>>>>>>>> return ksmTune with None >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> *[root@he ~]# tail -f /var/log/vdsm/connectivity.log * >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> 2015-11-26 15:02:02,632:DEBUG:recent_client:False >>>>>>>>>> 2015-11-26 15:04:44,975:DEBUG:recent_client:True >>>>>>>>>> 2015-11-26 15:05:15,039:DEBUG:recent_client:False >>>>>>>>>> 2015-11-26 15:07:23,311:DEBUG:recent_client:True >>>>>>>>>> 2015-11-26 15:08:25,774:DEBUG:recent_client:True, >>>>>>>>>> ovirtmgmt:(operstate:up speed:0 duplex:unknown), lo:(operstate:up >>>>>>>>>> speed:0 >>>>>>>>>> duplex:unknown), ;vdsmdummy;:(operstate:down speed:0 duplex:unknown), >>>>>>>>>> bond0:(operstate:down speed:0 duplex:unknown), eth0:(operstate:up >>>>>>>>>> speed:100 >>>>>>>>>> duplex:full) >>>>>>>>>> 2015-11-26 15:08:55,845:DEBUG:recent_client:False >>>>>>>>>> 2015-11-26 15:08:59,859:DEBUG:recent_client:True >>>>>>>>>> 2015-11-26 15:09:29,929:DEBUG:recent_client:False >>>>>>>>>> 2015-11-26 15:13:32,292:DEBUG:recent_client:True, >>>>>>>>>> ovirtmgmt:(operstate:up speed:0 duplex:unknown), lo:(operstate:up >>>>>>>>>> speed:0 >>>>>>>>>> duplex:unknown), ;vdsmdummy;:(operstate:down speed:0 duplex:unknown), >>>>>>>>>> bond0:(operstate:down speed:0 duplex:unknown), eth0:(operstate:up >>>>>>>>>> speed:100 >>>>>>>>>> duplex:full) >>>>>>>>>> 2015-11-26 15:14:02,363:DEBUG:recent_client:False >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> *[root@he ~]# tail -f /var/log/vdsm/mom.log * >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> 2015-11-26 15:13:30,581 - mom.Policy - INFO - Loaded policy >>>>>>>>>> '04-cputune' >>>>>>>>>> 2015-11-26 15:13:30,581 - mom.PolicyEngine - INFO - Policy Engine >>>>>>>>>> starting >>>>>>>>>> 2015-11-26 15:13:30,582 - mom.RPCServer - INFO - RPC Server is >>>>>>>>>> disabled >>>>>>>>>> 2015-11-26 15:13:40,618 - mom.Controllers.KSM - INFO - Updating >>>>>>>>>> KSM configuration: pages_to_scan:0 merge_across_nodes:8 run:0 >>>>>>>>>> sleep_millisecs:0 >>>>>>>>>> 2015-11-26 15:14:51,492 - mom.RPCServer - INFO - getStatistics() >>>>>>>>>> 2015-11-26 15:14:56,962 - mom.RPCServer - INFO - getStatistics() >>>>>>>>>> 2015-11-26 15:15:02,451 - mom.RPCServer - INFO - getStatistics() >>>>>>>>>> 2015-11-26 15:15:07,777 - mom.RPCServer - INFO - getStatistics() >>>>>>>>>> 2015-11-26 15:15:13,267 - mom.RPCServer - INFO - getStatistics() >>>>>>>>>> 2015-11-26 15:15:18,765 - mom.RPCServer - INFO - getStatistics() >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> On Thu, Nov 26, 2015 at 3:28 PM, Budur Nagaraju < >>>>>>>>>> nbud...@gmail.com> wrote: >>>>>>>>>> >>>>>>>>>>> Below are the logs, >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> [root@he ~]# tail -f /var/log/vdsm/vdsm.log >>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>> 15:16:05,622::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read) >>>>>>>>>>> Detected protocol xml from 127.0.0.1:50944 >>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>> 15:16:05,623::BindingXMLRPC::1173::XmlDetector::(handleSocket) xml >>>>>>>>>>> over >>>>>>>>>>> http detected from ('127.0.0.1', 50944) >>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>> 15:16:05,703::protocoldetector::187::vds.MultiProtocolAcceptor::(_add_connection) >>>>>>>>>>> Adding connection from 127.0.0.1:50945 >>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>> 15:16:06,101::protocoldetector::201::vds.MultiProtocolAcceptor::(_remove_connection) >>>>>>>>>>> Connection removed from 127.0.0.1:50945 >>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>> 15:16:06,101::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read) >>>>>>>>>>> Detected protocol xml from 127.0.0.1:50945 >>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>> 15:16:06,101::BindingXMLRPC::1173::XmlDetector::(handleSocket) xml >>>>>>>>>>> over >>>>>>>>>>> http detected from ('127.0.0.1', 50945) >>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>> 15:16:06,182::protocoldetector::187::vds.MultiProtocolAcceptor::(_add_connection) >>>>>>>>>>> Adding connection from 127.0.0.1:50946 >>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>> 15:16:06,710::protocoldetector::201::vds.MultiProtocolAcceptor::(_remove_connection) >>>>>>>>>>> Connection removed from 127.0.0.1:50946 >>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>> 15:16:06,711::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read) >>>>>>>>>>> Detected protocol xml from 127.0.0.1:50946 >>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>> 15:16:06,711::BindingXMLRPC::1173::XmlDetector::(handleSocket) xml >>>>>>>>>>> over >>>>>>>>>>> http detected from ('127.0.0.1', 50946) >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> On Thu, Nov 26, 2015 at 3:06 PM, Simone Tiraboschi < >>>>>>>>>>> stira...@redhat.com> wrote: >>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> On Thu, Nov 26, 2015 at 10:33 AM, Budur Nagaraju < >>>>>>>>>>>> nbud...@gmail.com> wrote: >>>>>>>>>>>> >>>>>>>>>>>>> I have done a fresh installation and now am getting the below >>>>>>>>>>>>> error, >>>>>>>>>>>>> >>>>>>>>>>>>> [ INFO ] Updating hosted-engine configuration >>>>>>>>>>>>> [ INFO ] Stage: Transaction commit >>>>>>>>>>>>> [ INFO ] Stage: Closing up >>>>>>>>>>>>> The following network ports should be opened: >>>>>>>>>>>>> tcp:5900 >>>>>>>>>>>>> tcp:5901 >>>>>>>>>>>>> udp:5900 >>>>>>>>>>>>> udp:5901 >>>>>>>>>>>>> An example of the required configuration for >>>>>>>>>>>>> iptables can be found at: >>>>>>>>>>>>> /etc/ovirt-hosted-engine/iptables.example >>>>>>>>>>>>> In order to configure firewalld, copy the files from >>>>>>>>>>>>> /etc/ovirt-hosted-engine/firewalld to >>>>>>>>>>>>> /etc/firewalld/services >>>>>>>>>>>>> and execute the following commands: >>>>>>>>>>>>> firewall-cmd -service hosted-console >>>>>>>>>>>>> [ INFO ] Creating VM >>>>>>>>>>>>> [ ERROR ] Failed to execute stage 'Closing up': Cannot set >>>>>>>>>>>>> temporary password for console connection. The VM may not have >>>>>>>>>>>>> been >>>>>>>>>>>>> created: please check VDSM logs >>>>>>>>>>>>> [ INFO ] Stage: Clean up >>>>>>>>>>>>> [ INFO ] Generating answer file >>>>>>>>>>>>> '/var/lib/ovirt-hosted-engine-setup/answers/answers-20151126145701.conf' >>>>>>>>>>>>> [ INFO ] Stage: Pre-termination >>>>>>>>>>>>> [ INFO ] Stage: Termination >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> [root@he ovirt]# tail -f /var/log/vdsm/ >>>>>>>>>>>>> backup/ connectivity.log mom.log >>>>>>>>>>>>> supervdsm.log vdsm.log >>>>>>>>>>>>> [root@he ovirt]# tail -f /var/log/vdsm/vdsm.log >>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>>>> 14:57:07,564::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read) >>>>>>>>>>>>> Detected protocol xml from 127.0.0.1:42741 >>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>>>> 14:57:07,564::BindingXMLRPC::1173::XmlDetector::(handleSocket) >>>>>>>>>>>>> xml over >>>>>>>>>>>>> http detected from ('127.0.0.1', 42741) >>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>>>> 14:57:07,644::protocoldetector::187::vds.MultiProtocolAcceptor::(_add_connection) >>>>>>>>>>>>> Adding connection from 127.0.0.1:42742 >>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>>>> 14:57:08,088::protocoldetector::201::vds.MultiProtocolAcceptor::(_remove_connection) >>>>>>>>>>>>> Connection removed from 127.0.0.1:42742 >>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>>>> 14:57:08,088::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read) >>>>>>>>>>>>> Detected protocol xml from 127.0.0.1:42742 >>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>>>> 14:57:08,088::BindingXMLRPC::1173::XmlDetector::(handleSocket) >>>>>>>>>>>>> xml over >>>>>>>>>>>>> http detected from ('127.0.0.1', 42742) >>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>>>> 14:57:08,171::protocoldetector::187::vds.MultiProtocolAcceptor::(_add_connection) >>>>>>>>>>>>> Adding connection from 127.0.0.1:42743 >>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>>>> 14:57:08,572::protocoldetector::201::vds.MultiProtocolAcceptor::(_remove_connection) >>>>>>>>>>>>> Connection removed from 127.0.0.1:42743 >>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>>>> 14:57:08,573::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read) >>>>>>>>>>>>> Detected protocol xml from 127.0.0.1:42743 >>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>>>> 14:57:08,573::BindingXMLRPC::1173::XmlDetector::(handleSocket) >>>>>>>>>>>>> xml over >>>>>>>>>>>>> http detected from ('127.0.0.1', 42743) >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> It failed before, can you please attach the whole VDSM logs? >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> On Thu, Nov 26, 2015 at 2:01 PM, Simone Tiraboschi < >>>>>>>>>>>>> stira...@redhat.com> wrote: >>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> On Thu, Nov 26, 2015 at 7:30 AM, Budur Nagaraju < >>>>>>>>>>>>>> nbud...@gmail.com> wrote: >>>>>>>>>>>>>> >>>>>>>>>>>>>>> Its a fresh setup ,I have deleted all the vms ,still am >>>>>>>>>>>>>>> facing same issues . >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>> Can you please paste the output of >>>>>>>>>>>>>> vdsClient -s 0 list >>>>>>>>>>>>>> ? >>>>>>>>>>>>>> thanks >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> On Thu, Nov 26, 2015 at 11:56 AM, Oved Ourfali < >>>>>>>>>>>>>>> oourf...@redhat.com> wrote: >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Hi >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Seems like you have existing VMs running on the host (you >>>>>>>>>>>>>>>> can check that by looking for qemu processes on your host). >>>>>>>>>>>>>>>> Is that a clean deployment, or was the host used before for >>>>>>>>>>>>>>>> running VMs? >>>>>>>>>>>>>>>> Perhaps you already ran the hosted engine setup, and the VM >>>>>>>>>>>>>>>> was left there? >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> CC-ing Sandro who is more familiar in that than me. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>>>> Oved >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> On Thu, Nov 26, 2015 at 7:07 AM, Budur Nagaraju < >>>>>>>>>>>>>>>> nbud...@gmail.com> wrote: >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> HI >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Getting below error while configuring Hosted engine, >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> root@he ~]# hosted-engine --deploy >>>>>>>>>>>>>>>>> [ INFO ] Stage: Initializing >>>>>>>>>>>>>>>>> [ INFO ] Generating a temporary VNC password. >>>>>>>>>>>>>>>>> [ INFO ] Stage: Environment setup >>>>>>>>>>>>>>>>> Continuing will configure this host for serving >>>>>>>>>>>>>>>>> as hypervisor and create a VM where you have to install oVirt >>>>>>>>>>>>>>>>> Engine >>>>>>>>>>>>>>>>> afterwards. >>>>>>>>>>>>>>>>> Are you sure you want to continue? (Yes, >>>>>>>>>>>>>>>>> No)[Yes]: yes >>>>>>>>>>>>>>>>> Configuration files: [] >>>>>>>>>>>>>>>>> Log file: >>>>>>>>>>>>>>>>> /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20151126102302-bkozgk.log >>>>>>>>>>>>>>>>> Version: otopi-1.3.2 (otopi-1.3.2-1.el6) >>>>>>>>>>>>>>>>> It has been detected that this program is >>>>>>>>>>>>>>>>> executed through an SSH connection without using screen. >>>>>>>>>>>>>>>>> Continuing with the installation may lead to >>>>>>>>>>>>>>>>> broken installation if the network connection fails. >>>>>>>>>>>>>>>>> It is highly recommended to abort the >>>>>>>>>>>>>>>>> installation and run it inside a screen session using command >>>>>>>>>>>>>>>>> "screen". >>>>>>>>>>>>>>>>> Do you want to continue anyway? (Yes, No)[No]: >>>>>>>>>>>>>>>>> yes >>>>>>>>>>>>>>>>> [WARNING] Cannot detect if hardware supports virtualization >>>>>>>>>>>>>>>>> [ INFO ] Bridge ovirtmgmt already created >>>>>>>>>>>>>>>>> [ INFO ] Stage: Environment packages setup >>>>>>>>>>>>>>>>> [ INFO ] Stage: Programs detection >>>>>>>>>>>>>>>>> [ INFO ] Stage: Environment setup >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> *[ ERROR ] The following VMs has been found: >>>>>>>>>>>>>>>>> 2b8d6d91-d838-44f6-ae3b-c92cda014280[ ERROR ] Failed to >>>>>>>>>>>>>>>>> execute stage >>>>>>>>>>>>>>>>> 'Environment setup': Cannot setup Hosted Engine with other >>>>>>>>>>>>>>>>> VMs running* >>>>>>>>>>>>>>>>> [ INFO ] Stage: Clean up >>>>>>>>>>>>>>>>> [ INFO ] Generating answer file >>>>>>>>>>>>>>>>> '/var/lib/ovirt-hosted-engine-setup/answers/answers-20151126102310.conf' >>>>>>>>>>>>>>>>> [ INFO ] Stage: Pre-termination >>>>>>>>>>>>>>>>> [ INFO ] Stage: Termination >>>>>>>>>>>>>>>>> [root@he ~]# >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>>>>> Users mailing list >>>>>>>>>>>>>>>>> Users@ovirt.org >>>>>>>>>>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>>> Users mailing list >>>>>>>>>>>>>>> Users@ovirt.org >>>>>>>>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>>> _______________________________________________ >>>>>>>> Users mailing list >>>>>>>> Users@ovirt.org >>>>>>>> http://lists.ovirt.org/mailman/listinfo/users >>>>>>>> >>>>>>>> >>>>>>> >>>>>>> >>>>>>> -- >>>>>>> Sandro Bonazzola >>>>>>> Better technology. Faster innovation. Powered by community >>>>>>> collaboration. >>>>>>> See how it works at redhat.com >>>>>>> >>>>>> >>>>>> >>>>> >>>>> _______________________________________________ >>>>> Users mailing list >>>>> Users@ovirt.org >>>>> http://lists.ovirt.org/mailman/listinfo/users >>>>> >>>>> >>> >> >
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users