* Ayal Baron <aba...@redhat.com> [2012-01-18 10:35]: > > > ----- Original Message ----- > > * Haim Ateya <hat...@redhat.com> [2012-01-18 10:15]: > > > On Wed 18 Jan 2012 06:09:46 PM IST, Ryan Harper wrote: > > > >* Haim Ateya<hat...@redhat.com> [2012-01-18 08:02]: > > > >>On Wed 18 Jan 2012 03:48:08 PM IST, Ryan Harper wrote: > > > >>>* Haim Ateya<hat...@redhat.com> [2012-01-18 07:13]: > > > >>>>On Wed 18 Jan 2012 02:59:01 PM IST, Ryan Harper wrote: > > > >>>>>I've created some f16 VMs that contain both ovirt-engine and a > > > >>>>>few > > > >>>>>to run vdsm as nodes. When I add in the VM host into the > > > >>>>>engine and it > > > >>>>>attempts to install vdsm (even though I've already installed > > > >>>>>vdsm) the > > > >>>>>install fails because the vdsm install script is checking to > > > >>>>>see if the > > > >>>>>host has virt capabilities; since I'm not running nested KVM, > > > >>>>>this > > > >>>>>fails. Is there a way to work around this can enable a VM to > > > >>>>>be a host > > > >>>>>in oVirt? I had heard in the past there was a way to create > > > >>>>>fake VMs > > > >>>>>when attempting to do ovirt-engine stress testing, wondering > > > >>>>>if that > > > >>>>>might be of help here. > > > >>>>> > > > >>>>>Also, are their vdsm rpms built for RHEL6.x available? > > > >>>>> > > > >>>>>Thanks! > > > >>>>> > > > >>>> > > > >>>>Hi Rayn, > > > >>>> > > > >>>>- login to your ovirt-engine machine > > > >>>>- edit > > > >>>>/usr/share/ovirt-engine/engine.ear/components.war/vds/vds_bootstrap.py > > > >>>>- comment out the following: > > > >>>> > > > >>>> 836 if not oDeploy.virtExplorer(random_num): > > > >>>> 837 logging.error('virtExplorer test failed') > > > >>>> 838 return False > > > >>>>- reinstall host > > > >>> > > > >>>So I'm getting further, but now the bootstrap.log has more > > > >>>errors below. > > > >>>If I follow the test day instructions, it indicates to install > > > >>>vdsm and > > > >>>includes instructions, but it's clear that ovirt-engine is > > > >>>configured by > > > >>>default to push out vdsm and install it. If I've already > > > >>>configured and > > > >>>installed vdsm on the node is there any way to not attempting to > > > >>>bootstrap > > > >>>vdsm > > > >>>at all and just attempt to have it connect? > > > >>> > > > >>> > > > >>> > > > >>>Wed, 18 Jan 2012 08:35:37 DEBUG<BSTRAP component='VDS PACKAGES' > > > >>>status='OK' result='qemu-kvm-tools' > > > >>>message='qemu-kvm-tools-0.15.1-3.fc16.x86_64 '/> > > > >>>Wed, 18 Jan 2012 08:35:37 DEBUG Basic configuration found, > > > >>>skipping > > > >>>this step > > > >>>Wed, 18 Jan 2012 08:35:37 DEBUG<BSTRAP component='CreateConf' > > > >>>status='OK' > > > >>>message='Basic configuration found, skipping this step'/> > > > >>>Wed, 18 Jan 2012 08:35:37 DEBUG<BSTRAP > > > >>>component='CheckLocalHostname' > > > >>>status='OK' message='Local hostname is correct.'/> > > > >>>Wed, 18 Jan 2012 08:35:37 DEBUG Bridge ovirtmgmt not found, > > > >>>need to > > > >>>create it. > > > >>>Wed, 18 Jan 2012 08:35:37 DEBUG getAddress Entry. > > > >>>url=http://ichigo-dom223.phx.austin.ibm.com:8080/Components/vds/ > > > >>>Wed, 18 Jan 2012 08:35:37 DEBUG getAddress return. > > > >>>address=ichigo-dom223.phx.austin.ibm.com port=8080 > > > >>>Wed, 18 Jan 2012 08:35:37 DEBUG makeBridge begin. > > > >>>Wed, 18 Jan 2012 08:35:37 DEBUG _getMGTIface: read host name: > > > >>>ichigo-dom223.phx.austin.ibm.com > > > >>>Wed, 18 Jan 2012 08:35:37 DEBUG _getMGTIface: using host name > > > >>>ichigo-dom223.phx.austin.ibm.com strIP= 192.168.68.223 > > > >>>Wed, 18 Jan 2012 08:35:37 DEBUG _getMGTIface > > > >>>IP=192.168.68.223 > > > >>>strIface=engine > > > >>>Wed, 18 Jan 2012 08:35:37 ERROR makeBridge found existing > > > >>>bridge > > > >>>named: > > > >>>engine > > > >>>Wed, 18 Jan 2012 08:35:37 ERROR makeBridge errored: out= > > > >>>err=None > > > >>>ret=None > > > >>>Wed, 18 Jan 2012 08:35:37 DEBUG makeBridge return. > > > >>>Wed, 18 Jan 2012 08:35:37 ERROR addNetwork error trying to > > > >>>add > > > >>>management bridge > > > >>>Wed, 18 Jan 2012 08:35:37 DEBUG<BSTRAP component='SetNetworking' > > > >>>status='FAIL' message='addNetwork error trying to add management > > > >>>bridge'/> > > > >>>Wed, 18 Jan 2012 08:35:37 DEBUG getAddress Entry. > > > >>>url=http://ichigo-dom223.phx.austin.ibm.com:8080/Components/vds/ > > > >>>Wed, 18 Jan 2012 08:35:37 DEBUG getAddress return. > > > >>>address=ichigo-dom223.phx.austin.ibm.com port=8080 > > > >>>Wed, 18 Jan 2012 08:35:37 DEBUG getRemoteFile start. IP = > > > >>>ichigo-dom223.phx.austin.ibm.com port = 8080 fileName = > > > >>>"/engine.ssh.key.txt" > > > >>>Wed, 18 Jan 2012 08:35:37 DEBUG /engine.ssh.key.txt failed in > > > >>>HTTPS. > > > >>>Retrying using HTTP. > > > >>>Traceback (most recent call last): > > > >>> File "/tmp/deployUtil.py", line 1334, in getRemoteFile > > > >>> conn.sock = getSSLSocket(sock, certPath) > > > >>> File "/tmp/deployUtil.py", line 1178, in getSSLSocket > > > >>> cert_reqs=ssl.CERT_REQUIRED) > > > >>> File "/usr/lib64/python2.7/ssl.py", line 372, in wrap_socket > > > >>> ciphers=ciphers) > > > >>> File "/usr/lib64/python2.7/ssl.py", line 132, in __init__ > > > >>> ciphers) > > > >>>SSLError: [Errno 185090050] _ssl.c:340: error:0B084002:x509 > > > >>>certificate > > > >>>routines:X509_load_cert_crl_file:system lib > > > >>>Wed, 18 Jan 2012 08:35:37 DEBUG getRemoteFile end. > > > >>>Wed, 18 Jan 2012 08:35:37 DEBUG handleSSHKey start > > > >>>Wed, 18 Jan 2012 08:35:37 DEBUG Failed to read > > > >>>/root/.ssh/authorized_keys > > > >>>Wed, 18 Jan 2012 08:35:37 DEBUG handleSSHKey: failed to chmod > > > >>>authorized_keys > > > >>>Traceback (most recent call last): > > > >>> File "/tmp/deployUtil.py", line 608, in handleSSHKey > > > >>> silentRestoreCon(P_ROOT_AUTH_KEYS) > > > >>> File "/tmp/deployUtil.py", line 576, in silentRestoreCon > > > >>> import selinux > > > >>> File > > > >>> "/usr/lib64/python2.7/site-packages/selinux/__init__.py", > > > >>> line > > > >>> 26, > > > >>> in<module> > > > >>> _selinux = swig_import_helper() > > > >>> File > > > >>> "/usr/lib64/python2.7/site-packages/selinux/__init__.py", > > > >>> line > > > >>> 22, > > > >>> in swig_import_helper > > > >>> _mod = imp.load_module('_selinux', fp, pathname, > > > >>> description) > > > >>>ImportError: > > > >>>/usr/lib64/python2.7/site-packages/selinux/_selinux.so: > > > >>>undefined symbol: selinux_check_access > > > >>>Wed, 18 Jan 2012 08:35:37 DEBUG handleSSHKey end > > > >>>Wed, 18 Jan 2012 08:35:37 DEBUG<BSTRAP component='SetSSHAccess' > > > >>>status='FAIL' message='Failed to write server~s SSH key.'/> > > > >>>Wed, 18 Jan 2012 08:35:37 ERROR setSSHAccess test failed > > > >>>Wed, 18 Jan 2012 08:35:37 DEBUG<BSTRAP component='RHEV_INSTALL' > > > >>>status='FAIL'/> > > > >>>Wed, 18 Jan 2012 08:35:37 DEBUG **** End VDS Validation **** > > > >>> > > > >>> > > > >>> > > > >>> > > > >>>>- add fake_kvm_support = True to your vdsm.conf under > > > >>>>/etc/vdsm/vdsm.conf > > > >>>>- restart vdsmd service > > > >>> > > > >> > > > >>please make sure selinux is set at least on permissive mode; > > > >> > > > >>sed -i 's/SELINUX=disabled/SELINUX=permissive/g' > > > >>/etc/sysconfig/selinux > > > >> > > > >>reboot and reinstall. > > > >> > > > >>anyhow, if this is the case, its a known issue and patch is > > > >>pending > > > >>upstream. > > > > > > > >I did this, but I was also able to just re-run the installer and > > > >bootstrap completed. However, now I have another issue. > > > > > > > >THe host is marked unresponsive in engine, engine.log shows a > > > >connectivity issue, but both hosts can ping and share data. > > > > > > > >. Stage completed. (Stage: Running second installation script on > > > >Host) > > > >2012-01-18 09:58:08,550 INFO > > > >[org.ovirt.engine.core.utils.hostinstall.MinaInstallWrapper] > > > >(pool-5-thread-49) RunSSHCommand returns true > > > >2012-01-18 09:58:08,550 INFO > > > > [org.ovirt.engine.core.bll.VdsInstaller] > > > >(pool-5-thread-49) FinishCommand ended:true > > > >2012-01-18 09:58:08,554 INFO > > > >[org.ovirt.engine.core.bll.InstallVdsCommand] (pool-5-thread-49) > > > >After > > > >Installation pool-5-thread-49 > > > >2012-01-18 09:58:08,555 INFO > > > >[org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand] > > > >(pool-5-thread-49) START, SetVdsStatusVDSCommand(vdsId = > > > >8c627fa8-41d8-11e1-8d2f-00fffe0000df, status=Reboot, > > > >nonOperationalReason=NONE), log id: 703c3cbd > > > >2012-01-18 09:58:08,560 INFO > > > >[org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand] > > > >(pool-5-thread-49) FINISH, SetVdsStatusVDSCommand, log id: > > > >703c3cbd > > > >2012-01-18 09:58:08,560 INFO > > > > [org.ovirt.engine.core.bll.VdsCommand] > > > >(pool-5-thread-50) Waiting 300 seconds, for server to finish > > > >reboot > > > >process. > > > >2012-01-18 10:03:08,561 INFO > > > >[org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand] > > > >(pool-5-thread-50) START, SetVdsStatusVDSCommand(vdsId = > > > >8c627fa8-41d8-11e1-8d2f-00fffe0000df, status=NonResponsive, > > > >nonOperationalReason=NONE), log id: 3e57bdd2 > > > >2012-01-18 10:03:08,570 INFO > > > >[org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand] > > > >(pool-5-thread-50) FINISH, SetVdsStatusVDSCommand, log id: > > > >3e57bdd2 > > > >2012-01-18 10:03:10,201 ERROR > > > >[org.ovirt.engine.core.vdsbroker.vdsbroker.VdsBrokerCommand] > > > >(QuartzScheduler_Worker-35) XML RPC error in command > > > >GetCapabilitiesVDS ( > > > >Vds: ichigo-dom224 ), the error was: > > > >java.util.concurrent.ExecutionException: > > > >java.lang.reflect.InvocationTargetException > > > > > > > > > > > >I can query vdsm on the on the node: > > > > > > > >[root@f16-node1 ~]# vdsClient -s 0 getVdsCaps > > > > HBAInventory = {'iSCSI': [{'InitiatorName': > > > > 'iqn.1994-05.com.redhat:2abcda43e16d'}], 'FC': []} > > > > ISCSIInitiatorName = iqn.1994-05.com.redhat:2abcda43e16d > > > > bondings = {'bond4': {'hwaddr': '00:00:00:00:00:00', 'cfg': > > > > {}, > > > > 'netmask': '', 'addr': '', 'slaves': []}, 'bond0': {'hwaddr': > > > > '00:00:00:00:00:00', 'cfg': {}, 'netmask': '', 'addr': '', > > > > 'slaves': > > > > []}, 'bond1': {'hwaddr': '00:00:00:00:00:00', 'cfg': {}, > > > > 'netmask': > > > > '', 'addr': '', 'slaves': []}, 'bond2': {'hwaddr': > > > > '00:00:00:00:00:00', 'cfg': {}, 'netmask': '', 'addr': '', > > > > 'slaves': > > > > []}, 'bond3': {'hwaddr': '00:00:00:00:00:00', 'cfg': {}, > > > > 'netmask': > > > > '', 'addr': '', 'slaves': []}} > > > > clusterLevels = ['3.0'] > > > > cpuCores = 1 > > > > cpuFlags = > > > > > > > > pge,clflush,sep,syscall,tsc,vmx,cmov,nx,constant_tsc,pat,sse4_1,lm,msr,fpu,fxsr,pae,nopl,mmx,cx8,mce,de,mca,pse,pni,popcnt,apic,sse,sse4_2,lahf_lm,sse2,hypervisor,up,ssse3,cx16,pse36,mtrr,x2apicmodel_486,model_pentium,model_pentium2,model_pentium3,model_pentiumpro,model_qemu32,model_coreduo,model_core2duo,model_n270,model_Conroe,model_Penryn,model_Nehalem,model_Opteron_G1 > > > > cpuModel = Intel(Fake) CPU > > > > cpuSockets = 1 > > > > cpuSpeed = 2800.482 > > > > emulatedMachines = ['pc-0.14', 'pc', 'fedora-13', 'pc-0.13', > > > > 'pc-0.12', 'pc-0.11', 'pc-0.10', 'isapc'] > > > > guestOverhead = 65 > > > > hooks = {} > > > > kvmEnabled = true > > > > management_ip = > > > > memSize = 7988 > > > > networks = {'ovirtmgmt': {'addr': '192.168.68.224', 'cfg': > > > > {'DEVICE': > > > > 'ovirtmgmt', 'DELAY': '0', 'BOOTPROTO': 'dhcp', 'TYPE': > > > > 'Bridge', > > > > 'ONBOOT': 'yes'}, 'ports': ['eth0'], 'netmask': > > > > '255.255.192.0', > > > > 'stp': 'off', 'gateway': '192.168.68.1'}} > > > > nics = {'eth0': {'hwaddr': '00:FF:FE:00:00:E0', 'netmask': > > > > '', > > > > 'speed': 0, 'addr': ''}} > > > > operatingSystem = {'release': '1', 'version': '16', 'name': > > > > 'Fedora'} > > > > packages2 = {'kernel': {'release': '7.fc16.x86_64', > > > > 'buildtime': > > > > 1320196248.0, 'version': '3.1.0'}, 'spice-server': > > > > {'release': > > > > '1.fc16', 'buildtime': '1321276111', 'version': '0.10.0'}, > > > > 'vdsm': > > > > {'release': '0.fc16', 'buildtime': '1326734129', 'version': > > > > '4.9.3.1'}, 'qemu-kvm': {'release': '3.fc16', 'buildtime': > > > > '1321651456', 'version': '0.15.1'}, 'libvirt': {'release': > > > > '4.fc16', > > > > 'buildtime': '1324326688', 'version': '0.9.6'}, 'qemu-img': > > > > {'release': '3.fc16', 'buildtime': '1321651456', 'version': > > > > '0.15.1'}} > > > > reservedMem = 321 > > > > software_revision = 0 > > > > software_version = 4.9 > > > > supportedProtocols = ['2.2', '2.3'] > > > > supportedRHEVMs = ['3.0'] > > > > uuid = 922F4AE6-8EEA-4B11-44C4-EA1E1D665AC2_00:FF:FE:00:00:E0 > > > > version_name = Snow Man > > > > vlans = {} > > > > vmTypes = ['kvm'] > > > > > > > > > > > > > > can you check if problem is solved if you run iptables -F ? > > > > It doesn't. > > can you also post the vdsm.log to see if the request made it and was rejected > for some reason?
yes, I think this is the issue: Thread-1016::ERROR::2012-01-18 11:36:50,986::SecureXMLRPCServer::73::root::(handle_error) client ('192.168.68.223', 58819) Traceback (most recent call last): File "/usr/lib64/python2.7/SocketServer.py", line 582, in process_request_thread self.finish_request(request, client_address) File "/usr/share/vdsm/SecureXMLRPCServer.py", line 66, in finish_request request.do_handshake() File "/usr/lib64/python2.7/ssl.py", line 296, in do_handshake self._sslobj.do_handshake() I have vdsm.conf with ssl=true, however, if I set ssl=false, then I cannot query vdsm from the localhost client: [root@f16-node1 vdsm]# vdsClient -s 0 getVdsCaps with ssl=false, that returns connection refused. -- Ryan Harper Software Engineer; Linux Technology Center IBM Corp., Austin, Tx ry...@us.ibm.com _______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users