Yes, I am using OVS as the switch type and I did not know that it was not supported officially.
The output of ovs-vsctl show is as follows: f634d53e-4849-488b-8454-6b1fafa7c6ac ovs_version: "2.6.90" I am attaching OVS switch logs below: /var/log/openvswitch/ovsdb-server.log 2017-02-06T09:46:07.788Z|00001|vlog|INFO|opened log file /var/log/openvswitch/ovsdb-server.log 2017-02-06T09:46:07.791Z|00002|ovsdb_server|INFO|ovsdb-server (Open vSwitch) 2.6.90 2017-02-06T09:46:17.802Z|00003|memory|INFO|2296 kB peak resident set size after 10.0 seconds 2017-02-06T09:46:17.802Z|00004|memory|INFO|cells:16 json-caches:1 monitors:1 sessions:1 ovs-vswitchd.log 2017-02-06T09:46:07.999Z|00001|vlog|INFO|opened log file /var/log/openvswitch/ovs-vswitchd.log 2017-02-06T09:46:08.036Z|00002|ovs_numa|INFO|Discovered 24 CPU cores on NUMA node 0 2017-02-06T09:46:08.036Z|00003|ovs_numa|INFO|Discovered 24 CPU cores on NUMA node 1 2017-02-06T09:46:08.036Z|00004|ovs_numa|INFO|Discovered 2 NUMA nodes and 48 CPU cores 2017-02-06T09:46:08.037Z|00005|reconnect|INFO|unix:/var/run/openvswitch/db.sock: connecting... 2017-02-06T09:46:08.037Z|00006|reconnect|INFO|unix:/var/run/openvswitch/db.sock: connected 2017-02-06T09:46:08.039Z|00007|bridge|INFO|ovs-vswitchd (Open vSwitch) 2.6.90 What should I do now? The engine says that "Host host2 does not comply with the cluster Default networks, the following networks are missing on host: 'ovirtmgmt'" What other logs should I attach? Thanks Shalabh Goel On Sun, Feb 5, 2017 at 1:10 PM, Edward Haas <eh...@redhat.com> wrote: > Based on what I can see, you used OVS as the switch type and it seems ovs > (openvswitch) is not properly installed on your host. > Make sure that you have ovs operational by issuing "ovs-vsctl show". > > You should note that OVS network support is not an official release > feature, and you should use it on 4.1 and up versions. > Fixes will be probably submitted to master (appearing in nightly builds). > > Next time please include the mailing-list in your replies and attach the > log files, it is less spamming. > > Thanks, > Edy. > > On Fri, Feb 3, 2017 at 5:07 AM, Shalabh Goel <shalabhgoe...@gmail.com> > wrote: > >> log from messages >> >> Feb 3 08:27:53 ovirtnode3 ovs-vsctl: >> ovs|00001|db_ctl_base|ERR|unix:/var/run/openvswitch/db.sock: >> database connection failed (No such file or directory) >> Feb 3 08:27:53 ovirtnode3 journal: vdsm vds ERROR Executing commands >> failed: ovs-vsctl: unix:/var/run/openvswitch/db.sock: database >> connection failed (No su >> ch file or directory)#012Traceback (most recent call last):#012 File >> "/usr/share/vdsm/API.py", line 1531, in setupNetworks#012 >> supervdsm.getProxy().setup >> Networks(networks, bondings, options)#012 File >> "/usr/lib/python2.7/site-packages/vdsm/supervdsm.py", line 53, in >> __call__#012 return callMethod()#012 Fi >> le "/usr/lib/python2.7/site-packages/vdsm/supervdsm.py", line 51, in >> <lambda>#012 **kwargs)#012 File "<string>", line 2, in >> setupNetworks#012 File "/usr >> /lib64/python2.7/multiprocessing/managers.py", line 773, in >> _callmethod#012 raise convert_to_error(kind, >> result)#012ConfigNetworkError: (21, 'Executing co >> mmands failed: ovs-vsctl: unix:/var/run/openvswitch/db.sock: database >> connection failed (No such file or directory)') >> >> Log from vdsm.log >> >> Reactor thread::INFO::2017-02-03 08:32:24,638::protocoldetector >> ::72::ProtocolDetector.AcceptorImpl::(handle_accept) Accepted connection >> from ::1:53082 >> Reactor thread::DEBUG::2017-02-03 08:32:24,643::protocoldetector >> ::91::ProtocolDetector.Detector::(__init__) Using required_size=11 >> Reactor thread::INFO::2017-02-03 08:32:24,643::protocoldetector >> ::127::ProtocolDetector.Detector::(handle_read) Detected protocol xml >> from ::1:53082 >> Reactor thread::DEBUG::2017-02-03 08:32:24,643::bindingxmlrpc::1 >> 317::XmlDetector::(handle_socket) xml over http detected from ('::1', >> 53082) >> BindingXMLRPC::INFO::2017-02-03 >> 08:32:24,643::xmlrpc::73::vds.XMLRPCServer::(handle_request) >> Starting request handler for ::1:53082 >> Thread-20::INFO::2017-02-03 08:32:24,644::xmlrpc::83::vds. >> XMLRPCServer::(_process_requests) Request handler for ::1:53082 started >> Thread-20::DEBUG::2017-02-03 >> 08:32:24,644::bindingxmlrpc::1263::vds::(wrapper) >> client [::1]::call getAllVmStats with () {} >> Thread-20::DEBUG::2017-02-03 >> 08:32:24,644::bindingxmlrpc::1270::vds::(wrapper) >> return getAllVmStats with {'status': {'message': 'Done', 'code': 0}, >> 'statsLis >> t': (suppressed)} >> Thread-20::INFO::2017-02-03 08:32:24,644::bindingxmlrpc::1297::vds::(wrapper) >> RPC call getAllVmStats finished (code=0) in 0.00 seconds >> Thread-20::INFO::2017-02-03 08:32:24,645::xmlrpc::91::vds. >> XMLRPCServer::(_process_requests) Request handler for ::1:53082 stopped >> periodic/0::DEBUG::2017-02-03 08:32:24,907::sampling::508::v >> irt.sampling.VMBulkSampler::(__call__) sampled timestamp 4347942.15 >> elapsed 0.010 acquired True d >> omains all >> periodic/2::DEBUG::2017-02-03 08:32:24,908::task::599::stora >> ge.TaskManager.Task::(_updateState) >> (Task='bb07f89c-6165-42bc-902f-1b7701b813fa') >> moving from sta >> te init -> state preparing >> periodic/2::INFO::2017-02-03 >> 08:32:24,908::logUtils::49::dispatcher::(wrapper) >> Run and protect: repoStats(options=None) >> periodic/2::INFO::2017-02-03 >> 08:32:24,908::logUtils::52::dispatcher::(wrapper) >> Run and protect: repoStats, Return response: {} >> periodic/2::DEBUG::2017-02-03 08:32:24,909::task::1195::stor >> age.TaskManager.Task::(prepare) (Task='bb07f89c-6165-42bc-902f-1b7701b813fa') >> finished: {} >> periodic/2::DEBUG::2017-02-03 08:32:24,909::task::599::stora >> ge.TaskManager.Task::(_updateState) >> (Task='bb07f89c-6165-42bc-902f-1b7701b813fa') >> moving from sta >> te preparing -> state finished >> periodic/2::DEBUG::2017-02-03 08:32:24,909::resourceManager: >> :910::storage.ResourceManager.Owner::(releaseAll) Owner.releaseAll >> requests {} resources {} >> periodic/2::DEBUG::2017-02-03 08:32:24,909::resourceManager: >> :947::storage.ResourceManager.Owner::(cancelAll) Owner.cancelAll >> requests {} >> periodic/2::DEBUG::2017-02-03 >> 08:32:24,909::task::997::storage.TaskManager.Task::(_decref) >> (Task='bb07f89c-6165-42bc-902f-1b7701b813fa') ref 0 aborting False >> jsonrpc/4::DEBUG::2017-02-03 08:32:26,653::__init__::532::j >> sonrpc.JsonRpcServer::(_handle_request) Calling 'Host.getAllVmStats' in >> bridge with {} >> jsonrpc/4::DEBUG::2017-02-03 08:32:26,653::__init__::557::j >> sonrpc.JsonRpcServer::(_handle_request) Return 'Host.getAllVmStats' in >> bridge with (suppressed) >> jsonrpc/4::INFO::2017-02-03 08:32:26,653::__init__::515::j >> sonrpc.JsonRpcServer::(_serveRequest) RPC call Host.getAllVmStats >> succeeded in 0.00 seconds >> jsonrpc/6::DEBUG::2017-02-03 08:32:27,706::__init__::532::j >> sonrpc.JsonRpcServer::(_handle_request) Calling 'Host.getStats' in >> bridge with {} >> jsonrpc/6::DEBUG::2017-02-03 08:32:27,706::__init__::532::j >> sonrpc.JsonRpcServer::(_handle_request) Calling 'Host.getStats' in >> bridge with {} >> jsonrpc/6::DEBUG::2017-02-03 08:32:27,707::task::599::stora >> ge.TaskManager.Task::(_updateState) >> (Task='9285329e-485b-4165-8bdc-4f411f394356') >> moving from state init -> state preparing >> jsonrpc/6::INFO::2017-02-03 08:32:27,707::logUtils::49::dispatcher::(wrapper) >> Run and protect: repoStats(options=None) >> jsonrpc/6::INFO::2017-02-03 08:32:27,707::logUtils::52::dispatcher::(wrapper) >> Run and protect: repoStats, Return response: {} >> jsonrpc/6::DEBUG::2017-02-03 08:32:27,707::task::1195::stor >> age.TaskManager.Task::(prepare) (Task='9285329e-485b-4165-8bdc-4f411f394356') >> finished: {} >> jsonrpc/6::DEBUG::2017-02-03 08:32:27,707::task::599::stora >> ge.TaskManager.Task::(_updateState) >> (Task='9285329e-485b-4165-8bdc-4f411f394356') >> moving from state preparing -> state finished >> jsonrpc/6::DEBUG::2017-02-03 08:32:27,708::resourceManager: >> :910::storage.ResourceManager.Owner::(releaseAll) Owner.releaseAll >> requests {} resources {} >> jsonrpc/6::DEBUG::2017-02-03 08:32:27,708::resourceManager: >> :947::storage.ResourceManager.Owner::(cancelAll) Owner.cancelAll >> requests {} >> jsonrpc/6::DEBUG::2017-02-03 >> 08:32:27,708::task::997::storage.TaskManager.Task::(_decref) >> (Task='9285329e-485b-4165-8bdc-4f411f394356') ref 0 aborting False >> jsonrpc/6::DEBUG::2017-02-03 08:32:27,712::__init__::557::j >> sonrpc.JsonRpcServer::(_handle_request) Return 'Host.getStats' in bridge >> with {'cpuStatistics': {'28': {'cpuUser': '0.13', 'nodeIndex': 0, 'cpuSys': >> '0.00', 'cpuIdle': '99.87'}, '29': {'cpuUser': '0.00', 'nodeIndex': 0, >> 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '24': {'cpuUser': '0.00', >> 'nodeIndex': 0, 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '25': {'cpuUser': >> '0.13', 'nodeIndex': 0, 'cpuSys': '0.07', 'cpuIdle': '99.80'}, '26': >> {'cpuUser': '0.00', 'nodeIndex': 0, 'cpuSys': '0.00', 'cpuIdle': '100.00'}, >> '27': {'cpuUser': '0.00', 'nodeIndex': 0, 'cpuSys': '0.00', 'cpuIdle': >> '100.00'}, '20': {'cpuUser': '0.00', 'nodeIndex': 1, 'cpuSys': '0.00', >> 'cpuIdle': '100.00'}, '21': {'cpuUser': '0.00', 'nodeIndex': 1, 'cpuSys': >> '0.00', 'cpuIdle': '100.00'}, '22': {'cpuUser': '0.00', 'nodeIndex': 1, >> 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '23': {'cpuUser': '0.00', >> 'nodeIndex': 1, 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '46': {'cpuUser': >> '0.00', 'nodeIndex': 1, 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '47': >> {'cpuUser': '0.00', 'nodeIndex': 1, 'cpuSys': '0.00', 'cpuIdle': '100.00'}, >> '44': {'cpuUser': '0.00', 'nodeIndex': 1, 'cpuSys': '0.00', 'cpuIdle': >> '100.00'}, '45': {'cpuUser': '0.00', 'nodeIndex': 1, 'cpuSys': '0.00', >> 'cpuIdle': '100.00'}, '42': {'cpuUser': '0.00', 'nodeIndex': 1, 'cpuSys': >> '0.00', 'cpuIdle': '100.00'}, '43': {'cpuUser': '0.00', 'nodeIndex': 1, >> 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '40': {'cpuUser': '0.00', >> 'nodeIndex': 1, 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '41': {'cpuUser': >> '0.00', 'nodeIndex': 1, 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '1': >> {'cpuUser': '0.13', 'nodeIndex': 0, 'cpuSys': '0.07', 'cpuIdle': '99.80'}, >> '0': {'cpuUser': '0.00', 'nodeIndex': 0, 'cpuSys': '0.00', 'cpuIdle': >> '100.00'}, '3': {'cpuUser': '0.07', 'nodeIndex': 0, 'cpuSys': '0.07', >> 'cpuIdle': '99.86'}, '2': {'cpuUser': '0.07', 'nodeIndex': 0, 'cpuSys': >> '0.00', 'cpuIdle': '99.93'}, '5': {'cpuUser': '0.00', 'nodeIndex': 0, >> 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '4': {'cpuUser': '0.00', >> 'nodeIndex': 0, 'cpuSys': '0.07', 'cpuIdle': '99.93'}, '7': {'cpuUser': >> '0.00', 'nodeIndex': 0, 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '6': >> {'cpuUser': '0.20', 'nodeIndex': 0, 'cpuSys': '0.00', 'cpuIdle': '99.80'}, >> '9': {'cpuUser': '0.00', 'nodeIndex': 0, 'cpuSys': '0.00', 'cpuIdle': >> '100.00'}, '8': {'cpuUser': '0.00', 'nodeIndex': 0, 'cpuSys': '0.00', >> 'cpuIdle': '100.00'}, '39': {'cpuUser': '0.00', 'nodeIndex': 1, 'cpuSys': >> '0.00', 'cpuIdle': '100.00'}, '38': {'cpuUser': '0.00', 'nodeIndex': 1, >> 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '11': {'cpuUser': '0.00', >> 'nodeIndex': 0, 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '10': {'cpuUser': >> '0.00', 'nodeIndex': 0, 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '13': >> {'cpuUser': '0.00', 'nodeIndex': 1, 'cpuSys': '0.00', 'cpuIdle': '100.00'}, >> '12': {'cpuUser': '0.00', 'nodeIndex': 1, 'cpuSys': '0.00', 'cpuIdle': >> '100.00'}, '15': {'cpuUser': '0.00', 'nodeIndex': 1, 'cpuSys': '0.00', >> 'cpuIdle': '100.00'}, '14': {'cpuUser': '0.00', 'nodeIndex': 1, 'cpuSys': >> '0.00', 'cpuIdle': '100.00'}, '17': {'cpuUser': '0.00', 'nodeIndex': 1, >> 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '16': {'cpuUser': '0.00', >> 'nodeIndex': 1, 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '19': {'cpuUser': >> '0.00', 'nodeIndex': 1, 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '18': >> {'cpuUser': '0.00', 'nodeIndex': 1, 'cpuSys': '0.00', 'cpuIdle': '100.00'}, >> '31': {'cpuUser': '0.00', 'nodeIndex': 0, 'cpuSys': '0.00', 'cpuIdle': >> '100.00'}, '30': {'cpuUser': '0.00', 'nodeIndex': 0, 'cpuSys': '0.00', >> 'cpuIdle': '100.00'}, '37': {'cpuUser': '0.00', 'nodeIndex': 1, 'cpuSys': >> '0.00', 'cpuIdle': '100.00'}, '36': {'cpuUser': '0.00', 'nodeIndex': 1, >> 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '35': {'cpuUser': '0.00', >> 'nodeIndex': 0, 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '34': {'cpuUser': >> '0.00', 'nodeIndex': 0, 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '33': >> {'cpuUser': '0.00', 'nodeIndex': 0, 'cpuSys': '0.00', 'cpuIdle': '100.00'}, >> '32': {'cpuUser': '0.00', 'nodeIndex': 0, 'cpuSys': '0.00', 'cpuIdle': >> '100.00'}}, 'numaNodeMemFree': {'1': {'memPercent': 3, 'memFree': '63780'}, >> '0': {'memPercent': 4, 'memFree': '63426'}}, 'memShared': 0, 'thpState': >> 'always', 'ksmMergeAcrossNodes': True, 'vmCount': 0, 'memUsed': '2', >> 'storageDomains': {}, 'incomingVmMigrations': 0, 'network': {'ib0': >> {'txErrors': '0', 'state': 'down', 'sampleTime': 1486090944.903051, 'name': >> 'ib0', 'tx': '0', 'txDropped': '0', 'rx': '0', 'rxErrors': '0', 'speed': >> '1000', 'rxDropped': '0'}, 'lo': {'txErrors': '0', 'state': 'up', >> 'sampleTime': 1486090944.903051, 'name': 'lo', 'tx': '56174388', >> 'txDropped': '0', 'rx': '56174388', 'rxErrors': '0', 'speed': '1000', >> 'rxDropped': '0'}, 'eno2': {'txErrors': '0', 'state': 'down', 'sampleTime': >> 1486090944.903051, 'name': 'eno2', 'tx': '0', 'txDropped': '0', 'rx': '0', >> 'rxErrors': '0', 'speed': '1000', 'rxDropped': '0'}, 'eno1': {'txErrors': >> '0', 'state': 'up', 'sampleTime': 1486090944.903051, 'name': 'eno1', 'tx': >> '26032823', 'txDropped': '0', 'rx': '174722400', 'rxErrors': '0', 'speed': >> '1000', 'rxDropped': '5624'}, ';vdsmdummy;': {'txErrors': '0', 'state': >> 'down', 'sampleTime': 1486090944.903051, 'name': ';vdsmdummy;', 'tx': '0', >> 'txDropped': '0', 'rx': '0', 'rxErrors': '0', 'speed': '1000', 'rxDropped': >> '0'}}, 'txDropped': '0', 'anonHugePages': '34', 'ksmPages': 100, >> 'elapsedTime': '287.81', 'cpuLoad': '0.07', 'cpuSys': '0.00', 'diskStats': >> {'/var/log': {'free': '15072'}, '/var/log/core': {'free': '15072'}, >> '/var/run/vdsm/': {'free': '64418'}, '/tmp': {'free': '914756'}}, >> 'cpuUserVdsmd': '0.13', 'netConfigDirty': 'True', 'memCommitted': 0, >> 'ksmState': False, 'vmMigrating': 0, 'ksmCpu': 0, 'memAvailable': 127200, >> 'bootTime': '1486037670', 'haStats': {'active': False, 'configured': False, >> 'score': 0, 'localMaintenance': False, 'globalMaintenance': False}, >> 'momStatus': 'active', 'rxDropped': '5624', 'outgoingVmMigrations': 0, >> 'swapTotal': 4095, 'swapFree': 4095, 'dateTime': '2017-02-03T03:02:27 GMT', >> 'cpuUser': '0.02', 'memFree': 127456, 'cpuIdle': '99.98', 'vmActive': 0, >> 'v2vJobs': {}, 'cpuSysVdsmd': '0.07'} >> jsonrpc/6::INFO::2017-02-03 08:32:27,713::__init__::515::j >> sonrpc.JsonRpcServer::(_serveRequest) RPC call Host.getStats succeeded >> in 0.00 seconds >> Reactor thread::INFO::2017-02-03 08:32:39,661::protocoldetector >> ::72::ProtocolDetector.AcceptorImpl::(handle_accept) Accepted connection >> from ::1:53084 >> >> supervdsm.log >> >> >> File "/usr/share/vdsm/supervdsmServer", line 95, in wrapper >> res = func(*args, **kwargs) >> File "/usr/share/vdsm/supervdsmServer", line 143, in umount >> timeout=timeout) >> File "/usr/lib/python2.7/site-packages/vdsm/storage/mount.py", line >> 281, in _umount >> _runcmd(cmd, timeout) >> File "/usr/lib/python2.7/site-packages/vdsm/storage/mount.py", line >> 297, in _runcmd >> raise MountError(rc, ";".join((out, err))) >> MountError: (32, ';umount: /rhev/data-center/mnt/10.1.235.6:_iso: >> mountpoint not found\n') >> MainProcess|jsonrpc/5::DEBUG::2017-02-03 08:26:59,495::supervdsmServer: >> :93::SuperVdsm.ServerCallback::(wrapper) call hbaRescan with () {} >> MainProcess|jsonrpc/5::DEBUG::2017-02-03 >> 08:26:59,495::commands::69::storage.HBA::(execCmd) >> /usr/bin/taskset --cpu-list 0-47 /usr/libexec/vdsm/fc-scan (cwd N >> one) >> MainProcess|jsonrpc/5::DEBUG::2017-02-03 08:26:59,579::supervdsmServer: >> :100::SuperVdsm.ServerCallback::(wrapper) return hbaRescan with None >> MainThread::DEBUG::2017-02-03 >> 08:27:26,367::supervdsmServer::311::SuperVdsm.Server::(main) >> Terminated normally >> MainThread::DEBUG::2017-02-03 >> 08:27:36,148::__init__::47::blivet::(register_device_format) >> registered device format class DeviceFormat as None >> MainThread::DEBUG::2017-02-03 >> 08:27:36,149::__init__::47::blivet::(register_device_format) >> registered device format class BIOSBoot as biosboot >> MainThread::DEBUG::2017-02-03 >> 08:27:36,150::__init__::47::blivet::(register_device_format) >> registered device format class DiskLabel as disklabel >> MainThread::DEBUG::2017-02-03 >> 08:27:36,150::__init__::47::blivet::(register_device_format) >> registered device format class DMRaidMember as dmraidmember >> MainThread::DEBUG::2017-02-03 >> 08:27:36,151::__init__::47::blivet::(register_device_format) >> registered device format class Ext2FS as ext2 >> MainThread::DEBUG::2017-02-03 >> 08:27:36,152::__init__::47::blivet::(register_device_format) >> registered device format class Ext3FS as ext3 >> MainThread::DEBUG::2017-02-03 >> 08:27:36,152::__init__::47::blivet::(register_device_format) >> registered device format class Ext4FS as ext4 >> MainThread::DEBUG::2017-02-03 >> 08:27:36,152::__init__::47::blivet::(register_device_format) >> registered device format class FATFS as vfat >> MainThread::DEBUG::2017-02-03 >> 08:27:36,152::__init__::47::blivet::(register_device_format) >> registered device format class EFIFS as efi >> MainThread::DEBUG::2017-02-03 >> 08:27:36,152::__init__::47::blivet::(register_device_format) >> registered device format class BTRFS as btrfs >> MainThread::DEBUG::2017-02-03 >> 08:27:36,153::__init__::47::blivet::(register_device_format) >> registered device format class GFS2 as gfs2 >> MainThread::DEBUG::2017-02-03 >> 08:27:36,153::__init__::47::blivet::(register_device_format) >> registered device format class JFS as jfs >> MainThread::DEBUG::2017-02-03 >> 08:27:36,153::__init__::47::blivet::(register_device_format) >> registered device format class ReiserFS as reiserfs >> MainThread::DEBUG::2017-02-03 >> 08:27:36,153::__init__::47::blivet::(register_device_format) >> registered device format class XFS as xfs >> MainThread::DEBUG::2017-02-03 >> 08:27:36,153::__init__::47::blivet::(register_device_format) >> registered device format class HFS as hfs >> MainThread::DEBUG::2017-02-03 >> 08:27:36,153::__init__::47::blivet::(register_device_format) >> registered device format class AppleBootstrapFS as appleboot >> MainThread::DEBUG::2017-02-03 >> 08:27:36,154::__init__::47::blivet::(register_device_format) >> registered device format class HFSPlus as hfs+ >> MainThread::DEBUG::2017-02-03 >> 08:27:36,154::__init__::47::blivet::(register_device_format) >> registered device format class MacEFIFS as macefi >> MainThread::DEBUG::2017-02-03 >> 08:27:36,154::__init__::47::blivet::(register_device_format) >> registered device format class NTFS as ntfs >> MainThread::DEBUG::2017-02-03 >> 08:27:36,154::__init__::47::blivet::(register_device_format) >> registered device format class NFS as nfs >> MainThread::DEBUG::2017-02-03 >> 08:27:36,154::__init__::47::blivet::(register_device_format) >> registered device format class NFSv4 as nfs4 >> MainThread::DEBUG::2017-02-03 >> 08:27:36,154::__init__::47::blivet::(register_device_format) >> registered device format class Iso9660FS as iso9660 >> MainThread::DEBUG::2017-02-03 >> 08:27:36,154::__init__::47::blivet::(register_device_format) >> registered device format class NoDevFS as nodev >> MainThread::DEBUG::2017-02-03 >> 08:27:36,155::__init__::47::blivet::(register_device_format) >> registered device format class DevPtsFS as devpts >> MainThread::DEBUG::2017-02-03 >> 08:27:36,155::__init__::47::blivet::(register_device_format) >> registered device format class ProcFS as proc >> MainThread::DEBUG::2017-02-03 >> 08:27:36,155::__init__::47::blivet::(register_device_format) >> registered device format class SysFS as sysfs >> MainThread::DEBUG::2017-02-03 >> 08:27:36,155::__init__::47::blivet::(register_device_format) >> registered device format class TmpFS as tmpfs >> MainThread::DEBUG::2017-02-03 >> 08:27:36,155::__init__::47::blivet::(register_device_format) >> registered device format class BindFS as bind >> MainThread::DEBUG::2017-02-03 >> 08:27:36,155::__init__::47::blivet::(register_device_format) >> registered device format class SELinuxFS as selinuxfs >> MainThread::DEBUG::2017-02-03 >> 08:27:36,155::__init__::47::blivet::(register_device_format) >> registered device format class USBFS as usbfs >> MainThread::DEBUG::2017-02-03 >> 08:27:36,155::__init__::47::blivet::(register_device_format) >> registered device format class EFIVarFS as efivarfs >> MainThread::DEBUG::2017-02-03 >> 08:27:36,179::__init__::47::blivet::(register_device_format) >> registered device format class LUKS as luks >> MainThread::DEBUG::2017-02-03 >> 08:27:36,181::__init__::47::blivet::(register_device_format) >> registered device format class LVMPhysicalVolume as lvmpv >> MainThread::DEBUG::2017-02-03 >> 08:27:36,181::__init__::47::blivet::(register_device_format) >> registered device format class MDRaidMember as mdmember >> MainThread::DEBUG::2017-02-03 >> 08:27:36,181::__init__::47::blivet::(register_device_format) >> registered device format class MultipathMember as multipath_member >> MainThread::DEBUG::2017-02-03 >> 08:27:36,181::__init__::47::blivet::(register_device_format) >> registered device format class PPCPRePBoot as prepboot >> MainThread::DEBUG::2017-02-03 >> 08:27:36,182::__init__::47::blivet::(register_device_format) >> registered device format class SwapSpace as swap >> MainThread::DEBUG::2017-02-03 >> 08:27:36,216::storage_log::69::blivet::(log_exception_info) >> IGNORED: Caught exception, continuing. >> MainThread::DEBUG::2017-02-03 >> 08:27:36,216::storage_log::72::blivet::(log_exception_info) >> IGNORED: Problem description: failed to get initiator name from >> iscsi firmware >> MainThread::DEBUG::2017-02-03 >> 08:27:36,216::storage_log::73::blivet::(log_exception_info) >> IGNORED: Begin exception details. >> MainThread::DEBUG::2017-02-03 >> 08:27:36,216::storage_log::76::blivet::(log_exception_info) >> IGNORED: Traceback (most recent call last): >> MainThread::DEBUG::2017-02-03 >> 08:27:36,216::storage_log::76::blivet::(log_exception_info) >> IGNORED: File "/usr/lib/python2.7/site-packages/blivet >> /iscsi.py", line 146, in __init__ >> MainThread::DEBUG::2017-02-03 >> 08:27:36,216::storage_log::76::blivet::(log_exception_info) >> IGNORED: initiatorname = libiscsi.get_firmware_initi >> ator_name() >> MainThread::DEBUG::2017-02-03 >> 08:27:36,216::storage_log::76::blivet::(log_exception_info) >> IGNORED: IOError: Unknown error >> MainThread::DEBUG::2017-02-03 >> 08:27:36,216::storage_log::77::blivet::(log_exception_info) >> IGNORED: End exception details. >> MainThread::DEBUG::2017-02-03 >> 08:27:36,227::supervdsmServer::271::SuperVdsm.Server::(main) >> Making sure I'm root - SuperVdsm >> MainThread::DEBUG::2017-02-03 >> 08:27:36,227::supervdsmServer::280::SuperVdsm.Server::(main) >> Parsing cmd args >> MainThread::DEBUG::2017-02-03 >> 08:27:36,227::supervdsmServer::283::SuperVdsm.Server::(main) >> Cleaning old socket /var/run/vdsm/svdsm.sock >> MainThread::DEBUG::2017-02-03 >> 08:27:36,227::supervdsmServer::287::SuperVdsm.Server::(main) >> Setting up keep alive thread >> MainThread::DEBUG::2017-02-03 >> 08:27:36,228::supervdsmServer::293::SuperVdsm.Server::(main) >> Creating remote object manager >> MainThread::DEBUG::2017-02-03 >> 08:27:36,228::fileUtils::246::storage.fileUtils::(chown) >> Changing owner for /var/run/vdsm/svdsm.sock, to (36:36) >> MainThread::DEBUG::2017-02-03 >> 08:27:36,228::supervdsmServer::303::SuperVdsm.Server::(main) >> Started serving super vdsm object >> sourceRoute::DEBUG::2017-02-03 08:27:36,229::sourceroutethrea >> d::79::root::(_subscribeToInotifyLoop) >> sourceRouteThread.subscribeToInotifyLoop >> started >> restore-net::INFO::2017-02-03 >> 08:27:37,160::vdsm-restore-net-config::465::root::(restore) >> networks already restored. doing nothing. >> MainProcess|hsm/init::DEBUG::2017-02-03 08:27:39,601::supervdsmServer: >> :93::SuperVdsm.ServerCallback::(wrapper) call hbaRescan with () {} >> MainProcess|hsm/init::DEBUG::2017-02-03 >> 08:27:39,602::commands::69::storage.HBA::(execCmd) >> /usr/bin/taskset --cpu-list 0-47 /usr/libexec/vdsm/fc-scan (cwd No >> ne) >> MainProcess|hsm/init::DEBUG::2017-02-03 08:27:39,691::supervdsmServer: >> :100::SuperVdsm.ServerCallback::(wrapper) return hbaRescan with None >> MainProcess|jsonrpc/4::DEBUG::2017-02-03 08:27:48,355::supervdsmServer: >> :93::SuperVdsm.ServerCallback::(wrapper) call network_caps with () {} >> MainProcess|jsonrpc/4::DEBUG::2017-02-03 >> 08:27:48,360::commands::69::root::(execCmd) >> /usr/bin/taskset --cpu-list 0-47 /sbin/ip route show to 0.0.0.0/0 table >> main (cwd None) >> MainProcess|jsonrpc/4::DEBUG::2017-02-03 >> 08:27:48,364::commands::93::root::(execCmd) >> SUCCESS: <err> = ''; <rc> = 0 >> MainProcess|jsonrpc/4::DEBUG::2017-02-03 >> 08:27:48,408::libvirtconnection::160::root::(get) >> trying to connect libvirt >> MainProcess|jsonrpc/4::DEBUG::2017-02-03 >> 08:27:48,450::commands::69::root::(execCmd) >> /usr/bin/taskset --cpu-list 0-47 /usr/sbin/tc qdisc show (cwd None) >> MainProcess|jsonrpc/4::DEBUG::2017-02-03 >> 08:27:48,453::commands::93::root::(execCmd) >> SUCCESS: <err> = ''; <rc> = 0 >> MainProcess|jsonrpc/4::DEBUG::2017-02-03 >> 08:27:48,454::commands::69::root::(execCmd) >> /usr/bin/taskset --cpu-list 0-47 /bin/systemctl --no-pager list-unit-fil >> es (cwd None) >> MainProcess|jsonrpc/4::DEBUG::2017-02-03 >> 08:27:48,530::commands::93::root::(execCmd) >> SUCCESS: <err> = ''; <rc> = 0 >> MainProcess|jsonrpc/4::DEBUG::2017-02-03 >> 08:27:48,530::commands::69::root::(execCmd) >> /usr/bin/taskset --cpu-list 0-47 /bin/systemctl status openvswitch.servi >> ce (cwd None) >> MainProcess|jsonrpc/4::DEBUG::2017-02-03 >> 08:27:48,561::commands::93::root::(execCmd) >> FAILED: <err> = ''; <rc> = 3 >> MainProcess|jsonrpc/4::DEBUG::2017-02-03 08:27:48,561::supervdsmServer: >> :100::SuperVdsm.ServerCallback::(wrapper) return network_caps with >> {'bridges': {}, 'bo >> ndings': {}, 'nameservers': ['10.1.1.8'], 'nics': {'ib0': >> {'ipv6autoconf': True, 'addr': '', 'ipv6gateway': '::', 'dhcpv6': False, >> 'ipv6addrs': [], 'mtu': '4 >> 092', 'dhcpv4': False, 'netmask': '', 'ipv4defaultroute': False, >> 'ipv4addrs': [], 'hwaddr': '80:00:02:08:fe:80:00:00:00:00 >> :00:00:0c:c4:7a:ff:ff:0f:d8:85', 's >> peed': 0, 'gateway': ''}, 'eno1': {'ipv6autoconf': False, 'addr': >> '10.1.235.5', 'ipv6gateway': '::', 'dhcpv6': False, 'ipv6addrs': [], 'mtu': >> '1500', 'dhcpv4 >> ': False, 'netmask': '255.255.0.0', 'ipv4defaultroute': True, >> 'ipv4addrs': ['10.1.235.5/16'], 'hwaddr': '0c:c4:7a:0f:6b:2a', 'speed': >> 1000, 'gateway': '10.1. >> 0.1'}, 'eno2': {'ipv6autoconf': True, 'addr': '', 'ipv6gateway': '::', >> 'dhcpv6': False, 'ipv6addrs': [], 'mtu': '1500', 'dhcpv4': False, >> 'netmask': '', 'ipv4 >> defaultroute': False, 'ipv4addrs': [], 'hwaddr': '0c:c4:7a:0f:6b:2b', >> 'speed': 0, 'gateway': ''}}, 'supportsIPv6': True, 'vlans': {}, 'networks': >> {}} >> MainProcess|jsonrpc/5::DEBUG::2017-02-03 08:27:53,690::supervdsmServer: >> :93::SuperVdsm.ServerCallback::(wrapper) call setupNetworks with ({}, >> {u'connectivityC >> heck': u'true', u'connectivityTimeout': 120}) {} >> MainProcess|jsonrpc/5::DEBUG::2017-02-03 >> 08:27:53,691::api::247::root::(setupNetworks) >> Setting up network according to configuration: networks:{u'ovirtmgmt': >> {u'ipv6autoconf': False, u'nic': u'eno1', u'ipaddr': u'10.1.235.5', >> u'netmask': u'255.255.0.0', u'mtu': 1500, u'switch': u'ovs', u'dhcpv6': >> False, u'STP': u'no', u'bridged': u'true', u'gateway': u'10.1.0.1', >> u'defaultRoute': True}}, bondings:{}, options:{u'connectivityCheck': >> u'true', u'connectivityTimeout': 120} >> MainProcess|jsonrpc/5::DEBUG::2017-02-03 >> 08:27:53,691::api::252::root::(setupNetworks) >> Validating configuration >> MainProcess|jsonrpc/5::DEBUG::2017-02-03 >> 08:27:53,694::commands::69::root::(execCmd) >> /usr/bin/taskset --cpu-list 0-47 /sbin/ip route show to 0.0.0.0/0 table >> main (cwd None) >> MainProcess|jsonrpc/5::DEBUG::2017-02-03 >> 08:27:53,698::commands::93::root::(execCmd) >> SUCCESS: <err> = ''; <rc> = 0 >> MainProcess|jsonrpc/5::DEBUG::2017-02-03 >> 08:27:53,771::commands::69::root::(execCmd) >> /usr/bin/taskset --cpu-list 0-47 /usr/sbin/tc qdisc show (cwd None) >> MainProcess|jsonrpc/5::DEBUG::2017-02-03 >> 08:27:53,774::commands::93::root::(execCmd) >> SUCCESS: <err> = ''; <rc> = 0 >> MainProcess|jsonrpc/5::DEBUG::2017-02-03 >> 08:27:53,774::vsctl::57::root::(commit) >> Executing commands: /usr/bin/ovs-vsctl --oneline --format=json -- list >> Bridge -- list Port -- list Interface >> MainProcess|jsonrpc/5::DEBUG::2017-02-03 >> 08:27:53,774::commands::69::root::(execCmd) >> /usr/bin/taskset --cpu-list 0-47 /usr/bin/ovs-vsctl --oneline --format=json >> -- list Bridge -- list Port -- list Interface (cwd None) >> MainProcess|jsonrpc/5::DEBUG::2017-02-03 >> 08:27:53,807::commands::93::root::(execCmd) >> FAILED: <err> = 'ovs-vsctl: unix:/var/run/openvswitch/db.sock: database >> connection failed (No such file or directory)\n'; <rc> = 1 >> MainProcess|jsonrpc/5::ERROR::2017-02-03 08:27:53,808::supervdsmServer: >> :97::SuperVdsm.ServerCallback::(wrapper) Error in setupNetworks >> Traceback (most recent call last): >> File "/usr/share/vdsm/supervdsmServer", line 95, in wrapper >> res = func(*args, **kwargs) >> File "/usr/lib/python2.7/site-packages/vdsm/network/api.py", line 254, >> in setupNetworks >> netswitch.validate(networks, bondings) >> File "/usr/lib/python2.7/site-packages/vdsm/network/netswitch.py", >> line 121, in validate >> ovs_switch.validate_network_setup(ovs_nets, ovs_bonds) >> File "/usr/lib/python2.7/site-packages/vdsm/network/ovs/switch.py", >> line 40, in validate_network_setup >> ovs_networks = info.create_netinfo(info.OvsInfo())['networks'] >> File "/usr/lib/python2.7/site-packages/vdsm/network/ovs/info.py", line >> 73, in __init__ >> ovs_db = OvsDB(driver.create()) >> File "/usr/lib/python2.7/site-packages/vdsm/network/ovs/info.py", line >> 64, in __init__ >> transaction.add(ifaces_command) >> File >> "/usr/lib/python2.7/site-packages/vdsm/network/ovs/driver/__init__.py", >> line 54, in __exit__ >> self.result = self.commit() >> File "/usr/lib/python2.7/site-packages/vdsm/network/ovs/driver/vsctl.py", >> line 63, in commit >> 'Executing commands failed: %s' % '\n'.join(err)) >> ConfigNetworkError: (21, 'Executing commands failed: ovs-vsctl: >> unix:/var/run/openvswitch/db.sock: database connection failed (No such >> file or directory)') >> MainProcess|Thread-2::DEBUG::2017-02-03 08:27:54,440::supervdsmServer: >> :93::SuperVdsm.ServerCallback::(wrapper) call ksmTune with () {} >> MainProcess|Thread-2::DEBUG::2017-02-03 08:27:54,440::supervdsmServer: >> :100::SuperVdsm.ServerCallback::(wrapper) return ksmTune with None >> MainProcess|jsonrpc/0::DEBUG::2017-02-03 08:30:00,918::supervdsmServer: >> :93::SuperVdsm.ServerCallback::(wrapper) call network_caps with () {} >> MainProcess|jsonrpc/0::DEBUG::2017-02-03 >> 08:30:00,922::commands::69::root::(execCmd) >> /usr/bin/taskset --cpu-list 0-47 /sbin/ip route show to 0.0.0.0/0 table >> main (cwd None) >> MainProcess|jsonrpc/0::DEBUG::2017-02-03 >> 08:30:00,927::commands::93::root::(execCmd) >> SUCCESS: <err> = ''; <rc> = 0 >> MainProcess|jsonrpc/0::DEBUG::2017-02-03 >> 08:30:01,013::commands::69::root::(execCmd) >> /usr/bin/taskset --cpu-list 0-47 /usr/sbin/tc qdisc show (cwd None) >> MainProcess|jsonrpc/0::DEBUG::2017-02-03 >> 08:30:01,016::commands::93::root::(execCmd) >> SUCCESS: <err> = ''; <rc> = 0 >> MainProcess|jsonrpc/0::DEBUG::2017-02-03 08:30:01,017::supervdsmServer: >> :100::SuperVdsm.ServerCallback::(wrapper) return network_caps with >> {'bridges': {}, 'bondings': {}, 'nameservers': ['10.1.1.8'], 'nics': >> {'ib0': {'ipv6autoconf': True, 'addr': '', 'ipv6gateway': '::', 'dhcpv6': >> False, 'ipv6addrs': [], 'mtu': '4092', 'dhcpv4': False, 'netmask': '', >> 'ipv4defaultroute': False, 'ipv4addrs': [], 'hwaddr': >> '80:00:02:08:fe:80:00:00:00:00:00:00:0c:c4:7a:ff:ff:0f:d8:85', 'speed': >> 0, 'gateway': ''}, 'eno1': {'ipv6autoconf': False, 'addr': '10.1.235.5', >> 'ipv6gateway': '::', 'dhcpv6': False, 'ipv6addrs': [], 'mtu': '1500', >> 'dhcpv4': False, 'netmask': '255.255.0.0', 'ipv4defaultroute': True, >> 'ipv4addrs': ['10.1.235.5/16'], 'hwaddr': '0c:c4:7a:0f:6b:2a', 'speed': >> 1000, 'gateway': '10.1.0.1'}, 'eno2': {'ipv6autoconf': True, 'addr': '', >> 'ipv6gateway': '::', 'dhcpv6': False, 'ipv6addrs': [], 'mtu': '1500', >> 'dhcpv4': False, 'netmask': '', 'ipv4defaultroute': False, 'ipv4addrs': [], >> 'hwaddr': '0c:c4:7a:0f:6b:2b', 'speed': 0, 'gateway': ''}}, 'supportsIPv6': >> True, 'vlans': {}, 'networks': {}} >> MainProcess|jsonrpc/1::DEBUG::2017-02-03 08:30:01,036::supervdsmServer: >> :93::SuperVdsm.ServerCallback::(wrapper) call getHardwareInfo with () {} >> MainProcess|jsonrpc/1::DEBUG::2017-02-03 08:30:01,066::supervdsmServer: >> :100::SuperVdsm.ServerCallback::(wrapper) return getHardwareInfo with >> {'systemProductName': 'X9DRT', 'systemSerialNumber': '0123456789', >> 'systemFamily': 'To be filled by O.E.M.', 'systemVersion': '0123456789', >> 'systemUUID': '00000000-0000-0000-0000-0CC47A0F6B2A', >> 'systemManufacturer': 'Supermicro'} >> (END) >> >> >> >> On Fri, Feb 3, 2017 at 12:05 AM, Edward Haas <eh...@redhat.com> wrote: >> >>> Hello Shalabh, >>> >>> Please provide the logs from your node: >>> - messages >>> - vdsm/vdsm.log, vdsm/supervdsm.log >>> >>> It may be that you are missing openvswitch installed, although VDSM >>> should not require it for its operation. >>> >>> Thanks, >>> Edy. >>> >>> >>> On Thu, Feb 2, 2017 at 2:10 PM, Shalabh Goel <shalabhgoe...@gmail.com> >>> wrote: >>> >>>> HI, >>>> >>>> I am getting the following error on my node after rebooting it. >>>> >>>> VDSM ovirtnode2 command HostSetupNetworksVDS failed: Executing commands >>>> failed: ovs-vsctl: unix:/var/run/openvswitch/db.sock: database >>>> connection failed (No such file or directory) >>>> >>>> >>>> To solve this, I am trying to restart ovsdb-server using the following >>>> command, >>>> >>>> ovsdb-server --remote=punix:*/var/run/openvswitch/db.sock* >>>> --remote=db:Open_vSwitch,Open_vSwitch,manager_options >>>> --private-key=db:Open_vSwitch,SSL,private_key--certificate=db:Open_vSwitch,SSL,certificate >>>> --bootstrap-ca-cert=db:Open_vSwitch,SSL,ca_cert --pidfile --detach >>>> >>>> But I am getting the following error. >>>> >>>> ovsdb-server: /var/run/openvswitch/ovsdb-server.pid.tmp: create failed >>>> (No such file or directory) >>>> >>>> How to restart the ovsdb-server?? Also ovirtmgmt network is missing >>>> from my node. It happened after I rebooted my node after it got upgraded to >>>> Ovirt 4.1 >>>> >>>> -- >>>> Shalabh Goel >>>> >>>> _______________________________________________ >>>> Users mailing list >>>> Users@ovirt.org >>>> http://lists.ovirt.org/mailman/listinfo/users >>>> >>>> >>> >> >> >> -- >> Shalabh Goel >> > > -- Shalabh Goel
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users