> On 28 Dec 2017, at 00:02, Blaster <blas...@556nato.com> wrote: > > Well, I've spent the last 2.5 days trying to get oVirt 4.2 up and running. > > I sneeze on it, vdsm has a conniption and there appears to be no way to > recover from it. > > 1) Install 4.2. Everything looks good. Start copying over some > data..accidently wipe out the master storage domain...It's gone. The only > method google could suggest was to re-initialize the data center. Great. > I'd love to! It's greyed out. Can't get it back...Try several hosted-engine > uninstall methods, including > /usr/sbin/ovirt-hosted-engine-cleanup and wiping out the storage. > > re-run hosted-engine --deploy > All I get over and over in the vdsm log file while waiting for vdsm to > become operational is.. > 2017-12-27 16:36:22,150-0600 ERROR (periodic/3) [virt.periodic.Operation] > <vdsm.virt.sampling.VMBulkstatsMonitor object at 0x397b250> operation failed > (periodic:215) > Traceback (most recent call last): > File "/usr/lib/python2.7/site-packages/vdsm/virt/periodic.py", line 213, in > __call__ > self._func() > File "/usr/lib/python2.7/site-packages/vdsm/virt/sampling.py", line 522, in > __call__ > self._send_metrics() > File "/usr/lib/python2.7/site-packages/vdsm/virt/sampling.py", line 538, in > _send_metrics > vm_sample.interval) > File "/usr/lib/python2.7/site-packages/vdsm/virt/vmstats.py", line 45, in > produce > networks(vm, stats, first_sample, last_sample, interval) > File "/usr/lib/python2.7/site-packages/vdsm/virt/vmstats.py", line 322, in > networks > if nic.name.startswith('hostdev'): > AttributeError: name not relevant to your issue, just fyi this error is not significant and fixed by 9a2f73a4384e1d72c3285ef88876e404ec8228ff now
> 2017-12-27 16:36:22,620-0600 INFO (periodic/1) [vdsm.api] START > repoStats(domains=()) from=internal, > task_id=94688cf1-a991-433e-9e22-7065ed5dc1bf (api:46) > 2017-12-27 16:36:22,620-0600 INFO (periodic/1) [vdsm.api] FINISH repoStats > return={} from=internal, task_id=94688cf1-a991-433e-9e22-7065ed5dc1bf (api:52) > 2017-12-27 16:36:22,621-0600 INFO (periodic/1) [vdsm.api] START > multipath_health() from=internal, > task_id=9c680369-8f2a-439e-8fe5-b2a1e33c0706 (api:46) > 2017-12-27 16:36:22,622-0600 INFO (periodic/1) [vdsm.api] FINISH > multipath_health return={} from=internal, > task_id=9c680369-8f2a-439e-8fe5-b2a1e33c0706 (api:52) > 2017-12-27 16:36:22,633-0600 ERROR (periodic/1) [root] failed to retrieve > Hosted Engine HA score '[Errno 2] No such file or directory'Is the Hosted > Engine setup finished? (api:196) > 2017-12-27 16:36:23,178-0600 INFO (vmrecovery) [vdsm.api] START > getConnectedStoragePoolsList(options=None) from=internal, > task_id=a7e48a2f-8cb7-4ec5-acd7-452c8f0c522b (api:46) > 2017-12-27 16:36:23,179-0600 INFO (vmrecovery) [vdsm.api] FINISH > getConnectedStoragePoolsList return={'poollist': []} from=internal, > task_id=a7e48a2f-8cb7-4ec5-acd7-452c8f0c522b (api:52) > 2017-12-27 16:36:23,179-0600 INFO (vmrecovery) [vds] recovery: waiting for > storage pool to go up (clientIF:643) > > sigh...reinstall 7.4 and do it all over again. > > 2) copying data to master storage pool. Didn't wipe it out this time, but > filled the volume instead. Environment freezes. > vdsm can't start...infinite loop waiting for storage pool again. Try clean > up and redeploy. Same problem as above. > 7.4 reinstall #2 here we go... > > 3)Up and running again. Forgot to add my NIC card. Shut it down. Boot back > up. vdsm sees new network interfaces. > for some reason, it switches ovirtmgmt over to one of the new interfaces > which doesn't have a cable > attached to it. Clean up ifcfg- files and reboot. ifcfg-ovirtmgmt is now > gone. recreate and reboot. Interface > comes alive, but vdsm is not starting. > supervdsm log shows: > Multiple southbound ports per network detected, ignoring this network for > the QoS report (network: ovirtmgmt, ports: ['enp3s0', 'enp4s0']) > restore-net::DEBUG::2017-12-27 13:10:39,815::cmdutils::150::root::(exec_cmd) > /usr/share/openvswitch/scripts/ovs-ctl status (cwd None) > restore-net::DEBUG::2017-12-27 13:10:39,856::cmdutils::158::root::(exec_cmd) > SUCCESS: <err> = ''; <rc> = 0 > restore-net::DEBUG::2017-12-27 13:10:39,863::vsctl::58::root::(commit) > Executing commands: /usr/bin/ovs-vsctl --oneline --format=json -- list Bridge > -- list Port -- list Interface > restore-net::DEBUG::2017-12-27 13:10:39,864::cmdutils::150::root::(exec_cmd) > /usr/bin/ovs-vsctl --oneline --format=json -- list Bridge -- list Port -- > list Interface (cwd None) > restore-net::DEBUG::2017-12-27 13:10:39,944::cmdutils::158::root::(exec_cmd) > SUCCESS: <err> = ''; <rc> = 0 > restore-net::ERROR::2017-12-27 > 13:10:39,954::restore_net_config::454::root::(restore) unified restoration > failed. > Traceback (most recent call last): > File "/usr/lib/python2.7/site-packages/vdsm/network/restore_net_config.py", > line 448, in restore > unified_restoration() > File "/usr/lib/python2.7/site-packages/vdsm/network/restore_net_config.py", > line 131, in unified_restoration > classified_conf = _classify_nets_bonds_config(available_config) > File "/usr/lib/python2.7/site-packages/vdsm/network/restore_net_config.py", > line 260, in _classify_nets_bonds_config > current_config = kernelconfig.KernelConfig(net_info) > File "/usr/lib/python2.7/site-packages/vdsm/network/kernelconfig.py", line > 44, in __init__ > for net, net_attr in self._analyze_netinfo_nets(netinfo): > File "/usr/lib/python2.7/site-packages/vdsm/network/kernelconfig.py", line > 57, in _analyze_netinfo_nets > attrs = _translate_netinfo_net(net, net_attr, netinfo, _routes) > File "/usr/lib/python2.7/site-packages/vdsm/network/kernelconfig.py", line > 99, in _translate_netinfo_net > raise MultipleSouthBoundNicsPerNetworkError(net, nics) > MultipleSouthBoundNicsPerNetworkError: ('ovirtmgmt', ['enp3s0', 'enp4s0']) > > Remove new nic. reboot. vdsm once again stuck waiting for storage pool to > come up. i’m not clear on what are you actually trying to do. You do a clean install with a new SD, and then you write over it? With what? data from your former installation, just plain file-level copy? Why? > > So this is where I'm at now. Stuck. once again. > > I've been running 3.6.3 All In One for many years because I've been concerned > about the complexity of the self hosted > configuration. Guess I was right. > > Google shows lots of other people also concerned about the stability of > oVirt..It's great when it runs, but > any little issue and you're basically reinstalling from scratch. > > _______________________________________________ > Users mailing list > Users@ovirt.org > http://lists.ovirt.org/mailman/listinfo/users
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users