> On 28 Dec 2017, at 00:02, Blaster <blas...@556nato.com> wrote:
> 
> Well, I've spent the last 2.5 days trying to get oVirt 4.2 up and running.
> 
> I sneeze on it, vdsm has a conniption and there appears to be no way to 
> recover from it.
> 
> 1) Install 4.2.  Everything looks good.  Start copying over some 
> data..accidently wipe out the master storage domain...It's gone.  The only 
> method google could suggest was to re-initialize the data center.  Great.  
> I'd love to!  It's greyed out.  Can't get it back...Try several hosted-engine 
> uninstall methods, including 
> /usr/sbin/ovirt-hosted-engine-cleanup and wiping out the storage.  
> 
> re-run hosted-engine --deploy
>   All I get over and over in the vdsm log file while waiting for vdsm to 
> become operational is..
> 2017-12-27 16:36:22,150-0600 ERROR (periodic/3) [virt.periodic.Operation] 
> <vdsm.virt.sampling.VMBulkstatsMonitor object at 0x397b250> operation failed 
> (periodic:215)
> Traceback (most recent call last):
>   File "/usr/lib/python2.7/site-packages/vdsm/virt/periodic.py", line 213, in 
> __call__
>     self._func()
>   File "/usr/lib/python2.7/site-packages/vdsm/virt/sampling.py", line 522, in 
> __call__
>     self._send_metrics()
>   File "/usr/lib/python2.7/site-packages/vdsm/virt/sampling.py", line 538, in 
> _send_metrics
>     vm_sample.interval)
>   File "/usr/lib/python2.7/site-packages/vdsm/virt/vmstats.py", line 45, in 
> produce
>     networks(vm, stats, first_sample, last_sample, interval)
>   File "/usr/lib/python2.7/site-packages/vdsm/virt/vmstats.py", line 322, in 
> networks
>     if nic.name.startswith('hostdev'):
> AttributeError: name
not relevant to your issue, just fyi
this error is not significant and fixed by 
9a2f73a4384e1d72c3285ef88876e404ec8228ff now

> 2017-12-27 16:36:22,620-0600 INFO  (periodic/1) [vdsm.api] START 
> repoStats(domains=()) from=internal, 
> task_id=94688cf1-a991-433e-9e22-7065ed5dc1bf (api:46)
> 2017-12-27 16:36:22,620-0600 INFO  (periodic/1) [vdsm.api] FINISH repoStats 
> return={} from=internal, task_id=94688cf1-a991-433e-9e22-7065ed5dc1bf (api:52)
> 2017-12-27 16:36:22,621-0600 INFO  (periodic/1) [vdsm.api] START 
> multipath_health() from=internal, 
> task_id=9c680369-8f2a-439e-8fe5-b2a1e33c0706 (api:46)
> 2017-12-27 16:36:22,622-0600 INFO  (periodic/1) [vdsm.api] FINISH 
> multipath_health return={} from=internal, 
> task_id=9c680369-8f2a-439e-8fe5-b2a1e33c0706 (api:52)
> 2017-12-27 16:36:22,633-0600 ERROR (periodic/1) [root] failed to retrieve 
> Hosted Engine HA score '[Errno 2] No such file or directory'Is the Hosted 
> Engine setup finished? (api:196)
> 2017-12-27 16:36:23,178-0600 INFO  (vmrecovery) [vdsm.api] START 
> getConnectedStoragePoolsList(options=None) from=internal, 
> task_id=a7e48a2f-8cb7-4ec5-acd7-452c8f0c522b (api:46)
> 2017-12-27 16:36:23,179-0600 INFO  (vmrecovery) [vdsm.api] FINISH 
> getConnectedStoragePoolsList return={'poollist': []} from=internal, 
> task_id=a7e48a2f-8cb7-4ec5-acd7-452c8f0c522b (api:52)
> 2017-12-27 16:36:23,179-0600 INFO  (vmrecovery) [vds] recovery: waiting for 
> storage pool to go up (clientIF:643)
> 
> sigh...reinstall 7.4 and do it all over again.
> 
> 2) copying data to master storage pool.  Didn't wipe it out this time, but 
> filled the volume instead.  Environment freezes.
>  vdsm can't start...infinite loop waiting for storage pool again.  Try clean 
> up and redeploy.  Same problem as above.
> 7.4 reinstall #2 here we go...
> 
> 3)Up and running again.  Forgot to add my NIC card. Shut it down.  Boot back 
> up.  vdsm sees new network interfaces.
> for some reason, it switches ovirtmgmt over to one of the new interfaces 
> which doesn't have a cable 
> attached to it.  Clean up ifcfg- files and reboot.  ifcfg-ovirtmgmt is now 
> gone.  recreate and reboot.  Interface
> comes alive, but vdsm is not starting.
> supervdsm log shows:
>  Multiple southbound ports per network detected, ignoring this network for 
> the QoS report (network: ovirtmgmt, ports: ['enp3s0', 'enp4s0'])
> restore-net::DEBUG::2017-12-27 13:10:39,815::cmdutils::150::root::(exec_cmd) 
> /usr/share/openvswitch/scripts/ovs-ctl status (cwd None)
> restore-net::DEBUG::2017-12-27 13:10:39,856::cmdutils::158::root::(exec_cmd) 
> SUCCESS: <err> = ''; <rc> = 0
> restore-net::DEBUG::2017-12-27 13:10:39,863::vsctl::58::root::(commit) 
> Executing commands: /usr/bin/ovs-vsctl --oneline --format=json -- list Bridge 
> -- list Port -- list Interface
> restore-net::DEBUG::2017-12-27 13:10:39,864::cmdutils::150::root::(exec_cmd) 
> /usr/bin/ovs-vsctl --oneline --format=json -- list Bridge -- list Port -- 
> list Interface (cwd None)
> restore-net::DEBUG::2017-12-27 13:10:39,944::cmdutils::158::root::(exec_cmd) 
> SUCCESS: <err> = ''; <rc> = 0
> restore-net::ERROR::2017-12-27 
> 13:10:39,954::restore_net_config::454::root::(restore) unified restoration 
> failed.
> Traceback (most recent call last):
>   File "/usr/lib/python2.7/site-packages/vdsm/network/restore_net_config.py", 
> line 448, in restore
>     unified_restoration()
>   File "/usr/lib/python2.7/site-packages/vdsm/network/restore_net_config.py", 
> line 131, in unified_restoration
>     classified_conf = _classify_nets_bonds_config(available_config)
>   File "/usr/lib/python2.7/site-packages/vdsm/network/restore_net_config.py", 
> line 260, in _classify_nets_bonds_config
>     current_config = kernelconfig.KernelConfig(net_info)
>   File "/usr/lib/python2.7/site-packages/vdsm/network/kernelconfig.py", line 
> 44, in __init__
>     for net, net_attr in self._analyze_netinfo_nets(netinfo):
>   File "/usr/lib/python2.7/site-packages/vdsm/network/kernelconfig.py", line 
> 57, in _analyze_netinfo_nets
>     attrs = _translate_netinfo_net(net, net_attr, netinfo, _routes)
>   File "/usr/lib/python2.7/site-packages/vdsm/network/kernelconfig.py", line 
> 99, in _translate_netinfo_net
>     raise MultipleSouthBoundNicsPerNetworkError(net, nics)
> MultipleSouthBoundNicsPerNetworkError: ('ovirtmgmt', ['enp3s0', 'enp4s0'])
> 
> Remove new nic.  reboot.  vdsm once again stuck waiting for storage pool to 
> come up.

i’m not clear on what are you actually trying to do. You do a clean install 
with a new SD, and then you write over it? With what? data from your former 
installation, just plain file-level copy? Why?
> 
> So this is where I'm at now.  Stuck.  once again.
> 
> I've been running 3.6.3 All In One for many years because I've been concerned 
> about the complexity of the self hosted
> configuration.  Guess I was right.
> 
> Google shows lots of other people also concerned about the stability of 
> oVirt..It's great when it runs, but 
> any little issue and you're basically reinstalling from scratch.
> 
> _______________________________________________
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users

_______________________________________________
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users

Reply via email to