Hi Joop, 1. I did disabled firewalld/iptbles on both nodes and engines. 2. Disabling selinux: when setting to permissive via /etc/selinux/config (and of course unpersist/persist them -> /config/etc........) doesn't work ( as write this I've just reboot one node with the above modification and when I try to ssh node I get "connection closed by..." - of course selinux is enabled; after setenforce 0 ssh is OK.The only way I manage to diasble selinux was
mount -o.rw,remount /run/initramfs/live edit grub.cfg and add selinux=0 to kernel line 3. one strange issue: after I reboot nodes, trying to ssh raise ssh warning WARNING: REMOTE HOST IDENTIFICATION HAS CHANGED! " so I have to remove host entry from .ssh/.known_host in order to connect 4. I am about to check name resolution 5.I have glusterfs also on both nodes. After I put nodes on maintenace, stop gluster volumes and rebooting nodes caused gluster volumes to dissapear from engine web interface Thanks On Fri, Jan 17, 2014 at 10:35 AM, noc <n...@nieuwland.nl> wrote: > On 17-1-2014 8:53, Gabi C wrote: > > 've been there! :-D > > I mean exactly same issuse you had on Centos, I had on Fedora 19. > Did you disable selinux on nodes? 'cause that's what is causing SSh > connection closing > > My setup: > > 1 engine on vmware - fedora 19, up-to-date > > > 2 nodes on IBM x series 3650 - fedora 19 based -oVirt Node - 3.0.3 - > 1.1.fc19 with nodes beig in glusterfs cluster also..... > > > Right now, I'm banging my head against "Operation Add-Disk failed to > complete." , message I have got after adding a new virtual machine and try > to addd its disk > > > On Fri, Jan 17, 2014 at 6:08 AM, Will Dennis (Live.com) < > willardden...@live.com> wrote: > >> Hi all, ready for a story? (well, more of a rant, but hopefully it will >> be a >> good UX tale, and may even be entertaining.) >> >> Had one of the groups come to me at work this week and request a OpenStack >> setup. When I sat down and discussed their needs, it turns out that they >> really only need a multi-hypervisor setup where they can spin up VMs for >> their research projects. The VMs should be fairly long-lived, and will >> have >> persistent storage. Their other request is that the storage should be >> local >> on the hypervisor nodes (they plan to use Intel servers with 8-10 2TB >> drives >> for VM storage on each node.) They desire this in order to keep the VM I/O >> local - they do not have a SAN of any sort anyhow, and they do not care >> about live migration, etc >> > @Will > If the installation ends, either with or without error, it will give you a > log location. Upload the log to a paste.bin and mail the link. > > @Gabi, > There should be more info in either the vdsm.log on the SPM server or in > the engine.log on the engine server, see above for lettings us know what > the error is. > > Having installed oVirt, probably dozens of times, I have some guidelines: > - temporarily disable firewalld/iptables (if all works, enable should > still work, scripts with rules are generated and location is given) > - make selinux permissive either via setenforce 0 (until next boot) or via > /etc/selinux/config ( survives reboots), dont disable it! > - make sure fqdn work in both directions between engine and host(s) > (either using /etc/hosts or DNS) > - make sure NetworkManager is disabled and network enabled > > Joop > > > _______________________________________________ > Users mailing list > Users@ovirt.org > http://lists.ovirt.org/mailman/listinfo/users > >
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users