Thanks Ruben for the reply. You are correct about the IP in the network, as soon as I reconfigured the vnet to be in the 10.0.0.X space instead of 192.168.100.X I could start ssh from outside. I tried this earlier but it did not work, I couldn't tell what happened in the meanwhile.
I will try the virtual router + alias or tun to give internet access to the VMs. Thanks, Giuliano On 6 May 2013 13:49, Ruben S. Montero <rsmont...@opennebula.org> wrote: > Hi > > The VM networking is correctly setup, as you can "ping each other". If you > want to access the VMs from the outside you need a NIC attached to that > network and a IP in the network. You can either create and interface alias, > or a tun device with a valid ip, attach it to the bridge, and then you > should be able to send packages to the vnet. > > This configuration is not the standard one, though. You will probably want > to assign a public IP or place a virtual router with a public IP to log in > the VMs. > > > Ruben > > > > > On Sat, May 4, 2013 at 9:11 PM, Giuliano Casale > <g.cas...@imperial.ac.uk>wrote: > >> Hi all, >> >> I have a working installation of OpenNebula 3.8.3, but despite many >> efforts I still cannot ssh from the host into the VMs. >> >> All hosts have: >> - an eth0 interface listening to the internal network of my institution >> - an eth1 interface for a private switched LAN only between the hosts >> that compose the OpenNebula cloud. The hosts have IPs 10.0.0.X on this >> network. >> >> On all the machines I have run the following commands: >> >> brctl addbr br1 >> brctl addif br1 eth1 >> ifconfig eth1 down >> ifconfig eth1 0.0.0.0 up >> ifconfig br1 10.0.0.X/24 up (where X=1,2,... depending on the host). >> >> I then defined in Sunstone a virtual network as follows: >> ============= >> ID : 0 >> NAME : vn-private >> USER : oneadmin >> GROUP : oneadmin >> CLUSTER : cluster >> TYPE : RANGED >> BRIDGE : br1 >> VLAN : No >> PHYSICAL DEVICE: >> VLAN ID : >> USED LEASES : 2 >> >> PERMISSIONS >> OWNER : um- >> GROUP : --- >> OTHER : --- >> >> VIRTUAL NETWORK TEMPLATE >> >> >> RANGE >> IP_START : 192.168.100.1 >> IP_END : 192.168.100.128 >> >> USED LEASES >> LEASE=[ IP="192.168.100.13", MAC="02:00:c0:a8:64:0d", USED="1", VID="14" ] >> LEASE=[ IP="192.168.100.37", MAC="02:00:c0:a8:64:25", USED="1", VID="41" ] >> ============= >> >> Now: using VNC, I see that the two VMs .13 and .37 can correctly ping >> each other.However, if I login to any host and do 'ssh >> root@192.168.100.13' it does not connect to the VMs. >> >> I tried then 'route add -net 192.168.100.0 netmask 255.255.255.0 dev >> br1' and 'route' shows correctly >> >> Destination Gateway Genmask Flags Metric Ref Use >> Iface >> 192.168.100.0 * 255.255.255.0 U 0 0 0 >> br1 >> >> but still ssh does not work. >> >> Can someone help me? My ultimate goals are: >> 1) to be able to ssh from the host into the VMs. >> 2) if possible, give internet connectivity to the VMs through eth0. >> >> Thanks, >> Giuliano >> _______________________________________________ >> Users mailing list >> Users@lists.opennebula.org >> http://lists.opennebula.org/listinfo.cgi/users-opennebula.org >> > > > > -- > Ruben S. Montero, PhD > Project co-Lead and Chief Architect > OpenNebula - The Open Source Solution for Data Center Virtualization > www.OpenNebula.org | rsmont...@opennebula.org | @OpenNebula >
_______________________________________________ Users mailing list Users@lists.opennebula.org http://lists.opennebula.org/listinfo.cgi/users-opennebula.org