About xen bridging, these can help;

brctl show all (bridge vif mappings)
ip show addr xenbr0 (bridge specific info)
brctl showmacs br0 (bridge mac mappings)

Wiki:
http://wiki.xen.org/wiki/Xen_FAQ_Networking
http://wiki.xen.org/wiki/XenNetworking

Regards.

________________________________________
From: John Burwell [jburw...@basho.com]
Sent: Saturday, December 15, 2012 8:35 PM
To: cloudstack-dev@incubator.apache.org
Subject: Re: SSVM Network Configuration Issue

Marcus,

That's what I thought.  The Xen physical bridge names are xenbr0 (to
eth0) and xenbr1 (to eth1).  Using basic network configuration, I set
the Xen network traffic labels for each to the appropriate bridge
device name.  I receive errors regarding invalid network device when
it attempts to create a VM.    Does anyone else know how determine the
mapping of physical devices to CloudStack Xen network traffic labels?

Thanks,
-John

On Dec 15, 2012, at 1:20 AM, Marcus Sorensen <shadow...@gmail.com> wrote:

> Vlans in advanced/KVM should only be required for the guest networks. If I
> create a bridge on physical eth0, name it 'br0', and a bridge on physical
> eth1, naming it 'br1', and then set my management network to label 'br0',
> my public network to 'br1', and my guest network to 'br1', it should use
> the bridges you asked for when connecting the system VMs for the specified
> traffic. I'd just leave the 'vlan' blank when specifying public and
> pod(management) IPs. In this scenario, the only place you need to enter
> vlans is on the guest, and it should create new tagged interfaces/bridges
> on eth1(per your label of br1) as new guest networks are brought online.
> This is how my dev VMs are usually set up.
>
>
> On Thu, Dec 6, 2012 at 11:03 AM, John Burwell <jburw...@basho.com> wrote:
>
>> Marcus,
>>
>> My question, more specifically, is are VLANs required to implement traffic
>> labels?  Also, can traffic labels be configured in Basic networking mode or
>> do I need to switch my configuration to Advanced?
>>
>> I am not disagreeing on the how DNS servers should be associated with
>> interfaces nor do I think a network operator should be required to make any
>> upstream router configuration changes.  I am simply saying that CloudStack
>> should not make assumptions about the gateways that have been specified.
>> The behavior I experienced of CloudStack attempting to
>> "correct" my configuration by injecting another route fails the rule of
>> least surprise and is based on incomplete knowledge.  In my opinion,
>> CloudStack (or any system of its ilk) should faithfully (or slavishly)
>> realize the routes on the system VM as specified.  If the configuration is
>> incorrect, networking will fail in an expected manner, and the operator can
>> adjust their environment as necessary.   Otherwise, there is an upstream
>> router configuration to which CloudStack has no visibility, but with which
>> it is completely compatible.  Essentially, I am asking CloudStack to do
>> less, assume I know what I am doing, and break in a manner consistent with
>> other network applications.
>>
>> Thanks,
>> -John
>>
>> On Dec 6, 2012, at 12:30 PM, Marcus Sorensen <shadow...@gmail.com> wrote:
>>
>>> Traffic labels essentially tell the system which physical network to use.
>>> So if you've allocated a vlan for a specific traffic type, it will first
>>> look at the tag associated with that traffic type, figure out which
>>> physical interface goes with that, and then create a tagged interface and
>>> bridge also on that physical.
>>>
>>> I guess we'll just have to disagree, I think the current behavior makes
>>> total sense.  To me, internal DNS should always use the management
>>> interface, since it's internally facing. There's no sane way to do that
>>> other than a static route on the system vm (it seems you're suggesting
>> that
>>> the network operator force something like this on the upstream router,
>>> which seems really strange to require everyone to create static routes on
>>> their public network to force specific IPs back into their internal
>>> networks, so correct me if I have the wrong impression).  Cloudstack is
>>> doing exactly what you tell it to. You told it that 10.0.3.2 should be
>>> accessible via your internal network by setting it as your internal DNS.
>>> The fact that a broken config doesn't work isn't CloudStack's fault.
>>>
>>> Note that internal DNS is just the default for the ssvm, public DNS is
>>> still offered as a backup, so had you not said that 10.0.3.2 was
>> available
>>> on your internal network (perhaps offering a dummy internal DNS
>>> address or 192.68.56.1),
>>> lookups would fall back to public and everything would work as expected
>> as
>>> well.
>>>
>>> There is also a global config called 'use.external.dns', but in setting
>>> this, restarting the management server, recreating system VMs, I don't
>> see
>>> a noticeable difference on any of this, so perhaps that would solve your
>>> issue as well but it's either broken or doesn't do what I thought it
>> would.
>>>
>>>
>>> On Thu, Dec 6, 2012 at 8:39 AM, John Burwell <jburw...@basho.com> wrote:
>>>
>>>> Marcus,
>>>>
>>>> Are traffic labels independent of VLANs?  I ask because my current XCP
>>>> network configuration is bridged, and I am not using Open vSwitch.
>>>>
>>>> I disagree on the routing issue.  CloudStack should do what's told
>> because
>>>> it does not have insight into or control of the configuration of the
>> routes
>>>> in the layers beneath it.  If CloudStack simply did as it was told, it
>>>> would fail as expected in a typical networking environment while
>> preserving
>>>> the flexibility of configuration expected by a network engineer.
>>>>
>>>> Thanks,
>>>> -John
>>>>
>>>> On Dec 6, 2012, at 10:35 AM, Marcus Sorensen <shadow...@gmail.com>
>> wrote:
>>>>
>>>>> I can't really tell you for xen, although it might be similar to KVM.
>>>>> During setup I would set a traffic label matching the name of the
>> bridge,
>>>>> for example if my public interface were eth0 and the bridge I had set
>> up
>>>>> was br0, I'd go to the zone network settings, find public traffic, and
>>>> set
>>>>> a label on it of "br0". Maybe someone more familiar with the xen setup
>>>> can
>>>>> help.
>>>>>
>>>>> On the DNS, it makes sense from the perspective that the ssvm has
>> access
>>>> to
>>>>> your internal networks, thus it uses your internal DNS. Its default
>>>> gateway
>>>>> is public. So if I have a DNS server on an internal network at
>>>>> 10.30.20.10/24, and my management network on 192.168.10.0/24, this
>> route
>>>>> has to be set in order for the DNS server to be reachable. You would
>>>> under
>>>>> normal circumstances not want to use a DNS server on public net as your
>>>>> internal DNS setting anyway, although I agree that the route insertion
>>>>> should have a bit more sanity checking and not set a static route to
>> your
>>>>> default gateway.
>>>>> On Dec 6, 2012 6:31 AM, "John Burwell" <jburw...@basho.com> wrote:
>>>>>
>>>>>> Marcus,
>>>>>>
>>>>>> I setup a small PowerDNS recursor on 192.168.56.15, configured the DNS
>>>> for
>>>>>> the management network to use it, and the route table in the SSVM is
>> now
>>>>>> correct.  However, this behavior does not seem correct.  At a minimum,
>>>> it
>>>>>> violates the rule of least surprise.  CloudStack shouldn't be adding
>>>>>> gateways that are not configured.  Therefore, I have entered a
>>>> defect[1] to
>>>>>> remove the behavior.
>>>>>>
>>>>>> With the route table fixed, I am now experiencing a new problem.  The
>>>>>> external NIC (10.0.3.0/24) on the SSVM is being connected to the
>>>> internal
>>>>>> NIC (192.168.56.0/24) on the host.  The host-only network
>>>> (192.168.56.15)
>>>>>> is configured on xenbr0 and the NAT network is configured on xenbr1.
>>>> As a
>>>>>> reference, the following is the contents of the
>> /etc/network/interfaces
>>>>>> file and ifconfig from devcloud host:
>>>>>>
>>>>>> root@zone1:/opt/cloudstack/apache-tomcat-6.0.32/bin# cat
>>>>>> /etc/network/interfaces
>>>>>> # The loopback network interface
>>>>>> auto lo
>>>>>> iface lo inet loopback
>>>>>>
>>>>>> auto eth0
>>>>>> iface eth0 inet manual
>>>>>>
>>>>>> allow-hotplug eth1
>>>>>> iface eth1 inet manual
>>>>>>
>>>>>> # The primary network interface
>>>>>> auto xenbr0
>>>>>> iface xenbr0 inet static
>>>>>> address 192.168.56.15
>>>>>> netmask 255.255.255.0
>>>>>> network 192.168.56.0
>>>>>> broadcast 192.168.56.255
>>>>>> dns_nameserver 192.168.56.15
>>>>>> bridge_ports eth0
>>>>>>
>>>>>> auto xenbr1
>>>>>> iface xenbr1 inet dhcp
>>>>>> bridge_ports eth1
>>>>>> dns_nameserver 8.8.8.8 8.8.4.4
>>>>>> post-up route add default gw 10.0.3.2
>>>>>>
>>>>>> root@zone1:/opt/cloudstack/apache-tomcat-6.0.32/bin# ifconfig
>>>>>> eth0      Link encap:Ethernet  HWaddr 08:00:27:7e:74:9c
>>>>>>        inet6 addr: fe80::a00:27ff:fe7e:749c/64 Scope:Link
>>>>>>        UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
>>>>>>        RX packets:777 errors:0 dropped:0 overruns:0 frame:0
>>>>>>        TX packets:188 errors:0 dropped:0 overruns:0 carrier:0
>>>>>>        collisions:0 txqueuelen:1000
>>>>>>        RX bytes:109977 (109.9 KB)  TX bytes:11900 (11.9 KB)
>>>>>>
>>>>>> eth1      Link encap:Ethernet  HWaddr 08:00:27:df:00:00
>>>>>>        inet6 addr: fe80::a00:27ff:fedf:0/64 Scope:Link
>>>>>>        UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
>>>>>>        RX packets:4129 errors:0 dropped:0 overruns:0 frame:0
>>>>>>        TX packets:3910 errors:0 dropped:0 overruns:0 carrier:0
>>>>>>        collisions:0 txqueuelen:1000
>>>>>>        RX bytes:478719 (478.7 KB)  TX bytes:2542459 (2.5 MB)
>>>>>>
>>>>>> lo        Link encap:Local Loopback
>>>>>>        inet addr:127.0.0.1  Mask:255.0.0.0
>>>>>>        inet6 addr: ::1/128 Scope:Host
>>>>>>        UP LOOPBACK RUNNING  MTU:16436  Metric:1
>>>>>>        RX packets:360285 errors:0 dropped:0 overruns:0 frame:0
>>>>>>        TX packets:360285 errors:0 dropped:0 overruns:0 carrier:0
>>>>>>        collisions:0 txqueuelen:0
>>>>>>        RX bytes:169128181 (169.1 MB)  TX bytes:169128181 (169.1 MB)
>>>>>>
>>>>>> vif1.0    Link encap:Ethernet  HWaddr fe:ff:ff:ff:ff:ff
>>>>>>        inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link
>>>>>>        UP BROADCAST RUNNING NOARP PROMISC  MTU:1500  Metric:1
>>>>>>        RX packets:6 errors:0 dropped:0 overruns:0 frame:0
>>>>>>        TX packets:152 errors:0 dropped:0 overruns:0 carrier:0
>>>>>>        collisions:0 txqueuelen:32
>>>>>>        RX bytes:292 (292.0 B)  TX bytes:9252 (9.2 KB)
>>>>>>
>>>>>> vif1.1    Link encap:Ethernet  HWaddr fe:ff:ff:ff:ff:ff
>>>>>>        inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link
>>>>>>        UP BROADCAST RUNNING NOARP PROMISC  MTU:1500  Metric:1
>>>>>>        RX packets:566 errors:0 dropped:0 overruns:0 frame:0
>>>>>>        TX packets:1405 errors:0 dropped:0 overruns:0 carrier:0
>>>>>>        collisions:0 txqueuelen:32
>>>>>>        RX bytes:44227 (44.2 KB)  TX bytes:173995 (173.9 KB)
>>>>>>
>>>>>> vif1.2    Link encap:Ethernet  HWaddr fe:ff:ff:ff:ff:ff
>>>>>>        inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link
>>>>>>        UP BROADCAST RUNNING NOARP PROMISC  MTU:1500  Metric:1
>>>>>>        RX packets:3 errors:0 dropped:0 overruns:0 frame:0
>>>>>>        TX packets:838 errors:0 dropped:0 overruns:0 carrier:0
>>>>>>        collisions:0 txqueuelen:32
>>>>>>        RX bytes:84 (84.0 B)  TX bytes:111361 (111.3 KB)
>>>>>>
>>>>>> vif4.0    Link encap:Ethernet  HWaddr fe:ff:ff:ff:ff:ff
>>>>>>        inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link
>>>>>>        UP BROADCAST RUNNING NOARP PROMISC  MTU:1500  Metric:1
>>>>>>        RX packets:64 errors:0 dropped:0 overruns:0 frame:0
>>>>>>        TX packets:197 errors:0 dropped:0 overruns:0 carrier:0
>>>>>>        collisions:0 txqueuelen:32
>>>>>>        RX bytes:10276 (10.2 KB)  TX bytes:18453 (18.4 KB)
>>>>>>
>>>>>> vif4.1    Link encap:Ethernet  HWaddr fe:ff:ff:ff:ff:ff
>>>>>>        inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link
>>>>>>        UP BROADCAST RUNNING NOARP PROMISC  MTU:1500  Metric:1
>>>>>>        RX packets:2051 errors:0 dropped:0 overruns:0 frame:0
>>>>>>        TX packets:2446 errors:0 dropped:0 overruns:0 carrier:0
>>>>>>        collisions:0 txqueuelen:32
>>>>>>        RX bytes:233914 (233.9 KB)  TX bytes:364243 (364.2 KB)
>>>>>>
>>>>>> vif4.2    Link encap:Ethernet  HWaddr fe:ff:ff:ff:ff:ff
>>>>>>        inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link
>>>>>>        UP BROADCAST RUNNING NOARP PROMISC  MTU:1500  Metric:1
>>>>>>        RX packets:3 errors:0 dropped:0 overruns:0 frame:0
>>>>>>        TX packets:582 errors:0 dropped:0 overruns:0 carrier:0
>>>>>>        collisions:0 txqueuelen:32
>>>>>>        RX bytes:84 (84.0 B)  TX bytes:74700 (74.7 KB)
>>>>>>
>>>>>> vif4.3    Link encap:Ethernet  HWaddr fe:ff:ff:ff:ff:ff
>>>>>>        inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link
>>>>>>        UP BROADCAST RUNNING NOARP PROMISC  MTU:1500  Metric:1
>>>>>>        RX packets:0 errors:0 dropped:0 overruns:0 frame:0
>>>>>>        TX packets:585 errors:0 dropped:0 overruns:0 carrier:0
>>>>>>        collisions:0 txqueuelen:32
>>>>>>        RX bytes:0 (0.0 B)  TX bytes:74826 (74.8 KB)
>>>>>>
>>>>>> xapi0     Link encap:Ethernet  HWaddr fe:ff:ff:ff:ff:ff
>>>>>>        inet addr:169.254.0.1  Bcast:169.254.255.255  Mask:255.255.0.0
>>>>>>        inet6 addr: fe80::c870:1aff:fec2:22b/64 Scope:Link
>>>>>>        UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
>>>>>>        RX packets:568 errors:0 dropped:0 overruns:0 frame:0
>>>>>>        TX packets:1132 errors:0 dropped:0 overruns:0 carrier:0
>>>>>>        collisions:0 txqueuelen:0
>>>>>>        RX bytes:76284 (76.2 KB)  TX bytes:109085 (109.0 KB)
>>>>>>
>>>>>> xenbr0    Link encap:Ethernet  HWaddr 08:00:27:7e:74:9c
>>>>>>        inet addr:192.168.56.15  Bcast:192.168.56.255
>>>> Mask:255.255.255.0
>>>>>>        inet6 addr: fe80::a00:27ff:fe7e:749c/64 Scope:Link
>>>>>>        UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
>>>>>>        RX packets:4162 errors:0 dropped:0 overruns:0 frame:0
>>>>>>        TX packets:3281 errors:0 dropped:0 overruns:0 carrier:0
>>>>>>        collisions:0 txqueuelen:0
>>>>>>        RX bytes:469199 (469.1 KB)  TX bytes:485688 (485.6 KB)
>>>>>>
>>>>>> xenbr1    Link encap:Ethernet  HWaddr 08:00:27:df:00:00
>>>>>>        inet addr:10.0.3.15  Bcast:10.0.3.255  Mask:255.255.255.0
>>>>>>        inet6 addr: fe80::a00:27ff:fedf:0/64 Scope:Link
>>>>>>        UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
>>>>>>        RX packets:4129 errors:0 dropped:0 overruns:0 frame:0
>>>>>>        TX packets:3114 errors:0 dropped:0 overruns:0 carrier:0
>>>>>>        collisions:0 txqueuelen:0
>>>>>>        RX bytes:404327 (404.3 KB)  TX bytes:2501443 (2.5 MB)
>>>>>>
>>>>>> These physical NICs on the host translate to the following Xen PIFs:
>>>>>>
>>>>>> root@zone1:/opt/cloudstack/apache-tomcat-6.0.32/bin# xe pif-list
>>>>>> uuid ( RO)                  : 207413c9-5058-7a40-6c96-2dab21057f30
>>>>>>              device ( RO): eth1
>>>>>>  currently-attached ( RO): true
>>>>>>                VLAN ( RO): -1
>>>>>>        network-uuid ( RO): 1679ddb1-5a21-b827-ab07-c16275d5ce72
>>>>>>
>>>>>>
>>>>>> uuid ( RO)                  : c0274787-e768-506f-3191-f0ac17b0c72b
>>>>>>              device ( RO): eth0
>>>>>>  currently-attached ( RO): true
>>>>>>                VLAN ( RO): -1
>>>>>>        network-uuid ( RO): 8ee927b1-a35d-ac10-4471-d7a6a475839a
>>>>>>
>>>>>> The following is the ifconfig from the SSVM:
>>>>>>
>>>>>> root@s-5-TEST:~# ifconfig
>>>>>> eth0      Link encap:Ethernet  HWaddr 0e:00:a9:fe:03:8b
>>>>>>        inet addr:169.254.3.139  Bcast:169.254.255.255
>>>> Mask:255.255.0.0
>>>>>>        UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
>>>>>>        RX packets:235 errors:0 dropped:0 overruns:0 frame:0
>>>>>>        TX packets:92 errors:0 dropped:0 overruns:0 carrier:0
>>>>>>        collisions:0 txqueuelen:1000
>>>>>>        RX bytes:21966 (21.4 KiB)  TX bytes:16404 (16.0 KiB)
>>>>>>        Interrupt:8
>>>>>>
>>>>>> eth1      Link encap:Ethernet  HWaddr 06:bc:62:00:00:05
>>>>>>        inet addr:192.168.56.104  Bcast:192.168.56.255
>>>>>> Mask:255.255.255.0
>>>>>>        UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
>>>>>>        RX packets:2532 errors:0 dropped:0 overruns:0 frame:0
>>>>>>        TX packets:2127 errors:0 dropped:0 overruns:0 carrier:0
>>>>>>        collisions:0 txqueuelen:1000
>>>>>>        RX bytes:341242 (333.2 KiB)  TX bytes:272183 (265.8 KiB)
>>>>>>        Interrupt:10
>>>>>>
>>>>>> eth2      Link encap:Ethernet  HWaddr 06:12:72:00:00:37
>>>>>>        inet addr:10.0.3.204  Bcast:10.0.3.255  Mask:255.255.255.0
>>>>>>        UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
>>>>>>        RX packets:600 errors:0 dropped:0 overruns:0 frame:0
>>>>>>        TX packets:3 errors:0 dropped:0 overruns:0 carrier:0
>>>>>>        collisions:0 txqueuelen:1000
>>>>>>        RX bytes:68648 (67.0 KiB)  TX bytes:126 (126.0 B)
>>>>>>        Interrupt:11
>>>>>>
>>>>>> eth3      Link encap:Ethernet  HWaddr 06:25:e2:00:00:15
>>>>>>        inet addr:192.168.56.120  Bcast:192.168.56.255
>>>>>> Mask:255.255.255.0
>>>>>>        UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
>>>>>>        RX packets:603 errors:0 dropped:0 overruns:0 frame:0
>>>>>>        TX packets:0 errors:0 dropped:0 overruns:0 carrier:0
>>>>>>        collisions:0 txqueuelen:1000
>>>>>>        RX bytes:68732 (67.1 KiB)  TX bytes:0 (0.0 B)
>>>>>>        Interrupt:12
>>>>>>
>>>>>> lo        Link encap:Local Loopback
>>>>>>        inet addr:127.0.0.1  Mask:255.0.0.0
>>>>>>        UP LOOPBACK RUNNING  MTU:16436  Metric:1
>>>>>>        RX packets:61 errors:0 dropped:0 overruns:0 frame:0
>>>>>>        TX packets:61 errors:0 dropped:0 overruns:0 carrier:0
>>>>>>        collisions:0 txqueuelen:0
>>>>>>        RX bytes:5300 (5.1 KiB)  TX bytes:5300 (5.1 KiB)
>>>>>>
>>>>>> Finally, the following are the vif params for the eth2 device on the
>>>> SSVM
>>>>>> depicting its connection to eth0 instead of eth1:
>>>>>>
>>>>>> root@zone1:/opt/cloudstack/apache-tomcat-6.0.32/bin# !1243
>>>>>> xe vif-param-list uuid=be44bb30-5700-b461-760e-10fe93079210
>>>>>> uuid ( RO)                        :
>> be44bb30-5700-b461-760e-10fe93079210
>>>>>>                   vm-uuid ( RO): 7958d91f-e52d-a25d-718c-7f831ae701d7
>>>>>>             vm-name-label ( RO): s-5-TEST
>>>>>>        allowed-operations (SRO): attach; unplug_force; unplug
>>>>>>        current-operations (SRO):
>>>>>>                    device ( RO): 2
>>>>>>                       MAC ( RO): 06:12:72:00:00:37
>>>>>>         MAC-autogenerated ( RO): false
>>>>>>                       MTU ( RO): 1500
>>>>>>        currently-attached ( RO): true
>>>>>>        qos_algorithm_type ( RW): ratelimit
>>>>>>      qos_algorithm_params (MRW): kbps: 25600
>>>>>>  qos_supported_algorithms (SRO):
>>>>>>              other-config (MRW): nicira-iface-id:
>>>>>> 3d68b9f8-98d1-4ac7-92d8-fb57cb8b0adc; nicira-vm-id:
>>>>>> 7958d91f-e52d-a25d-718c-7f831ae701d7
>>>>>>              network-uuid ( RO): 8ee927b1-a35d-ac10-4471-d7a6a475839a
>>>>>>        network-name-label ( RO): Pool-wide network associated with
>>>> eth0
>>>>>>               io_read_kbs ( RO): 0.007
>>>>>>              io_write_kbs ( RO): 0.000
>>>>>>
>>>>>> How do I configure CloudStack such that the guest network NIC on the
>> VM
>>>>>> will be connected to correct physical NIC?
>>>>>>
>>>>>> Thanks for your help,
>>>>>> -John
>>>>>>
>>>>>>
>>>>>> [1]: https://issues.apache.org/jira/browse/CLOUDSTACK-590
>>>>>>
>>>>>> On Dec 5, 2012, at 2:47 PM, Marcus Sorensen <shadow...@gmail.com>
>>>> wrote:
>>>>>>
>>>>>>> Yes, see your cmdline. internaldns1=10.0.3.2, so it is forcing the
>> use
>>>> of
>>>>>>> management network to route to 10.0.3.2 for DNS. that's where the
>> route
>>>>>> is
>>>>>>> coming from. you will want to use something on your management net
>> for
>>>>>>> internal DNS, or something other than that router.
>>>>>>>
>>>>>>>
>>>>>>> On Wed, Dec 5, 2012 at 11:59 AM, John Burwell <jburw...@basho.com>
>>>>>> wrote:
>>>>>>>
>>>>>>>> Anthony,
>>>>>>>>
>>>>>>>> I apologize for forgetting to response to the part of your answer
>> the
>>>>>>>> first part of the question.  I had set the management.network.cidr
>> and
>>>>>> host
>>>>>>>> global settings to 192.168.0.0/24 and 192.168.56.18 respectively.
>>>>>> Please
>>>>>>>> see the zone1.devcloud.cfg Marvin configuration attached to my
>>>> original
>>>>>>>> email for the actual setting, as well as, the network configurations
>>>>>> used
>>>>>>>> when this problem occurs.
>>>>>>>>
>>>>>>>> Thanks,
>>>>>>>> -John
>>>>>>>>
>>>>>>>> On Dec 5, 2012, at 12:46 PM, Anthony Xu <xuefei...@citrix.com>
>> wrote:
>>>>>>>>
>>>>>>>>> Hi join,
>>>>>>>>>
>>>>>>>>> Try following,
>>>>>>>>>
>>>>>>>>> Set global configuration management.network.cidr to your management
>>>>>>>> server CIDR, if this configuration is not available in UI, you can
>>>>>> change
>>>>>>>> it in DB directly.
>>>>>>>>>
>>>>>>>>> Restart management,
>>>>>>>>> Stop/Start SSVM and CPVM.
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> And could you post "cat /proc/cmdline" in SSVM?
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> Anthony
>>>>>>>>>
>>>>>>>>>> -----Original Message-----
>>>>>>>>>> From: John Burwell [mailto:jburw...@basho.com]
>>>>>>>>>> Sent: Wednesday, December 05, 2012 9:11 AM
>>>>>>>>>> To: cloudstack-dev@incubator.apache.org
>>>>>>>>>> Subject: Re: SSVM Network Configuration Issue
>>>>>>>>>>
>>>>>>>>>> All,
>>>>>>>>>>
>>>>>>>>>> I was wondering if anyone else is experiencing this problem when
>>>> using
>>>>>>>>>> secondary storage on a devcloud-style VM with a host-only and NAT
>>>>>>>>>> adapter.  One aspect of this issue that seems interesting is that
>>>>>>>>>> following route table from the SSVM:
>>>>>>>>>>
>>>>>>>>>> root@s-5-TEST:~# route
>>>>>>>>>> Kernel IP routing table
>>>>>>>>>> Destination     Gateway         Genmask         Flags Metric Ref
>>>>>> Use
>>>>>>>>>> Iface
>>>>>>>>>> 10.0.3.2        192.168.56.1    255.255.255.255 UGH   0      0
>>>>>> 0
>>>>>>>>>> eth1
>>>>>>>>>> 10.0.3.0        *               255.255.255.0   U     0      0
>>>>>> 0
>>>>>>>>>> eth2
>>>>>>>>>> 192.168.56.0    *               255.255.255.0   U     0      0
>>>>>> 0
>>>>>>>>>> eth1
>>>>>>>>>> 192.168.56.0    *               255.255.255.0   U     0      0
>>>>>> 0
>>>>>>>>>> eth3
>>>>>>>>>> link-local      *               255.255.0.0     U     0      0
>>>>>> 0
>>>>>>>>>> eth0
>>>>>>>>>> default         10.0.3.2        0.0.0.0         UG    0      0
>>>>>> 0
>>>>>>>>>> eth2
>>>>>>>>>>
>>>>>>>>>> In particular, the gateways for the management and guest networks
>> do
>>>>>>>>>> not match to the configuration provided to the management server
>>>> (i.e.
>>>>>>>>>> 10.0.3.2 is the gateway for the 10.0.3.0/24 network and
>>>> 192.168.56.1
>>>>>> is
>>>>>>>>>> the gateway for the 192.168.56.0/24 network).  With this
>>>>>> configuration,
>>>>>>>>>> the SSVM has a socket connection to the management server, but is
>> in
>>>>>>>>>> alert state.  Finally, when I remove the host-only NIC and use
>> only
>>>> a
>>>>>>>>>> NAT adapter the SSVM's networking works as expecting leading me to
>>>>>>>>>> believe that the segregated network configuration is at the root
>> of
>>>>>> the
>>>>>>>>>> problem.
>>>>>>>>>>
>>>>>>>>>> Until I can get the networking on the SSVM configured, I am unable
>>>> to
>>>>>>>>>> complete the testing of the S3-backed Secondary Storage
>> enhancement.
>>>>>>>>>>
>>>>>>>>>> Thank you for your help,
>>>>>>>>>> -John
>>>>>>>>>>
>>>>>>>>>> On Dec 3, 2012, at 4:46 PM, John Burwell <jburw...@basho.com>
>>>> wrote:
>>>>>>>>>>
>>>>>>>>>>> All,
>>>>>>>>>>>
>>>>>>>>>>> I am setting up a multi-zone devcloud configuration on VirtualBox
>>>>>>>>>> 4.2.4 using the Ubuntu 12.04.1 and Xen 4.1.  I have configured the
>>>>>> base
>>>>>>>>>> management server VM (zone1) to serve as both the zone1, as well
>> as,
>>>>>>>>>> the management server (running MySql) with eth0 as a host-only
>>>> adapter
>>>>>>>>>> and a static IP of 192.168.56.15 and eth1 as a NAT adapter (see
>> the
>>>>>>>>>> attached zone1-interfaces file for the exact network configuration
>>>> on
>>>>>>>>>> the VM).  The management and guest networks are configured as
>>>> follows:
>>>>>>>>>>>
>>>>>>>>>>> Zone 1
>>>>>>>>>>> Management: 192.168.56.100-149 gw 192.168.56.1 dns 10.0.3.2 (?)
>>>>>>>>>>> Guest: 10.0.3.200-10.0.3.220 gw 10.0.3.2 dns 8.8.8.8
>>>>>>>>>>> Zone 2
>>>>>>>>>>> Management: 192.168.56.150-200 gw 192.68.56.1 dns 10.0.3.2 (?)
>>>>>>>>>>> Guest: 10.0.3.221-240 gw 10.0.3.2 dns 8.8.8.8
>>>>>>>>>>>
>>>>>>>>>>> The management server deploys and starts without error.  I then
>>>>>>>>>> populate the configuration it using the attached Marvin
>>>> configuration
>>>>>>>>>> file (zone1.devcloud.cfg) and restart the management server in
>> order
>>>>>> to
>>>>>>>>>> allow the global configuration option changes to take effect.
>>>>>>>>>> Following the restart, the CPVM and SSVM start without error.
>>>>>>>>>> Unfortunately, they drop into alert status, and the SSVM is unable
>>>> to
>>>>>>>>>> connect outbound through the guest network (very important for my
>>>>>> tests
>>>>>>>>>> because I am testing S3-backed secondary storage).
>>>>>>>>>>>
>>>>>>>>>>> From the diagnostic checks I have performed on the management
>>>> server
>>>>>>>>>> and the SSVM, it appears that the daemon on the SSVM is connecting
>>>>>> back
>>>>>>>>>> to the management server.  I have attached a set of diagnostic
>>>>>>>>>> information from the management server
>>>> (mgmtsvr-zone1-diagnostics.log)
>>>>>>>>>> and SSVM server (ssvm-zone1-diagnostics.log) that includes the
>>>> results
>>>>>>>>>> of ifconfig, route, netstat and ping checks, as well as, other
>>>>>>>>>> information (e.g. the contents of /var/cache/cloud/cmdline on the
>>>>>> SSVM).
>>>>>>>>>> Finally, I have attached the vmops log from the management server
>>>>>>>>>> (vmops-zone1.log).
>>>>>>>>>>>
>>>>>>>>>>> What changes need to be made to management server configuration
>> in
>>>>>>>>>> order to start up an SSVM that can communicate with the secondary
>>>>>>>>>> storage NFS volumes, management server, and connect to hosts on
>> the
>>>>>>>>>> Internet?
>>>>>>>>>>>
>>>>>>>>>>> Thanks for your help,
>>>>>>>>>>> -John
>>>>>>>>>>>
>>>>>>>>>>> <ssvm-zone1-diagnostics.log>
>>>>>>>>>>> <vmops-zone1.tar.gz>
>>>>>>>>>>> <mgmtsvr-zone1-diagnostics.log>
>>>>>>>>>>> <zone1-interfaces>
>>>>>>>>>>> <zone1.devcloud.cfg>
>>
>>

Reply via email to