Hi Xarses, Actually it is multi-hypervisor environment. The error in the nova.log is:
NetworkNotFoundForBridge: Network could not be found for bridge br-int the above error disappears changing the mech driver order in /etc/neutron/plugins/ml2/ml2_conf.ini file from mechanism_drivers = openvswitch,dvs to mechanism_drivers =dvs, openvswitch On Mon, Jan 19, 2015 at 12:31 PM, Foss Geek <thefossg...@gmail.com> wrote: > Hi Xarses, > > Thanks for your time! > > I was not able to check my mail yesterday. Sorry for the delay. > > One of my colleague fixed this issue yesterday. I will understand the > issue and update this thread. > > -- > Thanks & Regards > E-Mail: thefossg...@gmail.com > IRC: neophy > Blog : http://lmohanphy.livejournal.com/ > > > > On Sat, Jan 17, 2015 at 1:17 AM, Andrew Woodward <xar...@gmail.com> wrote: > >> neophy, >> >> It seems like there are left overs that fuel was using in the config >> that would not be present when you installed neutron fresh. I'd >> compare the config files and start backing out bits you dont need. I'd >> start with the lines refrencing br-int, you dont need them on nodes >> that aren't using the ovs agent. >> >> Poke me on IRC if you need more help >> >> Xarses (GMT-8) >> >> On Fri, Jan 9, 2015 at 1:08 PM, Foss Geek <thefossg...@gmail.com> wrote: >> > Dear All, >> > >> > I am trying to integrate Openstack + vCenter + Neutron + VMware >> dvSwitch ML2 >> > Mechanism driver. >> > >> > I deployed a two node openstack environment (controller + compute with >> KVM) >> > with Neutron VLAN + KVM using fuel 5.1. Again I installed nova-compute >> using >> > yum in controller node and configured nova-compute in controller to >> point >> > vCenter. I am also using Neutron VLAN with VMware dvSwitch ML2 Mechanism >> > driver. My vCenter is properly configured as suggested by the doc: >> > >> https://www.mirantis.com/blog/managing-vmware-vcenter-resources-mirantis-openstack-5-0-part-1-create-vsphere-cluster/ >> > >> > I am able to create network from Horizon and I can see the same network >> > created in vCenter. When I try to create a VM I am getting the below >> error >> > in Horizon. >> > >> > Error: Failed to launch instance "test-01": Please try again later >> [Error: >> > No valid host was found. ]. >> > >> > Here is the error message from Instance Overview tab: >> > >> > Instance Overview >> > Info >> > Name >> > test-01 >> > ID >> > 309a1f47-83b6-4ab4-9d71-642a2000c8a1 >> > Status >> > Error >> > Availability Zone >> > nova >> > Created >> > Jan. 9, 2015, 8:16 p.m. >> > Uptime >> > 0 minutes >> > Fault >> > Message >> > No valid host was found. >> > Code >> > 500 >> > Details >> > File >> "/usr/lib/python2.6/site-packages/nova/scheduler/filter_scheduler.py", >> > line 108, in schedule_run_instance raise >> exception.NoValidHost(reason="") >> > Created >> > Jan. 9, 2015, 8:16 p.m >> > >> > Getting the below error in nova-all.log: >> > >> > >> > <183>Jan 9 20:16:23 node-18 nova-api 2015-01-09 20:16:23.135 31870 >> DEBUG >> > keystoneclient.middleware.auth_token >> > [req-c9ec0973-ff63-4ac3-a0f7-1d2d7b7aa470 ] Authenticating user token >> > __call__ >> > >> /usr/lib/python2.6/site-packages/keystoneclient/middleware/auth_token.py:676 >> > <183>Jan 9 20:16:23 node-18 nova-api 2015-01-09 20:16:23.136 31870 >> DEBUG >> > keystoneclient.middleware.auth_token >> > [req-c9ec0973-ff63-4ac3-a0f7-1d2d7b7aa470 ] Removing headers from >> request >> > environment: >> > >> X-Identity-Status,X-Domain-Id,X-Domain-Name,X-Project-Id,X-Project-Name,X-Project-Domain-Id,X-Project-Domain-Name,X-User-Id,X-User-Name,X-User-Domain-Id,X-User-Domain-Name,X-Roles,X-Service-Catalog,X-User,X-Tenant-Id,X-Tenant-Name,X-Tenant,X-Role >> > _remove_auth_headers >> > >> /usr/lib/python2.6/site-packages/keystoneclient/middleware/auth_token.py:733 >> > <183>Jan 9 20:16:23 node-18 nova-api 2015-01-09 20:16:23.137 31870 >> DEBUG >> > keystoneclient.middleware.auth_token >> > [req-c9ec0973-ff63-4ac3-a0f7-1d2d7b7aa470 ] Returning cached token >> > _cache_get >> > >> /usr/lib/python2.6/site-packages/keystoneclient/middleware/auth_token.py:1545 >> > <183>Jan 9 20:16:23 node-18 nova-api 2015-01-09 20:16:23.138 31870 >> DEBUG >> > keystoneclient.middleware.auth_token >> > [req-c9ec0973-ff63-4ac3-a0f7-1d2d7b7aa470 ] Storing token in cache store >> > >> /usr/lib/python2.6/site-packages/keystoneclient/middleware/auth_token.py:1460 >> > <183>Jan 9 20:16:23 node-18 nova-api 2015-01-09 20:16:23.139 31870 >> DEBUG >> > keystoneclient.middleware.auth_token >> > [req-c9ec0973-ff63-4ac3-a0f7-1d2d7b7aa470 ] Received request from user: >> > 4564fea80fa14e1daed160afa074d389 with project_id : >> > dd32714d9009495bb51276e284380d6a and roles: admin,_member_ >> > _build_user_headers >> > >> /usr/lib/python2.6/site-packages/keystoneclient/middleware/auth_token.py:996 >> > <183>Jan 9 20:16:23 node-18 nova-api 2015-01-09 20:16:23.141 31870 >> DEBUG >> > routes.middleware [req-05089e83-e4c1-4d90-b7c5-065226e55d91 ] Matched >> GET >> > >> /dd32714d9009495bb51276e284380d6a/servers/309a1f47-83b6-4ab4-9d71-642a2000c8a1 >> > __call__ /usr/lib/python2.6/site-packages/routes/middleware.py:100 >> > <183>Jan 9 20:16:23 node-18 nova-api 2015-01-09 20:16:23.142 31870 >> DEBUG >> > routes.middleware [req-05089e83-e4c1-4d90-b7c5-065226e55d91 ] Route >> path: >> > '/{project_id}/servers/:(id)', defaults: {'action': u'show', >> 'controller': >> > <nova.api.openstack.wsgi.Resource object at 0x43e2550>} __call__ >> > /usr/lib/python2.6/site-packages/routes/middleware.py:102 >> > <183>Jan 9 20:16:23 node-18 nova-api 2015-01-09 20:16:23.142 31870 >> DEBUG >> > routes.middleware [req-05089e83-e4c1-4d90-b7c5-065226e55d91 ] Match >> dict: >> > {'action': u'show', 'controller': <nova.api.openstack.wsgi.Resource >> object >> > at 0x43e2550>, 'project_id': u'dd32714d9009495bb51276e284380d6a', 'id': >> > u'309a1f47-83b6-4ab4-9d71-642a2000c8a1'} __call__ >> > /usr/lib/python2.6/site-packages/routes/middleware.py:103 >> > <183>Jan 9 20:16:23 node-18 nova-api 2015-01-09 20:16:23.143 31870 >> DEBUG >> > nova.api.openstack.wsgi [req-05089e83-e4c1-4d90-b7c5-065226e55d91 None] >> > Calling method '<bound method Controller.show of >> > <nova.api.openstack.compute.servers.Controller object at 0x4204290>>' >> > (Content-type='None', Accept='application/json') _process_stack >> > /usr/lib/python2.6/site-packages/nova/api/openstack/wsgi.py:945 >> > <183>Jan 9 20:16:23 node-18 nova-compute 2015-01-09 20:16:23.170 29111 >> > DEBUG nova.virt.vmwareapi.network_util >> > [req-27cf4cd7-9184-4d7e-b57a-19ef3caeef26 None] Network br-int not >> found on >> > host! get_network_with_the_name >> > /usr/lib/python2.6/site-packages/nova/virt/vmwareapi/network_util.py:80 >> > <179>Jan 9 20:16:23 node-18 nova-compute 2015-01-09 20:16:23.171 29111 >> > ERROR nova.compute.manager [req-27cf4cd7-9184-4d7e-b57a-19ef3caeef26 >> None] >> > [instance: 309a1f47-83b6-4ab4-9d71-642a2000c8a1] Instance failed to >> spawn >> > 2015-01-09 20:16:23.171 29111 TRACE nova.compute.manager [instance: >> > 309a1f47-83b6-4ab4-9d71-642a2000c8a1] Traceback (most recent call last): >> > 2015-01-09 20:16:23.171 29111 TRACE nova.compute.manager [instance: >> > 309a1f47-83b6-4ab4-9d71-642a2000c8a1] File >> > "/usr/lib/python2.6/site-packages/nova/compute/manager.py", line 1714, >> in >> > _spawn >> > 2015-01-09 20:16:23.171 29111 TRACE nova.compute.manager [instance: >> > 309a1f47-83b6-4ab4-9d71-642a2000c8a1] block_device_info) >> > 2015-01-09 20:16:23.171 29111 TRACE nova.compute.manager [instance: >> > 309a1f47-83b6-4ab4-9d71-642a2000c8a1] File >> > "/usr/lib/python2.6/site-packages/nova/virt/vmwareapi/driver.py", line >> 626, >> > in spawn >> > 2015-01-09 20:16:23.171 29111 TRACE nova.compute.manager [instance: >> > 309a1f47-83b6-4ab4-9d71-642a2000c8a1] admin_password, network_info, >> > block_device_info) >> > 2015-01-09 20:16:23.171 29111 TRACE nova.compute.manager [instance: >> > 309a1f47-83b6-4ab4-9d71-642a2000c8a1] File >> > "/usr/lib/python2.6/site-packages/nova/virt/vmwareapi/vmops.py", line >> 285, >> > in spawn >> > 2015-01-09 20:16:23.171 29111 TRACE nova.compute.manager [instance: >> > 309a1f47-83b6-4ab4-9d71-642a2000c8a1] vif_infos = _get_vif_infos() >> > 2015-01-09 20:16:23.171 29111 TRACE nova.compute.manager [instance: >> > 309a1f47-83b6-4ab4-9d71-642a2000c8a1] File >> > "/usr/lib/python2.6/site-packages/nova/virt/vmwareapi/vmops.py", line >> 276, >> > in _get_vif_infos >> > 2015-01-09 20:16:23.171 29111 TRACE nova.compute.manager [instance: >> > 309a1f47-83b6-4ab4-9d71-642a2000c8a1] self._is_neutron) >> > 2015-01-09 20:16:23.171 29111 TRACE nova.compute.manager [instance: >> > 309a1f47-83b6-4ab4-9d71-642a2000c8a1] File >> > "/usr/lib/python2.6/site-packages/nova/virt/vmwareapi/vif.py", line >> 146, in >> > get_network_ref >> > 2015-01-09 20:16:23.171 29111 TRACE nova.compute.manager [instance: >> > 309a1f47-83b6-4ab4-9d71-642a2000c8a1] network_ref = >> > get_neutron_network(session, network_name, cluster, vif) >> > 2015-01-09 20:16:23.171 29111 TRACE nova.compute.manager [instance: >> > 309a1f47-83b6-4ab4-9d71-642a2000c8a1] File >> > "/usr/lib/python2.6/site-packages/nova/virt/vmwareapi/vif.py", line >> 138, in >> > get_neutron_network >> > 2015-01-09 20:16:23.171 29111 TRACE nova.compute.manager [instance: >> > 309a1f47-83b6-4ab4-9d71-642a2000c8a1] raise >> > exception.NetworkNotFoundForBridge(bridge=bridge) >> > 2015-01-09 20:16:23.171 29111 TRACE nova.compute.manager [instance: >> > 309a1f47-83b6-4ab4-9d71-642a2000c8a1] NetworkNotFoundForBridge: Network >> > could not be found for bridge br-int >> > 2015-01-09 20:16:23.171 29111 TRACE nova.compute.manager [instance: >> > 309a1f47-83b6-4ab4-9d71-642a2000c8a1] >> > >> > >> > # cat /etc/neutron/plugins/ml2/ml2_conf.ini | grep -v ^# | grep -v ^$ >> > [ml2] >> > type_drivers = vlan >> > tenant_network_types = vlan >> > mechanism_drivers = openvswitch,dvs >> > [ml2_type_flat] >> > [ml2_type_vlan] >> > network_vlan_ranges = physnet1:3000:3999,physnet2 >> > [ml2_type_gre] >> > [ml2_type_vxlan] >> > [securitygroup] >> > firewall_driver = >> > neutron.agent.linux.iptables_firewall.OVSHybridIptablesFirewallDriver >> > enable_security_group = True >> > [agent] >> > l2_population=False >> > polling_interval=2 >> > arp_responder=False >> > [ovs] >> > enable_tunneling=False >> > integration_bridge=br-int >> > bridge_mappings=physnet1:br-ex >> > [ml2_vmware] >> > host_ip=<vcenter_ip> >> > host_username=root >> > host_password=<password> >> > wsdl_location=file:///opt/vmware/vimService.wsdl >> > task_poll_interval=5.0 >> > api_retry_count=10 >> > network_maps = physnet1:dvSwitch >> > >> > >> > # cat /etc/neutron/plugins/ml2/ml2_conf_vmware_dvs.ini | grep -v ^# | >> grep >> > -v ^$ >> > [ml2_vmware] >> > host_ip=<vcenter_ip> >> > host_username=root >> > host_password=<password> >> > wsdl_location=file:///opt/vmware/vimService.wsdl >> > task_poll_interval=5.0 >> > api_retry_count=10 >> > network_maps = physnet1:dvSwitch >> > >> > >> > # ovs-vsctl show >> > 80248645-469e-4b64-9408-7d26efce777f >> > Bridge "br-eth3" >> > Port "br-eth3" >> > Interface "br-eth3" >> > type: internal >> > Port "eth3" >> > Interface "eth3" >> > Bridge br-int >> > fail_mode: secure >> > Port "tape9c03794-63" >> > tag: 2 >> > Interface "tape9c03794-63" >> > type: internal >> > Port br-int >> > Interface br-int >> > type: internal >> > Port int-br-ex >> > Interface int-br-ex >> > Port int-br-prv >> > Interface int-br-prv >> > Bridge br-ex >> > Port "br-ex--br-eth2" >> > trunks: [0] >> > Interface "br-ex--br-eth2" >> > type: patch >> > options: {peer="br-eth2--br-ex"} >> > Port br-ex >> > Interface br-ex >> > type: internal >> > Port phy-br-ex >> > Interface phy-br-ex >> > Bridge br-storage >> > Port "br-storage--br-eth0" >> > Interface "br-storage--br-eth0" >> > type: patch >> > options: {peer="br-eth0--br-storage"} >> > Port br-storage >> > Interface br-storage >> > type: internal >> > Bridge br-mgmt >> > Port br-mgmt >> > Interface br-mgmt >> > type: internal >> > Port "br-mgmt--br-eth0" >> > Interface "br-mgmt--br-eth0" >> > type: patch >> > options: {peer="br-eth0--br-mgmt"} >> > Bridge "br-eth0" >> > Port "br-eth0" >> > Interface "br-eth0" >> > type: internal >> > Port "br-eth0--br-storage" >> > tag: 102 >> > Interface "br-eth0--br-storage" >> > type: patch >> > options: {peer="br-storage--br-eth0"} >> > Port "br-eth0--br-mgmt" >> > tag: 101 >> > Interface "br-eth0--br-mgmt" >> > type: patch >> > options: {peer="br-mgmt--br-eth0"} >> > Port "br-eth0--br-prv" >> > Interface "br-eth0--br-prv" >> > type: patch >> > options: {peer="br-prv--br-eth0"} >> > Port "br-eth0--br-fw-admin" >> > trunks: [0] >> > Interface "br-eth0--br-fw-admin" >> > type: patch >> > options: {peer="br-fw-admin--br-eth0"} >> > Port "eth0" >> > Interface "eth0" >> > Bridge "br-eth2" >> > Port "eth2" >> > Interface "eth2" >> > Port "br-eth2" >> > Interface "br-eth2" >> > type: internal >> > Port "br-eth2--br-ex" >> > trunks: [0] >> > Interface "br-eth2--br-ex" >> > type: patch >> > options: {peer="br-ex--br-eth2"} >> > Bridge "br-eth1" >> > Port "eth1" >> > Interface "eth1" >> > Port "br-eth1" >> > Interface "br-eth1" >> > type: internal >> > Bridge br-prv >> > Port "br-prv--br-eth0" >> > Interface "br-prv--br-eth0" >> > type: patch >> > options: {peer="br-eth0--br-prv"} >> > Port "qg-de0a02f9-d2" >> > Interface "qg-de0a02f9-d2" >> > type: internal >> > Port br-prv >> > Interface br-prv >> > type: internal >> > Port phy-br-prv >> > Interface phy-br-prv >> > Bridge br-fw-admin >> > Port br-fw-admin >> > Interface br-fw-admin >> > type: internal >> > Port "br-fw-admin--br-eth0" >> > trunks: [0] >> > Interface "br-fw-admin--br-eth0" >> > type: patch >> > options: {peer="br-eth0--br-fw-admin"} >> > ovs_version: "1.10.2" >> > >> > >> > # ip link >> > 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 16436 qdisc noqueue state UNKNOWN >> > link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 >> > 2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP >> qlen >> > 1000 >> > link/ether 14:fe:b5:0f:b6:79 brd ff:ff:ff:ff:ff:ff >> > 3: eth1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP >> qlen >> > 1000 >> > link/ether 14:fe:b5:0f:b6:7b brd ff:ff:ff:ff:ff:ff >> > 4: eth2: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP >> qlen >> > 1000 >> > link/ether 14:fe:b5:0f:b6:7d brd ff:ff:ff:ff:ff:ff >> > 5: eth3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP >> qlen >> > 1000 >> > link/ether 14:fe:b5:0f:b6:7f brd ff:ff:ff:ff:ff:ff >> > 6: ovs-system: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN >> > link/ether 6a:26:28:63:48:52 brd ff:ff:ff:ff:ff:ff >> > 7: br-eth3: <BROADCAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UNKNOWN >> > link/ether 14:fe:b5:0f:b6:7f brd ff:ff:ff:ff:ff:ff >> > 8: br-ex: <BROADCAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UNKNOWN >> > link/ether a6:3d:66:56:16:40 brd ff:ff:ff:ff:ff:ff >> > 9: br-eth1: <BROADCAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UNKNOWN >> > link/ether 14:fe:b5:0f:b6:7b brd ff:ff:ff:ff:ff:ff >> > 10: br-int: <BROADCAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UNKNOWN >> > link/ether 8e:6a:fb:1f:18:47 brd ff:ff:ff:ff:ff:ff >> > 14: br-fw-admin: <BROADCAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state >> > UNKNOWN >> > link/ether e6:a1:ea:f3:0f:45 brd ff:ff:ff:ff:ff:ff >> > 15: br-storage: <BROADCAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state >> UNKNOWN >> > link/ether 42:a0:c7:5e:45:4d brd ff:ff:ff:ff:ff:ff >> > 16: br-eth2: <BROADCAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state >> UNKNOWN >> > link/ether 14:fe:b5:0f:b6:7d brd ff:ff:ff:ff:ff:ff >> > 17: br-prv: <BROADCAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UNKNOWN >> > link/ether 16:23:fe:ec:eb:4f brd ff:ff:ff:ff:ff:ff >> > 19: br-eth0: <BROADCAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state >> UNKNOWN >> > link/ether 14:fe:b5:0f:b6:79 brd ff:ff:ff:ff:ff:ff >> > 20: br-mgmt: <BROADCAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state >> UNKNOWN >> > link/ether b6:9c:f9:60:a3:40 brd ff:ff:ff:ff:ff:ff >> > 22: virbr0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue >> state >> > UNKNOWN >> > link/ether 92:c8:0e:96:13:db brd ff:ff:ff:ff:ff:ff >> > 33: phy-br-prv: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc >> pfifo_fast >> > state UP qlen 1000 >> > link/ether b2:29:ee:f4:86:16 brd ff:ff:ff:ff:ff:ff >> > 34: int-br-prv: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc >> pfifo_fast >> > state UP qlen 1000 >> > link/ether 6e:c4:d3:3e:c2:11 brd ff:ff:ff:ff:ff:ff >> > 57: phy-br-ex: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc >> pfifo_fast >> > state UP qlen 1000 >> > link/ether 96:8b:87:06:4b:e3 brd ff:ff:ff:ff:ff:ff >> > 58: int-br-ex: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc >> pfifo_fast >> > state UP qlen 1000 >> > link/ether fe:08:e6:ba:bf:d3 brd ff:ff:ff:ff:ff:ff >> > >> > >> > # brctl show >> > bridge name bridge id STP enabled interfaces >> > virbr0 8000.000000000000 yes >> > >> > I guess I am missing some thing. >> > >> > It looks like my issue is similar to this : >> > >> https://ask.openstack.org/en/question/43594/vmware-neutron-bridging-problem/ >> > >> > I have configured br100 with VLAN ID 103 in vCenter. But I don't have >> br100 >> > in my controller node. Not sure how to create it in my controller. >> > >> > NOTE : I have another openstack environment which I deployed as vCenter >> > Environment using Fuel 5.1 and manually installed/configured Neutron + >> > VMware dvSwitch ML2 Mechanism driver. It works fine with same vCenter. >> > >> > Any help? >> > >> > I happy to provide more info if required. >> > >> > -- >> > Thanks & Regards >> > E-Mail: thefossg...@gmail.com >> > IRC: neophy >> > Blog : http://lmohanphy.livejournal.com/ >> > >> > >> > _______________________________________________ >> > OpenStack Development Mailing List (not for usage questions) >> > Unsubscribe: >> openstack-dev-requ...@lists.openstack.org?subject:unsubscribe >> > http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev >> > >> >> >> >> -- >> Andrew >> Mirantis >> Ceph community >> >> __________________________________________________________________________ >> OpenStack Development Mailing List (not for usage questions) >> Unsubscribe: >> openstack-dev-requ...@lists.openstack.org?subject:unsubscribe >> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev >> > > > > > > -- Thanks & Regards Mobile: 9444955058 E-Mail: thefossg...@gmail.com IRC: neophy Blog : http://lmohanphy.livejournal.com/
__________________________________________________________________________ OpenStack Development Mailing List (not for usage questions) Unsubscribe: openstack-dev-requ...@lists.openstack.org?subject:unsubscribe http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev