Hi Xarses, Thanks for your time!
I was not able to check my mail yesterday. Sorry for the delay. One of my colleague fixed this issue yesterday. I will understand the issue and update this thread. -- Thanks & Regards E-Mail: thefossg...@gmail.com IRC: neophy Blog : http://lmohanphy.livejournal.com/ On Sat, Jan 17, 2015 at 1:17 AM, Andrew Woodward <xar...@gmail.com> wrote: > neophy, > > It seems like there are left overs that fuel was using in the config > that would not be present when you installed neutron fresh. I'd > compare the config files and start backing out bits you dont need. I'd > start with the lines refrencing br-int, you dont need them on nodes > that aren't using the ovs agent. > > Poke me on IRC if you need more help > > Xarses (GMT-8) > > On Fri, Jan 9, 2015 at 1:08 PM, Foss Geek <thefossg...@gmail.com> wrote: > > Dear All, > > > > I am trying to integrate Openstack + vCenter + Neutron + VMware dvSwitch > ML2 > > Mechanism driver. > > > > I deployed a two node openstack environment (controller + compute with > KVM) > > with Neutron VLAN + KVM using fuel 5.1. Again I installed nova-compute > using > > yum in controller node and configured nova-compute in controller to point > > vCenter. I am also using Neutron VLAN with VMware dvSwitch ML2 Mechanism > > driver. My vCenter is properly configured as suggested by the doc: > > > https://www.mirantis.com/blog/managing-vmware-vcenter-resources-mirantis-openstack-5-0-part-1-create-vsphere-cluster/ > > > > I am able to create network from Horizon and I can see the same network > > created in vCenter. When I try to create a VM I am getting the below > error > > in Horizon. > > > > Error: Failed to launch instance "test-01": Please try again later > [Error: > > No valid host was found. ]. > > > > Here is the error message from Instance Overview tab: > > > > Instance Overview > > Info > > Name > > test-01 > > ID > > 309a1f47-83b6-4ab4-9d71-642a2000c8a1 > > Status > > Error > > Availability Zone > > nova > > Created > > Jan. 9, 2015, 8:16 p.m. > > Uptime > > 0 minutes > > Fault > > Message > > No valid host was found. > > Code > > 500 > > Details > > File > "/usr/lib/python2.6/site-packages/nova/scheduler/filter_scheduler.py", > > line 108, in schedule_run_instance raise exception.NoValidHost(reason="") > > Created > > Jan. 9, 2015, 8:16 p.m > > > > Getting the below error in nova-all.log: > > > > > > <183>Jan 9 20:16:23 node-18 nova-api 2015-01-09 20:16:23.135 31870 DEBUG > > keystoneclient.middleware.auth_token > > [req-c9ec0973-ff63-4ac3-a0f7-1d2d7b7aa470 ] Authenticating user token > > __call__ > > > /usr/lib/python2.6/site-packages/keystoneclient/middleware/auth_token.py:676 > > <183>Jan 9 20:16:23 node-18 nova-api 2015-01-09 20:16:23.136 31870 DEBUG > > keystoneclient.middleware.auth_token > > [req-c9ec0973-ff63-4ac3-a0f7-1d2d7b7aa470 ] Removing headers from request > > environment: > > > X-Identity-Status,X-Domain-Id,X-Domain-Name,X-Project-Id,X-Project-Name,X-Project-Domain-Id,X-Project-Domain-Name,X-User-Id,X-User-Name,X-User-Domain-Id,X-User-Domain-Name,X-Roles,X-Service-Catalog,X-User,X-Tenant-Id,X-Tenant-Name,X-Tenant,X-Role > > _remove_auth_headers > > > /usr/lib/python2.6/site-packages/keystoneclient/middleware/auth_token.py:733 > > <183>Jan 9 20:16:23 node-18 nova-api 2015-01-09 20:16:23.137 31870 DEBUG > > keystoneclient.middleware.auth_token > > [req-c9ec0973-ff63-4ac3-a0f7-1d2d7b7aa470 ] Returning cached token > > _cache_get > > > /usr/lib/python2.6/site-packages/keystoneclient/middleware/auth_token.py:1545 > > <183>Jan 9 20:16:23 node-18 nova-api 2015-01-09 20:16:23.138 31870 DEBUG > > keystoneclient.middleware.auth_token > > [req-c9ec0973-ff63-4ac3-a0f7-1d2d7b7aa470 ] Storing token in cache store > > > /usr/lib/python2.6/site-packages/keystoneclient/middleware/auth_token.py:1460 > > <183>Jan 9 20:16:23 node-18 nova-api 2015-01-09 20:16:23.139 31870 DEBUG > > keystoneclient.middleware.auth_token > > [req-c9ec0973-ff63-4ac3-a0f7-1d2d7b7aa470 ] Received request from user: > > 4564fea80fa14e1daed160afa074d389 with project_id : > > dd32714d9009495bb51276e284380d6a and roles: admin,_member_ > > _build_user_headers > > > /usr/lib/python2.6/site-packages/keystoneclient/middleware/auth_token.py:996 > > <183>Jan 9 20:16:23 node-18 nova-api 2015-01-09 20:16:23.141 31870 DEBUG > > routes.middleware [req-05089e83-e4c1-4d90-b7c5-065226e55d91 ] Matched GET > > > /dd32714d9009495bb51276e284380d6a/servers/309a1f47-83b6-4ab4-9d71-642a2000c8a1 > > __call__ /usr/lib/python2.6/site-packages/routes/middleware.py:100 > > <183>Jan 9 20:16:23 node-18 nova-api 2015-01-09 20:16:23.142 31870 DEBUG > > routes.middleware [req-05089e83-e4c1-4d90-b7c5-065226e55d91 ] Route path: > > '/{project_id}/servers/:(id)', defaults: {'action': u'show', > 'controller': > > <nova.api.openstack.wsgi.Resource object at 0x43e2550>} __call__ > > /usr/lib/python2.6/site-packages/routes/middleware.py:102 > > <183>Jan 9 20:16:23 node-18 nova-api 2015-01-09 20:16:23.142 31870 DEBUG > > routes.middleware [req-05089e83-e4c1-4d90-b7c5-065226e55d91 ] Match dict: > > {'action': u'show', 'controller': <nova.api.openstack.wsgi.Resource > object > > at 0x43e2550>, 'project_id': u'dd32714d9009495bb51276e284380d6a', 'id': > > u'309a1f47-83b6-4ab4-9d71-642a2000c8a1'} __call__ > > /usr/lib/python2.6/site-packages/routes/middleware.py:103 > > <183>Jan 9 20:16:23 node-18 nova-api 2015-01-09 20:16:23.143 31870 DEBUG > > nova.api.openstack.wsgi [req-05089e83-e4c1-4d90-b7c5-065226e55d91 None] > > Calling method '<bound method Controller.show of > > <nova.api.openstack.compute.servers.Controller object at 0x4204290>>' > > (Content-type='None', Accept='application/json') _process_stack > > /usr/lib/python2.6/site-packages/nova/api/openstack/wsgi.py:945 > > <183>Jan 9 20:16:23 node-18 nova-compute 2015-01-09 20:16:23.170 29111 > > DEBUG nova.virt.vmwareapi.network_util > > [req-27cf4cd7-9184-4d7e-b57a-19ef3caeef26 None] Network br-int not found > on > > host! get_network_with_the_name > > /usr/lib/python2.6/site-packages/nova/virt/vmwareapi/network_util.py:80 > > <179>Jan 9 20:16:23 node-18 nova-compute 2015-01-09 20:16:23.171 29111 > > ERROR nova.compute.manager [req-27cf4cd7-9184-4d7e-b57a-19ef3caeef26 > None] > > [instance: 309a1f47-83b6-4ab4-9d71-642a2000c8a1] Instance failed to spawn > > 2015-01-09 20:16:23.171 29111 TRACE nova.compute.manager [instance: > > 309a1f47-83b6-4ab4-9d71-642a2000c8a1] Traceback (most recent call last): > > 2015-01-09 20:16:23.171 29111 TRACE nova.compute.manager [instance: > > 309a1f47-83b6-4ab4-9d71-642a2000c8a1] File > > "/usr/lib/python2.6/site-packages/nova/compute/manager.py", line 1714, in > > _spawn > > 2015-01-09 20:16:23.171 29111 TRACE nova.compute.manager [instance: > > 309a1f47-83b6-4ab4-9d71-642a2000c8a1] block_device_info) > > 2015-01-09 20:16:23.171 29111 TRACE nova.compute.manager [instance: > > 309a1f47-83b6-4ab4-9d71-642a2000c8a1] File > > "/usr/lib/python2.6/site-packages/nova/virt/vmwareapi/driver.py", line > 626, > > in spawn > > 2015-01-09 20:16:23.171 29111 TRACE nova.compute.manager [instance: > > 309a1f47-83b6-4ab4-9d71-642a2000c8a1] admin_password, network_info, > > block_device_info) > > 2015-01-09 20:16:23.171 29111 TRACE nova.compute.manager [instance: > > 309a1f47-83b6-4ab4-9d71-642a2000c8a1] File > > "/usr/lib/python2.6/site-packages/nova/virt/vmwareapi/vmops.py", line > 285, > > in spawn > > 2015-01-09 20:16:23.171 29111 TRACE nova.compute.manager [instance: > > 309a1f47-83b6-4ab4-9d71-642a2000c8a1] vif_infos = _get_vif_infos() > > 2015-01-09 20:16:23.171 29111 TRACE nova.compute.manager [instance: > > 309a1f47-83b6-4ab4-9d71-642a2000c8a1] File > > "/usr/lib/python2.6/site-packages/nova/virt/vmwareapi/vmops.py", line > 276, > > in _get_vif_infos > > 2015-01-09 20:16:23.171 29111 TRACE nova.compute.manager [instance: > > 309a1f47-83b6-4ab4-9d71-642a2000c8a1] self._is_neutron) > > 2015-01-09 20:16:23.171 29111 TRACE nova.compute.manager [instance: > > 309a1f47-83b6-4ab4-9d71-642a2000c8a1] File > > "/usr/lib/python2.6/site-packages/nova/virt/vmwareapi/vif.py", line 146, > in > > get_network_ref > > 2015-01-09 20:16:23.171 29111 TRACE nova.compute.manager [instance: > > 309a1f47-83b6-4ab4-9d71-642a2000c8a1] network_ref = > > get_neutron_network(session, network_name, cluster, vif) > > 2015-01-09 20:16:23.171 29111 TRACE nova.compute.manager [instance: > > 309a1f47-83b6-4ab4-9d71-642a2000c8a1] File > > "/usr/lib/python2.6/site-packages/nova/virt/vmwareapi/vif.py", line 138, > in > > get_neutron_network > > 2015-01-09 20:16:23.171 29111 TRACE nova.compute.manager [instance: > > 309a1f47-83b6-4ab4-9d71-642a2000c8a1] raise > > exception.NetworkNotFoundForBridge(bridge=bridge) > > 2015-01-09 20:16:23.171 29111 TRACE nova.compute.manager [instance: > > 309a1f47-83b6-4ab4-9d71-642a2000c8a1] NetworkNotFoundForBridge: Network > > could not be found for bridge br-int > > 2015-01-09 20:16:23.171 29111 TRACE nova.compute.manager [instance: > > 309a1f47-83b6-4ab4-9d71-642a2000c8a1] > > > > > > # cat /etc/neutron/plugins/ml2/ml2_conf.ini | grep -v ^# | grep -v ^$ > > [ml2] > > type_drivers = vlan > > tenant_network_types = vlan > > mechanism_drivers = openvswitch,dvs > > [ml2_type_flat] > > [ml2_type_vlan] > > network_vlan_ranges = physnet1:3000:3999,physnet2 > > [ml2_type_gre] > > [ml2_type_vxlan] > > [securitygroup] > > firewall_driver = > > neutron.agent.linux.iptables_firewall.OVSHybridIptablesFirewallDriver > > enable_security_group = True > > [agent] > > l2_population=False > > polling_interval=2 > > arp_responder=False > > [ovs] > > enable_tunneling=False > > integration_bridge=br-int > > bridge_mappings=physnet1:br-ex > > [ml2_vmware] > > host_ip=<vcenter_ip> > > host_username=root > > host_password=<password> > > wsdl_location=file:///opt/vmware/vimService.wsdl > > task_poll_interval=5.0 > > api_retry_count=10 > > network_maps = physnet1:dvSwitch > > > > > > # cat /etc/neutron/plugins/ml2/ml2_conf_vmware_dvs.ini | grep -v ^# | > grep > > -v ^$ > > [ml2_vmware] > > host_ip=<vcenter_ip> > > host_username=root > > host_password=<password> > > wsdl_location=file:///opt/vmware/vimService.wsdl > > task_poll_interval=5.0 > > api_retry_count=10 > > network_maps = physnet1:dvSwitch > > > > > > # ovs-vsctl show > > 80248645-469e-4b64-9408-7d26efce777f > > Bridge "br-eth3" > > Port "br-eth3" > > Interface "br-eth3" > > type: internal > > Port "eth3" > > Interface "eth3" > > Bridge br-int > > fail_mode: secure > > Port "tape9c03794-63" > > tag: 2 > > Interface "tape9c03794-63" > > type: internal > > Port br-int > > Interface br-int > > type: internal > > Port int-br-ex > > Interface int-br-ex > > Port int-br-prv > > Interface int-br-prv > > Bridge br-ex > > Port "br-ex--br-eth2" > > trunks: [0] > > Interface "br-ex--br-eth2" > > type: patch > > options: {peer="br-eth2--br-ex"} > > Port br-ex > > Interface br-ex > > type: internal > > Port phy-br-ex > > Interface phy-br-ex > > Bridge br-storage > > Port "br-storage--br-eth0" > > Interface "br-storage--br-eth0" > > type: patch > > options: {peer="br-eth0--br-storage"} > > Port br-storage > > Interface br-storage > > type: internal > > Bridge br-mgmt > > Port br-mgmt > > Interface br-mgmt > > type: internal > > Port "br-mgmt--br-eth0" > > Interface "br-mgmt--br-eth0" > > type: patch > > options: {peer="br-eth0--br-mgmt"} > > Bridge "br-eth0" > > Port "br-eth0" > > Interface "br-eth0" > > type: internal > > Port "br-eth0--br-storage" > > tag: 102 > > Interface "br-eth0--br-storage" > > type: patch > > options: {peer="br-storage--br-eth0"} > > Port "br-eth0--br-mgmt" > > tag: 101 > > Interface "br-eth0--br-mgmt" > > type: patch > > options: {peer="br-mgmt--br-eth0"} > > Port "br-eth0--br-prv" > > Interface "br-eth0--br-prv" > > type: patch > > options: {peer="br-prv--br-eth0"} > > Port "br-eth0--br-fw-admin" > > trunks: [0] > > Interface "br-eth0--br-fw-admin" > > type: patch > > options: {peer="br-fw-admin--br-eth0"} > > Port "eth0" > > Interface "eth0" > > Bridge "br-eth2" > > Port "eth2" > > Interface "eth2" > > Port "br-eth2" > > Interface "br-eth2" > > type: internal > > Port "br-eth2--br-ex" > > trunks: [0] > > Interface "br-eth2--br-ex" > > type: patch > > options: {peer="br-ex--br-eth2"} > > Bridge "br-eth1" > > Port "eth1" > > Interface "eth1" > > Port "br-eth1" > > Interface "br-eth1" > > type: internal > > Bridge br-prv > > Port "br-prv--br-eth0" > > Interface "br-prv--br-eth0" > > type: patch > > options: {peer="br-eth0--br-prv"} > > Port "qg-de0a02f9-d2" > > Interface "qg-de0a02f9-d2" > > type: internal > > Port br-prv > > Interface br-prv > > type: internal > > Port phy-br-prv > > Interface phy-br-prv > > Bridge br-fw-admin > > Port br-fw-admin > > Interface br-fw-admin > > type: internal > > Port "br-fw-admin--br-eth0" > > trunks: [0] > > Interface "br-fw-admin--br-eth0" > > type: patch > > options: {peer="br-eth0--br-fw-admin"} > > ovs_version: "1.10.2" > > > > > > # ip link > > 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 16436 qdisc noqueue state UNKNOWN > > link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 > > 2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP > qlen > > 1000 > > link/ether 14:fe:b5:0f:b6:79 brd ff:ff:ff:ff:ff:ff > > 3: eth1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP > qlen > > 1000 > > link/ether 14:fe:b5:0f:b6:7b brd ff:ff:ff:ff:ff:ff > > 4: eth2: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP > qlen > > 1000 > > link/ether 14:fe:b5:0f:b6:7d brd ff:ff:ff:ff:ff:ff > > 5: eth3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP > qlen > > 1000 > > link/ether 14:fe:b5:0f:b6:7f brd ff:ff:ff:ff:ff:ff > > 6: ovs-system: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN > > link/ether 6a:26:28:63:48:52 brd ff:ff:ff:ff:ff:ff > > 7: br-eth3: <BROADCAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UNKNOWN > > link/ether 14:fe:b5:0f:b6:7f brd ff:ff:ff:ff:ff:ff > > 8: br-ex: <BROADCAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UNKNOWN > > link/ether a6:3d:66:56:16:40 brd ff:ff:ff:ff:ff:ff > > 9: br-eth1: <BROADCAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UNKNOWN > > link/ether 14:fe:b5:0f:b6:7b brd ff:ff:ff:ff:ff:ff > > 10: br-int: <BROADCAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UNKNOWN > > link/ether 8e:6a:fb:1f:18:47 brd ff:ff:ff:ff:ff:ff > > 14: br-fw-admin: <BROADCAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state > > UNKNOWN > > link/ether e6:a1:ea:f3:0f:45 brd ff:ff:ff:ff:ff:ff > > 15: br-storage: <BROADCAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state > UNKNOWN > > link/ether 42:a0:c7:5e:45:4d brd ff:ff:ff:ff:ff:ff > > 16: br-eth2: <BROADCAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UNKNOWN > > link/ether 14:fe:b5:0f:b6:7d brd ff:ff:ff:ff:ff:ff > > 17: br-prv: <BROADCAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UNKNOWN > > link/ether 16:23:fe:ec:eb:4f brd ff:ff:ff:ff:ff:ff > > 19: br-eth0: <BROADCAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UNKNOWN > > link/ether 14:fe:b5:0f:b6:79 brd ff:ff:ff:ff:ff:ff > > 20: br-mgmt: <BROADCAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UNKNOWN > > link/ether b6:9c:f9:60:a3:40 brd ff:ff:ff:ff:ff:ff > > 22: virbr0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue > state > > UNKNOWN > > link/ether 92:c8:0e:96:13:db brd ff:ff:ff:ff:ff:ff > > 33: phy-br-prv: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc > pfifo_fast > > state UP qlen 1000 > > link/ether b2:29:ee:f4:86:16 brd ff:ff:ff:ff:ff:ff > > 34: int-br-prv: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc > pfifo_fast > > state UP qlen 1000 > > link/ether 6e:c4:d3:3e:c2:11 brd ff:ff:ff:ff:ff:ff > > 57: phy-br-ex: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc > pfifo_fast > > state UP qlen 1000 > > link/ether 96:8b:87:06:4b:e3 brd ff:ff:ff:ff:ff:ff > > 58: int-br-ex: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc > pfifo_fast > > state UP qlen 1000 > > link/ether fe:08:e6:ba:bf:d3 brd ff:ff:ff:ff:ff:ff > > > > > > # brctl show > > bridge name bridge id STP enabled interfaces > > virbr0 8000.000000000000 yes > > > > I guess I am missing some thing. > > > > It looks like my issue is similar to this : > > > https://ask.openstack.org/en/question/43594/vmware-neutron-bridging-problem/ > > > > I have configured br100 with VLAN ID 103 in vCenter. But I don't have > br100 > > in my controller node. Not sure how to create it in my controller. > > > > NOTE : I have another openstack environment which I deployed as vCenter > > Environment using Fuel 5.1 and manually installed/configured Neutron + > > VMware dvSwitch ML2 Mechanism driver. It works fine with same vCenter. > > > > Any help? > > > > I happy to provide more info if required. > > > > -- > > Thanks & Regards > > E-Mail: thefossg...@gmail.com > > IRC: neophy > > Blog : http://lmohanphy.livejournal.com/ > > > > > > _______________________________________________ > > OpenStack Development Mailing List (not for usage questions) > > Unsubscribe: > openstack-dev-requ...@lists.openstack.org?subject:unsubscribe > > http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev > > > > > > -- > Andrew > Mirantis > Ceph community > > __________________________________________________________________________ > OpenStack Development Mailing List (not for usage questions) > Unsubscribe: openstack-dev-requ...@lists.openstack.org?subject:unsubscribe > http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev >
__________________________________________________________________________ OpenStack Development Mailing List (not for usage questions) Unsubscribe: openstack-dev-requ...@lists.openstack.org?subject:unsubscribe http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev