In one config, you have specified local_ip as 10.10.10.1 and in other as 192.168.3.3 . Doesn't they should belong to same network? As per the doc, it should be 10.10.10.3? Plus, these both belong to Data-Network, which is not controller-network communication but compute-network communication.
-Regards Rahul On Sat, Feb 23, 2013 at 12:53 AM, Aaron Rosen <[email protected]> wrote: > >From the network+controller node can you ping to 192.168.3.3 (just to > confirm there is ip connectivity between those). > > Your configs look fine to me. The issue you are having is that your > network+controller node doesn't have a tunnel to your HV node. I'd suggest > restarting the quantum-plugin-openvswitch-agent service on both nodes and > see if that does the trick in order to get the agent to add the tunnel for > you. Perhaps you edited this file and didn't restart the agent? > > Aaron > > On Fri, Feb 22, 2013 at 10:55 AM, Guilherme Russi < > [email protected]> wrote: > >> Here is my controller + network node: >> >> cat /etc/quantum/plugins/openvswitch/ovs_quantum_plugin.ini >> [DATABASE] >> # This line MUST be changed to actually run the plugin. >> # Example: >> # sql_connection = mysql://root:[email protected]:3306/ovs_quantum >> # Replace 127.0.0.1 above with the IP address of the database used by the >> # main quantum server. (Leave it as is if the database runs on this host.) >> sql_connection = mysql://quantum:password@localhost:3306/quantum >> # Database reconnection retry times - in event connectivity is lost >> # set to -1 implies an infinite retry count >> # sql_max_retries = 10 >> # Database reconnection interval in seconds - in event connectivity is >> lost >> reconnect_interval = 2 >> >> [OVS] >> # (StrOpt) Type of network to allocate for tenant networks. The >> # default value 'local' is useful only for single-box testing and >> # provides no connectivity between hosts. You MUST either change this >> # to 'vlan' and configure network_vlan_ranges below or change this to >> # 'gre' and configure tunnel_id_ranges below in order for tenant >> # networks to provide connectivity between hosts. Set to 'none' to >> # disable creation of tenant networks. >> # >> # Default: tenant_network_type = local >> # Example: tenant_network_type = gre >> tenant_network_type = gre >> >> # (ListOpt) Comma-separated list of >> # <physical_network>[:<vlan_min>:<vlan_max>] tuples enumerating ranges >> # of VLAN IDs on named physical networks that are available for >> # allocation. All physical networks listed are available for flat and >> # VLAN provider network creation. Specified ranges of VLAN IDs are >> # available for tenant network allocation if tenant_network_type is >> # 'vlan'. If empty, only gre and local networks may be created. >> # >> # Default: network_vlan_ranges = >> # Example: network_vlan_ranges = physnet1:1000:2999 >> >> # (BoolOpt) Set to True in the server and the agents to enable support >> # for GRE networks. Requires kernel support for OVS patch ports and >> # GRE tunneling. >> # >> # Default: enable_tunneling = False >> enable_tunneling = True >> >> # (ListOpt) Comma-separated list of <tun_min>:<tun_max> tuples >> # enumerating ranges of GRE tunnel IDs that are available for tenant >> # network allocation if tenant_network_type is 'gre'. >> # >> # Default: tunnel_id_ranges = >> # Example: tunnel_id_ranges = 1:1000 >> tunnel_id_ranges = 1:1000 >> >> # Do not change this parameter unless you have a good reason to. >> # This is the name of the OVS integration bridge. There is one per >> hypervisor. >> # The integration bridge acts as a virtual "patch bay". All VM VIFs are >> # attached to this bridge and then "patched" according to their network >> # connectivity. >> # >> # Default: integration_bridge = br-int >> integration_bridge = br-int >> >> # Only used for the agent if tunnel_id_ranges (above) is not empty for >> # the server. In most cases, the default value should be fine. >> # >> # Default: tunnel_bridge = br-tun >> tunnel_bridge = br-tun >> >> # Uncomment this line for the agent if tunnel_id_ranges (above) is not >> # empty for the server. Set local-ip to be the local IP address of >> # this hypervisor. >> # >> # Default: local_ip = >> local_ip = 10.10.10.1 >> >> >> And here is my compute node: >> >> cat /etc/quantum/plugins/openvswitch/ovs_quantum_plugin.ini >> [DATABASE] >> # This line MUST be changed to actually run the plugin. >> # Example: >> # sql_connection = mysql://root:[email protected]:3306/ovs_quantum >> # Replace 127.0.0.1 above with the IP address of the database used by the >> # main quantum server. (Leave it as is if the database runs on this host.) >> sql_connection = mysql://quantum:[email protected]:3306/quantum >> # Database reconnection retry times - in event connectivity is lost >> # set to -1 implies an infinite retry count >> # sql_max_retries = 10 >> # Database reconnection interval in seconds - in event connectivity is >> lost >> reconnect_interval = 2 >> >> [OVS] >> # (StrOpt) Type of network to allocate for tenant networks. The >> # default value 'local' is useful only for single-box testing and >> # provides no connectivity between hosts. You MUST either change this >> # to 'vlan' and configure network_vlan_ranges below or change this to >> # 'gre' and configure tunnel_id_ranges below in order for tenant >> # networks to provide connectivity between hosts. Set to 'none' to >> # disable creation of tenant networks. >> # >> # Default: tenant_network_type = local >> # Example: tenant_network_type = gre >> tenant_network_type = gre >> >> # (ListOpt) Comma-separated list of >> # <physical_network>[:<vlan_min>:<vlan_max>] tuples enumerating ranges >> # of VLAN IDs on named physical networks that are available for >> # allocation. All physical networks listed are available for flat and >> # VLAN provider network creation. Specified ranges of VLAN IDs are >> # available for tenant network allocation if tenant_network_type is >> # 'vlan'. If empty, only gre and local networks may be created. >> # >> # Default: network_vlan_ranges = >> # Example: network_vlan_ranges = physnet1:1000:2999 >> >> # (BoolOpt) Set to True in the server and the agents to enable support >> # for GRE networks. Requires kernel support for OVS patch ports and >> # GRE tunneling. >> # >> # Default: enable_tunneling = False >> enable_tunneling = True >> >> # (ListOpt) Comma-separated list of <tun_min>:<tun_max> tuples >> # enumerating ranges of GRE tunnel IDs that are available for tenant >> # network allocation if tenant_network_type is 'gre'. >> # >> # Default: tunnel_id_ranges = >> # Example: tunnel_id_ranges = 1:1000 >> tunnel_id_ranges = 1:1000 >> >> # Do not change this parameter unless you have a good reason to. >> # This is the name of the OVS integration bridge. There is one per >> hypervisor. >> # The integration bridge acts as a virtual "patch bay". All VM VIFs are >> # attached to this bridge and then "patched" according to their network >> # connectivity. >> # >> # Default: integration_bridge = br-int >> integration_bridge = br-int >> >> # Only used for the agent if tunnel_id_ranges (above) is not empty for >> # the server. In most cases, the default value should be fine. >> # >> # Default: tunnel_bridge = br-tun >> tunnel_bridge = br-tun >> >> # Uncomment this line for the agent if tunnel_id_ranges (above) is not >> # empty for the server. Set local-ip to be the local IP address of >> # this hypervisor. >> # >> # Default: local_ip = 10.0.0.3 >> local_ip = 192.168.3.3 >> >> The 10.10.10.1 is the Data Network from network controller (following the >> tutorial), and it is in the eth0:0 (I'm not pretty sure, but i guess the >> data from the VMs should communicate with this IP). >> >> >> >> >> >> 2013/2/22 Aaron Rosen <[email protected]> >> >>> Running with two nics for this should be fine for tunneling as ip >>> routing would handle which nic the packets should go out. From what you >>> pasted I see that one HV has a gre tunnel setup to 10.10.10.1 <-- Who is >>> that host? Can you attach >>> your /etc/quantum/plugins/openvswitch/ovs_quantum_plugin.ini on your nodes. >>> >>> I suspect the issue is a configuration issue in the [OVS] section. >>> You'll need something along the lines of this in that section: >>> [OVS] >>> local_ip = <ip address of HV> >>> enable_tunneling = True >>> tunnel_id_ranges = 1:1000 >>> tenant_network_type = gre >>> >>> >>> >>> >>> On Fri, Feb 22, 2013 at 9:54 AM, Guilherme Russi < >>> [email protected]> wrote: >>> >>>> Hello Aaron, >>>> >>>> Sorry about attaching the infos, about the quantum agent, is it the >>>> quantum-plugin-openvswitch-agent? If i was, the job is already ruunning at >>>> the controller and the compute nodes: >>>> >>>> service quantum-plugin-openvswitch-agent start >>>> start: Job is already running: quantum-plugin-openvswitch-agent >>>> >>>> Is there another thing i should do? I'm running my controller node and >>>> the network node at the same machine with 2 NICs, maybe can be a problem >>>> how i am making my network config? >>>> >>>> Thanks again. >>>> >>>> Guilherme. >>>> >>>> >>>> 2013/2/22 Aaron Rosen <[email protected]> >>>> >>>>> Hi Guilherme, >>>>> >>>>> (next time please paste these in the email rather than attaching, >>>>> thx). >>>>> >>>>> From the text in the attachment (show below). It seems like you are >>>>> not running the quantum-openvswitch-agent on your network node as there is >>>>> no GRE tunnel from that to your compute node. Once you have >>>>> quantum-openvswitch-agent running on all your machines you should be able >>>>> to run ovs-dpctl looking under br-tun and see a tunnel between each host. >>>>> >>>>> Aaron >>>>> >>>>> CONTROLLER + NETWORK NODE: >>>>> system@br-tun: >>>>> lookups: hit:0 missed:0 lost:0 >>>>> flows: 0 >>>>> port 0: br-tun (internal) >>>>> port 1: patch-int (patch: peer=patch-tun) >>>>> system@br-int: >>>>> lookups: hit:0 missed:0 lost:0 >>>>> flows: 0 >>>>> port 0: br-int (internal) >>>>> port 1: tap817d2f70-a0 (internal) >>>>> port 2: qr-ea64e9aa-31 (internal) >>>>> port 3: patch-tun (patch: peer=patch-int) >>>>> system@br-ex: >>>>> lookups: hit:0 missed:0 lost:0 >>>>> flows: 0 >>>>> port 0: br-ex (internal) >>>>> port 2: qg-95fe3fa1-d1 (internal) >>>>> >>>>> >>>>> COMPUTE NODES >>>>> >>>>> COMPUTE NODE 01: >>>>> ovs-dpctl show >>>>> system@br-int: >>>>> lookups: hit:380 missed:7590 lost:0 >>>>> flows: 0 >>>>> port 0: br-int (internal) >>>>> port 2: patch-tun (patch: peer=patch-int) >>>>> port 3: qvo981ae82e-d4 >>>>> port 6: qvoc9df3a96-5f >>>>> port 7: qvoc153ac28-ae >>>>> port 8: qvo722a5d05-e4 >>>>> system@br-tun: >>>>> lookups: hit:381 missed:7589 lost:0 >>>>> flows: 0 >>>>> port 0: br-tun (internal) >>>>> port 1: patch-int (patch: peer=patch-tun) >>>>> port 2: gre-1 (gre: key=flow, remote_ip=10.10.10.1) >>>>> >>>>> >>>>> On Fri, Feb 22, 2013 at 8:47 AM, Guilherme Russi < >>>>> [email protected]> wrote: >>>>> >>>>>> So guys, any idea about what am i missing? >>>>>> >>>>>> Thanks. >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> 2013/2/22 Guilherme Russi <[email protected]> >>>>>> >>>>>>> Hello Aaron, >>>>>>> >>>>>>> Here are the outputs. >>>>>>> >>>>>>> >>>>>>> Thanks. >>>>>>> >>>>>>> Guilherme. >>>>>>> >>>>>>> >>>>>>> 2013/2/21 Aaron Rosen <[email protected]> >>>>>>> >>>>>>>> The output to the following would be a good start: >>>>>>>> >>>>>>>> quantum net-list >>>>>>>> quantum port-list >>>>>>>> ovs-dpctl show (on all nodes) >>>>>>>> >>>>>>>> Also make sure the quantum-dhcp-agent is running on your network >>>>>>>> node. >>>>>>>> >>>>>>>> Aaron >>>>>>>> >>>>>>>> On Thu, Feb 21, 2013 at 11:23 AM, Guilherme Russi < >>>>>>>> [email protected]> wrote: >>>>>>>> >>>>>>>>> Sorry about that, I'm using Folsom release with quantum, I'm >>>>>>>>> installing the controller node and the network node in the same >>>>>>>>> physical >>>>>>>>> machine, I'm following this tutorial: >>>>>>>>> >>>>>>>>> >>>>>>>>> http://docs.openstack.org/folsom/basic-install/content/basic-install_controller.html >>>>>>>>> >>>>>>>>> Which config files do you need? >>>>>>>>> >>>>>>>>> Thanks. >>>>>>>>> >>>>>>>>> Guilherme. >>>>>>>>> >>>>>>>>> >>>>>>>>> 2013/2/21 Aaron Rosen <[email protected]> >>>>>>>>> >>>>>>>>>> Hi, >>>>>>>>>> >>>>>>>>>> You'll have to provide more information than this for anyone to >>>>>>>>>> help you: i.e are you using quantum or nova-network, if your using >>>>>>>>>> quantum >>>>>>>>>> which plugin, config files etc. >>>>>>>>>> >>>>>>>>>> Aaron >>>>>>>>>> >>>>>>>>>> On Thu, Feb 21, 2013 at 11:13 AM, Guilherme Russi < >>>>>>>>>> [email protected]> wrote: >>>>>>>>>> >>>>>>>>>>> Hello guys, >>>>>>>>>>> >>>>>>>>>>> I'm getting problem in my VMs' creation, they don't get IP, the >>>>>>>>>>> log piece shows: >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> Starting network... >>>>>>>>>>> udhcpc (v1.18.5) started >>>>>>>>>>> Sending discover... >>>>>>>>>>> Sending discover... >>>>>>>>>>> Sending discover... >>>>>>>>>>> No lease, failing >>>>>>>>>>> WARN: /etc/rc3.d/S40-network failed >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> Do you have any idea how I can solve it? >>>>>>>>>>> >>>>>>>>>>> Thank you so much. >>>>>>>>>>> >>>>>>>>>>> Guilherme. >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _______________________________________________ >>>>>>>>>>> Mailing list: https://launchpad.net/~openstack >>>>>>>>>>> Post to : [email protected] >>>>>>>>>>> Unsubscribe : https://launchpad.net/~openstack >>>>>>>>>>> More help : https://help.launchpad.net/ListHelp >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>> >> > > _______________________________________________ > Mailing list: https://launchpad.net/~openstack > Post to : [email protected] > Unsubscribe : https://launchpad.net/~openstack > More help : https://help.launchpad.net/ListHelp > >
_______________________________________________ Mailing list: https://launchpad.net/~openstack Post to : [email protected] Unsubscribe : https://launchpad.net/~openstack More help : https://help.launchpad.net/ListHelp

