Hi Jeremy
Nova-network has basically 3 modes (Flat, FlatDHCP and VLAN) while
neutron is designed to handle complex setups needing most flexibility.
And your setup seems to be complex so I think it would be better of with
Neutron networking.
Thanks,
Amit
On 05/30/2014 01:21 PM, Jeremy Utley wrote:
Hello list,
I'm working on a proof-of-concept openstack installation for my
company, and am looking for the answers to some questions that I
haven't been able to find the answer to. Hopefully someone here can
help point me in the right direction!
I've got a basic setup already in place - 2 controller nodes, 3
compute nodes (with 5 more ready to provision when we line everything
out). The main problems I'm facing right now is in the networking.
We need to integrate this openstack setup with the network resources
already in place in our setup, and I'm not exactly sure how to do so.
I'll start off with basically going over how we are currently setup.
1. Management and compute nodes are all connected via infiniband
networking, we are using this as our cluster management network
(10.30.x.x/16). This network has zero connectivity to anything else -
it's completely private for the cluster.
2. eth2 on all compute nodes are connected to each other via a vlan on
our switch, and bridged to br100 to be our openstack fixed network.
(10.18.x.x/16).
3. eth1 on all compute nodes is connected up to our existing colo
private network. We are currently defining this as our external
network. (10.5.20.0/22
<https://urldefense.proofpoint.com/v1/url?u=http://10.5.20.0/22&k=uWCMTgG0stZxwOEwDWvrOA%3D%3D%0A&r=SrS6LVctHAotDvfalKfFvzCQXOUI4d%2BwuHpBCRhEqKk%3D%0A&m=%2FeM6hckXh%2FT8ZrfqIm26OhTyCt9D11gaZDh881YmkC0%3D%0A&s=d031a1f31c5f5feb90b90502c1812948448bf241c1e2a7d539311a5369eaca81>)
- In our current setup (with nova-network and the FlatDHCPManager), I
have taken a block of IP's from this subnet and reserved them for use
as floating IP's for testing purposes - this is working perfectly
right now.
4. eth0 on all compute nodes is connected to our existing colo public
network. We have a /19 public allocation, broken up into numerous /24
and /25 segments to keep independent divisions of the company fully
segregated - each segment is a separate vlan on the public network
switches. In what we currently have setup, we are not utilizing this.
Ultimately, we'd like to have our cluster VM's connected to the fixed
network (on eth2), and treat both eth1 and eth0 as "public networks"
we can use floating IP's from. All VM's should connect to eth1 and be
able to have floating IP's assigned from that network to them, and
they should be able to connect to a single tagged vlan on eth0 as well.
From the reading I've done so far, I think what we are trying to do
might be too complicated for nova-network, since it depends on
defining a single interface as the public interface on the compute
nodes, and we might potentially have more. Am I interpreting that
correctly, or could we maybe accomplish this with nova-network
(perhaps using the VLANManager mode?)
If we have to switch to Neutron, can you run the neutron services on
each compute node? We have concerns about scale if we have to
implement a separate network node, as we could easily end up
saturating a full gig-e interface with this cluster in the future.
Plus, the extra cost expenditure of dedicated network nodes could end
up cost-prohibitive in the early stages of deployment.
Anyone got any suggestions for us?
Thank you for your time,
Jeremy Utley
_______________________________________________
Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
Post to : [email protected]
Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
_______________________________________________
Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
Post to : [email protected]
Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack