Network node:
root@network2:/# iptables -L -nv --line-numbers
Chain INPUT (policy ACCEPT 1286 packets, 351K bytes)
num pkts bytes target prot opt in out source
destination
1 1171 338K neutron-openvswi-INPUT all -- * * 0.0.0.0/0
0.0.0.0/0
Chain FORWARD (policy ACCEPT 0 packets, 0 bytes)
num pkts bytes target prot opt in out source
destination
1 0 0 neutron-filter-top all -- * * 0.0.0.0/0
0.0.0.0/0
2 0 0 neutron-openvswi-FORWARD all -- * * 0.0.0.0/0
0.0.0.0/0
Chain OUTPUT (policy ACCEPT 1124 packets, 180K bytes)
num pkts bytes target prot opt in out source
destination
1 988 164K neutron-filter-top all -- * * 0.0.0.0/0
0.0.0.0/0
2 988 164K neutron-openvswi-OUTPUT all -- * * 0.0.0.0/0
0.0.0.0/0
Chain neutron-filter-top (2 references)
num pkts bytes target prot opt in out source
destination
1 988 164K neutron-openvswi-local all -- * * 0.0.0.0/0
0.0.0.0/0
Chain neutron-openvswi-FORWARD (1 references)
num pkts bytes target prot opt in out source
destination
Chain neutron-openvswi-INPUT (1 references)
num pkts bytes target prot opt in out source
destination
Chain neutron-openvswi-OUTPUT (1 references)
num pkts bytes target prot opt in out source
destination
Chain neutron-openvswi-local (1 references)
num pkts bytes target prot opt in out source
destination
Chain neutron-openvswi-sg-chain (0 references)
num pkts bytes target prot opt in out source
destination
Chain neutron-openvswi-sg-fallback (0 references)
num pkts bytes target prot opt in out source
destination
1 0 0 DROP all -- * * 0.0.0.0/0
0.0.0.0/0
root@network2:/#
Controller node
root@controller2:/# iptables -L -nv --line-numbers
Chain INPUT (policy ACCEPT 25498 packets, 7540K bytes)
num pkts bytes target prot opt in out source
destination
1 25498 7540K nova-api-INPUT all -- * * 0.0.0.0/0
0.0.0.0/0
Chain FORWARD (policy ACCEPT 0 packets, 0 bytes)
num pkts bytes target prot opt in out source
destination
1 0 0 nova-filter-top all -- * * 0.0.0.0/0
0.0.0.0/0
2 0 0 nova-api-FORWARD all -- * * 0.0.0.0/0
0.0.0.0/0
Chain OUTPUT (policy ACCEPT 24216 packets, 7244K bytes)
num pkts bytes target prot opt in out source
destination
1 24216 7244K nova-filter-top all -- * * 0.0.0.0/0
0.0.0.0/0
2 24216 7244K nova-api-OUTPUT all -- * * 0.0.0.0/0
0.0.0.0/0
Chain nova-api-FORWARD (1 references)
num pkts bytes target prot opt in out source
destination
Chain nova-api-INPUT (1 references)
num pkts bytes target prot opt in out source
destination
1 0 0 ACCEPT tcp -- * * 0.0.0.0/0
172.22.14.22 tcp dpt:8775
Chain nova-api-OUTPUT (1 references)
num pkts bytes target prot opt in out source
destination
Chain nova-api-local (1 references)
num pkts bytes target prot opt in out source
destination
Chain nova-filter-top (2 references)
num pkts bytes target prot opt in out source
destination
1 24216 7244K nova-api-local all -- * * 0.0.0.0/0
0.0.0.0/0
root@controller2:/#
From: Uwe Sauter [mailto:[email protected]]
Sent: 29. júní 2015 22:46
To: Yngvi Páll Þorfinnsson; YANG LI
Cc: [email protected]
Subject: RE: [Openstack] network question on openstack installation
One more thing. Please provide
iptables -L -nv --line-numbers
for network and compute nodes.
Am 30. Juni 2015 00:25:45 MESZ, schrieb "Yngvi Páll Þorfinnsson"
<[email protected]<mailto:[email protected]>>:
In fact, I don‘t think I‘ll need more than one „external network“
so, am I on the wrong path here, i.e. when I‘m configuing the external network
as a VLAN ?
Best regards
Yngvi
From: Yngvi Páll Þorfinnsson
Sent: 29. júní 2015 21:57
To: Uwe Sauter; YANG LI
Cc: [email protected]<mailto:[email protected]>
Subject: Re: [Openstack] network question on openstack installation
OK, I only found one fresh error
Compute node; nova-compute.log ( as usually when I create instance)
grep ERR nova-compute.log
2015-06-29 21:11:11.801 4166 ERROR nova.compute.manager [-] [instance:
af901a2b-2462-4c19-b1f1-237371fd8177] Instance failed to spawn
I‘ve attached the neutron agent-show and neutron (sub)net-list in the attached
file.
Best regards
Yngvi
From: Uwe Sauter [mailto:[email protected]]
Sent: 29. júní 2015 21:34
To: Yngvi Páll Þorfinnsson; YANG LI
Cc: [email protected]<mailto:[email protected]>
Subject: RE: [Openstack] network question on openstack installation
Can you check for ERRORs in:
Network node: neutron server log, neutron openvswitch agent log, openvswitch log
Nova controller node: nova api log, nova scheduler log
Compute node: nova compute log, neutron openvswitch agent log, openvswitch log
Also please list again neutron agent-show for the different agents and neutron
net-show and neutron subnet-show for your (sub)networks.
Am 29. Juni 2015 23:24:48 MESZ, schrieb "Yngvi Páll Þorfinnsson"
<[email protected]<mailto:[email protected]>>:
Thanks a lot for your effort Uwe ;-)
It‘s relly helpful !
Now , I keep creating instances, and have the same error.
I still get a strange comment in the
Neutron server.log file when I try to create an instance:
2015-06-29 21:11:11.576 1960 DEBUG neutron.plugins.ml2.drivers.mech_openvswitch
[req-1e603e4b-61e6-4896-8f81-208daba8569b None] Checking segment:
{'segmentation_id': 1102L, 'physical_network': u'external', 'id':
u'11fab5ad-c457-4175-9e5a-f505fc0e072d', 'network_type': u'vlan'} for mappings:
{u'external': u'br-ex'} with tunnel_types: [u'gre'] check_segment_for_agent
/usr/lib/python2.7/dist-packages/neutron/plugins/ml2/drivers/mech_openvswitch.py:52
But still this segment is not listed for ‚neutron net-list‘ nor
‚neutron-subnet-list‘ !
root@controller2:/# source admin-openrc.sh
root@controller2:/#
root@controller2:/# neutron net-list
+--------------------------------------+-------------+-----------------------------------------------------+
| id | name | subnets
|
+--------------------------------------+-------------+-----------------------------------------------------+
| b43da44a-42d5-4b1f-91c2-d06a923deb29 | ext_net1101 |
c40fa8e3-cd8e-4566-ade6-5f3eabed121c 157.157.8.0/24 |
| 3446e54b-346f-45e5-89a2-1ec4eef251ab | demo-net |
2c79bb00-0ace-4319-8151-81210ee3dfb2 172.22.18.0/24 |
+--------------------------------------+-------------+-----------------------------------------------------+
root@controller2:/#
root@controller2:/# neutron subnet-list
+--------------------------------------+--------------------+----------------+---------------------------------------------------+
| id | name | cidr |
allocation_pools |
+--------------------------------------+--------------------+----------------+---------------------------------------------------+
| c40fa8e3-cd8e-4566-ade6-5f3eabed121c | ext_net1101-subnet | 157.157.8.0/24 |
{"start": "157.157.8.51", "end": "157.157.8.250"} |
| 2c79bb00-0ace-4319-8151-81210ee3dfb2 | demo-subnet | 172.22.18.0/24 |
{"start": "172.22.18.2", "end": "172.22.18.254"} |
+--------------------------------------+--------------------+----------------+---------------------------------------------------+
And the output of listing it is empty:
root@controller2:/# neutron net-show 11fab5ad-c457-4175-9e5a-f505fc0e072d
Unable to find network with name '11fab5ad-c457-4175-9e5a-f505fc0e072d'
root@controller2:/#
root@controller2:/# source demo-openrc.sh
root@controller2:/# neutron net-show 11fab5ad-c457-4175-9e5a-f505fc0e072d
Unable to find network with name '11fab5ad-c457-4175-9e5a-f505fc0e072d'
This is after I have dropped the neutron, and resynced....
Best regards
Yngvi
From: Uwe Sauter [mailto:[email protected]]
Sent: 29. júní 2015 21:16
To: Yngvi Páll Þorfinnsson; YANG LI
Cc: [email protected]<mailto:[email protected]>
Subject: RE: [Openstack] network question on openstack installation
Yes. Just keep in mind that if you extend your configuration with a new config
file, then you must change your init script / unit file to reference that file.
And it would probably be a good idea to re-sync the DB with that additional
file as an option. Or you keep your plugin configuration in a single file and
be happy with the current layout.
Am 29. Juni 2015 22:47:22 MESZ, schrieb "Yngvi Páll Þorfinnsson"
<[email protected]<mailto:[email protected]>>:
Hi Uwe
I just ran this some minutes ago, i.e. did the "population of db" again,
According to the manual.
Should'nt this be enough ?
root@controller2:/# su -s /bin/sh -c "neutron-db-manage --config-file
/etc/neutron/neutron.conf \
--config-file /etc/neutron/plugins/ml2/ml2_conf.ini upgrade juno" neutron
INFO [alembic.migration] Context impl MySQLImpl.
INFO [alembic.migration] Will assume non-transactional DDL.
INFO [alembic.migration] Running upgrade None -> havana, havana_initial
INFO [alembic.migration] Running upgrade havana -> e197124d4b9, add unique
constraint to members
INFO [alembic.migration] Running upgrade e197124d4b9 -> 1fcfc149aca4, Add a
unique constraint on (agent_type, host) columns to prevent a race
condition when an agent
entry is 'upserted'.
INFO [alembic.migration] Running upgrade 1fcfc149aca4 -> 50e86cb2637a,
nsx_mappings
INFO [alembic.migration] Running upgrade 50e86cb2637a -> 1421183d533f, NSX
DHCP/metadata support
INFO [alembic.migration] Running upgrade 1421183d533f -> 3d3cb89d84ee,
nsx_switch_mappings
INFO [alembic.migration] Running upgrade 3d3cb89d84ee -> 4ca36cfc898c,
nsx_router_mappings
INFO [alembic.migration] Running upgrade 4ca36cfc898c -> 27cc183af192,
ml2_vnic_type
INFO [alembic.migration] Running upgrade 27cc183af192 -> 50d5ba354c23, ml2
binding:vif_details
INFO [alembic.migration] Running upgrade 50d5ba354c23 -> 157a5d299379, ml2
binding:profile
INFO [alembic.migration] Running upgrade 157a5d299379 -> 3d2585038b95, VMware
NSX rebranding
INFO [alembic.migration] Running upgrade 3d2585038b95 -> abc88c33f74f, lb stats
INFO
[alembic.migration]
Running upgrade abc88c33f74f -> 1b2580001654,
nsx_sec_group_mapping
INFO [alembic.migration] Running upgrade 1b2580001654 -> e766b19a3bb,
nuage_initial
INFO [alembic.migration] Running upgrade e766b19a3bb -> 2eeaf963a447,
floatingip_status
INFO [alembic.migration] Running upgrade 2eeaf963a447 -> 492a106273f8, Brocade
ML2 Mech. Driver
INFO [alembic.migration] Running upgrade 492a106273f8 -> 24c7ea5160d7, Cisco
CSR VPNaaS
INFO [alembic.migration] Running upgrade 24c7ea5160d7 -> 81c553f3776c,
bsn_consistencyhashes
INFO [alembic.migration] Running upgrade 81c553f3776c -> 117643811bca, nec:
delete old ofc mapping tables
INFO [alembic.migration] Running upgrade 117643811bca -> 19180cf98af6,
nsx_gw_devices
INFO [alembic.migration] Running upgrade 19180cf98af6 -> 33dd0a9fa487,
embrane_lbaas_driver
INFO [alembic.migration] Running upgrade 33dd0a9fa487 -> 2447ad0e9585, Add
IPv6 Subnet properties
INFO [alembic.migration] Running upgrade 2447ad0e9585
-> 538732fa21e1, NEC Rename quantum_id to neutron_id
INFO [alembic.migration] Running upgrade 538732fa21e1 -> 5ac1c354a051, n1kv
segment allocs for cisco n1kv plugin
INFO [alembic.migration] Running upgrade 5ac1c354a051 -> icehouse, icehouse
INFO [alembic.migration] Running upgrade icehouse -> 54f7549a0e5f,
set_not_null_peer_address
INFO [alembic.migration] Running upgrade 54f7549a0e5f -> 1e5dd1d09b22,
set_not_null_fields_lb_stats
INFO [alembic.migration] Running upgrade 1e5dd1d09b22 -> b65aa907aec,
set_length_of_protocol_field
INFO [alembic.migration] Running upgrade b65aa907aec -> 33c3db036fe4,
set_length_of_description_field_metering
INFO [alembic.migration] Running upgrade 33c3db036fe4 -> 4eca4a84f08a, Remove
ML2 Cisco Credentials DB
INFO [alembic.migration] Running upgrade 4eca4a84f08a -> d06e871c0d5,
set_admin_state_up_not_null_ml2
INFO
[alembic.migration] Running upgrade d06e871c0d5 ->
6be312499f9, set_not_null_vlan_id_cisco
INFO [alembic.migration] Running upgrade 6be312499f9 -> 1b837a7125a9, Cisco
APIC Mechanism Driver
INFO [alembic.migration] Running upgrade 1b837a7125a9 -> 10cd28e692e9,
nuage_extraroute
INFO [alembic.migration] Running upgrade 10cd28e692e9 -> 2db5203cb7a9,
nuage_floatingip
INFO [alembic.migration] Running upgrade 2db5203cb7a9 -> 5446f2a45467,
set_server_default
INFO [alembic.migration] Running upgrade 5446f2a45467 -> db_healing, Include
all tables and make migrations unconditional.
INFO [alembic.migration] Context impl MySQLImpl.
INFO [alembic.migration] Will assume non-transactional DDL.
INFO [alembic.autogenerate.compare] Detected server default on column
'cisco_ml2_apic_epgs.provider'
INFO [alembic.autogenerate.compare] Detected removed index
'cisco_n1kv_vlan_allocations_ibfk_1' on 'cisco_n1kv_vlan_allocations'
INFO [alembic.autogenerate.compare] Detected server
default on column 'cisco_n1kv_vxlan_allocations.allocated'
INFO [alembic.autogenerate.compare] Detected removed index
'cisco_n1kv_vxlan_allocations_ibfk_1' on 'cisco_n1kv_vxlan_allocations'
INFO [alembic.autogenerate.compare] Detected removed index
'embrane_pool_port_ibfk_2' on 'embrane_pool_port'
INFO [alembic.autogenerate.compare] Detected removed index
'firewall_rules_ibfk_1' on 'firewall_rules'
INFO [alembic.autogenerate.compare] Detected removed index 'firewalls_ibfk_1'
on 'firewalls'
INFO [alembic.autogenerate.compare] Detected server default on column
'meteringlabelrules.excluded'
INFO [alembic.autogenerate.compare] Detected server default on column
'ml2_port_bindings.host'
INFO [alembic.autogenerate.compare] Detected added column
'nuage_routerroutes_mapping.destination'
INFO [alembic.autogenerate.compare] Detected added column
'nuage_routerroutes_mapping.nexthop'
INFO
[alembic.autogenerate.compare] Detected server default
on column 'poolmonitorassociations.status'
INFO [alembic.autogenerate.compare] Detected added index 'ix_quotas_tenant_id'
on '['tenant_id']'
INFO [alembic.autogenerate.compare] Detected NULL on column
'tz_network_bindings.phy_uuid'
INFO [alembic.autogenerate.compare] Detected NULL on column
'tz_network_bindings.vlan_id'
INFO [neutron.db.migration.alembic_migrations.heal_script] Detected removed
foreign key u'nuage_floatingip_pool_mapping_ibfk_2' on table
u'nuage_floatingip_pool_mapping'
INFO [alembic.migration] Running upgrade db_healing -> 3927f7f7c456, L3
extension distributed mode
INFO [alembic.migration] Running upgrade 3927f7f7c456 -> 2026156eab2f, L2
models to support DVR
INFO [alembic.migration] Running upgrade 2026156eab2f -> 37f322991f59,
removing_mapping_tables
INFO [alembic.migration] Running upgrade 37f322991f59 -> 31d7f831a591, add
constraint for routerid
INFO [alembic.migration] Running upgrade
31d7f831a591 -> 5589aa32bf80, L3 scheduler additions to support DVR
INFO [alembic.migration] Running upgrade 5589aa32bf80 -> 884573acbf1c, Drop
NSX table in favor of the extra_attributes one
INFO [alembic.migration] Running upgrade 884573acbf1c -> 4eba2f05c2f4, correct
Vxlan Endpoint primary key
INFO [alembic.migration] Running upgrade 4eba2f05c2f4 -> 327ee5fde2c7,
set_innodb_engine
INFO [alembic.migration] Running upgrade 327ee5fde2c7 -> 3b85b693a95f, Drop
unused servicedefinitions and servicetypes tables.
INFO [alembic.migration] Running upgrade 3b85b693a95f -> aae5706a396,
nuage_provider_networks
INFO [alembic.migration] Running upgrade aae5706a396 -> 32f3915891fd,
cisco_apic_driver_update
INFO [alembic.migration] Running upgrade 32f3915891fd -> 58fe87a01143,
cisco_csr_routing
INFO [alembic.migration] Running upgrade 58fe87a01143 -> 236b90af57ab,
ml2_type_driver_refactor_dynamic_segments
INFO
[alembic.migration] Running upgrade 236b90af57ab -> 86d6d9776e2b, Cisco APIC
Mechanism Driver
INFO [alembic.migration] Running upgrade 86d6d9776e2b -> 16a27a58e093,
ext_l3_ha_mode
INFO [alembic.migration] Running upgrade 16a27a58e093 -> 3c346828361e,
metering_label_shared
INFO [alembic.migration] Running upgrade 3c346828361e -> 1680e1f0c4dc, Remove
Cisco Nexus Monolithic Plugin
INFO [alembic.migration] Running upgrade 1680e1f0c4dc -> 544673ac99ab, add
router port relationship
INFO [alembic.migration] Running upgrade 544673ac99ab -> juno, juno
root@controller2:/#
-----Original Message-----
From: Uwe Sauter [mailto:uwe.sauter.de<http://uwe.sauter.de>@gmail.com]
Sent: 29. júní 2015 20:45
To: Yngvi Páll Þorfinnsson; YANG LI
Cc: [email protected]<mailto:[email protected]>
Subject: Re: [Openstack] network question
on
openstack installation
Go to http://docs.openstack.org/ , select the OpenStack version, then the
installation guide for your distribution and navigate to
6. Add a networking component
- OpenStack Networking (neutron)
- Install and configure controller node
and follow the database related stuff:
- create DB
- grant privileges to neutron user
- populate the DB (search for "neutron-db-manage")
Depending on your distribution, you probably need to use sudo.
The number of config files depends on the file layout you are working with, so
this is not an exact answer.
I did the following:
[on controller node]
cd /etc/neutron
cat plugins/ml2/ml2_conf.ini > plugin.ini cd plugins/ml2 rm -f ml2_config.ini
ln -s ../../plugin.ini ml2_conf.ini
[on network and compute nodes]
cd
/etc/neutron
cat plugins/ml2/ml2_conf.ini > plugin.ini cat
plugins/openvswitch/ovs_neutron_plugin.ini >> plugin.ini cd plugins/ml2 rm -f
ml2_config.ini ln -s ../../plugin.ini ml2_conf.ini cd ../openvswitch rm -f
ovs_neutron_plugin.ini ln -s ../../plugin.ini ovs_neutron_plugin.ini
Then I did that sed trick that is needed on RDO because of a packaging bug with
Juno (changes systemd unit file to load /etc/neutron/plugin.ini instead of
/etc/neutron/plugins/ml2/ml2_plugin.ini).
And then 'su -s /bin/sh -c "neutron-db-manage --config-file
/etc/neutron/neutron.conf --config-file /etc/neutron/plugin.ini upgrade juno"
neutron' on the controller node.
As far as I understood the configuration system used in Neutron, each config
file will be merged into the same configuration namespace; only the order will
determine, which value is the current one if they exist in more then one file
[] section.
So you could have just one ini file in addition to /etc/neutron/neutron.conf
(my way)
or
several. But then you
need to load them all when sync'ing the DB *and* starting the services.
Hopefully this brightens the dark side of Neutron configuration,
Uwe
Am 29.06.2015 um 22:19 schrieb Yngvi Páll Þorfinnsson:
HI Uwe
No, I did'nt drop the keystone ;-)
But is this the correct way to resync neutron ?
# neutron-db-manage --config-file /etc/neutron/neutron.conf \ #
--config-file /etc/neutron/plugins/ml2/ml2_plugin.ini
I mean, how many config files is necessary to have in the cmd ?
best regards
Yngvi
-----Original Message-----
From: Uwe Sauter [mailto:uwe.sauter.de<http://uwe.sauter.de>@gmail.com]
Sent: 29. júní 2015 18:08
To: YANG LI
Cc: [email protected]<mailto:[email protected]>
Subject: Re:
[Openstack] network question on openstack installation
It depends on your switch… some drop tagged packets on an access port, others
allow tagged packets, if the packet VLAN ID equals the configured VLAN ID.
I'd reconfigure the provider network type to "flat" but that's
personal taste. You could also reconfigure the switch port to be a
trunking port (with only one VLAN ID). Currently your in between those
configurations that might or might not work lateron…
Regards,
Uwe
Am 29.06.2015 um 19:42 schrieb YANG LI:
thank you, Uwe. our provider network actually is untagged, but I did
specified VLAN ID when I create our external network and everything still
works. will this cause issue later on?
eutron net-create --provider:network_type=vlan
—provider:segmentation_id=<vlan id>
--provider:physical_network=physnet1
--router:external=true public
Thanks,
Yang
On Jun 27, 2015, at 11:17 AM, Uwe Sauter
<uwe.sauter.de<http://uwe.sauter.de>@gmail.com
<mailto:uwe.sauter.de<http://uwe.sauter.de>@gmail.com>> wrote:
Hi Yang,
it depends on whether your provider network is tagged or untagged. If it is
untagged (the switch port is an "access"
port) then you don't specify the VLAN ID for the external network
(as it will get tagged by the switch). If the provider network is
tagged (the switch port is a "trunk" port) then you have to configure the VLAN
ID because the switch might refuse the traffic (depending if there is a default
VLAN ID defined on the switch's port).
Regards,
Uwe
Am 27.06.2015 um 13:47 schrieb YANG LI:
Thank you so much, James! This is so helpful. Another confusion I
have is about network_vlan_ranges. Is this network VLAN id range?
If so, does it has to match external network? For example, we only have one
external VLAN we can use as Our provider network and that VLAN id is 775
(xxx.xxx.xxx.0/26). Should I define network_vlan_ranges as following?
[ml2]
type_drivers=vlan
tenant_network_types = vlan
mechanism_drivers=openvswitch
#
[ml2_type_vlan]
#thistellsOpenstackthattheinternalname"physnet1"providesthevlanrang
e
100-199
network_vlan_ranges=physnet1:775
#
Thanks,
Yang
Sent from my iPhone
On Jun 26, 2015, at 8:54 AM, "James Denton"
<[email protected]
<mailto:[email protected]<mailto:[email protected]%20%3cmailto:[email protected]>>
<mailto:[email protected]>> wrote:
You can absolutely have instances in the same network span
different compute nodes. As an admin, you can run ‘nova show <instanceid>’ and
see the host in the output:
root@controller01:~# nova show
7bb18175-87da-4d1f-8dca-2ef07fee9d21
| grep host
| OS-EXT-SRV-ATTR:host | compute02
|
That info is not available to non-admin users by default.
James
On Jun 26, 2015, at 7:38 AM, YANG LI <[email protected]
<mailto:[email protected]<mailto:[email protected]%20%0b%20%3cmailto:[email protected]>>
<mailto:[email protected]>>
wrote:
Thanks, James for the explanation. it make more sense now.
<http://now.it/> it is possible that a instances on same tenant network reside
on different compute nodes right? how do I tell which compute node a instance
is on?
Thanks,
Yang
On Jun 24, 2015, at 10:27 AM, James Denton
<[email protected]
<mailto:[email protected]<mailto:[email protected]%20%3cmailto:[email protected]>>
<mailto:[email protected]>> wrote:
Hello.
all three nodes will have eth0 on management/api network. since
I am using ml2 plugin with vlan for tenant network, I think all
compute node should have eth1 as the second nic on provider
network. Is this
correct? I understand provider network is for instance to get external access
to internet, but how is instance live on compute1 communicate with instance
live on compute2? are they also go through provider network?
In short, yes. If you’re connecting instances to vlan “provider”
networks, traffic between instances on different compute nodes
will traverse the “provider bridge”, get tagged out eth1, and
hit the physical switching fabric. Your external gateway device could also sit
in that vlan, and the default route on the instance would direct external
traffic to that device.
In reality, every network has ‘provider’ attributes that
describe the network type, segmentation id, and bridge interface
(for vlan/flat only). So tenant networks that leverage vlans
would have provider attributes set by Neutron automatically
based on the configuration set in the ML2 config file. If you
use Neutron routers that connect to both ‘tenant’ vlan-based networks and
external ‘provider’ networks, all of that traffic could traverse the same
provider bridge on the controller/network node, but would be tagged accordingly
based on the network (ie. vlan 100 for external network, vlan 200 for tenant
network).
Hope that’s not too confusing!
James
On Jun 24, 2015, at 8:54 AM, YANG LI <[email protected]
<mailto:[email protected]<mailto:[email protected]%20%0b%20%3cmailto:[email protected]>>
<mailto:[email protected]>>
wrote:
I am working on install openstack from scratch, but get
confused with network part. I want to have one controller node, two compute
nodes.
the controller node will only handle following services:
glance-api
glance-registry
keystone
nova-api
nova-cert
nova-conductor
nova-consoleauth
nova-novncproxy
nova-scheduler
qpid
mysql
neutron-server
compute nodes will have following services:
neutron-dhcp-agent
neutron-l3-agent
neutron-metadata-agent
neutron-openvswitch-agent
neutron-ovs-cleanup
openvswtich
nova-compute
all three nodes will have eth0 on management/api network. since
I am using ml2 plugin with vlan for tenant network, I think all
compute node should have eth1 as the second nic on provider
network. Is this correct? I understand provider network is for instance to
get external access to internet, but how is instance live on compute1
communicate with instance live on compute2? are they also go through provider
network?
________________________________
Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
Post to :
[email protected]<mailto:[email protected]>
<mailto:[email protected]>
<mailto:[email protected]>
Unsubscribe :
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
________________________________
Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
Post to :
[email protected]<mailto:[email protected]>
<mailto:[email protected]>
Unsubscribe :
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
________________________________
Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
Post to :
[email protected]<mailto:[email protected]>
<mailto:[email protected]>
Unsubscribe :
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
________________________________
Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
Post to :
[email protected]<mailto:[email protected]>
Unsubscribe :
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
--
Diese Nachricht wurde von meinem Android-Mobiltelefon mit K-9 Mail gesendet.
_______________________________________________
Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
Post to : [email protected]
Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack