Thanks Ransy, I have created a tar file with the configuration and log files in it. Please download it using the following URL. I have pasted the output of the commands below.
https://www.dropbox.com/s/qyfcsn50060y304/confilesnlogs.tar *Controller node:* *root@controller:~# keystone endpoint-list* +----------------------------------+-----------+-------------------------------------------------+-------------------------------------------------+--------------------------------------------+Controller node | id | region | publicurl | internalurl | adminurl | +----------------------------------+-----------+-------------------------------------------------+-------------------------------------------------+--------------------------------------------+ | 2c9a1cb0fe8247d9b7716432cf459fe5 | RegionOne | http://192.168.2.225:8774/v2/$(tenant_id)s | http://192.168.2.225:8774/v2/$(tenant_id)s | http://192.168.2.225:8774/v2/$(tenant_id)s | | 2d306903ed3342a8aaaac7c5680c116f | RegionOne | http://192.168.2.225:9696/ | http://192.168.2.225:9696/ | http://192.168.2.225:9696/ | | 3848114f120f42bf819bc2443b28ac9e | RegionOne | http://192.168.2.225:8080/v1/AUTH_$(tenant_id)s | http://192.168.2.225:8080/v1/AUTH_$(tenant_id)s | http://192.168.2.225:8080/v1 | | 4955173b8d9e4d33ae4a5b29dc12c74d | RegionOne | http://192.168.2.225:8776/v1/$(tenant_id)s | http://192.168.2.225:8776/v1/$(tenant_id)s | http://192.168.2.225:8776/v1/$(tenant_id)s | | d313aa76bf854dde94f33a49a9f0c8ac | RegionOne | http://192.168.2.225:9292/v2 | http://192.168.2.225:9292/v2 | http://192.168.2.225:9292/v2 | | e5aa4ecf3cbe4dd5aba9b204c74fee6a | RegionOne | http://192.168.2.225:5000/v2.0 | http://192.168.2.225:5000/v2.0 | http://192.168.2.225:35357/v2.0 | | fba6f790e3b444c890d114f13cd32b37 | RegionOne | http://192.168.2.225:8773/services/Cloud | http://192.168.2.225:8773/services/Cloud | http://192.168.2.225:8773/services/Admin | +----------------------------------+-----------+-------------------------------------------------+-------------------------------------------------+--------------------------------------------+ *root@controller:~# ip link show* 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 16436 qdisc noqueue state UNKNOWN link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP qlen 1000 link/ether d4:ae:52:bb:aa:20 brd ff:ff:ff:ff:ff:ff 3: eth1: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN qlen 1000 link/ether d4:ae:52:bb:aa:21 brd ff:ff:ff:ff:ff:ff 4: eth0.2@eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UP link/ether d4:ae:52:bb:aa:20 brd ff:ff:ff:ff:ff:ff 5: br-int: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UNKNOWN link/ether ba:7a:e9:dc:2b:41 brd ff:ff:ff:ff:ff:ff 7: br-ex: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UNKNOWN link/ether 9a:41:c8:8a:9e:49 brd ff:ff:ff:ff:ff:ff 8: tapf71b5b86-5c: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UNKNOWN link/ether 2a:44:a3:d1:7d:f3 brd ff:ff:ff:ff:ff:ff 9: qr-4d088f3a-78: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UNKNOWN link/ether ca:5b:8d:4d:6d:fb brd ff:ff:ff:ff:ff:ff 10: qg-6f8374cb-cb: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UNKNOWN link/ether 0e:7f:dd:3a:80:bc brd ff:ff:ff:ff:ff:ff 27: br-tun: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN link/ether 8a:cf:ec:7c:15:40 brd ff:ff:ff:ff:ff:ff *root@controller:~# ip route show* default via 192.168.2.253 dev eth0.2 default via 192.168.2.253 dev eth0.2 metric 100 10.5.5.0/24 dev tapf71b5b86-5c proto kernel scope link src 10.5.5.2 10.5.5.0/24 dev qr-4d088f3a-78 proto kernel scope link src 10.5.5.1 10.10.10.0/24 dev eth0 proto kernel scope link src 10.10.10.1 192.168.2.0/24 dev eth0.2 proto kernel scope link src 192.168.2.225 192.168.2.0/24 dev qg-6f8374cb-cb proto kernel scope link src 192.168.2.151 192.168.2.0/24 dev br-ex proto kernel scope link src 192.168.2.225 *$ ip netns show (Did not return anything)* *root@controller:~# ovs-vsctl show* a566afae-d7a8-42a9-aefe-8b0f2f7054a3 Bridge br-tun Port "gre-4" Interface "gre-4" type: gre options: {in_key=flow, out_key=flow, remote_ip="10.10.10.12"} Port "gre-3" Interface "gre-3" type: gre options: {in_key=flow, out_key=flow, remote_ip="127.0.0.1"} Port patch-int Interface patch-int type: patch options: {peer=patch-tun} Port br-tun Interface br-tun type: internal Port "gre-1" Interface "gre-1" type: gre options: {in_key=flow, out_key=flow, remote_ip="10.0.0.3"} Bridge br-ex Port br-ex Interface br-ex type: internal Port "qg-6f8374cb-cb" Interface "qg-6f8374cb-cb" type: internal Port "br0" Interface "br0" Bridge br-int Port br-int Interface br-int type: internal Port "tapf71b5b86-5c" tag: 1 Interface "tapf71b5b86-5c" type: internal Port patch-tun Interface patch-tun type: patch options: {peer=patch-int} Port "qr-4d088f3a-78" tag: 1 Interface "qr-4d088f3a-78" type: internal ovs_version: "1.4.0+build0" *Compute node:* *root@cronus:~# ip link show* 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 16436 qdisc noqueue state UNKNOWN link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP qlen 1000 link/ether d4:ae:52:bb:a1:9d brd ff:ff:ff:ff:ff:ff 3: eth1: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN qlen 1000 link/ether d4:ae:52:bb:a1:9e brd ff:ff:ff:ff:ff:ff 4: eth0.2@eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UP link/ether d4:ae:52:bb:a1:9d brd ff:ff:ff:ff:ff:ff 5: br-int: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN link/ether ae:9b:43:09:af:40 brd ff:ff:ff:ff:ff:ff 9: qbr256f5ed2-43: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UP link/ether c6:c0:df:64:c6:99 brd ff:ff:ff:ff:ff:ff 10: qvo256f5ed2-43: <BROADCAST,MULTICAST,PROMISC,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast state UP qlen 1000 link/ether 76:25:8b:fd:90:3b brd ff:ff:ff:ff:ff:ff 11: qvb256f5ed2-43: <BROADCAST,MULTICAST,PROMISC,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast master qbr256f5ed2-43 state UP qlen 1000 link/ether c6:c0:df:64:c6:99 brd ff:ff:ff:ff:ff:ff 13: br-tun: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN link/ether be:8c:30:78:35:48 brd ff:ff:ff:ff:ff:ff 15: vnet0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast master qbr256f5ed2-43 state UNKNOWN qlen 500 link/ether fe:16:3e:57:ec:ff brd ff:ff:ff:ff:ff:ff *root@cronus:~# ip route show* default via 192.168.2.253 dev eth0.2 metric 100 10.10.10.0/24 dev eth0 proto kernel scope link src 10.10.10.12 192.168.2.0/24 dev eth0.2 proto kernel scope link src 192.168.2.234 *root@cronus:~# ovs-vsctl show* d85bc334-6d64-4a13-b851-d56b18ff1549 Bridge br-int Port "qvo0e743b01-89" tag: 4095 Interface "qvo0e743b01-89" Port "qvo256f5ed2-43" tag: 1 Interface "qvo256f5ed2-43" Port patch-tun Interface patch-tun type: patch options: {peer=patch-int} Port br-int Interface br-int type: internal Port "qvoee3d4131-2a" tag: 4095 Interface "qvoee3d4131-2a" Port "qvocbc816bd-3d" tag: 4095 Interface "qvocbc816bd-3d" Bridge br-tun Port br-tun Interface br-tun type: internal Port "gre-2" Interface "gre-2" type: gre options: {in_key=flow, out_key=flow, remote_ip="10.10.10.1"} Port "gre-1" Interface "gre-1" type: gre options: {in_key=flow, out_key=flow, remote_ip="10.0.0.3"} Port patch-int Interface patch-int type: patch options: {peer=patch-tun} Port "gre-3" Interface "gre-3" type: gre options: {in_key=flow, out_key=flow, remote_ip="127.0.0.1"} ovs_version: "1.4.0+build0" Thanks I appreciate your help. On 16 February 2013 16:49, Jean-Baptiste RANSY < jean-baptiste.ra...@alyseo.com> wrote: > Please provide files listed bellow : > > Controller Node : > /etc/nova/nova.conf > /etc/nova/api-paste.ini > /etc/quantum/l3_agent.ini > /etc/quantum/quantum.conf > /etc/quantum/dhcp_agent.ini > /etc/quantum/plugins/openvswitch/ovs_quantum_plugin.ini > /etc/quantum/api-paste.ini > /var/log/nova/*.log > /var/log/quantum/*.log > > Compute Node : > /etc/nova/nova.conf > /etc/nova/nova-compute.conf > /etc/nova/api-paste.ini > /etc/quantum/quantum.conf > /etc/quantum/plugins/openvswitch/ovs_quantum_plugin.ini > /var/log/nova/*.log > /var/log/quantum/*.log > > Plus, complete output of the following commands : > > Controller Node : > $ keystone endpoint-list > $ ip link show > $ ip route show > $ ip netns show > $ ovs-vsctl show > > Compute Node : > $ ip link show > $ ip route show > $ ovs-vsctl show > > Regards, > > Jean-Baptiste RANSY > > > > On 02/16/2013 05:32 PM, Chathura M. Sarathchandra Magurawalage wrote: > > Hello Jean, > > Thanks for your reply. > > I followed the instructions in > http://docs.openstack.org/folsom/basic-install/content/basic-install_network.html. > And my Controller and the Network-node is installed in the same physical > node. > > I am using Folsom but without Network namespaces. > > But in the website you have provided it states that "If you run both L3 > + DHCP services on the same node, you should enable namespaces to avoid > conflicts with routes :" > > But currently quantum-dhcp-agent and quantum-l3-agent are running in the > same node? > > Additionally the control node serves as a DHCP server for the local > network ( Don't know if that would make and difference) > > Any idea what the problem could be? > > > On 16 February 2013 16:21, Jean-Baptiste RANSY < > jean-baptiste.ra...@alyseo.com> wrote: > >> Hello Chathura, >> >> Are you using Folsom with Network Namespaces ? >> >> If yes, have a look here : >> http://docs.openstack.org/folsom/openstack-network/admin/content/ch_limitations.html >> >> >> Regards, >> >> Jean-Baptsite RANSY >> >> >> >> On 02/16/2013 05:01 PM, Chathura M. Sarathchandra Magurawalage wrote: >> >> Hello guys, >> >> The problem still exists. Any ideas? >> >> Thanks >> >> On 15 February 2013 14:37, Sylvain Bauza <sylvain.ba...@digimind.com>wrote: >> >>> Metadata API allows to fetch SSH credentials when booting (pubkey I >>> mean). >>> If a VM is unable to reach metadata service, then it won't be able to >>> get its public key, so you won't be able to connect, unless you >>> specifically go thru a Password authentication (provided password auth is >>> enabled in /etc/ssh/sshd_config, which is not the case with Ubuntu cloud >>> archive). >>> There is also a side effect, the boot process is longer as the instance >>> is waiting for the curl timeout (60sec.) to finish booting up. >>> >>> Re: Quantum, the metadata API is actually DNAT'd from Network node to >>> the Nova-api node (here 172.16.0.1 as internal management IP) : >>> Chain quantum-l3-agent-PREROUTING (1 references) >>> >>> target prot opt source destination >>> DNAT tcp -- 0.0.0.0/0 169.254.169.254 tcp >>> dpt:80 to:172.16.0.1:8775 >>> >>> >>> Anyway, the first step is to : >>> 1. grab the console.log >>> 2. access thru VNC to the desired instance >>> >>> Troubleshooting will be easier once that done. >>> >>> -Sylvain >>> >>> >>> >>> Le 15/02/2013 14:24, Chathura M. Sarathchandra Magurawalage a écrit : >>> >>>> Hello Guys, >>>> >>>> Not sure if this is the right port but these are the results: >>>> >>>> *Compute node:* >>>> >>>> >>>> root@computenode:~# netstat -an | grep 8775 >>>> tcp 0 0 0.0.0.0:8775 <http://0.0.0.0:8775> 0.0.0.0:* >>>> LISTEN >>>> >>>> *Controller: * >>>> >>>> >>>> root@controller:~# netstat -an | grep 8775 >>>> tcp 0 0 0.0.0.0:8775 <http://0.0.0.0:8775> 0.0.0.0:* >>>> LISTEN >>>> >>>> *Additionally I cant curl 169.254.169.254 from the compute node. I am >>>> not sure if this is related to not being able to PING the VM.* >>>> >>>> >>>> curl -v http://169.254.169.254 >>>> * About to connect() to 169.254.169.254 port 80 (#0) >>>> * Trying 169.254.169.254... >>>> >>>> Thanks for your help >>>> >>>> >>>> >>>> ----------------------------------------------------------------------------------------------------------------------------- >>>> Chathura Madhusanka Sarathchandra Magurawalage. >>>> 1NW.2.1, Desk 2 >>>> School of Computer Science and Electronic Engineering >>>> University Of Essex >>>> United Kingdom. >>>> >>>> Email: csar...@essex.ac.uk <mailto:csar...@essex.ac.uk> >>>> chathura.sarathchan...@gmail.com <mailto: >>>> 77.chath...@gmail.com> >>>> 77.chath...@gmail.com <mailto:77.chath...@gmail.com> >>>> >>>> >>>> >>>> On 15 February 2013 11:03, Anil Vishnoi <vishnoia...@gmail.com <mailto: >>>> vishnoia...@gmail.com>> wrote: >>>> >>>> If you are using ubuntu cloud image then the only way to log-in is >>>> to do ssh with the public key. For that you have to create ssh key >>>> pair and download the ssh key. You can create this ssh pair using >>>> horizon/cli. >>>> >>>> >>>> On Fri, Feb 15, 2013 at 4:27 PM, Sylvain Bauza >>>> <sylvain.ba...@digimind.com <mailto:sylvain.ba...@digimind.com>> >>>> >>>> wrote: >>>> >>>> >>>> Le 15/02/2013 11:42, Chathura M. Sarathchandra Magurawalage a >>>> écrit : >>>> >>>> >>>> How can I log into the VM from VNC? What are the >>>> credentials? >>>> >>>> >>>> You have multiple ways to get VNC access. The easiest one is >>>> thru Horizon. Other can be looking at the KVM command-line for >>>> the desired instance (on the compute node) and check the vnc >>>> port in use (assuming KVM as hypervisor). >>>> This is basic knowledge of Nova. >>>> >>>> >>>> >>>> nova-api-metadata is running fine in the compute node. >>>> >>>> >>>> Make sure the metadata port is avaible thanks to telnet or >>>> netstat, nova-api can be running without listening on metadata >>>> port. >>>> >>>> >>>> >>>> >>>> _______________________________________________ >>>> Mailing list: https://launchpad.net/~openstack >>>> <https://launchpad.net/%7Eopenstack> >>>> Post to : openstack@lists.launchpad.net >>>> <mailto:openstack@lists.launchpad.net> >>>> Unsubscribe : https://launchpad.net/~openstack >>>> <https://launchpad.net/%7Eopenstack> >>>> >>>> More help : https://help.launchpad.net/ListHelp >>>> >>>> >>>> >>>> >>>> -- Thanks & Regards >>>> --Anil Kumar Vishnoi >>>> >>>> >>>> >>> >> >> >> _______________________________________________ >> Mailing list: https://launchpad.net/~openstack >> >> Post to : openstack@lists.launchpad.net >> Unsubscribe : https://launchpad.net/~openstack >> More help : https://help.launchpad.net/ListHelp >> >> >> > >
_______________________________________________ Mailing list: https://launchpad.net/~openstack Post to : openstack@lists.launchpad.net Unsubscribe : https://launchpad.net/~openstack More help : https://help.launchpad.net/ListHelp