Re: [Openstack] error while creating instance on nfs share
I am using Folsom release. I have followed the instruction on that page. On 9 July 2013 10:57, JuanFra Rodriguez Cardoso < juanfra.rodriguez.card...@gmail.com> wrote: > Diablo release? I'd recommend you use the lastest release for live > migrations: > > > http://docs.openstack.org/trunk/openstack-compute/admin/content/configuring-live-migrations.html > > Regards, > --- > JuanFra > > > 2013/7/8 Chathura M. Sarathchandra Magurawalage <77.chath...@gmail.com> > >> Hello al, >> >> I followed the Openstack instructions for enabling vm migration ( >> http://docs.openstack.org/diablo/openstack-compute/admin/content/configuring-live-migrations.html >> ) >> >> But I can not get it working, since when I create an instance I get the >> following error on the compute node. >> >> 2013-07-08 17:01:34 9329 ERROR nova.openstack.common.rpc.amqp [-] >> Exception during message handling >> 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp Traceback >> (most recent call last): >> 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp File >> "/usr/lib/python2.7/dist-packages/nova/openstack/common/rpc/amqp.py", line >> 276, in _process_data >> 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp rval = >> self.proxy.dispatch(ctxt, version, method, **args) >> 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp File >> "/usr/lib/python2.7/dist-packages/nova/openstack/common/rpc/dispatcher.py", >> line 145, in dispatch >> 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp return >> getattr(proxyobj, method)(ctxt, **kwargs) >> 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp File >> "/usr/lib/python2.7/dist-packages/nova/exception.py", line 117, in wrapped >> 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp >> temp_level, payload) >> 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp File >> "/usr/lib/python2.7/contextlib.py", line 24, in __exit__ >> 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp >> self.gen.next() >> 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp File >> "/usr/lib/python2.7/dist-packages/nova/exception.py", line 92, in wrapped >> 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp return >> f(*args, **kw) >> 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp File >> "/usr/lib/python2.7/dist-packages/nova/compute/manager.py", line 176, in >> decorated_function >> 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp pass >> 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp File >> "/usr/lib/python2.7/contextlib.py", line 24, in __exit__ >> 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp >> self.gen.next() >> 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp File >> "/usr/lib/python2.7/dist-packages/nova/compute/manager.py", line 162, in >> decorated_function >> 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp return >> function(self, context, *args, **kwargs) >> 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp File >> "/usr/lib/python2.7/dist-packages/nova/compute/manager.py", line 197, in >> decorated_function >> 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp >> kwargs['instance']['uuid'], e, sys.exc_info()) >> 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp File >> "/usr/lib/python2.7/contextlib.py", line 24, in __exit__ >> 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp >> self.gen.next() >> 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp File >> "/usr/lib/python2.7/dist-packages/nova/compute/manager.py", line 191, in >> decorated_function >> 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp return >> function(self, context, *args, **kwargs) >> 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp File >> "/usr/lib/python2.7/dist-packages/nova/compute/manager.py", line 839, in >> run_instance >> 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp >> do_run_instance() >> 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp File >> "/usr/lib/python2.7/dist-packages/nova/utils.py", line 803, in inner >> 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp retval >> = f(*args, **kwargs) >> 2013-07-08 17:01:34
[Openstack] error while creating instance on nfs share
Hello al, I followed the Openstack instructions for enabling vm migration ( http://docs.openstack.org/diablo/openstack-compute/admin/content/configuring-live-migrations.html ) But I can not get it working, since when I create an instance I get the following error on the compute node. 2013-07-08 17:01:34 9329 ERROR nova.openstack.common.rpc.amqp [-] Exception during message handling 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp Traceback (most recent call last): 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp File "/usr/lib/python2.7/dist-packages/nova/openstack/common/rpc/amqp.py", line 276, in _process_data 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp rval = self.proxy.dispatch(ctxt, version, method, **args) 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp File "/usr/lib/python2.7/dist-packages/nova/openstack/common/rpc/dispatcher.py", line 145, in dispatch 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp return getattr(proxyobj, method)(ctxt, **kwargs) 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp File "/usr/lib/python2.7/dist-packages/nova/exception.py", line 117, in wrapped 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp temp_level, payload) 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp File "/usr/lib/python2.7/contextlib.py", line 24, in __exit__ 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp self.gen.next() 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp File "/usr/lib/python2.7/dist-packages/nova/exception.py", line 92, in wrapped 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp return f(*args, **kw) 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp File "/usr/lib/python2.7/dist-packages/nova/compute/manager.py", line 176, in decorated_function 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp pass 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp File "/usr/lib/python2.7/contextlib.py", line 24, in __exit__ 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp self.gen.next() 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp File "/usr/lib/python2.7/dist-packages/nova/compute/manager.py", line 162, in decorated_function 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp return function(self, context, *args, **kwargs) 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp File "/usr/lib/python2.7/dist-packages/nova/compute/manager.py", line 197, in decorated_function 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp kwargs['instance']['uuid'], e, sys.exc_info()) 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp File "/usr/lib/python2.7/contextlib.py", line 24, in __exit__ 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp self.gen.next() 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp File "/usr/lib/python2.7/dist-packages/nova/compute/manager.py", line 191, in decorated_function 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp return function(self, context, *args, **kwargs) 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp File "/usr/lib/python2.7/dist-packages/nova/compute/manager.py", line 839, in run_instance 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp do_run_instance() 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp File "/usr/lib/python2.7/dist-packages/nova/utils.py", line 803, in inner 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp retval = f(*args, **kwargs) 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp File "/usr/lib/python2.7/dist-packages/nova/compute/manager.py", line 838, in do_run_instance 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp admin_password, is_first_time, instance) 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp File "/usr/lib/python2.7/dist-packages/nova/compute/manager.py", line 529, in _run_instance 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp self._set_instance_error_state(context, instance['uuid']) 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp File "/usr/lib/python2.7/contextlib.py", line 24, in __exit__ 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp self.gen.next() 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp File "/usr/lib/python2.7/dist-packages/nova/compute/manager.py", line 517, in _run_instance 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp is_first_time, request_spec, filter_properties) 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp File "/usr/lib/python2.7/dist-packages/nova/compute/manager.py", line 503, in _run_instance 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp injected_files, admin_password) 2013-07-08 17:01:34 9329 TRACE nova.openstack.common.rpc.amqp File "/usr/lib/python2.7/dist
Re: [Openstack] wget: can't connect to remote host (169.254.169.254): No route to host, but can ping and ssh
Now I have spawned a Ubuntu VM and I can not ssh because it does not have the security group and the public keys injected, but I can still ping from the network node. On 24 May 2013 14:28, Chathura M. Sarathchandra Magurawalage < 77.chath...@gmail.com> wrote: > Hello everyone, > > I have am having a problem with my metadata service. My VM's log is as > follows. As you may see, it can not find the route to the metadata service > (169.254.169.254). But I can ssh and ping the VM's from the network node. > Although as far as I know, you need to have metadata imported to the > virtual machines, to be able to ping to them and to ssh in to them. > > I cant ping or ssh in to the VMs from the controller or compute node. > > Anyone any help? > > > [0.018022] Initializing cgroup subsys cpuacct > [0.020014] Initializing cgroup subsys memory > [0.021322] Initializing cgroup subsys devices > [0.022629] Initializing cgroup subsys freezer > [0.024007] Initializing cgroup subsys net_cls > [0.025310] Initializing cgroup subsys blkio > [0.026587] Initializing cgroup subsys perf_event > [0.028143] mce: CPU supports 10 MCE banks > [0.029686] SMP alternatives: switching to UP code > [0.132200] Freeing SMP alternatives: 24k freed > [0.133648] ACPI: Core revision 20110413 > [0.135949] ftrace: allocating 26075 entries in 103 pages > [0.144172] ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 > [0.145806] CPU0: Intel(R) Xeon(R) CPU X5355 @ 2.66GHz stepping > 07 > [0.148008] Performance Events: unsupported p6 CPU model 15 no PMU driver, > software events only. > [0.148008] Brought up 1 CPUs > [0.148012] Total of 1 processors activated (5320.05 BogoMIPS). > [0.150609] devtmpfs: initialized > [0.154178] print_constraints: dummy: > [0.155400] Time: 13:02:05 Date: 05/24/13 > [0.156087] NET: Registered protocol family 16 > [0.157512] ACPI: bus type pci registered > [0.158819] PCI: Using configuration type 1 for base access > [0.160810] bio: create slab at 0 > [0.165652] ACPI: Interpreter enabled > [0.166763] ACPI: (supports S0 S3 S4 S5) > [0.168414] ACPI: Using IOAPIC for interrupt routing > [0.173485] ACPI: No dock devices found. > [0.174646] HEST: Table not found. > [0.175698] PCI: Ignoring host bridge windows from ACPI; if necessary, use > "pci=use_crs" and report a bug > [0.176031] ACPI: PCI Root Bridge [PCI0] (domain [bus 00-ff]) > [0.183300] pci :00:01.3: quirk: [io 0xb000-0xb03f] claimed by PIIX4 > ACPI > [0.184026] pci :00:01.3: quirk: [io 0xb100-0xb10f] claimed by PIIX4 > SMB > [0.201264] pci:00: Unable to request _OSC control (_OSC support > mask: 0x1e) > [0.207189] ACPI: PCI Interrupt Link [LNKA] (IRQs 5 *10 11) > [0.209145] ACPI: PCI Interrupt Link [LNKB] (IRQs 5 *10 11) > [0.211331] ACPI: PCI Interrupt Link [LNKC] (IRQs 5 10 *11) > [0.213126] ACPI: PCI Interrupt Link [LNKD] (IRQs 5 10 *11) > [0.215291] ACPI: PCI Interrupt Link [LNKS] (IRQs 9) *0 > [0.217204] vgaarb: device added: > PCI::00:02.0,decodes=io+mem,owns=io+mem,locks=none > [0.220016] vgaarb: loaded > [0.220931] vgaarb: bridge control possible :00:02.0 > [0.222716] SCSI subsystem initialized > [0.224363] usbcore: registered new interface driver usbfs > [0.225873] usbcore: registered new interface driver hub > [0.227381] usbcore: registered new device driver usb > [0.228226] PCI: Using ACPI for IRQ routing > [0.229889] NetLabel: Initializing > [0.232021] NetLabel: domain hash size = 128 > [0.233289] NetLabel: protocols = UNLABELED CIPSOv4 > [0.234683] NetLabel: unlabeled traffic allowed by default > [0.236101] HPET: 3 timers in total, 0 timers will be used for per-cpu > timer > [0.237936] hpet0: at MMIO 0xfed0, IRQs 2, 8, 0 > [0.240017] hpet0: 3 comparators, 64-bit 100.00 MHz counter > [0.248162] Switching to clocksource kvm-clock > [0.250575] Switched to NOHz mode on CPU #0 > [0.260176] AppArmor: AppArmor Filesystem Enabled > [0.261587] pnp: PnP ACPI init > [0.262592] ACPI: bus type pnp registered > [0.264716] pnp: PnP ACPI: found 8 devices > [0.265919] ACPI: ACPI bus type pnp unregistered > [0.273657] NET: Registered protocol family 2 > [0.275023] IP route cache hash table entries: 4096 (order: 3, 32768 bytes) > [0.277174] TCP established hash table entries: 16384 (order: 6, 262144 > bytes) > [0.279627] TCP bind hash table entries: 16384 (order: 6, 262144 bytes) > [0.281735] TCP: Hash tables configured (established 16384 bind 16384) > [
[Openstack] wget: can't connect to remote host (169.254.169.254): No route to host, but can ping and ssh
Hello everyone, I have am having a problem with my metadata service. My VM's log is as follows. As you may see, it can not find the route to the metadata service (169.254.169.254). But I can ssh and ping the VM's from the network node. Although as far as I know, you need to have metadata imported to the virtual machines, to be able to ping to them and to ssh in to them. I cant ping or ssh in to the VMs from the controller or compute node. Anyone any help? [0.018022] Initializing cgroup subsys cpuacct [0.020014] Initializing cgroup subsys memory [0.021322] Initializing cgroup subsys devices [0.022629] Initializing cgroup subsys freezer [0.024007] Initializing cgroup subsys net_cls [0.025310] Initializing cgroup subsys blkio [0.026587] Initializing cgroup subsys perf_event [0.028143] mce: CPU supports 10 MCE banks [0.029686] SMP alternatives: switching to UP code [0.132200] Freeing SMP alternatives: 24k freed [0.133648] ACPI: Core revision 20110413 [0.135949] ftrace: allocating 26075 entries in 103 pages [0.144172] ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 [0.145806] CPU0: Intel(R) Xeon(R) CPU X5355 @ 2.66GHz stepping 07 [0.148008] Performance Events: unsupported p6 CPU model 15 no PMU driver, software events only. [0.148008] Brought up 1 CPUs [0.148012] Total of 1 processors activated (5320.05 BogoMIPS). [0.150609] devtmpfs: initialized [0.154178] print_constraints: dummy: [0.155400] Time: 13:02:05 Date: 05/24/13 [0.156087] NET: Registered protocol family 16 [0.157512] ACPI: bus type pci registered [0.158819] PCI: Using configuration type 1 for base access [0.160810] bio: create slab at 0 [0.165652] ACPI: Interpreter enabled [0.166763] ACPI: (supports S0 S3 S4 S5) [0.168414] ACPI: Using IOAPIC for interrupt routing [0.173485] ACPI: No dock devices found. [0.174646] HEST: Table not found. [0.175698] PCI: Ignoring host bridge windows from ACPI; if necessary, use "pci=use_crs" and report a bug [0.176031] ACPI: PCI Root Bridge [PCI0] (domain [bus 00-ff]) [0.183300] pci :00:01.3: quirk: [io 0xb000-0xb03f] claimed by PIIX4 ACPI [0.184026] pci :00:01.3: quirk: [io 0xb100-0xb10f] claimed by PIIX4 SMB [0.201264] pci:00: Unable to request _OSC control (_OSC support mask: 0x1e) [0.207189] ACPI: PCI Interrupt Link [LNKA] (IRQs 5 *10 11) [0.209145] ACPI: PCI Interrupt Link [LNKB] (IRQs 5 *10 11) [0.211331] ACPI: PCI Interrupt Link [LNKC] (IRQs 5 10 *11) [0.213126] ACPI: PCI Interrupt Link [LNKD] (IRQs 5 10 *11) [0.215291] ACPI: PCI Interrupt Link [LNKS] (IRQs 9) *0 [0.217204] vgaarb: device added: PCI::00:02.0,decodes=io+mem,owns=io+mem,locks=none [0.220016] vgaarb: loaded [0.220931] vgaarb: bridge control possible :00:02.0 [0.222716] SCSI subsystem initialized [0.224363] usbcore: registered new interface driver usbfs [0.225873] usbcore: registered new interface driver hub [0.227381] usbcore: registered new device driver usb [0.228226] PCI: Using ACPI for IRQ routing [0.229889] NetLabel: Initializing [0.232021] NetLabel: domain hash size = 128 [0.233289] NetLabel: protocols = UNLABELED CIPSOv4 [0.234683] NetLabel: unlabeled traffic allowed by default [0.236101] HPET: 3 timers in total, 0 timers will be used for per-cpu timer [0.237936] hpet0: at MMIO 0xfed0, IRQs 2, 8, 0 [0.240017] hpet0: 3 comparators, 64-bit 100.00 MHz counter [0.248162] Switching to clocksource kvm-clock [0.250575] Switched to NOHz mode on CPU #0 [0.260176] AppArmor: AppArmor Filesystem Enabled [0.261587] pnp: PnP ACPI init [0.262592] ACPI: bus type pnp registered [0.264716] pnp: PnP ACPI: found 8 devices [0.265919] ACPI: ACPI bus type pnp unregistered [0.273657] NET: Registered protocol family 2 [0.275023] IP route cache hash table entries: 4096 (order: 3, 32768 bytes) [0.277174] TCP established hash table entries: 16384 (order: 6, 262144 bytes) [0.279627] TCP bind hash table entries: 16384 (order: 6, 262144 bytes) [0.281735] TCP: Hash tables configured (established 16384 bind 16384) [0.283454] TCP reno registered [0.284478] UDP hash table entries: 256 (order: 1, 8192 bytes) [0.286046] UDP-Lite hash table entries: 256 (order: 1, 8192 bytes) [0.287800] NET: Registered protocol family 1 [0.289098] pci :00:00.0: Limiting direct PCI/PCI transfers [0.290692] pci :00:01.0: PIIX3: Enabling Passive Release [0.292282] pci :00:01.0: Activating ISA DMA hang workarounds [0.294366] audit: initializing netlink socket (disabled) [0.295853] type=2000 audit(1369400527.292:1): initialized [0.315902] Trying to unpack rootfs image as initramfs... [0.332248] HugeTLB registered 2 MB page size, pre-allocated 0 pages [0.340281] VFS: Disk quotas dquot_6.5.2 [0.341541] Dquot-cache hash
[Openstack] not able to create VMs with member user but only with admin
I have just done a fresh multi node openstack installation. I can not create VMs as a Member user. The VM end up getting ERROR state with the following error. u'message': u'ProcessExecutionError', u'code': 500, u'created': u'2013-05-19T10:56:58Z' Any help would be appreciated. ___ Mailing list: https://launchpad.net/~openstack Post to : openstack@lists.launchpad.net Unsubscribe : https://launchpad.net/~openstack More help : https://help.launchpad.net/ListHelp
Re: [Openstack] Allocating dynamic IP to the VMs
any idea, anyone? On 11 May 2013 10:37, Chathura M. Sarathchandra Magurawalage < 77.chath...@gmail.com> wrote: > Hello Sylvain, > > I am sorry I got caught up with another project in the past few weeks, > hence my late reply. Please bare with me. > > The floating ip is bound to the qg-550803ee-ce bridge. > > root@controller:~# ip a > 1: lo: mtu 16436 qdisc noqueue state UNKNOWN > link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 > inet 127.0.0.1/8 scope host lo > inet6 ::1/128 scope host >valid_lft forever preferred_lft forever > 2: eth0: mtu 1500 qdisc mq state UP qlen > 1000 > link/ether d4:ae:52:bb:aa:20 brd ff:ff:ff:ff:ff:ff > inet 192.168.2.225/24 brd 192.168.2.255 scope global eth0 > inet6 fe80::d6ae:52ff:febb:aa20/64 scope link >valid_lft forever preferred_lft forever > 3: eth1: mtu 1500 qdisc mq state UP qlen > 1000 > link/ether d4:ae:52:bb:aa:21 brd ff:ff:ff:ff:ff:ff > inet 10.10.10.1/24 brd 10.10.10.255 scope global eth1 > inet6 fe80::d6ae:52ff:febb:aa21/64 scope link >valid_lft forever preferred_lft forever > 31: br-int: mtu 1500 qdisc noop state DOWN > link/ether e6:fe:c6:5e:73:47 brd ff:ff:ff:ff:ff:ff > 32: br-ex: mtu 1500 qdisc noqueue state > UNKNOWN > link/ether f6:26:0f:0d:32:45 brd ff:ff:ff:ff:ff:ff > inet6 fe80::f426:fff:fe0d:3245/64 scope link >valid_lft forever preferred_lft forever > 33: tapd648cfe0-f6: mtu 1500 qdisc > noqueue state UNKNOWN > link/ether fa:16:3e:9a:2c:29 brd ff:ff:ff:ff:ff:ff > inet 10.5.5.2/24 brd 10.5.5.255 scope global tapd648cfe0-f6 > inet6 fe80::f816:3eff:fe9a:2c29/64 scope link >valid_lft forever preferred_lft forever > 34: br-tun: mtu 1500 qdisc noop state DOWN > link/ether 5e:2b:f9:ca:c6:40 brd ff:ff:ff:ff:ff:ff > 35: qr-9e818b07-92: mtu 1500 qdisc > noqueue state UNKNOWN > link/ether fa:16:3e:01:48:1d brd ff:ff:ff:ff:ff:ff > inet 10.5.5.1/24 brd 10.5.5.255 scope global qr-9e818b07-92 > inet6 fe80::f816:3eff:fe01:481d/64 scope link >valid_lft forever preferred_lft forever > 36: qg-550803ee-ce: mtu 1500 qdisc > noqueue state UNKNOWN > link/ether fa:16:3e:fc:87:1c brd ff:ff:ff:ff:ff:ff > inet 192.168.2.151/24 brd 192.168.2.255 scope global qg-550803ee-ce > inet 192.168.2.152/32 brd 192.168.2.152 scope global qg-550803ee-ce > inet6 fe80::f816:3eff:fefc:871c/64 scope link >valid_lft forever preferred_lft forever > > But I can not ping the floating ip. > > root@controller:~# quantum net-list -- --router:external True > > +-+--+---+ > | id | name | > subnets | > > ++---+---+ > | a83c3409-6c79-4bb7-9557-010f3b56024f | ext_net | > fb3439f4-2afa-4cdc-86a6-aaee2ce1a3a3 | > > ++---+---+ > > root@controller:~# quantum floatingip-list > > +---+--+++ > | id | > fixed_ip_address | floating_ip_address | port_id >| > > ++-+++ > | 46311a66-5793-43af-be74-612580e505ca | 10.5.5.3 | > 192.168.2.152 | 529f6c56-8037-489a-a120-b97675d1745f | > +---------- > -+-----+++ > > Any idea? > > > On 25 March 2013 16:09, Sylvain Bauza wrote: > >> Le 25/03/2013 16:00, Chathura M. Sarathchandra Magurawalage a écrit : >> >> >> I can not see anything going through qg- interface. >> >> >> Your iptables seem correct. >> Could you please "ip a" and make sure floating ip is bound to qg- ? >> >> Are you sure that floating IPs are working properly on your setup ? >> >> -Sylvain >> > > ___ Mailing list: https://launchpad.net/~openstack Post to : openstack@lists.launchpad.net Unsubscribe : https://launchpad.net/~openstack More help : https://help.launchpad.net/ListHelp
Re: [Openstack] Allocating dynamic IP to the VMs
Hello Sylvain, I am sorry I got caught up with another project in the past few weeks, hence my late reply. Please bare with me. The floating ip is bound to the qg-550803ee-ce bridge. root@controller:~# ip a 1: lo: mtu 16436 qdisc noqueue state UNKNOWN link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 inet 127.0.0.1/8 scope host lo inet6 ::1/128 scope host valid_lft forever preferred_lft forever 2: eth0: mtu 1500 qdisc mq state UP qlen 1000 link/ether d4:ae:52:bb:aa:20 brd ff:ff:ff:ff:ff:ff inet 192.168.2.225/24 brd 192.168.2.255 scope global eth0 inet6 fe80::d6ae:52ff:febb:aa20/64 scope link valid_lft forever preferred_lft forever 3: eth1: mtu 1500 qdisc mq state UP qlen 1000 link/ether d4:ae:52:bb:aa:21 brd ff:ff:ff:ff:ff:ff inet 10.10.10.1/24 brd 10.10.10.255 scope global eth1 inet6 fe80::d6ae:52ff:febb:aa21/64 scope link valid_lft forever preferred_lft forever 31: br-int: mtu 1500 qdisc noop state DOWN link/ether e6:fe:c6:5e:73:47 brd ff:ff:ff:ff:ff:ff 32: br-ex: mtu 1500 qdisc noqueue state UNKNOWN link/ether f6:26:0f:0d:32:45 brd ff:ff:ff:ff:ff:ff inet6 fe80::f426:fff:fe0d:3245/64 scope link valid_lft forever preferred_lft forever 33: tapd648cfe0-f6: mtu 1500 qdisc noqueue state UNKNOWN link/ether fa:16:3e:9a:2c:29 brd ff:ff:ff:ff:ff:ff inet 10.5.5.2/24 brd 10.5.5.255 scope global tapd648cfe0-f6 inet6 fe80::f816:3eff:fe9a:2c29/64 scope link valid_lft forever preferred_lft forever 34: br-tun: mtu 1500 qdisc noop state DOWN link/ether 5e:2b:f9:ca:c6:40 brd ff:ff:ff:ff:ff:ff 35: qr-9e818b07-92: mtu 1500 qdisc noqueue state UNKNOWN link/ether fa:16:3e:01:48:1d brd ff:ff:ff:ff:ff:ff inet 10.5.5.1/24 brd 10.5.5.255 scope global qr-9e818b07-92 inet6 fe80::f816:3eff:fe01:481d/64 scope link valid_lft forever preferred_lft forever 36: qg-550803ee-ce: mtu 1500 qdisc noqueue state UNKNOWN link/ether fa:16:3e:fc:87:1c brd ff:ff:ff:ff:ff:ff inet 192.168.2.151/24 brd 192.168.2.255 scope global qg-550803ee-ce inet 192.168.2.152/32 brd 192.168.2.152 scope global qg-550803ee-ce inet6 fe80::f816:3eff:fefc:871c/64 scope link valid_lft forever preferred_lft forever But I can not ping the floating ip. root@controller:~# quantum net-list -- --router:external True +-+--+---+ | id | name | subnets | ++---+---+ | a83c3409-6c79-4bb7-9557-010f3b56024f | ext_net | fb3439f4-2afa-4cdc-86a6-aaee2ce1a3a3 | ++---+---+ root@controller:~# quantum floatingip-list +---+--+++ | id | fixed_ip_address | floating_ip_address | port_id | ++-+++ | 46311a66-5793-43af-be74-612580e505ca | 10.5.5.3 | 192.168.2.152 | 529f6c56-8037-489a-a120-b97675d1745f | +-- -+-+++ Any idea? On 25 March 2013 16:09, Sylvain Bauza wrote: > Le 25/03/2013 16:00, Chathura M. Sarathchandra Magurawalage a écrit : > > > I can not see anything going through qg- interface. > > > Your iptables seem correct. > Could you please "ip a" and make sure floating ip is bound to qg- ? > > Are you sure that floating IPs are working properly on your setup ? > > -Sylvain > ___ Mailing list: https://launchpad.net/~openstack Post to : openstack@lists.launchpad.net Unsubscribe : https://launchpad.net/~openstack More help : https://help.launchpad.net/ListHelp
[Openstack] novnc not working (No such RPC function validate_console_port ) - fix
Hello, I get "Failed to connect to server (code: 1006)" error message on the dashboard when I try to view the VNC panel. I get the following error on the in the controller. /var/log/nova/nova-consoleauth.log 2013-03-26 16:55:49 16471 ERROR nova.openstack.common.rpc.amqp [-] Exception during message handling 2013-03-26 16:55:49 16471 TRACE nova.openstack.common.rpc.amqp Traceback (most recent call last): 2013-03-26 16:55:49 16471 TRACE nova.openstack.common.rpc.amqp File "/usr/lib/python2.7/dist-packages/nova/openstack/common/rpc/amqp.py", line 276, in _process_data 2013-03-26 16:55:49 16471 TRACE nova.openstack.common.rpc.amqp rval = self.proxy.dispatch(ctxt, version, method, **args) 2013-03-26 16:55:49 16471 TRACE nova.openstack.common.rpc.amqp File "/usr/lib/python2.7/dist-packages/nova/openstack/common/rpc/dispatcher.py", line 145, in dispatch 2013-03-26 16:55:49 16471 TRACE nova.openstack.common.rpc.amqp return getattr(proxyobj, method)(ctxt, **kwargs) 2013-03-26 16:55:49 16471 TRACE nova.openstack.common.rpc.amqp File "/usr/lib/python2.7/dist-packages/nova/consoleauth/manager.py", line 107, in check_token 2013-03-26 16:55:49 16471 TRACE nova.openstack.common.rpc.amqp if self._validate_token(context, token): 2013-03-26 16:55:49 16471 TRACE nova.openstack.common.rpc.amqp File "/usr/lib/python2.7/dist-packages/nova/consoleauth/manager.py", line 99, in _validate_token 2013-03-26 16:55:49 16471 TRACE nova.openstack.common.rpc.amqp token['console_type']) 2013-03-26 16:55:49 16471 TRACE nova.openstack.common.rpc.amqp File "/usr/lib/python2.7/dist-packages/nova/compute/rpcapi.py", line 267, in validate_console_port 2013-03-26 16:55:49 16471 TRACE nova.openstack.common.rpc.amqp None, instance)) 2013-03-26 16:55:49 16471 TRACE nova.openstack.common.rpc.amqp File "/usr/lib/python2.7/dist-packages/nova/openstack/common/rpc/proxy.py", line 80, in call 2013-03-26 16:55:49 16471 TRACE nova.openstack.common.rpc.amqp return rpc.call(context, self._get_topic(topic), msg, timeout) 2013-03-26 16:55:49 16471 TRACE nova.openstack.common.rpc.amqp File "/usr/lib/python2.7/dist-packages/nova/openstack/common/rpc/__init__.py", line 108, in call 2013-03-26 16:55:49 16471 TRACE nova.openstack.common.rpc.amqp return _get_impl().call(cfg.CONF, context, topic, msg, timeout) 2013-03-26 16:55:49 16471 TRACE nova.openstack.common.rpc.amqp File "/usr/lib/python2.7/dist-packages/nova/openstack/common/rpc/impl_kombu.py", line 718, in call 2013-03-26 16:55:49 16471 TRACE nova.openstack.common.rpc.amqp rpc_amqp.get_connection_pool(conf, Connection)) 2013-03-26 16:55:49 16471 TRACE nova.openstack.common.rpc.amqp File "/usr/lib/python2.7/dist-packages/nova/openstack/common/rpc/amqp.py", line 369, in call 2013-03-26 16:55:49 16471 TRACE nova.openstack.common.rpc.amqp rv = list(rv) 2013-03-26 16:55:49 16471 TRACE nova.openstack.common.rpc.amqp File "/usr/lib/python2.7/dist-packages/nova/openstack/common/rpc/amqp.py", line 337, in __iter__ 2013-03-26 16:55:49 16471 TRACE nova.openstack.common.rpc.amqp raise result 2013-03-26 16:55:49 16471 TRACE nova.openstack.common.rpc.amqp RemoteError: Remote error: AttributeError No such RPC function 'validate_console_port' 2013-03-26 16:55:49 16471 TRACE nova.openstack.common.rpc.amqp [u'Traceback (most recent call last):\n', u' File "/usr/lib/python2.7/dist-packages/nova/openstack/common/rpc/amqp.p y", line 276, in _process_data\nrval = self.proxy.dispatch(ctxt, version, method, **args)\n', u' File "/usr/lib/python2.7/dist-packages/nova/openstack/common/rpc/dispatcher.py ", line 148, in dispatch\nraise AttributeError("No such RPC function \'%s\'" % method)\n', u"AttributeError: No such RPC function 'validate_console_port'\n"]. 2013-03-26 16:55:49 16471 TRACE nova.openstack.common.rpc.amqp 2013-03-26 16:55:49 16471 ERROR nova.openstack.common.rpc.common [-] Returning exception Remote error: AttributeError No such RPC function 'validate_console_port' [u'Traceback (most recent call last):\n', u' File "/usr/lib/python2.7/dist-packages/nova/openstack/common/rpc/amqp.py", line 276, in _process_data\nrval = self.proxy.dispatch( ctxt, version, method, **args)\n', u' File "/usr/lib/python2.7/dist-packages/nova/openstack/common/rpc/dispatcher.py", line 148, in dispatch\nraise AttributeError("No such RPC function \'%s\'" % method)\n', u"AttributeError: No such RPC function 'validate_console_port'\n"]. to caller 2013-03-26 16:55:49 16471 ERROR nova.openstack.common.rpc.common [-] ['Traceback (most recent call last):\n', ' File "/usr/lib/python2.7/dist-packages/nova/openstack/common/rpc/am qp.py", line 276, in _process_data\nrval = self.proxy.dispatch(ctxt, version, method, **args)\n', ' File "/usr/lib/python2.7/dist-packages/nova/openstack/common/rpc/dispatcher .py", line 145, in dispatch\nreturn getattr(proxyobj, method)(ctxt, **kwargs)\n', ' File "/usr/lib/python2.7/dist-packages/nova/consoleauth/manager.py", line 107, in check_
Re: [Openstack] Allocating dynamic IP to the VMs
Thanks. root@controller:~# iptables -t nat -L -n Chain PREROUTING (policy ACCEPT) target prot opt source destination quantum-l3-agent-PREROUTING all -- 0.0.0.0/00.0.0.0/0 nova-api-PREROUTING all -- 0.0.0.0/00.0.0.0/0 Chain INPUT (policy ACCEPT) target prot opt source destination Chain OUTPUT (policy ACCEPT) target prot opt source destination quantum-l3-agent-OUTPUT all -- 0.0.0.0/00.0.0.0/0 nova-api-OUTPUT all -- 0.0.0.0/00.0.0.0/0 Chain POSTROUTING (policy ACCEPT) target prot opt source destination quantum-l3-agent-POSTROUTING all -- 0.0.0.0/00.0.0.0/0 nova-api-POSTROUTING all -- 0.0.0.0/00.0.0.0/0 quantum-postrouting-bottom all -- 0.0.0.0/00.0.0.0/0 nova-postrouting-bottom all -- 0.0.0.0/00.0.0.0/0 Chain nova-api-OUTPUT (1 references) target prot opt source destination Chain nova-api-POSTROUTING (1 references) target prot opt source destination Chain nova-api-PREROUTING (1 references) target prot opt source destination Chain nova-api-float-snat (1 references) target prot opt source destination Chain nova-api-snat (1 references) target prot opt source destination nova-api-float-snat all -- 0.0.0.0/00.0.0.0/0 Chain nova-postrouting-bottom (1 references) target prot opt source destination nova-api-snat all -- 0.0.0.0/00.0.0.0/0 Chain quantum-l3-agent-OUTPUT (1 references) target prot opt source destination DNAT all -- 0.0.0.0/0192.168.2.152to:10.5.5.3 Chain quantum-l3-agent-POSTROUTING (1 references) target prot opt source destination ACCEPT all -- 0.0.0.0/00.0.0.0/0! ctstate DNAT ACCEPT all -- 10.5.5.0/24 192.168.2.225 Chain quantum-l3-agent-PREROUTING (1 references) target prot opt source destination DNAT tcp -- 0.0.0.0/0169.254.169.254 tcp dpt:80 to: 192.168.2.225:8775 DNAT all -- 0.0.0.0/0192.168.2.152to:10.5.5.3 Chain quantum-l3-agent-float-snat (1 references) target prot opt source destination SNAT all -- 10.5.5.3 0.0.0.0/0 to:192.168.2.152 Chain quantum-l3-agent-snat (1 references) target prot opt source destination quantum-l3-agent-float-snat all -- 0.0.0.0/00.0.0.0/0 SNAT all -- 10.5.5.0/24 0.0.0.0/0 to:192.168.2.151 Chain quantum-postrouting-bottom (1 references) target prot opt source destination quantum-l3-agent-snat all -- 0.0.0.0/00.0.0.0/0 I can not see anything going through qg- interface. I have activated net.ipv4.ip_forward in /etc/sysctl.conf. On 25 March 2013 13:03, Sylvain Bauza wrote: > Le 25/03/2013 12:49, Chathura M. Sarathchandra Magurawalage a écrit : > > >> I have got one question on this. Does quantum directly request leases >> from the gateway of the physical network before reserving them to allocate >> to VMs? >> >> >> > Nope, not at all. It's up to the administrator to make sure the IP ranges > for Openstack are not pooled by any other DHCP server. There is (as per my > knowledge) no way to sync up in between quantum-l3-agent and other DHCP > servers. > > Actually, contrary to fixed ip networks in Quantum, floating IP networks > are not DHCP managed. Eligibility is made upon next IP address available in > Quantum mysql database and directly injected into iptables, that's it. > > -Sylvain > ___ Mailing list: https://launchpad.net/~openstack Post to : openstack@lists.launchpad.net Unsubscribe : https://launchpad.net/~openstack More help : https://help.launchpad.net/ListHelp
Re: [Openstack] Allocating dynamic IP to the VMs
Thanks Sylvain, I will check and get back to you on this. I have got one question on this. Does quantum directly request leases from the gateway of the physical network before reserving them to allocate to VMs? On 25 March 2013 10:53, Sylvain Bauza wrote: > The basic troubleshooting steps for L3 mapping are : > 1. make sure your DNAT/SNAT entries have been populated correctly (using > 'iptables -t nat -L -n ) > 2. monitor your qg-X interface making sure SNAT is working properly > (using 'tcpdump -i qg- -nn) and checking that you actually have *two* > TCP requests with the same id (the first one with the private IP, the > second one with the public IP) > 3. make sure you activated ip_forward in /etc/sysctl.conf and either > reboot or sysctl -w the value (and restart quantum-l3-agent in this case) > > > If these 3 steps are OK, then you have a gateway issue, not related to > Quantum. > > -Sylvain > > Le 24/03/2013 15:49, Chathura M. Sarathchandra Magurawalage a écrit : > > Thanks Sylvain, > > I have tried this, but does not seem to work. I can allocate the > floating ip to the VM but it is not accessible from the physical network. I > can not ping to it from the controller or any other physical nodes. > > Any idea? > > On 19 March 2013 16:14, Sylvain Bauza wrote: > >> As per >> http://docs.openstack.org/folsom/openstack-network/admin/content/demo_logical_network_config.htmlbut >> slightly modified as per CLI help, >> >> quantum net-create ext_net --tenant-id $TENANT_ID --router:external=True >> quantum subnet-create --ip_version 4 --allocation-pool >> start=192.168.2.151,end=192.168.2.240 \ >> --gateway 192.168.2.253 192.168.2.0/24 -- >> --enable_dhcp=False >> >> >> It will create ext_net subnet with preallocated IP range. For each VM, >> allocate floating IP from this pool and then associate it with the internal >> port. >> >> Hope it can helps, >> -Sylvain >> >> Le 19/03/2013 13:44, Chathura M. Sarathchandra Magurawalage a écrit : >> >> Thanks. >> >> its 192.168.2.0/24 >> >> free ip range: 192.168.2.151 192.168.2.240 >> >> gw/dhcp server: 192.168.2.253 >> >> >> On 19 March 2013 08:28, Sylvain Bauza wrote: >> >>> In that case, please refer to my previous e-mail : use floating IPs >>> bound to the same physical network. >>> That's up to you to know which IP pools are available inside your >>> network. Once you get one, create a external Quantum subnet defined with >>> this IP range. >>> >>> Sorry, I have feeling to explain again and again. If you still don't >>> catch the point, could you please then tell me your physical net/CIDR, your >>> free IP range and your gateway, and I'll mix you up the command to issue. >>> >>> -Sylvain >>> >>> Le 18/03/2013 18:02, Chathura M. Sarathchandra Magurawalage a écrit : >>> >>> Thanks Sylvain, >>> >>> There must be a way of doing this without having to do anything with my >>> default gateway of my physical network? . Even if I have to I do not wan to >>> do anything to the physical gateway. All I need is a way to let the VMs get >>> a dynamic IP from the physical network. How can I do this. For example this >>> can be done on virtual box using a bridge adapter which maps the VM in to >>> the physical network. >>> >>> On 18 March 2013 16:05, Sylvain Bauza wrote: >>> >>>> Could you please tell me your "physical network" CIDR ? >>>> Anyway, what you need is not requiring having a floating IP pool inside >>>> the same network, you can also play with static routing : if your physical >>>> host does have a default gw, you can create a static route from this gw to >>>> the VM network gateway. And on the VM network gateway, do the same... >>>> >>>> -Sylvain >>>> >>>> Le 18/03/2013 16:53, Chathura M. Sarathchandra Magurawalage a écrit : >>>> >>>> Hey Sylvain, >>>> >>>> Basically what I need is to have the VMs mapped to my physical >>>> network so that my physical hosts can directly access the VMs. How can I do >>>> this? >>>> >>>> Thanks. >>>> >>>> >>>> On 18 March 2013 15:50, Sylvain Bauza wrote: >>>> >>>>> Hi, >>>>> >>>>> I don't understand your business. Should you have a 192.168.1.0/24network >>>>> for man
Re: [Openstack] Allocating dynamic IP to the VMs
Thanks Sylvain, I have tried this, but does not seem to work. I can allocate the floating ip to the VM but it is not accessible from the physical network. I can not ping to it from the controller or any other physical nodes. Any idea? On 19 March 2013 16:14, Sylvain Bauza wrote: > As per > http://docs.openstack.org/folsom/openstack-network/admin/content/demo_logical_network_config.htmlbut > slightly modified as per CLI help, > > quantum net-create ext_net --tenant-id $TENANT_ID --router:external=True > quantum subnet-create --ip_version 4 --allocation-pool > start=192.168.2.151,end=192.168.2.240 \ > --gateway 192.168.2.253 192.168.2.0/24 -- --enable_dhcp=False > > > It will create ext_net subnet with preallocated IP range. For each VM, > allocate floating IP from this pool and then associate it with the internal > port. > > Hope it can helps, > -Sylvain > > Le 19/03/2013 13:44, Chathura M. Sarathchandra Magurawalage a écrit : > > Thanks. > > its 192.168.2.0/24 > > free ip range: 192.168.2.151 192.168.2.240 > > gw/dhcp server: 192.168.2.253 > > > On 19 March 2013 08:28, Sylvain Bauza wrote: > >> In that case, please refer to my previous e-mail : use floating IPs >> bound to the same physical network. >> That's up to you to know which IP pools are available inside your >> network. Once you get one, create a external Quantum subnet defined with >> this IP range. >> >> Sorry, I have feeling to explain again and again. If you still don't >> catch the point, could you please then tell me your physical net/CIDR, your >> free IP range and your gateway, and I'll mix you up the command to issue. >> >> -Sylvain >> >> Le 18/03/2013 18:02, Chathura M. Sarathchandra Magurawalage a écrit : >> >> Thanks Sylvain, >> >> There must be a way of doing this without having to do anything with my >> default gateway of my physical network? . Even if I have to I do not wan to >> do anything to the physical gateway. All I need is a way to let the VMs get >> a dynamic IP from the physical network. How can I do this. For example this >> can be done on virtual box using a bridge adapter which maps the VM in to >> the physical network. >> >> On 18 March 2013 16:05, Sylvain Bauza wrote: >> >>> Could you please tell me your "physical network" CIDR ? >>> Anyway, what you need is not requiring having a floating IP pool inside >>> the same network, you can also play with static routing : if your physical >>> host does have a default gw, you can create a static route from this gw to >>> the VM network gateway. And on the VM network gateway, do the same... >>> >>> -Sylvain >>> >>> Le 18/03/2013 16:53, Chathura M. Sarathchandra Magurawalage a écrit : >>> >>> Hey Sylvain, >>> >>> Basically what I need is to have the VMs mapped to my physical network >>> so that my physical hosts can directly access the VMs. How can I do this? >>> >>> Thanks. >>> >>> >>> On 18 March 2013 15:50, Sylvain Bauza wrote: >>> >>>> Hi, >>>> >>>> I don't understand your business. Should you have a 192.168.1.0/24network >>>> for management, you could also assign an external network with >>>> Quantum based on the same subnet (ie. 192.168.1.0/24). >>>> When creating a floating IP pool, Quantum does require at least 3 >>>> things : >>>> - the CIDR >>>> - the beginning and ending IPs >>>> - the external gateway >>>> >>>> So, based on what I previously said, you only need to create a >>>> 192.168.1.0/24 in Quantum with .1-.100 (for example) as the range, >>>> .254 being the external gateway. >>>> >>>> Thanks, >>>> -Sylvain >>>> >>>> Le 18/03/2013 16:29, Chathura M. Sarathchandra Magurawalage a écrit : >>>> >>>> anyone? >>>> >>>> On 17 March 2013 21:33, Chathura M. Sarathchandra Magurawalage < >>>> 77.chath...@gmail.com> wrote: >>>> >>>>> After reading a little bit more, I think I have found what I need. It >>>>> is a provider network that I need for the VMs so that they can get access >>>>> to the other resources in my main network ( such as other physical hosts >>>>> that are connected to the same network ). >>>>> >>>>> My question is, is it possible to do this alongside the use case >>>&
Re: [Openstack] Allocating dynamic IP to the VMs
Thanks. its 192.168.2.0/24 free ip range: 192.168.2.151 192.168.2.240 gw/dhcp server: 192.168.2.253 On 19 March 2013 08:28, Sylvain Bauza wrote: > In that case, please refer to my previous e-mail : use floating IPs > bound to the same physical network. > That's up to you to know which IP pools are available inside your network. > Once you get one, create a external Quantum subnet defined with this IP > range. > > Sorry, I have feeling to explain again and again. If you still don't catch > the point, could you please then tell me your physical net/CIDR, your free > IP range and your gateway, and I'll mix you up the command to issue. > > -Sylvain > > Le 18/03/2013 18:02, Chathura M. Sarathchandra Magurawalage a écrit : > > Thanks Sylvain, > > There must be a way of doing this without having to do anything with my > default gateway of my physical network? . Even if I have to I do not wan to > do anything to the physical gateway. All I need is a way to let the VMs get > a dynamic IP from the physical network. How can I do this. For example this > can be done on virtual box using a bridge adapter which maps the VM in to > the physical network. > > On 18 March 2013 16:05, Sylvain Bauza wrote: > >> Could you please tell me your "physical network" CIDR ? >> Anyway, what you need is not requiring having a floating IP pool inside >> the same network, you can also play with static routing : if your physical >> host does have a default gw, you can create a static route from this gw to >> the VM network gateway. And on the VM network gateway, do the same... >> >> -Sylvain >> >> Le 18/03/2013 16:53, Chathura M. Sarathchandra Magurawalage a écrit : >> >> Hey Sylvain, >> >> Basically what I need is to have the VMs mapped to my physical network >> so that my physical hosts can directly access the VMs. How can I do this? >> >> Thanks. >> >> >> On 18 March 2013 15:50, Sylvain Bauza wrote: >> >>> Hi, >>> >>> I don't understand your business. Should you have a 192.168.1.0/24network >>> for management, you could also assign an external network with >>> Quantum based on the same subnet (ie. 192.168.1.0/24). >>> When creating a floating IP pool, Quantum does require at least 3 things >>> : >>> - the CIDR >>> - the beginning and ending IPs >>> - the external gateway >>> >>> So, based on what I previously said, you only need to create a >>> 192.168.1.0/24 in Quantum with .1-.100 (for example) as the range, .254 >>> being the external gateway. >>> >>> Thanks, >>> -Sylvain >>> >>> Le 18/03/2013 16:29, Chathura M. Sarathchandra Magurawalage a écrit : >>> >>> anyone? >>> >>> On 17 March 2013 21:33, Chathura M. Sarathchandra Magurawalage < >>> 77.chath...@gmail.com> wrote: >>> >>>> After reading a little bit more, I think I have found what I need. It >>>> is a provider network that I need for the VMs so that they can get access >>>> to the other resources in my main network ( such as other physical hosts >>>> that are connected to the same network ). >>>> >>>> My question is, is it possible to do this alongside the use case that >>>> I have followed ( Provider router with private networks)? >>>> >>>> If so how can I do this? >>>> >>>> Thanks. >>>> >>>> >>>> On 16 March 2013 01:46, Chathura M. Sarathchandra Magurawalage < >>>> 77.chath...@gmail.com> wrote: >>>> >>>>> Hello, >>>>> >>>>> I want to know how I can allocate a dynamic IP to the VM from the >>>>> same network as the openstack hosts (controller/network-node/compute node) >>>>> network/management network . For example, in virtual box you can give your >>>>> VM an IP from the host's network using a Bridge adapter. How can I do this >>>>> in openstack? >>>>> >>>>> From what I understand floating IP's are used when you have a public >>>>> IP >>>>> (which is static) to be allocated to VM's. >>>>> >>>>> My openstack installation architecture: >>>>> >>>>> http://docs.openstack.org/folsom/basic-install/content/basic-install_architecture.html >>>>> >>>>> Quantum use case: >>>>> >>>>> http://docs.openstack.org/trunk/openstack-network/admin/content/use_cases_single_router.html >>>>> >>>> >>>> >>> >>> >>> ___ >>> Mailing list: https://launchpad.net/~openstack >>> Post to : openstack@lists.launchpad.net >>> Unsubscribe : https://launchpad.net/~openstack >>> More help : https://help.launchpad.net/ListHelp >>> >>> >>> >> >> > > ___ Mailing list: https://launchpad.net/~openstack Post to : openstack@lists.launchpad.net Unsubscribe : https://launchpad.net/~openstack More help : https://help.launchpad.net/ListHelp
Re: [Openstack] Allocating dynamic IP to the VMs
Thanks Sylvain, There must be a way of doing this without having to do anything with my default gateway of my physical network? . Even if I have to I do not wan to do anything to the physical gateway. All I need is a way to let the VMs get a dynamic IP from the physical network. How can I do this. For example this can be done on virtual box using a bridge adapter which maps the VM in to the physical network. On 18 March 2013 16:05, Sylvain Bauza wrote: > Could you please tell me your "physical network" CIDR ? > Anyway, what you need is not requiring having a floating IP pool inside > the same network, you can also play with static routing : if your physical > host does have a default gw, you can create a static route from this gw to > the VM network gateway. And on the VM network gateway, do the same... > > -Sylvain > > Le 18/03/2013 16:53, Chathura M. Sarathchandra Magurawalage a écrit : > > Hey Sylvain, > > Basically what I need is to have the VMs mapped to my physical network > so that my physical hosts can directly access the VMs. How can I do this? > > Thanks. > > > On 18 March 2013 15:50, Sylvain Bauza wrote: > >> Hi, >> >> I don't understand your business. Should you have a 192.168.1.0/24network >> for management, you could also assign an external network with >> Quantum based on the same subnet (ie. 192.168.1.0/24). >> When creating a floating IP pool, Quantum does require at least 3 things >> : >> - the CIDR >> - the beginning and ending IPs >> - the external gateway >> >> So, based on what I previously said, you only need to create a >> 192.168.1.0/24 in Quantum with .1-.100 (for example) as the range, .254 >> being the external gateway. >> >> Thanks, >> -Sylvain >> >> Le 18/03/2013 16:29, Chathura M. Sarathchandra Magurawalage a écrit : >> >> anyone? >> >> On 17 March 2013 21:33, Chathura M. Sarathchandra Magurawalage < >> 77.chath...@gmail.com> wrote: >> >>> After reading a little bit more, I think I have found what I need. It is >>> a provider network that I need for the VMs so that they can get access to >>> the other resources in my main network ( such as other physical hosts that >>> are connected to the same network ). >>> >>> My question is, is it possible to do this alongside the use case that >>> I have followed ( Provider router with private networks)? >>> >>> If so how can I do this? >>> >>> Thanks. >>> >>> >>> On 16 March 2013 01:46, Chathura M. Sarathchandra Magurawalage < >>> 77.chath...@gmail.com> wrote: >>> >>>> Hello, >>>> >>>> I want to know how I can allocate a dynamic IP to the VM from the >>>> same network as the openstack hosts (controller/network-node/compute node) >>>> network/management network . For example, in virtual box you can give your >>>> VM an IP from the host's network using a Bridge adapter. How can I do this >>>> in openstack? >>>> >>>> From what I understand floating IP's are used when you have a public >>>> IP >>>> (which is static) to be allocated to VM's. >>>> >>>> My openstack installation architecture: >>>> >>>> http://docs.openstack.org/folsom/basic-install/content/basic-install_architecture.html >>>> >>>> Quantum use case: >>>> >>>> http://docs.openstack.org/trunk/openstack-network/admin/content/use_cases_single_router.html >>>> >>> >>> >> >> >> ___ >> Mailing list: https://launchpad.net/~openstack >> Post to : openstack@lists.launchpad.net >> Unsubscribe : https://launchpad.net/~openstack >> More help : https://help.launchpad.net/ListHelp >> >> >> > > ___ Mailing list: https://launchpad.net/~openstack Post to : openstack@lists.launchpad.net Unsubscribe : https://launchpad.net/~openstack More help : https://help.launchpad.net/ListHelp
Re: [Openstack] Allocating dynamic IP to the VMs
Hey Sylvain, Basically what I need is to have the VMs mapped to my physical network so that my physical hosts can directly access the VMs. How can I do this? Thanks. On 18 March 2013 15:50, Sylvain Bauza wrote: > Hi, > > I don't understand your business. Should you have a 192.168.1.0/24network for > management, you could also assign an external network with > Quantum based on the same subnet (ie. 192.168.1.0/24). > When creating a floating IP pool, Quantum does require at least 3 things : > - the CIDR > - the beginning and ending IPs > - the external gateway > > So, based on what I previously said, you only need to create a > 192.168.1.0/24 in Quantum with .1-.100 (for example) as the range, .254 > being the external gateway. > > Thanks, > -Sylvain > > Le 18/03/2013 16:29, Chathura M. Sarathchandra Magurawalage a écrit : > > anyone? > > On 17 March 2013 21:33, Chathura M. Sarathchandra Magurawalage < > 77.chath...@gmail.com> wrote: > >> After reading a little bit more, I think I have found what I need. It is >> a provider network that I need for the VMs so that they can get access to >> the other resources in my main network ( such as other physical hosts that >> are connected to the same network ). >> >> My question is, is it possible to do this alongside the use case that I >> have followed ( Provider router with private networks)? >> >> If so how can I do this? >> >> Thanks. >> >> >> On 16 March 2013 01:46, Chathura M. Sarathchandra Magurawalage < >> 77.chath...@gmail.com> wrote: >> >>> Hello, >>> >>> I want to know how I can allocate a dynamic IP to the VM from the same >>> network as the openstack hosts (controller/network-node/compute node) >>> network/management network . For example, in virtual box you can give your >>> VM an IP from the host's network using a Bridge adapter. How can I do this >>> in openstack? >>> >>> From what I understand floating IP's are used when you have a public >>> IP >>> (which is static) to be allocated to VM's. >>> >>> My openstack installation architecture: >>> >>> http://docs.openstack.org/folsom/basic-install/content/basic-install_architecture.html >>> >>> Quantum use case: >>> >>> http://docs.openstack.org/trunk/openstack-network/admin/content/use_cases_single_router.html >>> >> >> > > > ___ > Mailing list: https://launchpad.net/~openstack > Post to : openstack@lists.launchpad.net > Unsubscribe : https://launchpad.net/~openstack > More help : https://help.launchpad.net/ListHelp > > > ___ Mailing list: https://launchpad.net/~openstack Post to : openstack@lists.launchpad.net Unsubscribe : https://launchpad.net/~openstack More help : https://help.launchpad.net/ListHelp
Re: [Openstack] Allocating dynamic IP to the VMs
anyone? On 17 March 2013 21:33, Chathura M. Sarathchandra Magurawalage < 77.chath...@gmail.com> wrote: > After reading a little bit more, I think I have found what I need. It is > a provider network that I need for the VMs so that they can get access to > the other resources in my main network ( such as other physical hosts that > are connected to the same network ). > > My question is, is it possible to do this alongside the use case that I > have followed ( Provider router with private networks)? > > If so how can I do this? > > Thanks. > > > On 16 March 2013 01:46, Chathura M. Sarathchandra Magurawalage < > 77.chath...@gmail.com> wrote: > >> Hello, >> >> I want to know how I can allocate a dynamic IP to the VM from the same >> network as the openstack hosts (controller/network-node/compute node) >> network/management network . For example, in virtual box you can give your >> VM an IP from the host's network using a Bridge adapter. How can I do this >> in openstack? >> >> From what I understand floating IP's are used when you have a public IP >> (which is static) to be allocated to VM's. >> >> My openstack installation architecture: >> >> http://docs.openstack.org/folsom/basic-install/content/basic-install_architecture.html >> >> Quantum use case: >> >> http://docs.openstack.org/trunk/openstack-network/admin/content/use_cases_single_router.html >> > > ___ Mailing list: https://launchpad.net/~openstack Post to : openstack@lists.launchpad.net Unsubscribe : https://launchpad.net/~openstack More help : https://help.launchpad.net/ListHelp
Re: [Openstack] Allocating dynamic IP to the VMs
After reading a little bit more, I think I have found what I need. It is a provider network that I need for the VMs so that they can get access to the other resources in my main network ( such as other physical hosts that are connected to the same network ). My question is, is it possible to do this alongside the use case that I have followed ( Provider router with private networks)? If so how can I do this? Thanks. On 16 March 2013 01:46, Chathura M. Sarathchandra Magurawalage < 77.chath...@gmail.com> wrote: > Hello, > > I want to know how I can allocate a dynamic IP to the VM from the same > network as the openstack hosts (controller/network-node/compute node) > network/management network . For example, in virtual box you can give your > VM an IP from the host's network using a Bridge adapter. How can I do this > in openstack? > > From what I understand floating IP's are used when you have a public IP > (which is static) to be allocated to VM's. > > My openstack installation architecture: > > http://docs.openstack.org/folsom/basic-install/content/basic-install_architecture.html > > Quantum use case: > > http://docs.openstack.org/trunk/openstack-network/admin/content/use_cases_single_router.html > ___ Mailing list: https://launchpad.net/~openstack Post to : openstack@lists.launchpad.net Unsubscribe : https://launchpad.net/~openstack More help : https://help.launchpad.net/ListHelp
Re: [Openstack] ssh from VM to VM
I solved the issue by copying the rsa public key of the first VM to the second VM. Thought I did not have to do this. Thanks. On 16 March 2013 12:34, Pranav wrote: > I think you need not exchange key pairs for Cirros image. > Regards, > Pranav > > > On Sat, Mar 16, 2013 at 4:32 PM, Chathura M. Sarathchandra Magurawalage < > 77.chath...@gmail.com> wrote: > >> Thanks for your reply. >> >> I have inserted PasswordAuthentication yes to the ssh config file. All >> VMs have the same metadata including the ssh public key of the controller. >> So I cant see why only cirros vms can do this. >> >> Still does not work. >> >> >> >> On 16 March 2013 06:24, Aaron Rosen wrote: >> >>> I suspect that that host 10.5.5.6 has ssh configured for >>> PasswordAuthentication set to no and you don't have your public key of the >>> host you are on, in the authorized_key file of 10.5.5.6. >>> >>> Aaron >>> >>> On Fri, Mar 15, 2013 at 7:26 PM, Chathura M. Sarathchandra >>> Magurawalage <77.chath...@gmail.com> wrote: >>> >>>> Hello, >>>> >>>> I can't ssh from Ubuntu cloud VM to other VM. I get following >>>> >>>> ubuntu@master:~$ ssh cirros@10.5.5.6 -v >>>> OpenSSH_5.9p1 Debian-5ubuntu1, OpenSSL 1.0.1 14 Mar 2012 >>>> debug1: Reading configuration data /etc/ssh/ssh_config >>>> debug1: /etc/ssh/ssh_config line 19: Applying options for * >>>> debug1: Connecting to 10.5.5.6 [10.5.5.6] port 22. >>>> debug1: Connection established. >>>> debug1: identity file /home/ubuntu/.ssh/id_rsa type -1 >>>> debug1: identity file /home/ubuntu/.ssh/id_rsa-cert type -1 >>>> debug1: identity file /home/ubuntu/.ssh/id_dsa type -1 >>>> debug1: identity file /home/ubuntu/.ssh/id_dsa-cert type -1 >>>> debug1: identity file /home/ubuntu/.ssh/id_ecdsa type -1 >>>> debug1: identity file /home/ubuntu/.ssh/id_ecdsa-cert type -1 >>>> debug1: Remote protocol version 2.0, remote software version >>>> OpenSSH_5.9p1 Debian-5ubuntu1 >>>> debug1: match: OpenSSH_5.9p1 Debian-5ubuntu1 pat OpenSSH* >>>> debug1: Enabling compatibility mode for protocol 2.0 >>>> debug1: Local version string SSH-2.0-OpenSSH_5.9p1 Debian-5ubuntu1 >>>> debug1: SSH2_MSG_KEXINIT sent >>>> debug1: SSH2_MSG_KEXINIT received >>>> debug1: kex: server->client aes128-ctr hmac-md5 none >>>> debug1: kex: client->server aes128-ctr hmac-md5 none >>>> debug1: sending SSH2_MSG_KEX_ECDH_INIT >>>> debug1: expecting SSH2_MSG_KEX_ECDH_REPLY >>>> debug1: Server host key: ECDSA >>>> 7b:8f:6a:ee:ba:e5:0a:c5:04:01:ca:bd:e5:38:69:55 >>>> debug1: Host '10.5.5.6' is known and matches the ECDSA host key. >>>> debug1: Found key in /home/ubuntu/.ssh/known_hosts:4 >>>> debug1: ssh_ecdsa_verify: signature correct >>>> debug1: SSH2_MSG_NEWKEYS sent >>>> debug1: expecting SSH2_MSG_NEWKEYS >>>> debug1: SSH2_MSG_NEWKEYS received >>>> debug1: Roaming not allowed by server >>>> debug1: SSH2_MSG_SERVICE_REQUEST sent >>>> debug1: SSH2_MSG_SERVICE_ACCEPT received >>>> debug1: Authentications that can continue: publickey >>>> debug1: Next authentication method: publickey >>>> debug1: Trying private key: /home/ubuntu/.ssh/id_rsa >>>> debug1: Trying private key: /home/ubuntu/.ssh/id_dsa >>>> debug1: Trying private key: /home/ubuntu/.ssh/id_ecdsa >>>> debug1: No more authentication methods to try. >>>> Permission denied (publickey). >>>> >>>> But I can ssh from to my Cirros VMs. Also I can ssh from Ubuntu VM to >>>> Cirros VM. >>>> >>>> Any Idea? >>>> >>>> Thanks. >>>> >>>> ___ >>>> Mailing list: https://launchpad.net/~openstack >>>> Post to : openstack@lists.launchpad.net >>>> Unsubscribe : https://launchpad.net/~openstack >>>> More help : https://help.launchpad.net/ListHelp >>>> >>>> >>> >> >> ___ >> Mailing list: https://launchpad.net/~openstack >> Post to : openstack@lists.launchpad.net >> Unsubscribe : https://launchpad.net/~openstack >> More help : https://help.launchpad.net/ListHelp >> >> > ___ Mailing list: https://launchpad.net/~openstack Post to : openstack@lists.launchpad.net Unsubscribe : https://launchpad.net/~openstack More help : https://help.launchpad.net/ListHelp
Re: [Openstack] ssh from VM to VM
Thanks for your reply. I have inserted PasswordAuthentication yes to the ssh config file. All VMs have the same metadata including the ssh public key of the controller. So I cant see why only cirros vms can do this. Still does not work. On 16 March 2013 06:24, Aaron Rosen wrote: > I suspect that that host 10.5.5.6 has ssh configured for > PasswordAuthentication set to no and you don't have your public key of the > host you are on, in the authorized_key file of 10.5.5.6. > > Aaron > > On Fri, Mar 15, 2013 at 7:26 PM, Chathura M. Sarathchandra Magurawalage < > 77.chath...@gmail.com> wrote: > >> Hello, >> >> I can't ssh from Ubuntu cloud VM to other VM. I get following >> >> ubuntu@master:~$ ssh cirros@10.5.5.6 -v >> OpenSSH_5.9p1 Debian-5ubuntu1, OpenSSL 1.0.1 14 Mar 2012 >> debug1: Reading configuration data /etc/ssh/ssh_config >> debug1: /etc/ssh/ssh_config line 19: Applying options for * >> debug1: Connecting to 10.5.5.6 [10.5.5.6] port 22. >> debug1: Connection established. >> debug1: identity file /home/ubuntu/.ssh/id_rsa type -1 >> debug1: identity file /home/ubuntu/.ssh/id_rsa-cert type -1 >> debug1: identity file /home/ubuntu/.ssh/id_dsa type -1 >> debug1: identity file /home/ubuntu/.ssh/id_dsa-cert type -1 >> debug1: identity file /home/ubuntu/.ssh/id_ecdsa type -1 >> debug1: identity file /home/ubuntu/.ssh/id_ecdsa-cert type -1 >> debug1: Remote protocol version 2.0, remote software version >> OpenSSH_5.9p1 Debian-5ubuntu1 >> debug1: match: OpenSSH_5.9p1 Debian-5ubuntu1 pat OpenSSH* >> debug1: Enabling compatibility mode for protocol 2.0 >> debug1: Local version string SSH-2.0-OpenSSH_5.9p1 Debian-5ubuntu1 >> debug1: SSH2_MSG_KEXINIT sent >> debug1: SSH2_MSG_KEXINIT received >> debug1: kex: server->client aes128-ctr hmac-md5 none >> debug1: kex: client->server aes128-ctr hmac-md5 none >> debug1: sending SSH2_MSG_KEX_ECDH_INIT >> debug1: expecting SSH2_MSG_KEX_ECDH_REPLY >> debug1: Server host key: ECDSA >> 7b:8f:6a:ee:ba:e5:0a:c5:04:01:ca:bd:e5:38:69:55 >> debug1: Host '10.5.5.6' is known and matches the ECDSA host key. >> debug1: Found key in /home/ubuntu/.ssh/known_hosts:4 >> debug1: ssh_ecdsa_verify: signature correct >> debug1: SSH2_MSG_NEWKEYS sent >> debug1: expecting SSH2_MSG_NEWKEYS >> debug1: SSH2_MSG_NEWKEYS received >> debug1: Roaming not allowed by server >> debug1: SSH2_MSG_SERVICE_REQUEST sent >> debug1: SSH2_MSG_SERVICE_ACCEPT received >> debug1: Authentications that can continue: publickey >> debug1: Next authentication method: publickey >> debug1: Trying private key: /home/ubuntu/.ssh/id_rsa >> debug1: Trying private key: /home/ubuntu/.ssh/id_dsa >> debug1: Trying private key: /home/ubuntu/.ssh/id_ecdsa >> debug1: No more authentication methods to try. >> Permission denied (publickey). >> >> But I can ssh from to my Cirros VMs. Also I can ssh from Ubuntu VM to >> Cirros VM. >> >> Any Idea? >> >> Thanks. >> >> ___ >> Mailing list: https://launchpad.net/~openstack >> Post to : openstack@lists.launchpad.net >> Unsubscribe : https://launchpad.net/~openstack >> More help : https://help.launchpad.net/ListHelp >> >> > ___ Mailing list: https://launchpad.net/~openstack Post to : openstack@lists.launchpad.net Unsubscribe : https://launchpad.net/~openstack More help : https://help.launchpad.net/ListHelp
[Openstack] ssh from VM to VM
Hello, I can't ssh from Ubuntu cloud VM to other VM. I get following ubuntu@master:~$ ssh cirros@10.5.5.6 -v OpenSSH_5.9p1 Debian-5ubuntu1, OpenSSL 1.0.1 14 Mar 2012 debug1: Reading configuration data /etc/ssh/ssh_config debug1: /etc/ssh/ssh_config line 19: Applying options for * debug1: Connecting to 10.5.5.6 [10.5.5.6] port 22. debug1: Connection established. debug1: identity file /home/ubuntu/.ssh/id_rsa type -1 debug1: identity file /home/ubuntu/.ssh/id_rsa-cert type -1 debug1: identity file /home/ubuntu/.ssh/id_dsa type -1 debug1: identity file /home/ubuntu/.ssh/id_dsa-cert type -1 debug1: identity file /home/ubuntu/.ssh/id_ecdsa type -1 debug1: identity file /home/ubuntu/.ssh/id_ecdsa-cert type -1 debug1: Remote protocol version 2.0, remote software version OpenSSH_5.9p1 Debian-5ubuntu1 debug1: match: OpenSSH_5.9p1 Debian-5ubuntu1 pat OpenSSH* debug1: Enabling compatibility mode for protocol 2.0 debug1: Local version string SSH-2.0-OpenSSH_5.9p1 Debian-5ubuntu1 debug1: SSH2_MSG_KEXINIT sent debug1: SSH2_MSG_KEXINIT received debug1: kex: server->client aes128-ctr hmac-md5 none debug1: kex: client->server aes128-ctr hmac-md5 none debug1: sending SSH2_MSG_KEX_ECDH_INIT debug1: expecting SSH2_MSG_KEX_ECDH_REPLY debug1: Server host key: ECDSA 7b:8f:6a:ee:ba:e5:0a:c5:04:01:ca:bd:e5:38:69:55 debug1: Host '10.5.5.6' is known and matches the ECDSA host key. debug1: Found key in /home/ubuntu/.ssh/known_hosts:4 debug1: ssh_ecdsa_verify: signature correct debug1: SSH2_MSG_NEWKEYS sent debug1: expecting SSH2_MSG_NEWKEYS debug1: SSH2_MSG_NEWKEYS received debug1: Roaming not allowed by server debug1: SSH2_MSG_SERVICE_REQUEST sent debug1: SSH2_MSG_SERVICE_ACCEPT received debug1: Authentications that can continue: publickey debug1: Next authentication method: publickey debug1: Trying private key: /home/ubuntu/.ssh/id_rsa debug1: Trying private key: /home/ubuntu/.ssh/id_dsa debug1: Trying private key: /home/ubuntu/.ssh/id_ecdsa debug1: No more authentication methods to try. Permission denied (publickey). But I can ssh from to my Cirros VMs. Also I can ssh from Ubuntu VM to Cirros VM. Any Idea? Thanks. ___ Mailing list: https://launchpad.net/~openstack Post to : openstack@lists.launchpad.net Unsubscribe : https://launchpad.net/~openstack More help : https://help.launchpad.net/ListHelp
[Openstack] Allocating dynamic IP to the VMs
Hello, I want to know how I can allocate a dynamic IP to the VM from the same network as the openstack hosts (controller/network-node/compute node) network/management network . For example, in virtual box you can give your VM an IP from the host's network using a Bridge adapter. How can I do this in openstack? >From what I understand floating IP's are used when you have a public IP (which is static) to be allocated to VM's. My openstack installation architecture: http://docs.openstack.org/folsom/basic-install/content/basic-install_architecture.html Quantum use case: http://docs.openstack.org/trunk/openstack-network/admin/content/use_cases_single_router.html ___ Mailing list: https://launchpad.net/~openstack Post to : openstack@lists.launchpad.net Unsubscribe : https://launchpad.net/~openstack More help : https://help.launchpad.net/ListHelp
Re: [Openstack] Cant ping private or floating IP
Thanks. I would be more concerned about the SIOCDELRT error above. Do you try to > manually remove a network route at bootup ? Seems like the 'route del' is > failing because the route is not already existing. > > I am not doing doing anything that I am aware of. > > As already said, you absolutely need VNC support for investigating. Could > you please fix your VNC setup which is incorrect ? > But VNC works fine. Its just that it VM hangs on the boot up it wont come to the log in prompt, I can't log into it. :( On 20 February 2013 13:46, Sylvain Bauza wrote: > Le 20/02/2013 14:04, Chathura M. Sarathchandra Magurawalage a écrit : > > There are apparently two instances running in the compute node but nova > just see only one. Probably when I have deleted an instance earlier it had > not deleted the instance properly. > > root@controller:~# nova list > > +--+++---+ > | ID | Name | Status | Networks >| > > +--+++---+ > | 42e18cd5-de6f-4181-b238-320fe37ef6f1 | master | ACTIVE | > demo-net=10.5.5.3 | > > +--+++---+ > > > virsh -c qemu+ssh://root@computenode/system list > root@computenode's password: > IdName State > > 14instance-002c running > 18instance-001e running > > > > You should have seen at 'sudo virsh list --all', plus looking at > /etc/libvirt/qemu/*.xml to check how many instances were defined. > I do suspect also that for some reason (probably nova-compute down), a > clean-up of 2c probably didn't work. Anyway, this is fixed as you > mention. > > > Then I have deleted all instances and created a new one. But still cant > ping or ssh the new VM. > > > > > > > > > > > >function='0x0'\ > /> > > > Starting network... > udhcpc (v1.18.5) started > Sending discover... > Sending select for 10.5.5.3... > Lease of 10.5.5.3 obtained, lease time 120 > deleting routers > route: SIOCDELRT: No such process > adding dns 8.8.8.8 > > > > The DHCP reply is correctly received by the instance from the network node > to the compute node. This is not a network issue (at least for IP > assignation). > I would be more concerned about the SIOCDELRT error above. Do you try to > manually remove a network route at bootup ? Seems like the 'route del' is > failing because the route is not already existing. > > > As already said, you absolutely need VNC support for investigating. Could > you please fix your VNC setup which is incorrect ? > > >keymap='en-u\ > s'> > > > > > Try in nova-compute.conf : > vncserver_proxyclient_address= > vncserver_listen= > and in nova.conf : > novncproxy_base_url=http://:6080/vnc_auto.html > > and restart nova-compute. > > > > On 20 February 2013 11:57, Sylvain Bauza wrote: > >> Could you please paste : >> - /etc/libvirt/qemu/.xml >> - ip a show vnet0 >> - brctl show >> >> Sounds like your virtual device is not created. Could you please launch a >> new VM and paste /var/log/nova/nova-compute.log ? >> >> Thanks, >> -Sylvain >> >> > ___ Mailing list: https://launchpad.net/~openstack Post to : openstack@lists.launchpad.net Unsubscribe : https://launchpad.net/~openstack More help : https://help.launchpad.net/ListHelp
Re: [Openstack] Cant ping private or floating IP
Yes definitely I will post it here for future reference for anybody. On 18 February 2013 13:28, Guilherme Russi wrote: > Got it, I have one virtual interface too, to make the management and VM > conf part. If you find anything, let me know, please. > > Thanks. > > > > 2013/2/18 Chathura M. Sarathchandra Magurawalage <77.chath...@gmail.com> > >> I have only got 1 NIC but got two virtual interface for two different >> networks. I have got network node in the same physical machine too. >> >> >> On 18 February 2013 13:15, Guilherme Russi wrote: >> >>> How did you install your controller node? I mean, mine I have 2 NICs and >>> I installed the network node at the same physical machine. >>> >>> >>> 2013/2/18 Chathura M. Sarathchandra Magurawalage <77.chath...@gmail.com> >>> >>> Hello Guilherme, >>>> >>>> No, I am still having the problem :( >>>> >>>> >>>> On 18 February 2013 13:01, Guilherme Russi >>>> wrote: >>>> >>>>> Hello Chathura, >>>>> >>>>> Have succeeded with your network? I'm having problems with mine too. >>>>> >>>>> Thanks. >>>>> >>>>> Guilherme. >>>>> >>>>> >>>>> 2013/2/17 Chathura M. Sarathchandra Magurawalage < >>>>> 77.chath...@gmail.com> >>>>> >>>>> Hope you had a good night sleep :) >>>>>> >>>>>> Yes sure I will be on irc. my nickname is chathura77 >>>>>> >>>>>> Thanks >>>>>> >>>>>> On 17 February 2013 13:15, Jean-Baptiste RANSY < >>>>>> jean-baptiste.ra...@alyseo.com> wrote: >>>>>> >>>>>>> ping >>>>>>> >>>>>>> Are you on IRC ? >>>>>>> >>>>>>> JB >>>>>>> >>>>>>> >>>>>>> >>>>>>> On 02/17/2013 04:07 AM, Jean-Baptiste RANSY wrote: >>>>>>> >>>>>>> Add Cirros Image to Glance :) >>>>>>> >>>>>>> Username: cirros >>>>>>> Password: cubswin:) >>>>>>> >>>>>>> >>>>>>> http://docs.openstack.org/trunk/openstack-compute/install/apt/content/uploading-to-glance.html >>>>>>> >>>>>>> to join your VM, it's a bit dirty but you can : >>>>>>> - put your computer in the same subnet as your controller ( >>>>>>> 192.168.2.0/24) >>>>>>> - then adds a static route to the subnet of your VM. (ip route add >>>>>>> 10.5.5.0/24 gw 192.168.2.151) >>>>>>> (192.168.2.151 is the quantum gateway) >>>>>>> >>>>>>> I'm going to sleep, we will continue tomorrow. >>>>>>> >>>>>>> JB >>>>>>> >>>>>>> PS : You also should get some sleep :) >>>>>>> >>>>>>> >>>>>>> On 02/17/2013 03:53 AM, Chathura M. Sarathchandra Magurawalage wrote: >>>>>>> >>>>>>> oh that's weird. >>>>>>> >>>>>>> I still get this error. couldnt this be because I cannot ping the >>>>>>> VM in the first place?. Because as far as I know metadata takes care of >>>>>>> ssh >>>>>>> keys. But what if you cant reach the VM in the first place? >>>>>>> >>>>>>> no instance data found in start-local >>>>>>> >>>>>>> ci-info: lo: 1 127.0.0.1 255.0.0.0 . >>>>>>> >>>>>>> ci-info: eth0 : 1 10.5.5.3255.255.255.0 fa:16:3e:a7:28:25 >>>>>>> >>>>>>> ci-info: route-0: 0.0.0.0 10.5.5.1 0.0.0.0 eth0 >>>>>>> UG >>>>>>> >>>>>>> ci-info: route-1: 10.5.5.00.0.0.0 255.255.255.0 eth0 >>>>>>> U >>>>>>> >>>>>>> cloud-init start running: Sun, 17 Feb 2013 02:45:35 +. up 3.51 >>>>>>> seconds >>>>>>> >>>>>>> 2013-02-17 02:48:25,840 - util.py[WARNING]: >>>&
Re: [Openstack] Cant ping private or floating IP
I have only got 1 NIC but got two virtual interface for two different networks. I have got network node in the same physical machine too. On 18 February 2013 13:15, Guilherme Russi wrote: > How did you install your controller node? I mean, mine I have 2 NICs and I > installed the network node at the same physical machine. > > > 2013/2/18 Chathura M. Sarathchandra Magurawalage <77.chath...@gmail.com> > > Hello Guilherme, >> >> No, I am still having the problem :( >> >> >> On 18 February 2013 13:01, Guilherme Russi wrote: >> >>> Hello Chathura, >>> >>> Have succeeded with your network? I'm having problems with mine too. >>> >>> Thanks. >>> >>> Guilherme. >>> >>> >>> 2013/2/17 Chathura M. Sarathchandra Magurawalage <77.chath...@gmail.com> >>> >>> Hope you had a good night sleep :) >>>> >>>> Yes sure I will be on irc. my nickname is chathura77 >>>> >>>> Thanks >>>> >>>> On 17 February 2013 13:15, Jean-Baptiste RANSY < >>>> jean-baptiste.ra...@alyseo.com> wrote: >>>> >>>>> ping >>>>> >>>>> Are you on IRC ? >>>>> >>>>> JB >>>>> >>>>> >>>>> >>>>> On 02/17/2013 04:07 AM, Jean-Baptiste RANSY wrote: >>>>> >>>>> Add Cirros Image to Glance :) >>>>> >>>>> Username: cirros >>>>> Password: cubswin:) >>>>> >>>>> >>>>> http://docs.openstack.org/trunk/openstack-compute/install/apt/content/uploading-to-glance.html >>>>> >>>>> to join your VM, it's a bit dirty but you can : >>>>> - put your computer in the same subnet as your controller ( >>>>> 192.168.2.0/24) >>>>> - then adds a static route to the subnet of your VM. (ip route add >>>>> 10.5.5.0/24 gw 192.168.2.151) >>>>> (192.168.2.151 is the quantum gateway) >>>>> >>>>> I'm going to sleep, we will continue tomorrow. >>>>> >>>>> JB >>>>> >>>>> PS : You also should get some sleep :) >>>>> >>>>> >>>>> On 02/17/2013 03:53 AM, Chathura M. Sarathchandra Magurawalage wrote: >>>>> >>>>> oh that's weird. >>>>> >>>>> I still get this error. couldnt this be because I cannot ping the VM >>>>> in the first place?. Because as far as I know metadata takes care of ssh >>>>> keys. But what if you cant reach the VM in the first place? >>>>> >>>>> no instance data found in start-local >>>>> >>>>> ci-info: lo: 1 127.0.0.1 255.0.0.0 . >>>>> >>>>> ci-info: eth0 : 1 10.5.5.3255.255.255.0 fa:16:3e:a7:28:25 >>>>> >>>>> ci-info: route-0: 0.0.0.0 10.5.5.10.0.0.0 eth0 >>>>> UG >>>>> >>>>> ci-info: route-1: 10.5.5.00.0.0.0 255.255.255.0 eth0 U >>>>> >>>>> cloud-init start running: Sun, 17 Feb 2013 02:45:35 +. up 3.51 seconds >>>>> >>>>> 2013-02-17 02:48:25,840 - util.py[WARNING]: >>>>> 'http://169.254.169.254/2009-04-04/meta-data/instance-id' failed >>>>> [50/120s]: url error [timed out] >>>>> >>>>> 2013-02-17 02:49:16,893 - util.py[WARNING]: >>>>> 'http://169.254.169.254/2009-04-04/meta-data/instance-id' failed >>>>> [101/120s]: url error [timed out] >>>>> >>>>> 2013-02-17 02:49:34,912 - util.py[WARNING]: >>>>> 'http://169.254.169.254/2009-04-04/meta-data/instance-id' failed >>>>> [119/120s]: url error [timed out] >>>>> >>>>> 2013-02-17 02:49:35,913 - DataSourceEc2.py[CRITICAL]: giving up on md >>>>> after 120 seconds >>>>> >>>>> >>>>> >>>>> no instance data found in start >>>>> >>>>> Skipping profile in /etc/apparmor.d/disable: usr.sbin.rsyslogd >>>>> >>>>> * Starting AppArmor profiles [80G >>>>> [74G[ OK ] >>>>> >>>>> >>>>> >>>>> On 17 February 2013 02:41, Jean-Baptiste RANSY < >>>>&g
Re: [Openstack] Cant ping private or floating IP
oh that's weird. I still get this error. couldnt this be because I cannot ping the VM in the first place?. Because as far as I know metadata takes care of ssh keys. But what if you cant reach the VM in the first place? no instance data found in start-local ci-info: lo: 1 127.0.0.1 255.0.0.0 . ci-info: eth0 : 1 10.5.5.3255.255.255.0 fa:16:3e:a7:28:25 ci-info: route-0: 0.0.0.0 10.5.5.10.0.0.0 eth0 UG ci-info: route-1: 10.5.5.00.0.0.0 255.255.255.0 eth0 U cloud-init start running: Sun, 17 Feb 2013 02:45:35 +. up 3.51 seconds 2013-02-17 02:48:25,840 - util.py[WARNING]: 'http://169.254.169.254/2009-04-04/meta-data/instance-id' failed [50/120s]: url error [timed out] 2013-02-17 02:49:16,893 - util.py[WARNING]: 'http://169.254.169.254/2009-04-04/meta-data/instance-id' failed [101/120s]: url error [timed out] 2013-02-17 02:49:34,912 - util.py[WARNING]: 'http://169.254.169.254/2009-04-04/meta-data/instance-id' failed [119/120s]: url error [timed out] 2013-02-17 02:49:35,913 - DataSourceEc2.py[CRITICAL]: giving up on md after 120 seconds no instance data found in start Skipping profile in /etc/apparmor.d/disable: usr.sbin.rsyslogd * Starting AppArmor profiles [80G [74G[ OK ] On 17 February 2013 02:41, Jean-Baptiste RANSY < jean-baptiste.ra...@alyseo.com> wrote: > For me, it's normal that you are not able to curl 169.254.169.254 from > your compute and controller nodes : Same thing on my side, but my VM get > their metadata. > > Try to lunch an instance. > > JB > > > > On 02/17/2013 03:35 AM, Chathura M. Sarathchandra Magurawalage wrote: > > root@computernode:~# curl -v http://169.254.169.254 > * About to connect() to 169.254.169.254 port 80 (#0) > * Trying 169.254.169.254... > > root@controller:~# curl -v http://169.254.169.254 > * About to connect() to 169.254.169.254 port 80 (#0) > * Trying 169.254.169.254... > > > root@athena:~# iptables -L -n -v > Chain INPUT (policy ACCEPT 59009 packets, 22M bytes) > pkts bytes target prot opt in out source > destination > 59493 22M quantum-l3-agent-INPUT all -- * * 0.0.0.0/0 >0.0.0.0/0 > 59493 22M nova-api-INPUT all -- * * 0.0.0.0/0 > 0.0.0.0/0 > 484 73533 ACCEPT 47 -- * * 0.0.0.0/0 > 0.0.0.0/0 > > Chain FORWARD (policy ACCEPT 707 packets, 47819 bytes) > pkts bytes target prot opt in out source > destination > 707 47819 quantum-filter-top all -- * * 0.0.0.0/0 >0.0.0.0/0 > 707 47819 quantum-l3-agent-FORWARD all -- * * 0.0.0.0/0 > 0.0.0.0/0 > 707 47819 nova-filter-top all -- * * 0.0.0.0/0 > 0.0.0.0/0 > 707 47819 nova-api-FORWARD all -- * * 0.0.0.0/0 > 0.0.0.0/0 > > Chain OUTPUT (policy ACCEPT 56022 packets, 22M bytes) > pkts bytes target prot opt in out source > destination > 56022 22M quantum-filter-top all -- * * 0.0.0.0/0 >0.0.0.0/0 > 56022 22M quantum-l3-agent-OUTPUT all -- * * 0.0.0.0/0 > 0.0.0.0/0 > 56022 22M nova-filter-top all -- * * 0.0.0.0/0 > 0.0.0.0/0 > 56022 22M nova-api-OUTPUT all -- * * 0.0.0.0/0 > 0.0.0.0/0 > > Chain nova-api-FORWARD (1 references) > pkts bytes target prot opt in out source > destination > > Chain nova-api-INPUT (1 references) > pkts bytes target prot opt in out source > destination > 0 0 ACCEPT tcp -- * * 0.0.0.0/0 > 192.168.2.225tcp dpt:8775 > > Chain nova-api-OUTPUT (1 references) > pkts bytes target prot opt in out source > destination > > Chain nova-api-local (1 references) > pkts bytes target prot opt in out source > destination > > Chain nova-filter-top (2 references) > pkts bytes target prot opt in out source > destination > 56729 22M nova-api-local all -- * * 0.0.0.0/0 > 0.0.0.0/0 > > Chain quantum-filter-top (2 references) > pkts bytes target prot opt in out source > destination > 56729 22M quantum-l3-agent-local all -- * * 0.0.0.0/0 >0.0.0.0/0 > > Chain quantum-l3-agent-FORWARD (1 references) > pkts bytes target prot opt in out source > destination > > Chain quantum-l3-agent-INPUT (1 references) > pkts bytes target prot opt in out source > destination > 0 0 ACCEPT tcp -- * * 0.0.0.0/0 > 192.168.2.225tcp dpt:8775 > > Chain quantum-l3-agent-OUTPUT (1 references) > pkts bytes target prot opt in out sourc
Re: [Openstack] Cant ping private or floating IP
out source destination Chain quantum-l3-agent-POSTROUTING (1 references) pkts bytes target prot opt in out source destination 3726 247K ACCEPT all -- !qg-6f8374cb-cb !qg-6f8374cb-cb 0.0.0.0/0 0.0.0.0/0! ctstate DNAT 0 0 ACCEPT all -- * * 10.5.5.0/24 192.168.2.225 Chain quantum-l3-agent-PREROUTING (1 references) pkts bytes target prot opt in out source destination 0 0 DNAT tcp -- * * 0.0.0.0/0 169.254.169.254 tcp dpt:80 to:192.168.2.225:8775 Chain quantum-l3-agent-float-snat (1 references) pkts bytes target prot opt in out source destination Chain quantum-l3-agent-snat (1 references) pkts bytes target prot opt in out source destination 0 0 quantum-l3-agent-float-snat all -- * * 0.0.0.0/0 0.0.0.0/0 0 0 SNAT all -- * * 10.5.5.0/24 0.0.0.0/0to:192.168.2.151 Chain quantum-postrouting-bottom (1 references) pkts bytes target prot opt in out source destination 0 0 quantum-l3-agent-snat all -- * * 0.0.0.0/0 0.0.0.0/0 thanks. On 17 February 2013 02:25, Jean-Baptiste RANSY < jean-baptiste.ra...@alyseo.com> wrote: > Controller node : > # iptables -L -n -v > # iptables -L -n -v -t nat > > > > On 02/17/2013 03:18 AM, Chathura M. Sarathchandra Magurawalage wrote: > > You should be able to curl 169.254.169.254 from compute node, which I cant > at the moment. > > I have got the bridge set up in the l3_agent.ini > > > ___ Mailing list: https://launchpad.net/~openstack Post to : openstack@lists.launchpad.net Unsubscribe : https://launchpad.net/~openstack More help : https://help.launchpad.net/ListHelp
Re: [Openstack] Cant ping private or floating IP
erface "qg-6f8374cb-cb" type: internal Port "br0" Interface "br0" Bridge br-int Port br-int Interface br-int type: internal Port "tapf71b5b86-5c" tag: 1 Interface "tapf71b5b86-5c" type: internal Port patch-tun Interface patch-tun type: patch options: {peer=patch-int} Port "qr-4d088f3a-78" tag: 1 Interface "qr-4d088f3a-78" type: internal ovs_version: "1.4.0+build0" *Compute node:* *root@cronus:~# ip link show* 1: lo: mtu 16436 qdisc noqueue state UNKNOWN link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 2: eth0: mtu 1500 qdisc mq state UP qlen 1000 link/ether d4:ae:52:bb:a1:9d brd ff:ff:ff:ff:ff:ff 3: eth1: mtu 1500 qdisc noop state DOWN qlen 1000 link/ether d4:ae:52:bb:a1:9e brd ff:ff:ff:ff:ff:ff 4: eth0.2@eth0: mtu 1500 qdisc noqueue state UP link/ether d4:ae:52:bb:a1:9d brd ff:ff:ff:ff:ff:ff 5: br-int: mtu 1500 qdisc noop state DOWN link/ether ae:9b:43:09:af:40 brd ff:ff:ff:ff:ff:ff 9: qbr256f5ed2-43: mtu 1500 qdisc noqueue state UP link/ether c6:c0:df:64:c6:99 brd ff:ff:ff:ff:ff:ff 10: qvo256f5ed2-43: mtu 1500 qdisc pfifo_fast state UP qlen 1000 link/ether 76:25:8b:fd:90:3b brd ff:ff:ff:ff:ff:ff 11: qvb256f5ed2-43: mtu 1500 qdisc pfifo_fast master qbr256f5ed2-43 state UP qlen 1000 link/ether c6:c0:df:64:c6:99 brd ff:ff:ff:ff:ff:ff 13: br-tun: mtu 1500 qdisc noop state DOWN link/ether be:8c:30:78:35:48 brd ff:ff:ff:ff:ff:ff 15: vnet0: mtu 1500 qdisc pfifo_fast master qbr256f5ed2-43 state UNKNOWN qlen 500 link/ether fe:16:3e:57:ec:ff brd ff:ff:ff:ff:ff:ff *root@cronus:~# ip route show* default via 192.168.2.253 dev eth0.2 metric 100 10.10.10.0/24 dev eth0 proto kernel scope link src 10.10.10.12 192.168.2.0/24 dev eth0.2 proto kernel scope link src 192.168.2.234 *root@cronus:~# ovs-vsctl show* d85bc334-6d64-4a13-b851-d56b18ff1549 Bridge br-int Port "qvo0e743b01-89" tag: 4095 Interface "qvo0e743b01-89" Port "qvo256f5ed2-43" tag: 1 Interface "qvo256f5ed2-43" Port patch-tun Interface patch-tun type: patch options: {peer=patch-int} Port br-int Interface br-int type: internal Port "qvoee3d4131-2a" tag: 4095 Interface "qvoee3d4131-2a" Port "qvocbc816bd-3d" tag: 4095 Interface "qvocbc816bd-3d" Bridge br-tun Port br-tun Interface br-tun type: internal Port "gre-2" Interface "gre-2" type: gre options: {in_key=flow, out_key=flow, remote_ip="10.10.10.1"} Port "gre-1" Interface "gre-1" type: gre options: {in_key=flow, out_key=flow, remote_ip="10.0.0.3"} Port patch-int Interface patch-int type: patch options: {peer=patch-tun} Port "gre-3" Interface "gre-3" type: gre options: {in_key=flow, out_key=flow, remote_ip="127.0.0.1"} ovs_version: "1.4.0+build0" Thanks I appreciate your help. On 16 February 2013 16:49, Jean-Baptiste RANSY < jean-baptiste.ra...@alyseo.com> wrote: > Please provide files listed bellow : > > Controller Node : > /etc/nova/nova.conf > /etc/nova/api-paste.ini > /etc/quantum/l3_agent.ini > /etc/quantum/quantum.conf > /etc/quantum/dhcp_agent.ini > /etc/quantum/plugins/openvswitch/ovs_quantum_plugin.ini > /etc/quantum/api-paste.ini > /var/log/nova/*.log > /var/log/quantum/*.log > > Compute Node : > /etc/nova/nova.conf > /etc/nova/nova-compute.conf > /etc/nova/api-paste.ini > /etc/quantum/quantum.conf > /etc/quantum/plugins/openvswitch/ovs_quantum_plugin.ini > /var/log/nova/*.log > /var/log/quantum/*.log > > Plus, complete output of the following commands : > > Controller Node : > $ keystone endpoint-list > $ ip link show > $ ip route show > $ ip netns show > $ ovs-vsctl show > > Compute Node : > $ ip link show > $ ip route show > $ ovs-vsctl show > > Regards, > > Jean-Baptiste RANSY > > > > On 02/16/2013 05:32 PM, Chathura M. Sarathchandra Magurawalage wrote: > > Hello Jean, > > Thanks for your reply. > > I followed the instructions in > http://docs.openstack.org/folsom/basic-install/content/basic-install_network.html. > And my Controll
Re: [Openstack] Cant ping private or floating IP
Hello Jean, Thanks for your reply. I followed the instructions in http://docs.openstack.org/folsom/basic-install/content/basic-install_network.html. And my Controller and the Network-node is installed in the same physical node. I am using Folsom but without Network namespaces. But in the website you have provided it states that "If you run both L3 + DHCP services on the same node, you should enable namespaces to avoid conflicts with routes :" But currently quantum-dhcp-agent and quantum-l3-agent are running in the same node? Additionally the control node serves as a DHCP server for the local network ( Don't know if that would make and difference) Any idea what the problem could be? On 16 February 2013 16:21, Jean-Baptiste RANSY < jean-baptiste.ra...@alyseo.com> wrote: > Hello Chathura, > > Are you using Folsom with Network Namespaces ? > > If yes, have a look here : > http://docs.openstack.org/folsom/openstack-network/admin/content/ch_limitations.html > > > Regards, > > Jean-Baptsite RANSY > > > > On 02/16/2013 05:01 PM, Chathura M. Sarathchandra Magurawalage wrote: > > Hello guys, > > The problem still exists. Any ideas? > > Thanks > > On 15 February 2013 14:37, Sylvain Bauza wrote: > >> Metadata API allows to fetch SSH credentials when booting (pubkey I mean). >> If a VM is unable to reach metadata service, then it won't be able to get >> its public key, so you won't be able to connect, unless you specifically go >> thru a Password authentication (provided password auth is enabled in >> /etc/ssh/sshd_config, which is not the case with Ubuntu cloud archive). >> There is also a side effect, the boot process is longer as the instance >> is waiting for the curl timeout (60sec.) to finish booting up. >> >> Re: Quantum, the metadata API is actually DNAT'd from Network node to the >> Nova-api node (here 172.16.0.1 as internal management IP) : >> Chain quantum-l3-agent-PREROUTING (1 references) >> >> target prot opt source destination >> DNAT tcp -- 0.0.0.0/0169.254.169.254 tcp >> dpt:80 to:172.16.0.1:8775 >> >> >> Anyway, the first step is to : >> 1. grab the console.log >> 2. access thru VNC to the desired instance >> >> Troubleshooting will be easier once that done. >> >> -Sylvain >> >> >> >> Le 15/02/2013 14:24, Chathura M. Sarathchandra Magurawalage a écrit : >> >>> Hello Guys, >>> >>> Not sure if this is the right port but these are the results: >>> >>> *Compute node:* >>> >>> >>> root@computenode:~# netstat -an | grep 8775 >>> tcp0 0 0.0.0.0:8775 <http://0.0.0.0:8775> 0.0.0.0:* >>> LISTEN >>> >>> *Controller: * >>> >>> >>> root@controller:~# netstat -an | grep 8775 >>> tcp0 0 0.0.0.0:8775 <http://0.0.0.0:8775> 0.0.0.0:* >>> LISTEN >>> >>> *Additionally I cant curl 169.254.169.254 from the compute node. I am >>> not sure if this is related to not being able to PING the VM.* >>> >>> >>> curl -v http://169.254.169.254 >>> * About to connect() to 169.254.169.254 port 80 (#0) >>> * Trying 169.254.169.254... >>> >>> Thanks for your help >>> >>> >>> >>> - >>> Chathura Madhusanka Sarathchandra Magurawalage. >>> 1NW.2.1, Desk 2 >>> School of Computer Science and Electronic Engineering >>> University Of Essex >>> United Kingdom. >>> >>> Email: csar...@essex.ac.uk <mailto:csar...@essex.ac.uk> >>> chathura.sarathchan...@gmail.com <mailto:77.chath...@gmail.com >>> > >>> 77.chath...@gmail.com <mailto:77.chath...@gmail.com> >>> >>> >>> >>> On 15 February 2013 11:03, Anil Vishnoi >> vishnoia...@gmail.com>> wrote: >>> >>> If you are using ubuntu cloud image then the only way to log-in is >>> to do ssh with the public key. For that you have to create ssh key >>> pair and download the ssh key. You can create this ssh pair using >>> horizon/cli. >>> >>> >>> On Fri, Feb 15, 2013 at 4:27 PM, Sylvain Bauza >>> mailto:sylvain.ba...@digimind.com>> >>> >>> wrote: >>> >>> >>> Le 15/02/2013 11:42, Cha
Re: [Openstack] Cant ping private or floating IP
Hello guys, The problem still exists. Any ideas? Thanks On 15 February 2013 14:37, Sylvain Bauza wrote: > Metadata API allows to fetch SSH credentials when booting (pubkey I mean). > If a VM is unable to reach metadata service, then it won't be able to get > its public key, so you won't be able to connect, unless you specifically go > thru a Password authentication (provided password auth is enabled in > /etc/ssh/sshd_config, which is not the case with Ubuntu cloud archive). > There is also a side effect, the boot process is longer as the instance is > waiting for the curl timeout (60sec.) to finish booting up. > > Re: Quantum, the metadata API is actually DNAT'd from Network node to the > Nova-api node (here 172.16.0.1 as internal management IP) : > Chain quantum-l3-agent-PREROUTING (1 references) > > target prot opt source destination > DNAT tcp -- 0.0.0.0/0169.254.169.254 tcp dpt:80 > to:172.16.0.1:8775 > > > Anyway, the first step is to : > 1. grab the console.log > 2. access thru VNC to the desired instance > > Troubleshooting will be easier once that done. > > -Sylvain > > > > Le 15/02/2013 14:24, Chathura M. Sarathchandra Magurawalage a écrit : > >> Hello Guys, >> >> Not sure if this is the right port but these are the results: >> >> *Compute node:* >> >> >> root@computenode:~# netstat -an | grep 8775 >> tcp0 0 0.0.0.0:8775 <http://0.0.0.0:8775> 0.0.0.0:* >> LISTEN >> >> *Controller: * >> >> >> root@controller:~# netstat -an | grep 8775 >> tcp0 0 0.0.0.0:8775 <http://0.0.0.0:8775> 0.0.0.0:* >> LISTEN >> >> *Additionally I cant curl 169.254.169.254 from the compute node. I am not >> sure if this is related to not being able to PING the VM.* >> >> >> curl -v http://169.254.169.254 >> * About to connect() to 169.254.169.254 port 80 (#0) >> * Trying 169.254.169.254... >> >> Thanks for your help >> >> >> --**--** >> --**--**- >> Chathura Madhusanka Sarathchandra Magurawalage. >> 1NW.2.1, Desk 2 >> School of Computer Science and Electronic Engineering >> University Of Essex >> United Kingdom. >> >> Email: csar...@essex.ac.uk <mailto:csar...@essex.ac.uk> >> >> chathura.sarathchandra@gmail.**com> 77.chath...@gmail.com> >> 77.chath...@gmail.com <mailto:77.chath...@gmail.com> >> >> >> >> On 15 February 2013 11:03, Anil Vishnoi > vishnoia...@gmail.com>**> wrote: >> >> If you are using ubuntu cloud image then the only way to log-in is >> to do ssh with the public key. For that you have to create ssh key >> pair and download the ssh key. You can create this ssh pair using >> horizon/cli. >> >> >> On Fri, Feb 15, 2013 at 4:27 PM, Sylvain Bauza >> > <mailto:sylvain.bauza@**digimind.com >> >> >> >> wrote: >> >> >> Le 15/02/2013 11:42, Chathura M. Sarathchandra Magurawalage a >> écrit : >> >> >> How can I log into the VM from VNC? What are the credentials? >> >> >> You have multiple ways to get VNC access. The easiest one is >> thru Horizon. Other can be looking at the KVM command-line for >> the desired instance (on the compute node) and check the vnc >> port in use (assuming KVM as hypervisor). >> This is basic knowledge of Nova. >> >> >> >> nova-api-metadata is running fine in the compute node. >> >> >> Make sure the metadata port is avaible thanks to telnet or >> netstat, nova-api can be running without listening on metadata >> port. >> >> >> >> >> __**_ >> Mailing list: >> https://launchpad.net/~**openstack<https://launchpad.net/~openstack> >> >> <https://launchpad.net/%**7Eopenstack<https://launchpad.net/%7Eopenstack> >> > >> Post to : openstack@lists.launchpad.net >> >> <mailto:openstack@lists.**launchpad.net >> > >> Unsubscribe : >> https://launchpad.net/~**openstack<https://launchpad.net/~openstack> >> >> <https://launchpad.net/%**7Eopenstack<https://launchpad.net/%7Eopenstack> >> > >> >> More help : >> https://help.launchpad.net/**ListHelp<https://help.launchpad.net/ListHelp> >> >> >> >> >> -- Thanks & Regards >> --Anil Kumar Vishnoi >> >> >> > ___ Mailing list: https://launchpad.net/~openstack Post to : openstack@lists.launchpad.net Unsubscribe : https://launchpad.net/~openstack More help : https://help.launchpad.net/ListHelp
Re: [Openstack] Cant ping private or floating IP
Thanks for your reply. first of all I do not see the following rule in my iptables target prot opt source destination DNAT tcp -- 0.0.0.0/0169.254.169.254 tcp dpt:80 to: x.x.x.x:8775 <http://172.16.0.1:8775/> Please find the console log at the beginning of the post. Since I am using Ubuntu cloud image I am not able to log in to it through VNC console. I can't even ping. This is the main problem. Any help will greatly appreciated. On 15 February 2013 14:37, Sylvain Bauza wrote: > Metadata API allows to fetch SSH credentials when booting (pubkey I mean). > If a VM is unable to reach metadata service, then it won't be able to get > its public key, so you won't be able to connect, unless you specifically go > thru a Password authentication (provided password auth is enabled in > /etc/ssh/sshd_config, which is not the case with Ubuntu cloud archive). > There is also a side effect, the boot process is longer as the instance is > waiting for the curl timeout (60sec.) to finish booting up. > > Re: Quantum, the metadata API is actually DNAT'd from Network node to the > Nova-api node (here 172.16.0.1 as internal management IP) : > Chain quantum-l3-agent-PREROUTING (1 references) > > target prot opt source destination > DNAT tcp -- 0.0.0.0/0169.254.169.254 tcp dpt:80 > to:172.16.0.1:8775 > > > Anyway, the first step is to : > 1. grab the console.log > 2. access thru VNC to the desired instance > > Troubleshooting will be easier once that done. > > -Sylvain > > > > Le 15/02/2013 14:24, Chathura M. Sarathchandra Magurawalage a écrit : > >> Hello Guys, >> >> Not sure if this is the right port but these are the results: >> >> *Compute node:* >> >> >> root@computenode:~# netstat -an | grep 8775 >> tcp0 0 0.0.0.0:8775 <http://0.0.0.0:8775> 0.0.0.0:* >> LISTEN >> >> *Controller: * >> >> >> root@controller:~# netstat -an | grep 8775 >> tcp0 0 0.0.0.0:8775 <http://0.0.0.0:8775> 0.0.0.0:* >> LISTEN >> >> *Additionally I cant curl 169.254.169.254 from the compute node. I am not >> sure if this is related to not being able to PING the VM.* >> >> >> curl -v http://169.254.169.254 >> * About to connect() to 169.254.169.254 port 80 (#0) >> * Trying 169.254.169.254... >> >> Thanks for your help >> > > On 15 February 2013 11:03, Anil Vishnoi > vishnoia...@gmail.com>**> wrote: >> >> If you are using ubuntu cloud image then the only way to log-in is >> to do ssh with the public key. For that you have to create ssh key >> pair and download the ssh key. You can create this ssh pair using >> horizon/cli. >> >> >> On Fri, Feb 15, 2013 at 4:27 PM, Sylvain Bauza >> > <mailto:sylvain.bauza@**digimind.com >> >> >> >> wrote: >> >> >> Le 15/02/2013 11:42, Chathura M. Sarathchandra Magurawalage a >> écrit : >> >> >> How can I log into the VM from VNC? What are the credentials? >> >> >> You have multiple ways to get VNC access. The easiest one is >> thru Horizon. Other can be looking at the KVM command-line for >> the desired instance (on the compute node) and check the vnc >> port in use (assuming KVM as hypervisor). >> This is basic knowledge of Nova. >> >> >> >> nova-api-metadata is running fine in the compute node. >> >> >> Make sure the metadata port is avaible thanks to telnet or >> netstat, nova-api can be running without listening on metadata >> port. >> >> >> >> >> __**_ >> Mailing list: >> https://launchpad.net/~**openstack<https://launchpad.net/~openstack> >> >> <https://launchpad.net/%**7Eopenstack<https://launchpad.net/%7Eopenstack> >> > >> Post to : openstack@lists.launchpad.net >> >> <mailto:openstack@lists.**launchpad.net >> > >> Unsubscribe : >> https://launchpad.net/~**openstack<https://launchpad.net/~openstack> >> >> <https://launchpad.net/%**7Eopenstack<https://launchpad.net/%7Eopenstack> >> > >> >> More help : >> https://help.launchpad.net/**ListHelp<https://help.launchpad.net/ListHelp> >> >> >> >> >> -- Thanks & Regards >> --Anil Kumar Vishnoi >> >> >> > ___ Mailing list: https://launchpad.net/~openstack Post to : openstack@lists.launchpad.net Unsubscribe : https://launchpad.net/~openstack More help : https://help.launchpad.net/ListHelp
Re: [Openstack] Cant ping private or floating IP
Hello Anil, I can not ssh into the VM so I cant do ifconfig from vm. I am using quantum and, quantum-plugin-openvswitch-agent, quantum-dhcp-agent, quantum-l3-agent as described in the guide. Thanks. - Chathura Madhusanka Sarathchandra Magurawalage. 1NW.2.1, Desk 2 School of Computer Science and Electronic Engineering University Of Essex United Kingdom. Email: csar...@essex.ac.uk chathura.sarathchan...@gmail.com <77.chath...@gmail.com> 77.chath...@gmail.com On 15 February 2013 07:34, Anil Vishnoi wrote: > Did your VM got ip address ? Can you paste the output of ifconfig from > your vm. Are you using nova-network or quantum ? If quantum - which plugin > are you using ? > > > On Fri, Feb 15, 2013 at 4:28 AM, Chathura M. Sarathchandra Magurawalage < > 77.chath...@gmail.com> wrote: > >> Hello, >> >> I followed the folsom basic install instructions in >> http://docs.openstack.org/folsom/basic-install/content/basic-install_operate.html >> >> But now I am not able to ping either the private or the floating ip of >> the instances. >> >> Can someone please help? >> >> Instance log: >> >> [0.00] Initializing cgroup subsys cpuset >> [0.00] Initializing cgroup subsys cpu >> [0.00] Linux version 3.2.0-37-virtual (buildd@allspice) (gcc version >> 4.6.3 (Ubuntu/Linaro 4.6.3-1ubuntu5) ) #58-Ubuntu SMP Thu Jan 24 15:48:03 >> UTC 2013 (Ubuntu 3.2.0-37.58-virtual 3.2.35) >> [0.00] Command line: BOOT_IMAGE=/boot/vmlinuz-3.2.0-37-virtual >> root=LABEL=cloudimg-rootfs ro console=ttyS0 >> [0.00] KERNEL supported cpus: >> [0.00] Intel GenuineIntel >> [0.00] AMD AuthenticAMD >> [0.00] Centaur CentaurHauls >> [0.00] BIOS-provided physical RAM map: >> [0.00] BIOS-e820: - 0009bc00 (usable) >> [0.00] BIOS-e820: 0009bc00 - 000a (reserved) >> [0.00] BIOS-e820: 000f - 0010 (reserved) >> [0.00] BIOS-e820: 0010 - 7fffd000 (usable) >> [0.00] BIOS-e820: 7fffd000 - 8000 (reserved) >> [0.00] BIOS-e820: feffc000 - ff00 (reserved) >> [0.00] BIOS-e820: fffc - 0001 (reserved) >> [0.00] NX (Execute Disable) protection: active >> [0.00] DMI 2.4 present. >> [0.00] No AGP bridge found >> [0.00] last_pfn = 0x7fffd max_arch_pfn = 0x4 >> [0.00] x86 PAT enabled: cpu 0, old 0x70406, new 0x7010600070106 >> [0.00] found SMP MP-table at [880fdae0] fdae0 >> [0.00] init_memory_mapping: -7fffd000 >> [0.00] RAMDISK: 3776c000 - 37bae000 >> [0.00] ACPI: RSDP 000fd980 00014 (v00 BOCHS ) >> [0.00] ACPI: RSDT 7fffd7b0 00034 (v01 BOCHS BXPCRSDT >> 0001 BXPC 0001) >> [0.00] ACPI: FACP 7f80 00074 (v01 BOCHS BXPCFACP >> 0001 BXPC 0001) >> [0.00] ACPI: DSDT 7fffd9b0 02589 (v01 BXPC BXDSDT >> 0001 INTL 20100528) >> [0.00] ACPI: FACS 7f40 00040 >> [0.00] ACPI: SSDT 7fffd910 0009E (v01 BOCHS BXPCSSDT >> 0001 BXPC 0001) >> [0.00] ACPI: APIC 7fffd830 00072 (v01 BOCHS BXPCAPIC >> 0001 BXPC 0001) >> [0.00] ACPI: HPET 7fffd7f0 00038 (v01 BOCHS BXPCHPET >> 0001 BXPC 0001) >> [0.00] No NUMA configuration found >> [0.00] Faking a node at -7fffd000 >> [0.00] Initmem setup node 0 -7fffd000 >> [0.00] NODE_DATA [7fff8000 - 7fffcfff] >> [0.00] kvm-clock: Using msrs 4b564d01 and 4b564d00 >> [0.00] kvm-clock: cpu 0, msr 0:1cf7681, boot clock >> [0.00] Zone PFN ranges: >> [0.00] DMA 0x0010 -> 0x1000 >> [0.00] DMA320x1000 -> 0x0010 >> [0.00] Normal empty >> [0.00] Movable zone start PFN for each node >> [0.00] early_node_map[2] active PFN ranges >> [0.00] 0: 0x0010 -> 0x009b >> [0.00] 0: 0x0100 -> 0x0007fffd >> [0.00] ACPI: PM-Timer IO Port: 0xb008 >> [0.00] ACPI: LAPIC (acpi_id[0x00] lapic_id[0x00] enabled) >> [0.00] ACPI: IOAPIC (id[0x01]
[Openstack] Cant ping private or floating IP
Hello, I followed the folsom basic install instructions in http://docs.openstack.org/folsom/basic-install/content/basic-install_operate.html But now I am not able to ping either the private or the floating ip of the instances. Can someone please help? Instance log: [0.00] Initializing cgroup subsys cpuset [0.00] Initializing cgroup subsys cpu [0.00] Linux version 3.2.0-37-virtual (buildd@allspice) (gcc version 4.6.3 (Ubuntu/Linaro 4.6.3-1ubuntu5) ) #58-Ubuntu SMP Thu Jan 24 15:48:03 UTC 2013 (Ubuntu 3.2.0-37.58-virtual 3.2.35) [0.00] Command line: BOOT_IMAGE=/boot/vmlinuz-3.2.0-37-virtual root=LABEL=cloudimg-rootfs ro console=ttyS0 [0.00] KERNEL supported cpus: [0.00] Intel GenuineIntel [0.00] AMD AuthenticAMD [0.00] Centaur CentaurHauls [0.00] BIOS-provided physical RAM map: [0.00] BIOS-e820: - 0009bc00 (usable) [0.00] BIOS-e820: 0009bc00 - 000a (reserved) [0.00] BIOS-e820: 000f - 0010 (reserved) [0.00] BIOS-e820: 0010 - 7fffd000 (usable) [0.00] BIOS-e820: 7fffd000 - 8000 (reserved) [0.00] BIOS-e820: feffc000 - ff00 (reserved) [0.00] BIOS-e820: fffc - 0001 (reserved) [0.00] NX (Execute Disable) protection: active [0.00] DMI 2.4 present. [0.00] No AGP bridge found [0.00] last_pfn = 0x7fffd max_arch_pfn = 0x4 [0.00] x86 PAT enabled: cpu 0, old 0x70406, new 0x7010600070106 [0.00] found SMP MP-table at [880fdae0] fdae0 [0.00] init_memory_mapping: -7fffd000 [0.00] RAMDISK: 3776c000 - 37bae000 [0.00] ACPI: RSDP 000fd980 00014 (v00 BOCHS ) [0.00] ACPI: RSDT 7fffd7b0 00034 (v01 BOCHS BXPCRSDT 0001 BXPC 0001) [0.00] ACPI: FACP 7f80 00074 (v01 BOCHS BXPCFACP 0001 BXPC 0001) [0.00] ACPI: DSDT 7fffd9b0 02589 (v01 BXPC BXDSDT 0001 INTL 20100528) [0.00] ACPI: FACS 7f40 00040 [0.00] ACPI: SSDT 7fffd910 0009E (v01 BOCHS BXPCSSDT 0001 BXPC 0001) [0.00] ACPI: APIC 7fffd830 00072 (v01 BOCHS BXPCAPIC 0001 BXPC 0001) [0.00] ACPI: HPET 7fffd7f0 00038 (v01 BOCHS BXPCHPET 0001 BXPC 0001) [0.00] No NUMA configuration found [0.00] Faking a node at -7fffd000 [0.00] Initmem setup node 0 -7fffd000 [0.00] NODE_DATA [7fff8000 - 7fffcfff] [0.00] kvm-clock: Using msrs 4b564d01 and 4b564d00 [0.00] kvm-clock: cpu 0, msr 0:1cf7681, boot clock [0.00] Zone PFN ranges: [0.00] DMA 0x0010 -> 0x1000 [0.00] DMA320x1000 -> 0x0010 [0.00] Normal empty [0.00] Movable zone start PFN for each node [0.00] early_node_map[2] active PFN ranges [0.00] 0: 0x0010 -> 0x009b [0.00] 0: 0x0100 -> 0x0007fffd [0.00] ACPI: PM-Timer IO Port: 0xb008 [0.00] ACPI: LAPIC (acpi_id[0x00] lapic_id[0x00] enabled) [0.00] ACPI: IOAPIC (id[0x01] address[0xfec0] gsi_base[0]) [0.00] IOAPIC[0]: apic_id 1, version 17, address 0xfec0, GSI 0-23 [0.00] ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) [0.00] ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) [0.00] ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) [0.00] ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) [0.00] ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) [0.00] Using ACPI (MADT) for SMP configuration information [0.00] ACPI: HPET id: 0x8086a201 base: 0xfed0 [0.00] SMP: Allowing 1 CPUs, 0 hotplug CPUs [0.00] PM: Registered nosave memory: 0009b000 - 0009c000 [0.00] PM: Registered nosave memory: 0009c000 - 000a [0.00] PM: Registered nosave memory: 000a - 000f [0.00] PM: Registered nosave memory: 000f - 0010 [0.00] Allocating PCI resources starting at 8000 (gap: 8000:7effc000) [0.00] Booting paravirtualized kernel on KVM [0.00] setup_percpu: NR_CPUS:64 nr_cpumask_bits:64 nr_cpu_ids:1 nr_node_ids:1 [0.00] PERCPU: Embedded 28 pages/cpu @88007fc0 s82880 r8192 d23616 u2097152 [0.00] kvm-clock: cpu 0, msr 0:7fc13681, primary cpu clock [0.00] KVM setup async PF for cpu 0 [0.00] kvm-stealtime: cpu 0, msr 7fc0dd40 [0.00] Built 1 zonelists in Node order, mobility grouping on. Total pages: 515971 [0.00] Policy zone: DMA32 [0.00] Kernel command line: BOOT_IMA
[Openstack] Undo changes made by Quantum script
Hello, I am new to openstack and I have accidently inserted wrong Ip addresses in the Quantum script ( https://raw.github.com/EmilienM/openstack-folsom-guide/master/scripts/quantum-networking.sh) in openstack folsom guide ( http://docs.openstack.org/folsom/basic-install/content/basic-install_network.html). How can I undo the changes that was made by the script? Thanks, Chathura ___ Mailing list: https://launchpad.net/~openstack Post to : openstack@lists.launchpad.net Unsubscribe : https://launchpad.net/~openstack More help : https://help.launchpad.net/ListHelp