Re: [Openstack] PREROUTING 169.254.169.254 rule shoud not on Compute node.......

2011-05-10 Thread Narayan Desai
For what it's worth, we're running in a configuration similar to the
one in the attached diagram using VlanManager. When we moved the
nova-network service off of the machine with nova-api, we needed to
add an additional prerouting rule on the network server that prevented
the traffic from being sent out via NAT (which caused the source
address to be lost, resulting in a metadata resolution error). Once
the packets arrive at the api service with the correct source address,
they need a route back, via the nova-network server in order to get
the response packets onto the correct vlan. With a single nova-network
server, a static route will do. With multiple nova-network instances
on different systems, things get a little more complicated. We ended
up setting up route distribution via quagga between the nova-api
server, and the nova-network servers. This ensures that nova-api knows
which nova-network instance to use to reach any particular project
network.
 -nld

On Tue, May 10, 2011 at 9:08 PM, 郭耀謙 tonyt...@gmail.com wrote:
 Hello , guys
 There's a problem while separate instance's network and nova-management
 network.
 EX.
 Nova management network : 192.168.1.0/24  eth0
 Instance network               :  10.0.0.0/12      eth1 bridge to br100
 During cloud-setup :
 Instance try to retrieve metadata from 169.254.169.254.
 Instances(10.0.0.0/12) request 169.254.169.254 PREROUTING from
 gateway(nova-network).
 But If PREROUTING rule is already been set on nova-Compute node, instance
 request will be redirected on VM host instead of nova-network host.
 So If your topology is like A diadram from StackOps , Plz Check iptables
 rule on Compute nodes.
 -A PREROUTING -d 169.254.169.254/32 -p tcp -m tcp --dport 80 -j DNAT
 --to-destination 192.168.1.2:8773
 And del this rule , your instance will get metadata correctly




 ___
 Mailing list: https://launchpad.net/~openstack
 Post to     : openstack@lists.launchpad.net
 Unsubscribe : https://launchpad.net/~openstack
 More help   : https://help.launchpad.net/ListHelp



___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] PREROUTING 169.254.169.254 rule shoud not on Compute node.......

2011-05-11 Thread Narayan Desai
We're using quagga for route exchange (via ospf) between the network
servers, and the api server. The api server peers with each network
server. This is about as simple a quagga setup as you can have.
Basically, when nova-network starts up a new project network, quagga
begins to advertise that network into ospf. The api server sees that
advertisement and can use that to reach the node (for metadata
requests).

Incidentally, you might want to lock down your POSTROUTING rule to
ensure that vms can't hit your management network. IIRC, we were able
to use a similar rule, except with a specific host and port as opposed
to a blanket rule for the whole network.
 -nld

2011/5/10 郭耀謙 tonyt...@gmail.com:
 #iptables -t nat -A  nova-network-POSTROUTING -s 10.0.0.0/12 -d
 192.168.1.0/24 -j ACCEPT
 That's what I did on nova-network host .
 btw , I always isolate nova-network.
 I'm interesting about quagga in your environment.

 2011/5/11 Narayan Desai narayan.de...@gmail.com

 For what it's worth, we're running in a configuration similar to the
 one in the attached diagram using VlanManager. When we moved the
 nova-network service off of the machine with nova-api, we needed to
 add an additional prerouting rule on the network server that prevented
 the traffic from being sent out via NAT (which caused the source
 address to be lost, resulting in a metadata resolution error). Once
 the packets arrive at the api service with the correct source address,
 they need a route back, via the nova-network server in order to get
 the response packets onto the correct vlan. With a single nova-network
 server, a static route will do. With multiple nova-network instances
 on different systems, things get a little more complicated. We ended
 up setting up route distribution via quagga between the nova-api
 server, and the nova-network servers. This ensures that nova-api knows
 which nova-network instance to use to reach any particular project
 network.
  -nld

 On Tue, May 10, 2011 at 9:08 PM, 郭耀謙 tonyt...@gmail.com wrote:
  Hello , guys
  There's a problem while separate instance's network and nova-management
  network.
  EX.
  Nova management network : 192.168.1.0/24  eth0
  Instance network   :  10.0.0.0/12  eth1 bridge to br100
  During cloud-setup :
  Instance try to retrieve metadata from 169.254.169.254.
  Instances(10.0.0.0/12) request 169.254.169.254 PREROUTING from
  gateway(nova-network).
  But If PREROUTING rule is already been set on nova-Compute node,
  instance
  request will be redirected on VM host instead of nova-network host.
  So If your topology is like A diadram from StackOps , Plz Check iptables
  rule on Compute nodes.
  -A PREROUTING -d 169.254.169.254/32 -p tcp -m tcp --dport 80 -j DNAT
  --to-destination 192.168.1.2:8773
  And del this rule , your instance will get metadata correctly
 
 
 
 
  ___
  Mailing list: https://launchpad.net/~openstack
  Post to : openstack@lists.launchpad.net
  Unsubscribe : https://launchpad.net/~openstack
  More help   : https://help.launchpad.net/ListHelp
 
 


 ___
 Mailing list: https://launchpad.net/~openstack
 Post to : openstack@lists.launchpad.net
 Unsubscribe : https://launchpad.net/~openstack
 More help   : https://help.launchpad.net/ListHelp



___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Real World Nova Deployments

2011-05-16 Thread Narayan Desai
We've got a system comprised of 336 compute nodes, a head node with
everything else except for the network and volume pieces, 12 volume
servers, and 2 network servers.

We're using mysql. We've deployed using VlanManager. We deploy using a
custom node imaging system here for base builds and bcfg2 for
configuration management.
 -nld

On Mon, May 16, 2011 at 3:43 PM, Chuck Short chuck.sh...@canonical.com wrote:
 Hi,

 As interest continues to grow in the openstack project and with nova in
 particular. I would like to get a better idea of how people are
 deploying multi machine environemnts and hopefully start a discussion
 about it:

 * Are you using mysql or postgresql or sqllite?
 * Are you using flat networking or something else?
 * How are you deploying nova?

 Regards
 chuck

 ___
 Mailing list: https://launchpad.net/~openstack
 Post to     : openstack@lists.launchpad.net
 Unsubscribe : https://launchpad.net/~openstack
 More help   : https://help.launchpad.net/ListHelp


___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] about vlan and switch

2011-07-21 Thread Narayan Desai
We had to preconfigure the vlan tags and set all network ports for
nova-compute nodes to trunk them in advance on our switching gear.
(BNT and Juniper both, but I've also needed to do it on Cisco gear) I
think that is a pretty common requirement for managed switches.
 -nld

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Monitoring rabbitmq messages

2011-08-11 Thread Narayan Desai
When we were having rabbitmq problems, we would use rabbitmqctl
list_queues to see what the queue depth for each nova service was.
While this doesn't show what the throughput is, it does let you know
when things start to get backed up.
 -nld

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] OpenStack + RDMA + Infiniband

2011-10-03 Thread Narayan Desai
I suspect that the original poster was looking for instance access
(mediated in some way) to IB gear. When we were trying to figure out
how to best use our IB gear inside of openstack, we decided that it
was too risky to try exposing IB at the verbs layer to instances
directly, since the security model doesn't appear to have a good way
to prevent administrative commands from being issued from untrusted
instances.

We decided to use to IB as fast plumbing for data movement (using
IPoIB) and have ended up with pretty nice I/O performance to the
volume service, etc. We haven't managed to use it for much more than
that at this point.
 -nld

On Mon, Oct 3, 2011 at 11:44 AM, Caitlin Bestler
caitlin.best...@nexenta.com wrote:
 Nick Khamis asked:

 I was wondering if any effort has been made regarding supporting
 infiniband
devices to support I/O Virtualization, RDMA etc..? If so, can you
 please direct
me to the latest documentation.


 Adding RDMA support to OpenStack will be a challenge with the current
 software architecture.

 What is really needed is adding zero-copy I/O to eventlet socket I/O.
 Explicit RDMA semantics
 being visible to OpenStack code would only be possible if *every* device
 in a cluster supported
 RDMA, and that would happen so seldom that the code taking explicit
 advantage of RDMA would
 Become a maintenance liability.

 But refactoring the eventlet sockets I/O to use I/O virtualization when
 available is distinctly possible.
 It would be major work, but would greatly reduce the memory bandwidth
 consumption. What you
 would need to do is convince one of the high speed NIC vendors (10 GbE
 or InfiniBand) that it was
 worthwhile to invest in some Python libraries.


 ___
 Mailing list: https://launchpad.net/~openstack
 Post to     : openstack@lists.launchpad.net
 Unsubscribe : https://launchpad.net/~openstack
 More help   : https://help.launchpad.net/ListHelp


___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] OpenStack + RDMA + Infiniband

2011-10-04 Thread Narayan Desai
On Mon, Oct 3, 2011 at 4:21 PM, Caitlin Bestler
caitlin.best...@nexenta.com wrote:


 Narayan Desai wrote:


 I suspect that the original poster was looking for instance access
 (mediated in some way) to IB gear.
 When we were trying to figure out how to best use our IB gear inside
 of openstack, we decided that
 it was too risky to try exposing IB at the verbs layer to instances
 directly, since the security model
 doesn't appear to have a good way to prevent administrative commands
 from being issued from
 untrusted instances.

 We decided to use to IB as fast plumbing for data movement (using
 IPoIB) and have ended up with
 pretty nice I/O performance to the volume service, etc. We haven't
 managed to use it for much more
 than that at this point.

 There's no reason to expect use of IPoIB to end up providing better
 TCP/IP service for large bulk data
 transfer than you would get from a quality Ethernet NIC. But if you have
 an existing IB infrastructure
 it is certainly worth considering. You should experiment to see whether
 you get better performance
 under load form IPoIB in connected mode as opposed to trying SDP.

I suppose that is true, if your link speeds are the same. We're
getting (without much effort) 3 GB/s over IPoIB (connected mode, etc).

 Either IPoIB or SDP should be accessible via a standard sockets
 interface,meaning they could be
 plugged in without modifying the Python code or Python libraries.

Yeah, that is exactly what we did. We used addresses on the IPoIB
layer 3 network to get all of our I/O traffic going over that instead
of ethernet.

 The response to congestion by an IB network is different than the
 response from a TCP network,
 and the response of a TCP network simulated over IPoIB is something else
 entirely. So you'd want
 to do your evaluation with realistic traffic patterns.

Yeah, in our case, the system was specced like an HPC cluster, so the
management network is pretty anemic compared with QDR.
 -nld

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Does Nova works with several netwoks and VLANs ?

2011-11-06 Thread Narayan Desai
This looks more or less right. We have been running a setup like you
are describing here for quite a while, and we've found it to be stable
(and easier to setup than a lot of the other network options, IMO).

When debugging this sort of setup, trunking setup problems on the
switch are often the culprit. Here is a list of things to verify after
firing up an instance on each project network:
on the nova-network server:
 - is there a vlanX for each project vlan
 - is the brXX interface configured with the appropriate gateway
address for the project network
 - is the bridging setup properly (ie does vlanXX belong to bridge brXX)
 - is there a dnsmasq process running for each project network (it
will be a pair of processes, iirc)

on the nova-compute node:
 - is there a vlanX interface for each project vlan that has an
instance on the local machine?
 - brXX should have no ip address settings
 - is vlanXX connected to brXX?
 - is vnetX connected to brXX?

Another thing to check is if the instance is getting an ip address
from dhcp or not.

If you suspect trunking setup problems, you can add an ip address on
the project network to brXX on the nova-compute node, and try pinging
the .1 gateway address. If that doesn't work, your trunking setup is
broken. Make sure that the vlan is setup tagged on the switch port the
compute node is attached to, and that it is also passed upstream
properly in your network.

You should also check your nova-network logs.

Hope this helps.
 -nld

On Sat, Nov 5, 2011 at 7:27 PM, Razique Mahroua
razique.mahr...@gmail.com wrote:
 Hi all,
 I'm trying to create an advanced architecture with Nova, that uses several
 projects per user, and multiple networks (one network per project)
 The networks have their own vlan :
 id   IPv4               IPv6           start address   DNS1           DNS2
         VlanID         project         uuid
 12   192.168.2.0/24     None           192.168.2.3     None           None
         100             first_project   None
 13   192.168.3.0/27     None           192.168.3.3     None           None
         50             another_project None
 It looks like while the first project runs wells (creates instances have
 connectivity and can be reached), while the instances created via the second
 project are unreacheable.
 Both have the right SG rules, and both networks create the rights VLANS :
 # /proc/net/vlan/config
 VLAN Dev name | VLAN ID
 Name-Type: VLAN_NAME_TYPE_PLUS_VID_NO_PAD
 vlan100        | 100  | eth0
 vlan50         | 50  | eth0
 # brctl show
 br100 8000.02163e137a78 no vlan100
 br50 8000.02163e447ed8 no vlan50
 # ip route sh
 192.168.3.0/27 dev br50  proto kernel  scope link  src 192.168.3.1
 192.168.2.0/24 dev br100  proto kernel  scope link  src 192.168.2.1

 - Can I have several VLANS per physical interface ?
 - Do I need to create everytime a new project, or can I create all my
 networks and link them to the same project, but be able to specify which
 network to use everytime I spawn a new instance ?
 - Is it possible to allow some communication between the VLANS (for
 instance, hosts into the VLAN 50 should only be able to contact hosts into
 VLAN100 on port 443) ? Does the security groups can manage per VLAN rules ?
 Thanks, i'm a bit desperate here :)



 ___
 Mailing list: https://launchpad.net/~openstack
 Post to     : openstack@lists.launchpad.net
 Unsubscribe : https://launchpad.net/~openstack
 More help   : https://help.launchpad.net/ListHelp



___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


[Openstack] Expectations for tracebacks

2011-12-15 Thread Narayan Desai
Hello all. We've recently upgraded our cactus system to more recent
code. In the process of doing this, we've started logging whenever we
get tracebacks out of any of the openstack components we are running.
Some of these are clearly bugs, while others correspond to normal
operational conditions (like the case where all compute resources are
occupied). My question is, what is the project position on tracebacks?
Should they be considered to be expected, or should we be treating
them as bugs and filing tickets/patches for them?
thanks.
 -nld

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Please stop the devstack non-sense!

2012-03-21 Thread Narayan Desai
Ghe, while you're right that these two workloads are different, deployers
need developers to use a representative environment during development, or
the code doesn't work when it hits real deployments. We've now been bitten
during our initial deployment of cactus, our upgrade to diablo, and our
recent tests preparing for the essex upgrade because we can't run our
management infrastructure on a single system.

During cactus, we had issues when we tried to run multiple nova-network
servers distinctly from the api service. IIRC during the diablo release, we
had issues with keystone and horizon. This time, we had issues with
glance/keystone integration. I'm not saying that things haven't improved,
it just seems that each release has a new issue caused by the assumption
that all services will be running on the same host.

As we get more users with large deployments, these sorts of issues will
only become a bigger deal, and will hinder large scale adoption.
 -nld

On Wed, Mar 21, 2012 at 3:26 AM, Ghe Rivero ghe.riv...@stackops.com wrote:

 Hi,
 we are facing two differents problems here. We have developers and
 final users, and both of them with different expectations about what to get
 from OpenStack. Developers wants an easy way to test untested code, new
 cool-probably-broken features and be able to change immediately - devstack
 is the perfect tool for this . On the other hand, final users just want a
 working easy to deploy system, without care if the latest
 cool-probably-broken feature is included (I bet they prefer it to not be).
 But the truth is that OpenStack can't ignore any of them. Nobody will use a
 program which is hard to install, deploy, test or develop on it. Some
 consensus will be necessary in the user side (I think we all agree that
 development is ok with devstack) As Justin pointed before,
 http://summit.openstack.org/sessions/view/26, can be a good starting
 point, defining a common set of minimums that every package should comply
 with (file/dir locations and perms, minimal contents of config files, users
 created, python external modules requiriments/ minimal versions, ...), so
 when someone complaint about something, we know that the installation has
 some minimal standards that it follows (just a quick idea that just came to
 my mind to help debugging users installations,  a simple script, that use
 paste.openstack.org, and post there the config files, daemons runnings,
 last lines of some logs, version of pkgs installed...)

 Ghe Rivero

 On Wed, Mar 21, 2012 at 5:20 AM, Joshua Harlow harlo...@yahoo-inc.comwrote:

  Another idea:

 *http://meego.gitorious.org/meego-developer-tools/spectacle
 *
 That python code seems to be able to take a yaml defintion and generate
 either rpm specfiles or debian pkg files.

 It might be forked or extended (or both) and used to generate the initial
 set of package definitions for openstack in a non-pkg specific format...


 On 3/20/12 11:01 AM, Justin Santa Barbara jus...@fathomdb.com wrote:

 Hi Thomas,

 I think devstack has done a lot for the developer's use-case, but I
 believe we should also have a official / semi-official project that does
 some sort of packaging to help the production use-case.  I've proposed a
 summit discussion: http://summit.openstack.org/sessions/view/26

 The background: I want a semi-production deployment, but as a developer I
 still want to be able to edit the code (which makes packages inconvenient).
  devstack is orientated towards e.g. wiping the databases.

 I'm hoping that all the various OS packagers can work together, or at
 least tell us what sucks.  As a community, we should solve these problems
 once, and the OpenStack project shouldn't treat them as externalities.
  I've been doing some initial coding here:
 https://github.com/justinsb/openstack-simple-config

 The first use case I'm trying to solve is single node installation of
 OpenStack that is as easy as possible, but also isn't painting the user
 into the corner.  Think apt-get openstack, then the user finds they like
 it and grows to a 4 node cluster, all the way up to a 100 node cluster.  So
 it uses KVM, FlatManager, config drive injection, Postgresql, etc. - I'm
 afraid it is still quite opinionated!  I have Keystone, Glance  Nova
 installing.  I'm using supervisord to avoid any OS dependencies/flamewars,
 but I would imagine that any OS packager could move it to their preferred
 init.d flavor easily.  Swift is next on my list - I was facing the problem
 that the number of replicas isn't changeable, though I have a patch for
 that now.

 If you'd like to work together, I'd love to collaborate (and that holds
 for anyone doing packaging).  I'm hanging out in #openstack-packaging

 Justin




 ___
 Mailing list: https://launchpad.net/~openstack
 Post to : openstack@lists.launchpad.net
 Unsubscribe : https://launchpad.net/~openstack
 More help   : https://help.launchpad.net/ListHelp




 --
 Ghe 

Re: [Openstack] OpenStack and routing configuration

2012-04-20 Thread Narayan Desai
We needed to setup something similar when we split out the
nova-network service to a different host than nova-api in cactus, so
that instances could get to the metadata service. It was pretty simple
to make quagga work, but then we needed to add a rule to bypass NAT.
Since this was just for the metadata service, it was a tightly scoped
rule, but in your case, you'll probably want something that covers
your whole intranet. You'll probably still want NAT for internet bound
traffic.
 -nld

2012/4/19 Jérôme Gallard jeronimo...@gmail.com:
 Hi everyone,

 I would like to know if someone has already tried to setup routing
 configurations with nova-network.

 From my understanding, nova-network only deals with NAT, but I am
 thinking about configurations to directly routes my virtual networks
 on my intranet.

 I am thinking to develop a new driver for the network manager which
 can configure a RIP router (Quagga, ...).

 Any feedbacks ?

 Thanks.

 Regards,
 Jérôme

 ___
 Mailing list: https://launchpad.net/~openstack
 Post to     : openstack@lists.launchpad.net
 Unsubscribe : https://launchpad.net/~openstack
 More help   : https://help.launchpad.net/ListHelp

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Sharing disk: OCFS2 or GFS2?

2012-04-24 Thread Narayan Desai
As far as I know, the current volume service doesn't support
connecting the same volume to multiple instances at the same time, so
neither of these can work directly through nova apis.
 -nld

On Tue, Apr 24, 2012 at 4:44 AM, Daniel Martinez danie...@gmail.com wrote:
 Hello everyone.

 My setup is simple. A volumen shared by iSCSI from our Netapp storage
 that I would like to use it for the instances running on our
 nova-compute nodes.

 Has anyone tried to use OCFS2 or GFS2 as FS via iSCSI mounted on the
 nova-computes as a sharing disk and running the instances into it?

 The plan b is to create a volumen by node and to format it as ext3 or ext4.

 Any recomendations?

 Thanks!

 Regards,
 Daniel

 ___
 Mailing list: https://launchpad.net/~openstack
 Post to     : openstack@lists.launchpad.net
 Unsubscribe : https://launchpad.net/~openstack
 More help   : https://help.launchpad.net/ListHelp

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] SAN and Fibrechannel with OpenStack

2012-05-16 Thread Narayan Desai
I'm not sure that it would be particularly easy to make nova-volume
support clustered filesystems; the current model only supports
attaching a volume to a single instance at a time. Aside from that, it
shouldn't be too hard to use fc as the data path instead of iscsi.
We're looking at using iSER in a similar capacity.
 -nld

On Wed, May 16, 2012 at 4:05 AM, Wolfgang Hennerbichler
wolfgang.hennerbich...@risc-software.at wrote:
 dear openstack godfathers;

 I do plan to migrate from crappy vmware and some ibm based cloud stack to
 openstack and kvm. here's the thing: I am lucky enough to have decent
 hardware, all the compute nodes are interconnected via fibre channel.
 so I don't want and don't need  iscsi. do you think I can make it with
 something like clvm? I read through the docs of openstack, but I am not
 really sure now if I can make clvm fly without hacking around in openstack
 (and nova-volume) too much, especially when it comes to live migration and
 so on...

 I realize OpenStack was not built for SAN and FC, but I would really like to
 hear your opinions on that.
 Thanks,

 Wolfgang
 --
 DI (FH) Wolfgang Hennerbichler
 Software Development
 Unit Advanced Computing Technologies
 RISC Software GmbH
 A company of the Johannes Kepler University Linz

 ___
 Mailing list: https://launchpad.net/~openstack
 Post to     : openstack@lists.launchpad.net
 Unsubscribe : https://launchpad.net/~openstack
 More help   : https://help.launchpad.net/ListHelp

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] dhcp is not leasing an ip address in vlan mode

2012-05-30 Thread Narayan Desai
This sounds like it might be working properly. In VLAN mode, all
instances are connected to one of the project vlans. The .1 address
(gateway, dhcp, etc) exists on an interface on the nova-network node
(or one of them, in the case that you are running multiple. This
interface is bridged to a tagged interface on the appropriate vlan
tag. On the nova-compute nodes, a vnet interface for the instance is
bridged to the vlan tagged interface. On the compute node, there isn't
an IP interface on this network, so the private IP for instances isn't
reachable, even if the instance is running on the same node.

The canonical test for correct network function is if an instance is
reachable via ping from the nova-network server that is currently
serving the instance's project network.
hth
 -nld

On Wed, May 30, 2012 at 5:42 PM, Vijay vija...@yahoo.com wrote:
 Hello,
 I am trying install Essex in VLAN mode on multiple compute nodes.

 I am able to lauch instances on controller (which also runs nova-compute)
 and ping/ssh those instances.
 I am able to launch instances on compute only node. However, I cannot ping
 the VM launched  on compute only node.
 When i did the euca-get-console-output on that instance, I see that it is
 not getting an IP leased from DHCP .. Because of that it is not able to
 reach metadata server.
 Any help is appreciated.

 Console output is
 udhcpc (v1.17.2) started
 Sending discover...
 Sending discover...
 Sending discover...
 No lease, forking to background
 starting DHCP forEthernet interface eth0  [  OK  ]
 cloud-setup: checking
 http://169.254.169.254/2009-04-04/meta-data/instance-id
 wget: can't connect to remote host (169.254.169.254): Network is unreachable
 cloud-setup: failed 1/30: up 17.71. request failed
 nova.conf:
 --dhcpbridge_flagfile=/etc/nova/nova.conf
 --dhcpbridge=/usr/local/bin/nova-dhcpbridge
 --logdir=/var/log/nova
 --state_path=/var/lib/nova
 --lock_path=/var/lock/nova
 --force_dhcp_release=True
 --use_deprecated_auth
 --iscsi_helper=tgtadm
 --verbose
 --vncserver_listen=0.0.0.0
 --sql_connection=mysql://novadbadmin:novasecret@192.168.198.85/nova
 --daemonize
 --s3_host=192.168.198.85
 --rabbit_host=192.168.198.85
 --cc_host=192.168.198.85
 --ospi_host=192.168.198.85
 --ec2_host=192.168.198.85
 --ec2_url=http://192.168.198.85:8773/services/Cloud
 --nova_url=http://192.168.198.85:8774/v1.1/

 # VLAN mode
 --flat_interface=eth1
 --flat_injected=False
 --flat_network_bridge=br100
 --flat_network_dhcp_start=192.168.4.2

 --network_manager=nova.network.manager.VlanManager
 --vlan_interface=eth1
 --public_interface=vlan100
 --allow_same_net_traffic=True
 --fixed_range=192.168.4.0/24
 --network_size=256
 --FAKE_subdomain=ec2
 --routing_source_ip=192.168.198.85
 --glance_api_servers=192.168.198.85:9292
 --image_service=nova.image.glance.GlanceImageService
 --iscsi_ip_prefix=192.168.
 --connection_type=libvirt
 --libvirt_type=qemu

 # Keystone
 --auth_strategy=keystone
 --api_paste_config=/etc/nova/api-paste.ini
 --keystone_ec2_url=http://192.168.198.85:5000/v2.0/ec2tokens




 Thanks,
 -vj

 ___
 Mailing list: https://launchpad.net/~openstack
 Post to     : openstack@lists.launchpad.net
 Unsubscribe : https://launchpad.net/~openstack
 More help   : https://help.launchpad.net/ListHelp


___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] dhcp is not leasing an ip address in vlan mode

2012-06-01 Thread Narayan Desai
Along these lines, it seemed simplest if we just ran nova-api-metadata on
each nova-network server. That way we could get rid of the awful ospf setup
for metadata and just let the network aggregation do its job.
 -nld

On Fri, Jun 1, 2012 at 1:58 PM, Vishvananda Ishaya vishvana...@gmail.comwrote:

 yes it can.  The best way is to run nova-api-metadata on every host so the
 request can go locally.  Alternatively you can set the metadata_host config
 option on your compute hosts to the ip of a nova-api server somewhere else.

 you might have to be careful which interface the ip metadata_host is on.
 It defaults to my_ip, but i have seen it do odd things if the metadata_host
 is on a different ethernet device than the vms, so you might have to
 manually set it to a different ip.

 Vish

 On Jun 1, 2012, at 9:11 AM, Vijay wrote:

 I did have a problem in vlan trunking on the switch. I fixed it. Now, I am
 able to ping and ssh the instance that is launched on the compute node from
 the controller. However, when I look into euca-get-console-output of that
 instance on compute node, I still see that it is not able to connect
 to 169.254.169.254 (metadata service). But, I see a private ip address
 getting leased correctly. Because of this I am able to ping and ssh
 successfully from CONTROLLER ONLY (not from compute node).
 I am not sure if this is the correct behavior. But, in case of flatDHCP
 this metadata connection should be successful. Otherwise, instances cannot
 be pinged/sshed in flatDHCP mode.

 Can VLAN be run in multi-host mode like it is done in flatDHCP mode as
 suggested by Sergio Ariel below? (with multi_host set to true and running
 nova-network running)


 euca-get-console-output log

 Sending discover...
 Sending select for 192.168.4.5...
 Lease of 192.168.4.5 obtained, lease time 120
 starting DHCP forEthernet interface eth0  [  OK  ]
 cloud-setup: checking
 http://169.254.169.254/2009-04-04/meta-data/instance-id
 wget: can't connect to remote host (169.254.169.254): Connection timed out
 cloud-setup: failed 1/30: up 9.84. request failed

 Thanks,
 -vj
   *From:* Sergio Ariel de la Campa Saiz saca...@gmv.com
 *To:* Vishvananda Ishaya vishvana...@gmail.com; Vijay vija...@yahoo.com

 *Cc:* openstack@lists.launchpad.net openstack@lists.launchpad.net
 *Sent:* Friday, June 1, 2012 5:12 AM
 *Subject:* RE: [Openstack] dhcp is not leasing an ip address in vlan mode
 **
  Hi:

 I had a similar problem as Vijay:

 Network controller assigns a private ip address to the vm launched oncompute 
 node.
 However, I still cannot ping this ip address from the network(controllernode).
 I am running nova-network service only on the controller.

 can't connect to remote host (169.254.169.254): Network is unreachable
 I solved it when I installed nova-network in all my compute nodes.
 I don´t use NAT but only routing, so each node is the default gateway to
  instances that are running on it.

 I don´t know if this workaround is good for you, but it is the best I got.

 Regards


   *Sergio Ariel *
 *de la Campa Saiz*
 GMV-SES Infraestructura /
 GMV-SES Infrastructure



  *GMV*
 Isaac Newton, 11
 P.T.M. Tres Cantos
 E-28760 Madrid
 Tel.
 +34 91 807 21 00
 Fax
 +34 91 807 21 99
  www.gmv.com




  *De:* openstack-bounces+sacampa=gmv@lists.launchpad.net [
 openstack-bounces+sacampa=gmv@lists.launchpad.net] En nombre de
 Vishvananda Ishaya [vishvana...@gmail.com]
 *Enviado el:* viernes, 01 de junio de 2012 8:35
 *Para:* Vijay
 *CC:* openstack@lists.launchpad.net
 *Asunto:* Re: [Openstack] dhcp is not leasing an ip address in vlan mode
 **
  do you see sent and received packets on the vlan?  I would suspect that
 you actually don't have the vlans trunked on the ports so the packets
 aren't making it across the switch.
 **
 Vish
 **
 On May 31, 2012, at 9:53 AM, Vijay wrote:
 **

  Thanks for the reply. Network controller assigns a private ip address to
 the vm launched on compute node. However, I still cannot ping this ip
 address from the network(controller node). I am running nova-network
 service only on the controller.

 Thanks,**-vj**
   *From:* Narayan Desai narayan.de...@gmail.com
 *To:* Vijay vija...@yahoo.com
 *Cc:* openstack@lists.launchpad.net openstack@lists.launchpad.net
 *Sent:* Wednesday, May 30, 2012 5:28 PM
 *Subject:* Re: [Openstack] dhcp is not leasing an ip address in vlan mode
 **This sounds like it might be working properly. In VLAN mode, all**instances
 are connected to one of the project vlans. The .1 address**(gateway,
 dhcp, etc) exists on an interface on the nova-network node**(or one of
 them, in the case that you are running multiple. This**interface is
 bridged to a tagged interface on the appropriate vlan**tag. On the
 nova-compute nodes, a vnet interface for the instance is**bridged to the
 vlan tagged interface. On the compute node, there isn't**an IP interface
 on this network, so the private IP for instances isn't**reachable, even
 if the instance is running on the same

Re: [Openstack] ZFS/ZVol + iscsi for volume

2012-06-10 Thread Narayan Desai
We're definitely interested in this sort of thing. So much so that
we've already hacked support into nova-volume to run directly on top
of an illumos box with zfs. ;)

We've only gotten the basics working, and we haven't done any serious
torture testing of it yet. Our real goal is to get things working
properly using iSER. We tried it on Stuart Stent's recommendation
after the summit and it performs considerably better than stock iscsi
over infiniband.

Our eventual plan is to get the code ready for upstream merge, but it
isn't nearly ready for that yet.

Ping me offline and I can send you the code, if you're interested.
(this is very much a caveat emptor thing at the moment; we're not done
any hardening work yet)
 -nld

On Sun, Jun 10, 2012 at 4:15 AM, Nicolas de BONFILS
openst...@nicolas2bonfils.com wrote:
 Hi Stackers,

 I got an idea/question about iscsi volume.

 Actually the workflow is :

 create an LVM pool
 create a volume in this pool (with nova command or horizon UI)
 share it with an iscsi target (the compute/volume node)
 attach it with an iscsi initiator (the VM)

 I use zfs for other project, and I'm very interested in using it everywhere
 (when possible). One functionality zfs bring is zvol : it allows to create a
 block device from a zfs pool (a zfs property allow to share this device with
 iscsi).
 With this, a new workflow will be :

 create a zfs pool (zpool create name device)
 create a zvol block device (zfs create -V size poolname/volumename)
 set the zfs property to share it with iscsi (zfs set shareiscsi=on
 poolname/volumename)
 attach it with an iscsi initiator (the VM)

 From what I read, the Nexenta guys do a lot of work around zfs, but for
 volume use I only found code to plug a Nexenta san (I do not have the
 hardware to build a san), but no solution to use zfs drive/pool directly.
 So my 2 questions are :

 are some people interested in it ?
 does it seem possible to do it (maybe some one already try) ? Nova volume
 driver [github.com] got subclasses for specific cases, maybe a new zfs one.



 Please discuss, submit ideas !

 --- Nicolas


 ___
 Mailing list: https://launchpad.net/~openstack
 Post to     : openstack@lists.launchpad.net
 Unsubscribe : https://launchpad.net/~openstack
 More help   : https://help.launchpad.net/ListHelp


___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] ZFS/ZVol + iscsi for volume

2012-06-11 Thread Narayan Desai
How integrated is the network target support for zfs on freebsd? One
of the most compelling features (IMHO) of ZFS on illumos is the whole
comstar stack. On the zfs linux port at least, there are just
integration hooks out to the standard linux methods (kernel-nfs, etc)
for nfs, iscsi, etc.

I'm really interested to hear how this goes on FreeBSD. I've been
playing with zfs on linux as well, but the results haven't been so
good for me.
 -nld

On Mon, Jun 11, 2012 at 10:05 AM, Alberto Molina Coballes
alb.mol...@gmail.com wrote:
 What about Debian/KFreeBSD for nova-volume?

 Debian testing (wheezy) with FreeBSD kernel provides support for both
 OpenStack (nova-volume) and ZFS. We are deploying a little cloud for
 testing purposes and it will probably be one of the alternatives we will
 test for storage.

 Cheers




 ___
 Mailing list: https://launchpad.net/~openstack
 Post to     : openstack@lists.launchpad.net
 Unsubscribe : https://launchpad.net/~openstack
 More help   : https://help.launchpad.net/ListHelp

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] instances loosing IP address while running, due to No DHCPOFFER

2012-06-14 Thread Narayan Desai
I vaguely recall Vish mentioning a bug in dnsmasq that had a somewhat
similar problem. (it had to do with lease renewal problems on ip
aliases or something like that).

This issue was particularly pronounced with windows VMs, apparently.
 -nld

On Thu, Jun 14, 2012 at 6:02 PM, Christian Parpart tra...@gmail.com wrote:
 Hey,

 thanks for your reply. Unfortunately there was no process restart in
 nova-network nor in dnsmasq,
 both processes seem to have been up for about 2 and 3 days.

 However, why is the default dhcp_lease_time value equal 120s? Not having
 this one overridden
 causes the clients to actually re-acquire a new DHCP lease every 42 seconds
 (at least on my nodes),
 which is completely ridiculous.
 OTOH, I took a look at the sources (linux_net.py) and found out, why the
 max_lease_time is
 set to 2048, because that is the size of my network.
 So why is the max lease time the size of my network?
 I've written a tiny patch to allow overriding this value in nova.conf, and
 will submit it to launchpad
 soon - and hope it'll be accepted and then also applied to essex, since this
 is a very straight forward
 few-liner helpful thing.

 Nevertheless, that does not clarify on why now I had 2 (well, 3 actually)
 instances getting
 no DHCP replies/offers after some hours/days anymore.

 The one host that caused issues today (a few hours ago), I fixed it by hard
 rebooting the instance,
 however, just about 40 minutes later, it again forgot its IP, so one might
 say, that it
 maybe did not get any reply from the dhcp server (dnsmasq) almost right
 after it got
 a lease on instance boot.

 So long,
 Christian.

 On Thu, Jun 14, 2012 at 10:55 PM, Nathanael Burton
 nathanael.i.bur...@gmail.com wrote:

 Has nova-network been restarted? There was an issue where nova-network was
 signalling dnsmasq which would cause dnsmasq to stop responding to requests
 yet appear to be running fine.

 You can see if killing dnsmasq, restarting nova-network, and rebooting an
 instance allows it to get a dhcp address again ...

 Nate

 On Jun 14, 2012 4:46 PM, Christian Parpart tra...@gmail.com wrote:

 Hey all,

 I feel really sad with saying this, now, that we have quite a few
 instances in producgtion
 since about 5 days at least, I now have encountered the second instance
 loosing its
 IP address due to No DHCPOFFER (as of syslog in the instance).

 I checked the logs in the central nova-network and gateway node and found
 dnsmasq still to reply on requests from all the other instances and it
 even
 got the request from the instance in question and even sent an OFFER, as
 of what
 I can tell by now (i'm investigating / posting logs asap), but while it
 seemed
 that the dnsmasq sends an offer, the instances says it didn't receive one
 - wtf?

 Please tell me what I can do to actually *fix* this issue, since this is
 by far very fatal.

 One chance I'd see (as a workaround) is, to let created instanced
 retrieve
 its IP via dhcp, but then reconfigure /etc/network/instances to continue
 with
 static networking setup. However, I'd just like the dhcp thingy to get
 fixed.

 I'm very open to any kind of helping comments, :)

 So long,
 Christian.


 ___
 Mailing list: https://launchpad.net/~openstack
 Post to     : openstack@lists.launchpad.net
 Unsubscribe : https://launchpad.net/~openstack
 More help   : https://help.launchpad.net/ListHelp



 ___
 Mailing list: https://launchpad.net/~openstack
 Post to     : openstack@lists.launchpad.net
 Unsubscribe : https://launchpad.net/~openstack
 More help   : https://help.launchpad.net/ListHelp


___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Performance metrics

2012-06-21 Thread Narayan Desai
On Thu, Jun 21, 2012 at 11:16 AM, Rick Jones rick.jon...@hp.com wrote:
 On 06/20/2012 08:09 PM, Huang Zhiteng wrote:

 By 'network scaling', do you mean the aggregated throughput
 (bandwidth, packets/sec) of the entire cloud (or part of it)? I think
 picking up 'netperf' as micro benchmark is just 1st step, there's more
 work needs to be done.

 Indeed. A great deal more.


 For OpenStack network, there's 'inter-cloud' and
 'cloud-to-external-world' throughput.  If we care about the
 performance for end user, then reason numbers (for network scaling)
 should be captured inside VM instances.  For example, spawn 1,000 VM
 instances across cloud, then pair them to do 'netperf' tests in
 order to measure 'inter-cloud' network throughput.


 That would certainly be an interesting test yes.

We did a bunch of similar tests to determine the overhead caused by
kvm and limitations of the nova network architecture. We found that
VMs themselves were able to consistently saturate the network link
available to the host system, whether it was 1GE or 10GE, with
relatively modern node and network hardware. With the default
VLANManager network setup, there isn't much you can do to scale your
outbound connectivity beyond the hardware you can reasonably drive
with a single node, but using multi-host nova-network, we were able to
run a bunch of nodes in parallel, scaling up our outbound bandwidth
linearly. We managed to get 10 nodes, with a single VM per node, each
running 4 TCP streams, up to 99 gigabits on a dedicated cross country
link. There was a bunch of tuning that we needed to do, but it wasn't
anything particularly outlandish compared with the tuning needed for
doing this with bare metal. We've been meaning to do a full writeup,
but haven't had time yet.
 -nld

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Performance metrics

2012-06-21 Thread Narayan Desai
On Thu, Jun 21, 2012 at 4:21 PM, Rick Jones rick.jon...@hp.com wrote:


 TSO and GRO can cover a multitude of path-length sins :)

Along with a 64 MB TCP window ;)

 That is one of the reasons netperf does more than just bulk transfer :)
  When I was/am measuring scaling of an SMP node I would use aggregate,
 burst-mode, single-byte netperf TCP_RR tests to maximize the packets per
 second while minimizing the actual bandwidth consumed.

Yeah, for a completely random workload pps is the limiting factor.
We're lucky in that our primary high bandwidth use case is wide area
data transfer where we can fill our jumbo packets. (kudos to the nova
folks on that count; we were completely able to configure jumbo frames
without having to hack anything up)

We were actually more thrilled to be able to build relatively fat
single streams, due to our workload.
 -nld

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] HA inside VMs (via Corosync/Pacemaker)

2012-06-30 Thread Narayan Desai
On Sat, Jun 30, 2012 at 3:06 AM, Christian Parpart tra...@gmail.com wrote:
 Hm, Pacemaker/Corosync *inside* the VM will add the Service-IP to the local
 ethernet
 interface, and thus, the outside OpenStack components do not know about.

 Using a dedicated floating IP pool for service IPs might feel like a great
 solution, but
 OpenStack is not the one to manage who gets what IP - but Corosync/Pacemaker
 inside
 the KVM instances. :-)

 Anyone an idea how to solve this?

It sounds like you want to add explicit support to pacemaker to deal
with openstack fixed addresses. Then you could run with rfc1918
floating addresses, and then have pacemaker/corosync reassign the
(external) fixed address when consensus changes.

Think of the openstack fixed address control plane in a similar way to
ifconfig. You should even be able to script it up yourself; you'd need
to add your openstack creds to the HA images though.
 -nld

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] [HPC] Openstack HPC telecon

2012-07-06 Thread Narayan Desai
On Fri, Jul 6, 2012 at 9:51 AM, John Paul Walters jwalt...@isi.edu wrote:

 Does something like the first Monday of the month at 4:00pm EDT (UTC-4) work? 
  I'm just throwing out that time as something that seems to broadly work on 
 my end, but I'd welcome any input from others.

That generally works fine for me.
 -nld

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] [HPC] Openstack HPC telecon

2012-07-06 Thread Narayan Desai
On Fri, Jul 6, 2012 at 11:52 AM, Stefano Maffulli stef...@openstack.org wrote:
 On 07/06/2012 07:51 AM, John Paul Walters wrote:
 One of the outputs of the design summit was that folks are
 interested in participating in a monthly (or so) telecon to express
 feature requests, best practices, etc.  I'd like to get this process
 started. For those of you who are interested, what's the preferred
 format? IRC, telephone, something else?

I've been using google hangouts for this sort of thing lately, and I
think it would be a great solution, provided we don't overrun the
maximum group size. (iirc it is 10, but being raised)

There are also provisions for recording hangouts.
 -nld

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] [nova] [cinder] Nova-volume vs. Cinder in Folsom

2012-07-11 Thread Narayan Desai
I also vote for option 1, but the migration path really needs to be
solid and well documented.
 -nld

On Wed, Jul 11, 2012 at 10:52 AM, Andrew Clay Shafer
a...@parvuscaptus.com wrote:
 One vote for option 1.

 Remove Volumes


 ___
 Mailing list: https://launchpad.net/~openstack
 Post to : openstack@lists.launchpad.net
 Unsubscribe : https://launchpad.net/~openstack
 More help   : https://help.launchpad.net/ListHelp


___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] [nova] [cinder] Nova-volume vs. Cinder in Folsom

2012-07-11 Thread Narayan Desai
On Wed, Jul 11, 2012 at 1:49 PM, Adam Gandelman ad...@canonical.com wrote:
 On 07/11/2012 09:22 AM, Narayan Desai wrote:

 I also vote for option 1, but the migration path really needs to be
 solid and well documented.
   -nld


 I feel the same.  I think documented and tested migration paths are of
 utmost importance here.  Unlike the Keystone - Keystone Light migration,
 people are actually using Nova volume in production right now.   There were
 some facilities added in the later Keystone to help with migration, but
 AFAICS Keystone adoption at the time wasn't to the point where the migration
 path really mattered--not enough to have any bugs being rasied, at least.

Thinking more about this, it seems to me that we need to have pretty
detailed doc about this. Particularly since cinder isn't in wide use
yet, I think that a lot of people are running weird configurations for
volume storage. I suspect that having an automagic upgrade path won't
be feasible in these cases, but there needs to be enough doc (and
tools) to get old volumes migrated in some fashion.
 -nld

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] [nova] [cinder] Nova-volume vs. Cinder in Folsom

2012-07-12 Thread Narayan Desai
On Thu, Jul 12, 2012 at 2:38 PM, Vishvananda Ishaya
vishvana...@gmail.com wrote:

 Agreed, I'm a developer, so I'm clearly biased towards what is easier for 
 developers. It will be a significant effort to have to maintain the 
 nova-volume code, so I want to be sure it is necessary. End users really 
 shouldn't care about this, so the other community members who are impacted 
 are operators.

 I really would like more feedback on how painful it will be for operators if 
 we force them to migrate. We have one clear response from Chuck, which is 
 very helpful. Is there anyone else out there running nova-volume that would 
 prefer to keep it when they move to folsom?

I think that the long term maintenance or removal of nova-volume in
its existing form is orthogonal to the actual issue of continuity from
one release to the next.

At this point, we've now run cactus, diablo and are in testing with
essex. Each of these has effectively been a flag day for us; we build
the new system, migrate users, images, etc, and let users do a bunch
of manual migration of volume data, etc, while running both systems in
parallel. This hasn't been as painful as it sounds because our
understanding of best practices for running openstack is moving pretty
quickly and each system has been quite different from the previous.

The lack of an effective process to move from one major release to the
next is the major issue here in my mind. It would be fantastic if
(some day, ha ha ha) you could apt-get upgrade from folsom to grizzly,
but i think that is likely to be more trouble than it is worth. A
reasonable compromise would be a well documented process as well as
tools to aid in the process. Each real deployment will have a
substantial set of local customizations, particularly if they are
running at any sort of scale. While it won't be feasible to support
any upgrade with these customizations, tools for the process (which
may only be used a straw man in complex cases) would go a long way.
 -nld

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Release Upgrades (was [nova] [cinder] Nova-volume vs. Cinder in Folsom)

2012-07-12 Thread Narayan Desai
On Thu, Jul 12, 2012 at 4:36 PM, Vishvananda Ishaya
vishvana...@gmail.com wrote:

 Upgrading has been painful and we are striving to improve this process
 as much as possible.

I think that this needs to be a core value of the developer community,
if Openstack is going to become pervasive.

 I would like to take this a bit further. Documentation is a great first step,
 but I would actually like to have an actual Jenkins test that does the upgrade
 from essex to Folsom with live resources created. I think this the only way
 that we can be sure that the upgrade is working properly.

I don't want to dampen enthusiasm around this issue at all, but I
think this goal is pretty difficult to achieve, just due to the
existing complexity in real deployments. I'm also worried that this
would take away from a high level upgrade documentation goal.


 The first version of this doesn't even have to be on a full cluster. I'm 
 thinking
 something as simple as:

 * configure devstack to checkout stable/essex from all of the projects
 * run the system, launch some instances and volumes
 * terminate the workers
 * upgrade all of the code to folsom
 * do any manual upgrade steps (nova-manage db sync, cinder migrate, etc.)
 * run all the workers
 * make sure the existing data still works and new api commands run

 The manual upgrade steps should be contained in a single script so that the
 distress can use it to help make their package upgrades and deployers can
 use it for reference when upgrading their clusters.

 This is something we can start working on today and we can run after each
 commit. Then we will immediately know if we do something that breaks
 upgradability, and we will have a testable documented process for upgrading.

Having a testable process for upgrading is definitely a start. I guess
that I am more or less resigned to upgrades being pretty effort
intensive, at least in the near term, so my personal bias is towards
getting pretty extreme documentation done first.
 -nld

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


[Openstack] Weird nova-network bridging problem with precise/essex

2012-07-16 Thread Narayan Desai
We're running into what looks like a linux bridging bug, which causes
both substantial (20-40%) packet loss, and DNS to fail about that same
fraction of the time. We're running essex on precise, with dedicated
nova-network servers and VLANManager. On either of our nova-network
servers, we see the same behavior. When tracking this down, I found
the following, when tcpdump'ing along the path between vm instance and
n-net gateway.

The packets appear to make it to the nova-network server, and are
properly pulled out of dot1q tagging:
root@m5-p:~# tcpdump -K -p -i vlan200 -v -vv udp port 53
tcpdump: WARNING: vlan200: no IPv4 address assigned
tcpdump: listening on vlan200, link-type EN10MB (Ethernet), capture
size 65535 bytes
20:34:02.377711 IP (tos 0x0, ttl 64, id 59761, offset 0, flags [none],
proto UDP (17), length 60)
10.0.0.3.54937  10.0.0.1.domain: 52874+ A? www.google.com. (32)
20:34:07.377942 IP (tos 0x0, ttl 64, id 59762, offset 0, flags [none],
proto UDP (17), length 60)10.0.0.3.54937  10.0.0.1.domain: 52874+
A? www.google.com. (32)
20:34:12.378248 IP (tos 0x0, ttl 64, id 59763, offset 0, flags [none],
proto UDP (17), length 60)10.0.0.3.54937  10.0.0.1.domain: 52874+
A? www.google.com. (32)
20:34:12.378428 IP (tos 0x0, ttl 64, id 0, offset 0, flags [DF], proto
UDP (17), length 170)10.0.0.1.domain  10.0.0.3.54937: 52874 q: A?
www.google.com. 6/0/0 www.google.com. [1d3h55m19s] CNAME
www.l.google.com., www.l.google.com. [1m33s] A 74.125.225.209,
www.l.google.com. [1m33s] A 74.125.225.208, www.l.google.com. [1m33s]
A 74.125.225.212, www.l.google.com. [1m33s] A 74.125.225.211,
www.l.google.com. [1m33s] A 74.125.225.210 (142)

But some packets don't make it all of the way to the bridged interface:
root@m5-p:~# brctl show
bridge name bridge id   STP enabled interfaces
br200   8000.fa163e18927b   no  vlan200

root@m5-p:~# tcpdump -K -p -i br200 -v -vv udp port 53
tcpdump: listening on br200, link-type EN10MB (Ethernet), capture size
65535 bytes
20:34:12.378264 IP (tos 0x0, ttl 64, id 59763, offset 0, flags [none],
proto UDP (17), length 60)
10.0.0.3.54937  10.0.0.1.domain: 52874+ A? www.google.com. (32)
20:34:12.378424 IP (tos 0x0, ttl 64, id 0, offset 0, flags [DF], proto
UDP (17), length 170)
10.0.0.1.domain  10.0.0.3.54937: 52874 q: A? www.google.com.
6/0/0 www.google.com. [1d3h55m19s] CNAME www.l.google.com.,
www.l.google.com. [1m33s] A 74.125.225.209, www.l.google.com. [1m33s]
A 74.125.225.208, www.l.google.com. [1m33s] A 74.125.225.212,
www.l.google.com. [1m33s] A 74.125.225.211, www.l.google.com. [1m33s]
A 74.125.225.210 (142)

I can't find any way that ipfilter could be implicated in this; there
aren't deny rules that are hitting.

Oddly enough, this seems to cause no loss in icmp traffic, even with ping -f.

So far, searching hasn't netted very much. I've found this similar
sounding ubuntu bug report, but it looks like no one is working on it:
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/986043

We're at 3.2.0-24, and there is a 3.2.0-25, but it is reported to not
fix this issue, and neither are 3.4 kernels.

It seems sad to try backrevving to an onieric kernel, but that is on
my list for tomorrow.  If this is a kernel bug, it will make the
precise default kernel unusable for nova-network servers with dot1q
(or whatever the appropriate feature interaction is).

Does this ring any bells, or is there another course of action I should attempt?
thanks in advance for any suggestions.
 -nld

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Solaris Fails as Guest VM

2012-07-17 Thread Narayan Desai
I suspect that you need the right solaris (more likely illumos) bits
to get guest side support for virtio. We tried a while ago and the
default openindiana at the time didn't work.
 -nld

On Tue, Jul 17, 2012 at 7:43 PM, Joshua j...@root.bz wrote:
 I have tried with both KVM and qemu. Solaris starts to boot and hits
 grub then cycles boot. Anyone experienced this?

 ___
 Mailing list: https://launchpad.net/~openstack
 Post to : openstack@lists.launchpad.net
 Unsubscribe : https://launchpad.net/~openstack
 More help   : https://help.launchpad.net/ListHelp

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Solaris Fails as Guest VM

2012-07-18 Thread Narayan Desai
On Wed, Jul 18, 2012 at 7:38 PM, Michael March mma...@gmail.com wrote:
 I don't follow Solaris that closely but I vaguely remember the Joyent folks
 ported all of KVM to Solaris, right?  Or am I just missing the whole point
 here?

They did, and it is a fairly impressive piece of work. Their focus (as
a hosting provider) was running kvm virtual machines in a illumos
host. That support is completely separate from the guest drivers to
make virtio work for block and network devices. afaict, network device
support is complete, but block device support isn't finished yet. This
means you can't run illumos/solaris as a guest using kvm with virtio.
 -nld

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] quota question

2012-07-20 Thread Narayan Desai
On Fri, Jul 20, 2012 at 4:38 AM, Eoghan Glynn egl...@redhat.com wrote:

 Hi Narayan,

 I had the idea previously of applying a weighting function to the
 resource usage being allocated from the quota, as opposed to simply
 counting raw instances.

 The notion I had in mind was more related to image usage in glance,
 where the image footprint can vary very widely. However I think it
 could be useful for some nova resources also.

 Now for some resource types, for example say volumes, usage can be
 controlled along multiple axes (i.e. number of volumes and total size),
 so that gives more flexibility.

 But if I'm hearing you correctly, you'd want to apply a lower weighting
 to instances that are scheduled onto one of the higher-memory compute
 nodes, and vice versa a higher weighting to instances that happen to
 be run on lower-memory nodes.

 Does that sum it up, or have I misunderstood?

I think you've got it. I hadn't really asked with a particular
solution in mind, i was mainly looking for ideas.

I think that weighting would help. Effectively we need to discount
memory usage on the bigmem nodes, or something like that.

The harder part is that we need to be able to specify
independent/orthogonal quota constraints on different flavors. It
would be really useful to be able to say basically, you can have 2TB
of memory from this flavor, and 4TB of memory from that flavor. This
would allow saying something like you can have up to 3 1TB instances,
and independently have up to 3TB of small instances as well.

 BTW what kind of nova-scheduler config are you using?

We're using the filter scheduler. We've defined a bunch of custom
flavors, in addition to the stock ones, that allow us to fill up all
of our node types. So for each node type, we define flavors for the
complete node (minus a GB of memory for the hypervisor), and 3/4, 1/2,
1/4, and 1/8, 1/16, and 1/32 of the node. We've used a machine type
prefix for each one. The compute nodes are IBM idataplex, so we have
idp.{100,75,50,25,12,6,3}. We've done this for each machine type, so
we have idp.*, mem.*, gpu.*, etc. Each machine type has a unique
hostname prefix (cc for the idp nodes, cm for the bigmem nodes, cg for
gpu nodes, etc), and the filter scheduler is setup to route requests
for these custom flavors only to nodes with the appropriate hostname
prefix. This isn't an ideal solution, but it minimizes risk of
fragmentation. (With the default flavors, we'd see a lot of cases
where there was idle capacity left on the nodes that wasn't usable
because the ratio was wrong for the default flavors)

So far, this scheduling scheme has worked pretty well, aside from
leaving some instances in a weird state when you try to start a bunch
(20-50) at a time. I haven't had time to track that down yet.
 -nld

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Weird nova-network bridging problem with precise/essex

2012-07-20 Thread Narayan Desai
Just for the record, we found the issue. There was some filtering
being applied in the bridge code which randomly (?) dropped some DNS
requests. Setting:
net.bridge.bridge-nf-call-arptables = 0
net.bridge.bridge-nf-call-iptables = 0
net.bridge.bridge-nf-call-ip6tables = 0

completely resolved the problem.

I've written up full details here:
http://buriedlede.blogspot.com/2012/07/debugging-networking-problems-with.html
 -nld

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Weird nova-network bridging problem with precise/essex

2012-07-21 Thread Narayan Desai
On Sat, Jul 21, 2012 at 6:47 AM, Xu (Simon) Chen xche...@gmail.com wrote:
 Narayan,

 If you do  net.bridge.bridge-nf-call-iptables = 0 on the network controller,
 does floating IP still work? For each tenant/network, a subnet is created,
 and the nova-network has a .1 gateway configured on the bridge with the vlan
 interface plugged in.

 The packets from VMs are actually sent to the bridge for NATting. But if you
 doesn't allow the bridges to call iptables, it might break public access all
 together. Don't know, maybe I'm not understanding the sysctl flag
 correctly... Maybe it only applies to the packet transiting the bridge, not
 impacting the ones destined to the nova-network?

Do you mean floating (private) or fixed (public) IPs? I suspect that
you mean fixed. Fixed IPs worked regardless of this setting.

The crux of the issue was that packets transiting the bridge (ie being
moved from vlan200 to the virtual br200) were hitting filtering rules.
It looks to me like the sysctls only apply to traffic moving across
the bridge (ie exactly between vlan200 and br200), but don't bypass
iptables entirely. I don't think that should effect NAT/SNAT in any
case.
 -nld

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] growing fixed-ip network

2012-08-02 Thread Narayan Desai
On Thu, Aug 2, 2012 at 8:42 AM, Christoph Kluenter c...@iphh.net wrote:
 * Am Thu, Aug 02 2012 at 09:24:55 -0400 , schrieb Ravi Jagannathan:
 It should hop on to the next subnet block if available ( assuming that in
 LAN its a private address scheme ) .
 We only use routable IPs. thats why we have some nets which can't be 
 subnetted.
 What difference does it make if its private adress space ?

The major reason to use private address space is that there is likely
a lot more of it than you have in public address space. if you have
effectively unlimited fixed_ip space, you can give each project a lot.
For example, we give each project a /23. While a user could
potentially still run out of address space on our system, it hasn't
happened yet with our workload.
 -nld

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Multi-Hypervisor Support

2012-08-22 Thread Narayan Desai
We've managed to get things working by hardwiring the filtering
scheduler to route instances to particular hosts that are running nova
compute with different virtualization layers. (in our case, kvm and
lxc for GPUs)
 -nld

On Wed, Aug 22, 2012 at 12:34 PM, Michael J Fork mjf...@us.ibm.com wrote:
 Looking at the Hypervisor section of the documentation,
 http://docs.openstack.org/trunk/openstack-compute/admin/content/selecting-a-hypervisor.html,
 it states

 You cannot configure more than one virtualization type on the compute
 nodes, so the hypervisor selection is for the entire installation.

 Is this statement still true with the recent filter additions, specifically
 the ImagePropertiesFilter and ComputeCapabilitiesFilter?

 Thanks.

 Michael

 -
 Michael Fork
 Cloud Architect, Emerging Solutions
 IBM Systems  Technology Group


 ___
 Mailing list: https://launchpad.net/~openstack
 Post to : openstack@lists.launchpad.net
 Unsubscribe : https://launchpad.net/~openstack
 More help   : https://help.launchpad.net/ListHelp


___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Migration

2012-08-29 Thread Narayan Desai
On Wed, Aug 29, 2012 at 12:19 PM, Joshua Harlow harlo...@yahoo-inc.com wrote:
 Perhaps we should also have a CHANGELOG file to explain the major
 features/changes...

 Perhaps a 'MIGRATION' file as well that explains how to migrate from
 version - 1?

I think that this would be a great start.

In the long term, I would like to be able to do rolling upgrades
across a system. It is important in the short term to understand
dependencies between components as it pertains to versioning; ie, is
it possible to run a folsom keystone with essex nova? Or could an
essex nova-compute interact with a folsom nova-api? What are the hard
and soft requirements for version mixing? I've had decent success
performing rolling upgrades between versions of nova-compute in terms
of the ubuntu maintenance releases of essex, but I suspect in general
this is way too high of a bar to expect for all components. Even
building a decent mental map of interdependencies would help a lot
when planning maintenance activities.
 -nld

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] question about multi-host network configuration

2012-09-17 Thread Narayan Desai
Sure, we've been running in that sort of configuration since bexar.
The only tricky part is that you need to make sure that you run
nova-api-metadata on each nova-network server, and you need to make
sure that floating IPs can get to the appropriate fixed addresses (ie
if a fixed address is not routed from the nova-network instance where
the floating address is served from). That last bit pretty much
requires setting up something like ospf for route exchange between
your nova-network servers. (fixed networks aren't pegged to one n-net
instance or the other, so they can move over time)
 -nld

On Mon, Sep 17, 2012 at 9:38 AM, Xin Zhao xz...@bnl.gov wrote:
 Hello,

 Our openstack cluster has 2 racks of physical nodes, I want to set up two
 openstack network service hosts, each serving one rack.
 Can I do that?

 Thanks,
 Xin


 ___
 Mailing list: https://launchpad.net/~openstack
 Post to : openstack@lists.launchpad.net
 Unsubscribe : https://launchpad.net/~openstack
 More help   : https://help.launchpad.net/ListHelp


___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] multi-process-api-service

2012-09-27 Thread Narayan Desai
On Thu, Sep 27, 2012 at 2:20 PM, Nandavar, Divakar Padiyar (STSD)
divakar.padiyar-nanda...@hp.com wrote:
 From the information available in the blueprint for 
 multi-process-api-service I see that implementation has been completed and 
 would be available as part of Folsom release  
 https://blueprints.launchpad.net/nova/+spec/multi-process-api-service

 I see the advantage in back porting this to ESSEX as well.  Any thoughts on 
 this?   Were there any discussion earlier on this topic for back porting to 
 ESSEX?

I'm not sure this is particularly needed for Essex. We're successfully
running a bunch of discrete nova-api process across a few systems with
Essex, load balanced using nginx, and it works great. This requires a
little more work on the setup side, but it has proven stable for us.
(it also allows us to service any of our machines that service
nova-api requests without taking a user-visible hit, which is an added
plus)
 -nld

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


[Openstack] High bandwidth networking with openstack

2012-11-27 Thread Narayan Desai
I've finally finished my writeup describing the experiments that we
performed using Openstack to drive a wide area 100 gigabit network.
I've included all of the details for configuration and tuning, as well
as speculation why we're seeing such good numbers.

tl;dr: you can push a whole lot of network bandwidth out of an
openstack deployment with kvm/virtio; we hit 98 gigabits peak with 10
instances across 50 ms of latency.

The post is here:
http://buriedlede.blogspot.com/2012/11/driving-100-gigabit-network-with.html

I'm happy to discuss this in more detail, if there are additional questions.
 -nld

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] How to Solve InfinitBand Adaptor Driver under Ubuntu 12.04?

2012-12-25 Thread Narayan Desai
We're using IB (QDR connectX and connectX2) on our system. It turns
out that the drivers included in version 3.2 of the linux kernel are
fine. I've built a ppa for updated management tools though; all of
those bits are ancient in precise. The ppa is here:
http://launchpad.net/~narayan-desai/+archive/infiniband

One word of warning; you'll only be able to use IPOIB for connectivity
between the hypervisor nodes and the openstack service nodes; getting
to VMs requires full ethernet compatibility, which IPOIB doesn't
provide. You will be able to use the IPOIB for fast image transfers
and really fast volume storage if you have it. Turning on iSER helps a
lot with volume bandwidth, but requires a patch to nova compute.
 -nld

On Tue, Dec 25, 2012 at 7:03 AM,  changliwe...@gmail.com wrote:
 Hi,
 Merry Christmas !
 Have you even tried to use infinitband network in OpenStack  base on Ubuntu
 12.04?
 In recent two days, I  want to use IPoIB in  Ubuntu, but the big problem is
 the driver of IB,
 The OFED src is special for enterprise distribution only, such as (RHEL and
 SLES ), even the so called
 source code  archive is a rpmbuild-tree-like structure. Then the question is
 how to solve the driver problem  in ubuntu 12.04.
 

 ___
 Mailing list: https://launchpad.net/~openstack
 Post to : openstack@lists.launchpad.net
 Unsubscribe : https://launchpad.net/~openstack
 More help   : https://help.launchpad.net/ListHelp


___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Structure Hybrid Compute Nodes

2013-01-22 Thread Narayan Desai
We have the same basic problems. We have 4 different types of systems
integrated into our system. They all have different ratios of cpu to
memory, and we have some specialized hardware on one class of nodes.

We ended up setting up a series of chassis specific instance
definitions. We then use the filter scheduler to only accept hosts
with a particular configuration for the associated instance types.
(for example, our large memory nodes will only run instances with the
mem. prefix, and won't run generic instance types or ones for other
hardware types -- it seemed silly to waste large memory resources on
1GB ram vms)

It is pretty simple to write filter rules for the filter scheduler to
set this up. You just need instance types and a hostname convention
that you can use to matchmake.

While this seems a little hacky, it works well.
 -nld


On Tue, Jan 22, 2013 at 6:47 AM, Danilo Perogil dpero...@dualtec.com.br wrote:
 Hello everybody,

 Do you know of a way to set to work with OpenStack Computes Nodes containing
 different hardware?
 In my case, I did tests with different hardware, example 1 compute node with
 24 processors and 128GB memory VS 1 compute node with 8 processors and 16GB
 of memory. In the test I saw he only makes division number of instances
 between total compute nodes, but it does a calculation for use by node.

 It is possible to make it work with some kind of calculation to not overload
 a node?

 Sorry my english google translate hahahah

 ___
 Mailing list: https://launchpad.net/~openstack
 Post to : openstack@lists.launchpad.net
 Unsubscribe : https://launchpad.net/~openstack
 More help   : https://help.launchpad.net/ListHelp


___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] [Openstack-operators] Help with simplest Quantum setup possible...

2013-03-14 Thread Narayan Desai
Make sure that the metadata server has a route back to the VM. Traffic
hitting that NAT rule ensures that data is flowing properly in one
direction, but you need to make sure bits can flow back to establish a
tcp connection. We had this problem running multiple nova-network
servers.
 -nld

On Thu, Mar 14, 2013 at 2:28 PM, Martinx - ジェームズ
thiagocmarti...@gmail.com wrote:
 Daneyon,

  Thanks! But, I'm not using L3, or GRE, or namespaces...

  I read the
 http://docs.openstack.org/trunk/openstack-network/admin/content/ch_limitations.html
 and I don't think my setup (Single Flat, no GRE, no L3) fits that case...

  From my understandings, I'm not using overlapping IP's, since it is
 disabled by default... Anyway, after forcing `allow_overlapping_ips to
 False', nothing happens...

  From the compute node, I can connect to metadata by running: `telnet
 10.32.14.232 8775' and / or `telnet 10.32.14.234 8775', both answer but,
 doesn't work in the end...

  I tried with metadata @ 10.32.14.232 (controller) and @ 10.32.14.234
 (compute node with `nova-api-metadata') but, same broken result.

  From within my instance, when I run: `telnet 169.254.169.254 80', I can see
 the packages passing trough the NAT rule `169.254.169.254  tcp dpt:80
 to:10.32.14.232:8775' but, the telnet connection times up.

  Do I need the package `quantum-plugin-metaplugin' instead
 `nova-api-metadata'? If yes, where, controller or on each compute node?

  BTW, if I set `use_namespaces = False' in dhcp_agent.ini, my Instances
 doesn't boots up anymore.

  Maybe it is time to go with Grizzly RC1?

  Anyway, I would like to finish everything I can with Quantum in Folsom,
 before going to try Grizzly...

  I really appreciate your help!!

 Regards,
 Thiago


 On 14 March 2013 15:21, Daneyon Hansen (danehans) daneh...@cisco.com
 wrote:


 Take a look at the limitations section of the Quantum docs. You will need
 to set allow overlapping IP's to false and make sure you have bidirectional
 network connectivity between the metadata ip and the tenant private
 network(s).  The L3 agent.ini also should be configured with the metadata
 ip.  The docs cover these topics in more detail.

 Regards,
 Daneyon Hansen

 On Mar 14, 2013, at 12:11 PM, Martinx - ジェームズ
 thiagocmarti...@gmail.com wrote:

 Hi!

  My Quantum Single Flat is almost working!

  My Instance get an IPv4 without any problem but, the metadata isn't
 working...

  What can I do?

  My compute node have the `nova-api-metadata' package, nova.conf have
 `metadata_host = 10.32.14.234', the iptables `169.254.169.254' NAT rule is
 there too...

  The Instance log shows:

 Starting network...
 udhcpc (v1.20.1) started
 Sending discover...
 Sending select for 10.33.14.3...
 Lease of 10.33.14.3 obtained, lease time 120
 deleting routers
 route: SIOCDELRT: No such process
 adding dns 8.8.4.4
 adding dns 8.8.8.8
 cirros-ds 'net' up at 1.84
 checking http://169.254.169.254/2009-04-04/instance-id
 failed 1/20: up 1.87. request failed
 failed 2/20: up 13.98. request failed
 failed 3/20: up 26.02. request failed
 failed 4/20: up 38.06. request failed
 failed 5/20: up 50.09. request failed
 failed 6/20: up 62.13. request failed
 failed 7/20: up 74.16. request failed
 failed 8/20: up 86.20. request failed
 failed 9/20: up 98.24. request failed
 failed 10/20: up 110.27. request failed
 failed 11/20: up 122.31. request failed
 failed 12/20: up 134.34. request failed
 failed 13/20: up 146.38. request failed
 failed 14/20: up 158.41. request failed
 failed 15/20: up 170.45. request failed
 failed 16/20: up 182.49. request failed
 failed 17/20: up 194.52. request failed
 failed 18/20: up 206.56. request failed
 failed 19/20: up 218.59. request failed
 failed 20/20: up 230.63. request failed
 failed to read iid from metadata. tried 20
 no results found for mode=net. up 242.66. searched: nocloud configdrive
 ec2
 failed to get instance-id of datasource
 Starting dropbear sshd: generating rsa key... generating dsa key... OK

 Any clue?

 Thanks!
 Thiago


 On 13 March 2013 10:10, Daneyon Hansen (danehans) daneh...@cisco.com
 wrote:



 Regards,
 Daneyon Hansen


 On Mar 13, 2013, at 1:04 AM, Martinx - ジェームズ
 thiagocmarti...@gmail.com wrote:

 Guys,

  Here:
 http://docs.openstack.org/trunk/openstack-network/admin/content/demo_flat_installions.html

  ...at the step 1. Install the L2 agent. but, what is the name of the
 Ubuntu package (I'm using Folsom)?

  ...apt-get install what?


 quantum-plugin-openvswitch-agent


  I thought that the Quantum L2 agent was the package
 `quantum-plugin-linuxbridge' but, this doc uses OpenvSwitch... So, I'm
 confused now...


 As you know Quantum uses plugins to realize the network abstractions.
 Some plugins (such as OVS) use an agent to collect information from the
 Quantum db to implement the logical data model on the local vswitch.


  The following packages quantum-server quantum-plugin-openvswitch
 quantum-plugin-openvswitch-agent quantum-dhcp-agent will be enough for 

Re: [Openstack] Non responding instance

2013-04-08 Thread Narayan Desai
And if a nova reboot fails, you can always fall back to issuing virsh
commands on the node behind nova's back.
 -nld

On Mon, Apr 8, 2013 at 8:28 PM, Blair Bethwaite
blair.bethwa...@gmail.com wrote:
 Dave,

 Have you tried rebooting it (via OpenStack dashboard/CLI/API)? Obviously
 you'll lose memory state, but the VM will boot from the same (current)
 virtual disks, so if those disks are ok the instance will have all
 previously saved/altered data.


 On 8 April 2013 19:09, Dave Pigott dave.pig...@linaro.org wrote:

 Hi all,

 I've got an instance that's been running for several months now, and
 suddenly we can not ping or telnet to it. It shows up in the list as Active
 and Running, but just isn't responding. We earlier had a disk full problem
 on that node, but this happened before that occurred, so I don't think
 that's responsible (but I could be wrong). The instance is used extensively
 so I'd prefer to fix it rather than have to kill it and spin up another
 instance. FYI all other instances in the cloud are fine and we can ssh and
 ping them.

 Any ideas what I can do to get this instance active again?

 Thanks

 Dave
 Lava Lab Lead
 Linaro Ltd
 ___
 Mailing list: https://launchpad.net/~openstack
 Post to : openstack@lists.launchpad.net
 Unsubscribe : https://launchpad.net/~openstack
 More help   : https://help.launchpad.net/ListHelp




 --
 Cheers,
 ~Blairo

 ___
 Mailing list: https://launchpad.net/~openstack
 Post to : openstack@lists.launchpad.net
 Unsubscribe : https://launchpad.net/~openstack
 More help   : https://help.launchpad.net/ListHelp


___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] What's the package flow in vlan mode when ping the floating ip between vms.

2013-04-09 Thread Narayan Desai
This will depend on whether the VMs are in the same tenant network or
not. Assuming they are on the same L2 and L3 network, then the packets
will transit either the linux bridge, or openvswitch, depending on how
you have things configured. Note that network filtering rules will be
processed on this path in some cases.

If you are talking about flows between VMs that are on different
tenant networks (both ends using RFC1918 addresses), nova-network
would setup gateways for each network in the host, and it should be
routing, i think. (this is supposition on my part -- we've only done
some basic experiments with multi-host)
 -nld

On Tue, Apr 9, 2013 at 7:41 PM, Lei Zhang zhang.lei@gmail.com wrote:
 Hi JuanFra,

 I read this article. But It doesn't explain what I concern.
 What I want to know is that what's the packet flow when ping the floating ip
 between the VMs which located in the same physical machine?


 On Tue, Apr 9, 2013 at 9:56 PM, JuanFra Rodriguez Cardoso
 juanfra.rodriguez.card...@gmail.com wrote:

 Could this Mirantis' post help you?

 http://www.mirantis.com/blog/vlanmanager-network-flow-analysis/

 ---
 JuanFra


 2013/4/9 Lei Zhang zhang.lei@gmail.com

 I am using: Folsom, nova-network, vlanManage, multi_host=ture

 I want to know what's the package flow between the vms which located in
 the same physical machine.

 Here is the typology. I want to know want happen if I ping 192.168.3.47
 in vm7.

 When the NAT is applied?

|vm: fixed: 10.0.104.5, floating:
 192.168.3.56|-|vnet15||br104||vnet14|| vm7: fixed:
 10.0.104.6, floating: 192.168.3.57|

 if the net.bridge.bridge-nf-call-iptables=0, I can not see the NAT
 properly.

 [root@openstack-node1 ~]# tcpdump -i vnet14 -l -ne icmp -v
 tcpdump: WARNING: vnet14: no IPv4 address assigned
 tcpdump: listening on vnet14, link-type EN10MB (Ethernet), capture size
 65535 bytes
 17:11:10.275818 fa:16:3e:3e:ab:8f  fa:16:3e:72:8f:b4, ethertype IPv4
 (0x0800), length 98: (tos 0x0, ttl 64, id 0, offset 0, flags [DF], proto
 ICMP (1), length 84)
 10.0.104.6  192.168.3.56: ICMP echo request, id 31090, seq 1, length
 64
 17:11:10.277546 fa:16:3e:22:31:09  fa:16:3e:3e:ab:8f, ethertype IPv4
 (0x0800), length 98: (tos 0x0, ttl 64, id 64703, offset 0, flags [none],
 proto ICMP (1), length 84)
 10.0.104.5  10.0.104.6: ICMP echo reply, id 31090, seq 1, length 64

 When net.bridge.bridge-nf-call-iptables=1, I found the NAT properly. But
 I also found it is said the bridge filter should be disabled in
 http://docs.openstack.org/trunk/openstack-compute/admin/content/network-troubleshooting.html.
 Is this true?

 [root@openstack-node1 ~]# tcpdump -i vnet14 -l -ne icmp -v
 tcpdump: WARNING: vnet14: no IPv4 address assigned
 tcpdump: listening on vnet14, link-type EN10MB (Ethernet), capture size
 65535 bytes
 17:13:40.507577 fa:16:3e:3e:ab:8f  fa:16:3e:72:8f:b4, ethertype IPv4
 (0x0800), length 98: (tos 0x0, ttl 64, id 0, offset 0, flags [DF], proto
 ICMP (1), length 84)
 10.0.104.6  192.168.3.56: ICMP echo request, id 42098, seq 1, length
 64
 17:13:40.508034 fa:16:3e:22:31:09  fa:16:3e:3e:ab:8f, ethertype IPv4
 (0x0800), length 98: (tos 0x0, ttl 64, id 64704, offset 0, flags [none],
 proto ICMP (1), length 84)
 192.168.3.56  10.0.104.6: ICMP echo reply, id 42098, seq 1, length
 64

 --
 Lei Zhang

 Blog: http://jeffrey4l.github.com
 twitter/weibo: @jeffrey4l

 ___
 Mailing list: https://launchpad.net/~openstack
 Post to : openstack@lists.launchpad.net
 Unsubscribe : https://launchpad.net/~openstack
 More help   : https://help.launchpad.net/ListHelp





 --
 Lei Zhang

 Blog: http://jeffrey4l.github.com
 twitter/weibo: @jeffrey4l

 ___
 Mailing list: https://launchpad.net/~openstack
 Post to : openstack@lists.launchpad.net
 Unsubscribe : https://launchpad.net/~openstack
 More help   : https://help.launchpad.net/ListHelp


___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Share Glance between cells or regions

2013-05-15 Thread Narayan Desai
+1.

We're going to be running a bunch of parallel deployments of openstack for
the purpose of experimentation in system design. it would be nice to be
able to share glance and keystone between instances.
 -nld


On Wed, May 15, 2013 at 1:46 PM, John Paul Walters jwalt...@isi.edu wrote:

 Hi,

 We're looking at setting up a geographically distributed OpenStack
 installation, and we're considering either cells or regions.  We'd like to
 share a single Glance install between our regions (or cells), so the same
 images can be spawned anywhere.  From here:

 http://docs.openstack.org/trunk/openstack-ops/content/segregate_cloud.html

 it's not clear whether that's possible.  Can anyone shed some light on
 this?  Is it possible in regions OR cells (or both)?  Is there a better
 solution that I'm not thinking of?

 thanks,
 JP
 ___
 Mailing list: https://launchpad.net/~openstack
 Post to : openstack@lists.launchpad.net
 Unsubscribe : https://launchpad.net/~openstack
 More help   : https://help.launchpad.net/ListHelp

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp