[openstack-dev] [Devstack] [Cinder] Registering 3rd party driver as default

2014-08-14 Thread Amit Das
Hi folks,

I have been trying to run devstack with my cinder driver as the default
volume_driver but with no luck.

Devstack seems to register the lvm driver as the default always.

I have tried below approaches:

   1. directly modifying the /etc/cinder/cinder.conf file
   2. creating a driver file @ ./devstack/lib/cinder_plugins/driver-name
   1. ref - https://review.openstack.org/#/c/68726/


This is my localrc details:
http://paste.openstack.org/show/94822/

I run ./unstack.sh  then FORCE=yes ./stack.sh

This is the cinder.conf that is generated after running above stack.sh. I
comment out the [lvmdriver-1] section manually *(not sure if this section
needs to be commented)*

http://paste.openstack.org/show/94841/

These are portions of c-sch  c-vol logs after restarting them in their
respective screens.

http://paste.openstack.org/show/94842/

Regards,
Amit
*CloudByte Inc.* http://www.cloudbyte.com/
___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [Ceilometer] Add a pollster plugin

2014-08-14 Thread Duan, Li-Gong (Gary@HPServers-Core-OE-PSC)
Thanks a lot, Doug. It does work after configuring pipeline.yaml.



Regards,

Gary

On Aug 12, 2014, at 4:11 AM, Duan, Li-Gong (Gary at 
HPServers-Core-OE-PSChttp://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev)
 li-gong.duan at 
hp.comhttp://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev 
wrote:



 Hi Folks,



 Is there any best practices or good way to debug whether new pollster plugin 
 work fine for ceilometer?



 I'd like to add a new pollster plugin into Ceilometer by

  - adding a new item under enterypoint/ceilometer.poll.central in 
 setup.cfg file

  - adding the implementation code inheriting plugin.CentralPollster.



 But when I sudo python setup.py install and restart ceilometer-related 
 services in devstack, NO new metering is displayed upon ceilometer 
 meter-list and I expect that there should be a new metering showing the item 
 defined in setup.cfg.

 Is there any other source/config files I need to modify or add?



You need to define a pipeline [1] to include the data from your new pollster 
and schedule it to be run.



Doug



[1] http://docs.openstack.org/developer/ceilometer/configuration.html#pipelines





 Thanks in advance,

 Gary



___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [Mistral] Mistral milestone 0.1 blueprint review

2014-08-14 Thread Renat Akhmerov
Thanks,

Dmitri, I left my comments in the document you shared. Please take a loot at 
them. We also need to define a reasonable date for 0.1 release which is now set 
to Sept 11 according to my very raw estimates. We may want to move it a little 
bit.

Renat Akhmerov
@ Mirantis Inc.



On 14 Aug 2014, at 08:30, Dmitri Zimine dzim...@stackstorm.com wrote:

 Renat, team: 
 
 The current plan of record on 0.1 here, 
 https://launchpad.net/mistral/+milestone/0.1
 
 Some suggested adjustments: 
 https://docs.google.com/a/stackstorm.com/document/d/1Ukd6SDk9tMpn8kpti-kXaOzVhl8Jm1YWTh3Kw1A6lgQ/edit#
 
 Pls use the doc to comment/disccuss and let’s decide by next mistral iRC 
 meeting on Mon. 
 
 Regards, Dmitri. 

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


[openstack-dev] [nova] Nova API meeting

2014-08-14 Thread Christopher Yeoh
Hi,

Just a reminder that the weekly Nova API meeting is being held tomorrow
Friday UTC . 

We encourage cloud operators and those who use the REST API such as
SDK developers and others who and are interested in the future of the
API to participate. 

In other timezones the meeting is at:

EST 20:00 (Thu)
Japan 09:00 (Fri)
China 08:00 (Fri)
ACDT 9:30 (Fri)

The proposed agenda and meeting details are here: 

https://wiki.openstack.org/wiki/Meetings/NovaAPI

Please feel free to add items to the agenda. 

Chris

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


[openstack-dev] [OpenStack-Dev][Cinder] Cinder Core nomination

2014-08-14 Thread Boring, Walter
Hey guys,
   I wanted to pose a nomination for Cinder core.

Xing Yang.
She has been active in the cinder community for many releases and has worked on 
several drivers as well as other features for cinder itself.   She has been 
doing an awesome job doing reviews and helping folks out in the 
#openstack-cinder irc channel for a long time.   I think she would be a good 
addition to the core team.


Walt
___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [Neutron] Simple proposal for stabilizing new features in-tree

2014-08-14 Thread Sridar Kandaswamy (skandasw)
Hi Aaron:


There is a certain fear of another cascading chain of emails so with hesitation 
i send this email out. :-)


1) I could not agree with u more on the issue with the logs and the pain with 
debugging issues here. Yes for sure bugs do keep popping up but often times, 
(speaking for fwaas) - given the L3 agent interactions - there are a multitude 
of reasons for a failure. An L3Agent crash or a router issue - also manifests 
itself as an fwaas issue - i think ur first paste is along those lines (perhaps 
i could be wrong without much context here).


The L3Agent - service coexistence is far from ideal - but this experience has 
lead to two proposals - a vendor proposal [1] that actually tries to address 
such agent limitations and collaboration on another community proposal[2] to 
enable the L3 Agent to be more suited to hosting services. Hopefully [2] will 
get picked up in K and should help provide the necessary infrastructure to 
clean up the reference implementation.


2) Regarding ur point on the  FWaaS API - the intent of the feature by design 
was to keep the service abstraction separate from how it is inserted in the 
network - to keep this vendor/technology neutral. The first priority, post 
Havana was to address  service insertion to get away from the all routers model 
[3] but did not get the blessings needed. Now with a redrafted proposal for 
Juno[4] again an effort is being made to address this now for the second time 
in the 2 releases post H.


In general, I would make a request that before we decide to go ahead and start 
moving things out into an incubator area - more discussion is needed.  We don’t 
want  to land up in a situation in K-3 where we find out that this model does 
not quite work for whatever reason. Also i wonder about the hoops to get out 
from incubation. As vendors who try to align with the community and upstream 
our work - we don’t want to land up waiting more cycles - there is quite a bit 
of frustration that is felt here too.


Lets also think about the impacts on feature velocity, somehow that keeps 
popping into my head every time i buy a book from a certain online retailer. :-)


[1] https://review.openstack.org/#/c/90729/

[2] https://review.openstack.org/#/c/91532/

[3] https://review.openstack.org/#/c/62599/

[4] https://review.openstack.org/#/c/93128/


Thanks


Sridar


From: Aaron Rosen aaronoro...@gmail.commailto:aaronoro...@gmail.com
Reply-To: OpenStack List 
openstack-dev@lists.openstack.orgmailto:openstack-dev@lists.openstack.org
Date: Wednesday, August 13, 2014 at 3:56 PM
To: OpenStack List 
openstack-dev@lists.openstack.orgmailto:openstack-dev@lists.openstack.org
Subject: Re: [openstack-dev] [Neutron] Simple proposal for stabilizing new 
features in-tree

Hi,

I've been thinking a good bit on this on the right way to move forward with 
this and in general the right way new services should be added. Yesterday I was 
working on a bug that was causing some problems in the openstack infra. We 
tracked down the issue then I uploaded a patch for it. A little while later 
after jenkins voted back a -1 so I started looking through the logs to see what 
the source of the failure was (which was actually unrelated to my patch). The 
random failure in the fwaas/vpn/l3-agent code which all outputs to the same log 
file that contains many traces for every run even successful ones. In one skim 
of this log file I was able to spot 4 [1]bugs which shows  these new 
experimental services that we've added to neutron have underlying problems 
(even though they've been in the tree for 2 releases+ now). This puts a huge 
strain on the whole openstack development community as they are always recheck 
no bug'ing due to neutron failures.

If you look at the fwaas work that was done. This merged over two releases ago 
and still does not have a complete API as there is no concept of where 
enforcement should be done. Today, enforcement is done across all of a tenant's 
routers making it more or less useless imho and we're just carrying it along in 
the tree (and it's causing us problems)!

I think Mark's idea of neutron-incubator[2] is a great step forward to 
improving neutron.

We can easily move these things out of the neutron source tree and we can plug 
these things in here:
https://github.com/openstack/neutron/blob/master/etc/neutron.conf#L52
https://github.com/openstack/neutron/blob/master/etc/neutron.conf#L72
(GASP: We have seen shipped our own API's here to customers before we were able 
to upstream them).

This allows us to decouple these experimental things from the neutron core and 
allows us to release these components on their own making things more 
modular/maintainable and stable (I think these things might even be better long 
term living out of the tree).  Most importantly though it doesn't put a burden 
on everyone else.


Best,

Aaron


[1]
http://paste.openstack.org/show/94664/
http://paste.openstack.org/show/94670/

[openstack-dev] [heat] heat docker multi host scheduling support

2014-08-14 Thread Malawade, Abhijeet
Hi all,

I am trying to use heat to create docker containers. I have configured 
heat-docker plugin.
I am also able to create stack using heat successfully.

To start container on different host we need to provide 'docker_endpoint' in 
template. For this we have to provide host address where container will run in 
template.

Is there any way to schedule docker container on available hosts using 
heat-docker plugin without giving 'docker_endpoint' in template file.
Is heat-docker plugin supports managing docker hosts cluster with scheduling 
logic.

Please let me know your suggestions on the same.

Thanks,
Abhijeet

__
Disclaimer:This email and any attachments are sent in strictest confidence for 
the sole use of the addressee and may contain legally privileged, confidential, 
and proprietary data.  If you are not the intended recipient, please advise the 
sender by replying promptly to this email and then delete and destroy this 
email and any attachments without any further use, copying or forwarding___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [Nova] Concerns around the Extensible Resource Tracker design - revert maybe?

2014-08-14 Thread Nikola Đipanov
On 08/13/2014 06:05 PM, Sylvain Bauza wrote:
 
 Le 13/08/2014 12:21, Sylvain Bauza a écrit :

 Le 12/08/2014 22:06, Sylvain Bauza a écrit :

 Le 12/08/2014 18:54, Nikola Đipanov a écrit :
 On 08/12/2014 04:49 PM, Sylvain Bauza wrote:
 (sorry for reposting, missed 2 links...)

 Hi Nikola,

 Le 12/08/2014 12:21, Nikola Đipanov a écrit :
 Hey Nova-istas,

 While I was hacking on [1] I was considering how to approach the fact
 that we now need to track one more thing (NUMA node utilization)
 in our
 resources. I went with - I'll add it to compute nodes table
 thinking
 it's a fundamental enough property of a compute host that it
 deserves to
 be there, although I was considering  Extensible Resource Tracker
 at one
 point (ERT from now on - see [2]) but looking at the code - it did
 not
 seem to provide anything I desperately needed, so I went with
 keeping it
 simple.

 So fast-forward a few days, and I caught myself solving a problem
 that I
 kept thinking ERT should have solved - but apparently hasn't, and I
 think it is fundamentally a broken design without it - so I'd really
 like to see it re-visited.

 The problem can be described by the following lemma (if you take
 'lemma'
 to mean 'a sentence I came up with just now' :)):

 
 Due to the way scheduling works in Nova (roughly: pick a host
 based on
 stale(ish) data, rely on claims to trigger a re-schedule), _same
 exact_
 information that scheduling service used when making a placement
 decision, needs to be available to the compute service when
 testing the
 placement.
 

 This is not the case right now, and the ERT does not propose any
 way to
 solve it - (see how I hacked around needing to be able to get
 extra_specs when making claims in [3], without hammering the DB). The
 result will be that any resource that we add and needs user supplied
 info for scheduling an instance against it, will need a buggy
 re-implementation of gathering all the bits from the request that
 scheduler sees, to be able to work properly.
 Well, ERT does provide a plugin mechanism for testing resources at the
 claim level. This is the plugin responsibility to implement a test()
 method [2.1] which will be called when test_claim() [2.2]

 So, provided this method is implemented, a local host check can be
 done
 based on the host's view of resources.


 Yes - the problem is there is no clear API to get all the needed
 bits to
 do so - especially the user supplied one from image and flavors.
 On top of that, in current implementation we only pass a hand-wavy
 'usage' blob in. This makes anyone wanting to use this in conjunction
 with some of the user supplied bits roll their own
 'extract_data_from_instance_metadata_flavor_image' or similar which is
 horrible and also likely bad for performance.

 I see your concern where there is no interface for user-facing
 resources like flavor or image metadata.
 I also think indeed that the big 'usage' blob is not a good choice
 for long-term vision.

 That said, I don't think as we say in French to throw the bath
 water... ie. the problem is with the RT, not the ERT (apart the
 mention of third-party API that you noted - I'll go to it later below)
 This is obviously a bigger concern when we want to allow users to
 pass
 data (through image or flavor) that can affect scheduling, but
 still a
 huge concern IMHO.
 And here is where I agree with you : at the moment, ResourceTracker
 (and
 consequently Extensible RT) only provides the view of the resources
 the
 host is knowing (see my point above) and possibly some other resources
 are missing.
 So, whatever your choice of going with or without ERT, your patch [3]
 still deserves it if we want not to lookup DB each time a claim goes.


 As I see that there are already BPs proposing to use this IMHO broken
 ERT ([4] for example), which will surely add to the proliferation of
 code that hacks around these design shortcomings in what is already a
 messy, but also crucial (for perf as well as features) bit of Nova
 code.
 Two distinct implementations of that spec (ie. instances and flavors)
 have been proposed [2.3] [2.4] so reviews are welcome. If you see the
 test() method, it's no-op thing for both plugins. I'm open to comments
 because I have the stated problem : how can we define a limit on
 just a
 counter of instances and flavors ?

 Will look at these - but none of them seem to hit the issue I am
 complaining about, and that is that it will need to consider other
 request data for claims, not only data available for on instances.

 Also - the fact that you don't implement test() in flavor ones tells me
 that the implementation is indeed racy (but it is racy atm as well) and
 two requests can indeed race for the same host, and since no claims are
 done, both can succeed. This is I believe (at least in case of single
 flavor hosts) unlikely to happen in practice, but you get the idea.

 Agreed, these 2 patches probably require another iteration, in
 particular how we make sure that it 

Re: [openstack-dev] [all] 3rd Party CI vs. Gerrit

2014-08-14 Thread Christopher Yeoh
On Wed, 13 Aug 2014 18:52:05 -0400
Jay Pipes jaypi...@gmail.com wrote:

 On 08/13/2014 06:35 PM, Russell Bryant wrote:
  On 08/13/2014 06:23 PM, Mark McLoughlin wrote:
  On Wed, 2014-08-13 at 12:05 -0700, James E. Blair wrote:
  cor...@inaugust.com (James E. Blair) writes:
 
  Sean Dague s...@dague.net writes:
 
  This has all gone far enough that someone actually wrote a
  Grease Monkey script to purge all the 3rd Party CI content out
  of Jenkins UI. People are writing mail filters to dump all the
  notifications. Dan Berange filters all them out of his gerrit
  query tools.
 
  I should also mention that there is a pending change to do
  something similar via site-local Javascript in our Gerrit:
 
 https://review.openstack.org/#/c/95743/
 
  I don't think it's an ideal long-term solution, but if it works,
  we may have some immediate relief without all having to install
  greasemonkey scripts.
 
  You may have noticed that this has merged, along with a further
  change that shows the latest results in a table format.  (You may
  need to force-reload in your browser to see the change.)
 
  Beautiful! Thank you so much to everyone involved.
 
  +1!  Love this.
 
 Indeed. Amazeballs.


Agreed! This is a really nice improvement

Chris

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [Neutron] Simple proposal for stabilizing new features in-tree

2014-08-14 Thread Wuhongning
FWaas can't seamlessly work with DVR yet. A BP [1] has been submitted, but it 
can only handle NS traffic, leaving W-E untouched. If we implement the WE 
firewall in DVR, the iptable might be applied at a per port basis, so there are 
some overlapping with SG (Can we image a packet run into iptable hook twice 
between VM and the wire, for both ingress and egress directions?).

Maybe the overall service plugins (including service extension in ML2) needs 
some cleaning up, It seems that Neutron is just built from separate single 
blocks.

[1]  
http://git.openstack.org/cgit/openstack/neutron-specs/tree/specs/juno/neutron-dvr-fwaas.rst


From: Sridar Kandaswamy (skandasw) [skand...@cisco.com]
Sent: Thursday, August 14, 2014 3:07 PM
To: OpenStack Development Mailing List (not for usage questions)
Subject: Re: [openstack-dev] [Neutron] Simple proposal for stabilizing new 
features in-tree


Hi Aaron:


There is a certain fear of another cascading chain of emails so with hesitation 
i send this email out. :-)


1) I could not agree with u more on the issue with the logs and the pain with 
debugging issues here. Yes for sure bugs do keep popping up but often times, 
(speaking for fwaas) - given the L3 agent interactions - there are a multitude 
of reasons for a failure. An L3Agent crash or a router issue - also manifests 
itself as an fwaas issue - i think ur first paste is along those lines (perhaps 
i could be wrong without much context here).


The L3Agent - service coexistence is far from ideal - but this experience has 
lead to two proposals - a vendor proposal [1] that actually tries to address 
such agent limitations and collaboration on another community proposal[2] to 
enable the L3 Agent to be more suited to hosting services. Hopefully [2] will 
get picked up in K and should help provide the necessary infrastructure to 
clean up the reference implementation.


2) Regarding ur point on the  FWaaS API - the intent of the feature by design 
was to keep the service abstraction separate from how it is inserted in the 
network - to keep this vendor/technology neutral. The first priority, post 
Havana was to address  service insertion to get away from the all routers model 
[3] but did not get the blessings needed. Now with a redrafted proposal for 
Juno[4] again an effort is being made to address this now for the second time 
in the 2 releases post H.


In general, I would make a request that before we decide to go ahead and start 
moving things out into an incubator area - more discussion is needed.  We don’t 
want  to land up in a situation in K-3 where we find out that this model does 
not quite work for whatever reason. Also i wonder about the hoops to get out 
from incubation. As vendors who try to align with the community and upstream 
our work - we don’t want to land up waiting more cycles - there is quite a bit 
of frustration that is felt here too.


Lets also think about the impacts on feature velocity, somehow that keeps 
popping into my head every time i buy a book from a certain online retailer. :-)


[1] https://review.openstack.org/#/c/90729/

[2] https://review.openstack.org/#/c/91532/

[3] https://review.openstack.org/#/c/62599/

[4] https://review.openstack.org/#/c/93128/


Thanks


Sridar


From: Aaron Rosen aaronoro...@gmail.commailto:aaronoro...@gmail.com
Reply-To: OpenStack List 
openstack-dev@lists.openstack.orgmailto:openstack-dev@lists.openstack.org
Date: Wednesday, August 13, 2014 at 3:56 PM
To: OpenStack List 
openstack-dev@lists.openstack.orgmailto:openstack-dev@lists.openstack.org
Subject: Re: [openstack-dev] [Neutron] Simple proposal for stabilizing new 
features in-tree

Hi,

I've been thinking a good bit on this on the right way to move forward with 
this and in general the right way new services should be added. Yesterday I was 
working on a bug that was causing some problems in the openstack infra. We 
tracked down the issue then I uploaded a patch for it. A little while later 
after jenkins voted back a -1 so I started looking through the logs to see what 
the source of the failure was (which was actually unrelated to my patch). The 
random failure in the fwaas/vpn/l3-agent code which all outputs to the same log 
file that contains many traces for every run even successful ones. In one skim 
of this log file I was able to spot 4 [1]bugs which shows  these new 
experimental services that we've added to neutron have underlying problems 
(even though they've been in the tree for 2 releases+ now). This puts a huge 
strain on the whole openstack development community as they are always recheck 
no bug'ing due to neutron failures.

If you look at the fwaas work that was done. This merged over two releases ago 
and still does not have a complete API as there is no concept of where 
enforcement should be done. Today, enforcement is done across all of a tenant's 
routers making it more or less useless imho and we're just carrying it along in 
the tree 

Re: [openstack-dev] [neutron][ml2] Mech driver as out-of-tree add-on

2014-08-14 Thread Mathieu Rohon
Hi,

I would like to add that it would be harder for the community to help
maintaining drivers.
such a work [1] wouldn't have occured with an out of tree ODL driver.

[1] https://review.openstack.org/#/c/96459/

On Wed, Aug 13, 2014 at 1:09 PM, Robert Kukura kuk...@noironetworks.com wrote:
 One thing to keep in mind is that the ML2 driver API does sometimes change,
 requiring updates to drivers. Drivers that are in-tree get updated along
 with the driver API change. Drivers that are out-of-tree must be updated by
 the owner.

 -Bob


 On 8/13/14, 6:59 AM, ZZelle wrote:

 Hi,


 The important thing to understand is how to integrate with neutron through
 stevedore/entrypoints:

 https://github.com/dave-tucker/odl-neutron-drivers/blob/master/setup.cfg#L32-L34


 Cedric


 On Wed, Aug 13, 2014 at 12:17 PM, Dave Tucker d...@dtucker.co.uk wrote:

 I've been working on this for OpenDaylight
 https://github.com/dave-tucker/odl-neutron-drivers

 This seems to work for me (tested Devstack w/ML2) but YMMV.

 -- Dave

 ___
 OpenStack-dev mailing list
 OpenStack-dev@lists.openstack.org
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev




 ___
 OpenStack-dev mailing list
 OpenStack-dev@lists.openstack.org
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev



 ___
 OpenStack-dev mailing list
 OpenStack-dev@lists.openstack.org
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


[openstack-dev] [nova][rally] nova-network, unable to associate floating ip

2014-08-14 Thread Li Tianqing
Hello,
Recently we use rally to test our cloud. We run boot-run-commnd-delete 
task. The configuration is this:
{
VMTasks.boot_runcommand_delete: [
{
args: {
flavor: {
name: m1.small
},
image: {
name: ubuntu-12-04-raw-rally-test
},
script: /home/rally/doc/samples/ec_script/ubuntu_ls_test.sh,
interpreter: bash,
username: root,
floating_network: LTQ,
use_floatingip: true,
availability_zone: dell420
},
runner: {
type: constant,
times: 1000,
concurrency: 40,
timeout: 6000
},
context: {
users: {
tenants: 1,
users_per_tenant: 1
},
quotas: {
 nova: {
 instances: -1,
 cores: -1,
 ram: -1,
 fixed_ips: -1,
 floating_ips: -1,
 }
 },
}
}
]
}
There are almose 39 Exceptions for unable to associate floating ips.
The output of rally is
Traceback (most recent call last):\n  File 
\/opt/rally/local/lib/python2.7/site-packages/rally/benchmark/runners/base.py\,
 line 62, in _run_scenario_once\nmethod_name)(**kwargs) or {}\n  File 
\/opt/rally/local/lib/python2.7/site-packages/rally/benchmark/scenarios/vm/vmtasks.py\,
 line 80, in boot_runcommand_delete\nself._associate_floating_ip(server, 
floating_ip)\n  File 
\/opt/rally/local/lib/python2.7/site-packages/rally/benchmark/scenarios/utils.py\,
 line 139, in func_atomic_actions\nf = func(self, *args, **kwargs)\n  File 
\/opt/rally/local/lib/python2.7/site-packages/rally/benchmark/scenarios/nova/utils.py\,
 line 402, in _associate_floating_ip\nserver.add_floating_ip(address, 
fixed_address=fixed_address)\n  File 
\/opt/rally/local/lib/python2.7/site-packages/novaclient/v1_1/servers.py\, 
line 123, in add_floating_ip\nself.manager.add_floating_ip(self, address, 
fixed_address)\n  File 
\/opt/rally/local/lib/python2.7/site-packages/novaclient/v1_1/servers.py\, 
line 631, in add_floating_ip\nself._action('addFloatingIp', server, 
{'address': address})\n  File 
\/opt/rally/local/lib/python2.7/site-packages/novaclient/v1_1/servers.py\, 
line 1190, in _action\nreturn self.api.client.post(url, body=body)\n  File 
\/opt/rally/local/lib/python2.7/site-packages/novaclient/client.py\, line 
485, in post\nreturn self._cs_request(url, 'POST', **kwargs)\n  File 
\/opt/rally/local/lib/python2.7/site-packages/novaclient/client.py\, line 
459, in _cs_request\n**kwargs)\n  File 
\/opt/rally/local/lib/python2.7/site-packages/novaclient/client.py\, line 
441, in _time_request\nresp, body = self.request(url, method, **kwargs)\n  
File \/opt/rally/local/lib/python2.7/site-packages/novaclient/client.py\, 
line 435, in request\nraise exceptions.from_response(resp, body, url, 
method)\nBadRequest: Error. Unable to associate floating ip (HTTP 400) 
(Request-ID: req-7b85f1cb-664d-4c98-b140-66f8b8d2ebe4)\n


I do not why associatiting floating ip is failed. I see in /var/log/sys.log
found this:
Aug 14 16:13:26 compute-node10 dnsmasq[21117]: failed to load names from 
/var/lib/nova/networks/nova-br100.hosts: No such file or directory
Aug 14 16:13:26 compute-node10 dnsmasq-dhcp[21117]: read 
/var/lib/nova/networks/nova-br100.conf
Aug 14 16:13:26 compute-node10 ceph-mon: 2014-08-14 16:13:26.952565 
7f7b2d65a700  1 mon.compute-node10@5(peon).paxos(paxos active c 
4947462..4947989) is_readable now=2014-08-14 16:13:26.952569 
lease_expire=2014-08-14 16:13:31.935450 has v0 lc 4947989
Aug 14 16:13:27 compute-node10 ceph-mon: 2014-08-14 16:13:27.425576 
7f7b2d65a700  1 mon.compute-node10@5(peon).paxos(paxos active c 
4947462..4947990) is_readable now=2014-08-14 16:13:27.425578 
lease_expire=2014-08-14 16:13:32.408369 has v0 lc 4947990
Aug 14 16:13:28 compute-node10 ceph-mon: 2014-08-14 16:13:28.368894 
7f7b2d65a700  1 mon.compute-node10@5(peon).paxos(paxos active c 
4947462..4947991) is_readable now=2014-08-14 16:13:28.368895 
lease_expire=2014-08-14 16:13:33.350551 has v0 lc 4947991
Aug 14 16:13:28 compute-node10 ceph-mon: 2014-08-14 16:13:28.504170 
7f7b2d65a700  1 mon.compute-node10@5(peon).paxos(paxos active c 
4947462..4947992) is_readable now=2014-08-14 16:13:28.504171 
lease_expire=2014-08-14 16:13:33.486123 has v0 lc 4947992
Aug 14 16:13:29 compute-node10 ceph-mon: 2014-08-14 16:13:29.593776 
7f7b2d65a700  1 mon.compute-node10@5(peon).paxos(paxos active c 
4947462..4947993) is_readable now=2014-08-14 16:13:29.593777 
lease_expire=2014-08-14 16:13:34.575913 has v0 lc 4947993




Although sys.log says that no address available, 

Re: [openstack-dev] [nova][core] Expectations of core reviewers

2014-08-14 Thread Daniel P. Berrange
On Thu, Aug 14, 2014 at 09:24:36AM +1000, Michael Still wrote:
 On Thu, Aug 14, 2014 at 3:09 AM, Dan Smith d...@danplanet.com wrote:
  I'm not questioning the value of f2f - I'm questioning the idea of
  doing f2f meetings sooo many times a year. OpenStack is very much
  the outlier here among open source projects - the vast majority of
  projects get along very well with much less f2f time and a far
  smaller % of their contributors attend those f2f meetings that do
  happen. So I really do question what is missing from OpenStack's
  community interaction that makes us believe that having 4 f2f
  meetings a year is critical to our success.
 
  How many is too many? So far, I have found the midcycles to be extremely
  productive -- productive in a way that we don't see at the summits, and
  I think other attendees agree. Obviously if budgets start limiting them,
  then we'll have to deal with it, but I don't want to stop meeting
  preemptively.
 
 I agree they're very productive. Let's pick on the nova v3 API case as
 an example... We had failed as a community to reach a consensus using
 our existing discussion mechanisms (hundreds of emails, at least three
 specs, phone calls between the various parties, et cetera), yet at the
 summit and then a midcycle meetup we managed to nail down an agreement
 on a very contentious and complicated topic.

We thought we had agreement on v3 API after Atlanta f2f summit and 
after Hong Kong f2f too. So I wouldn't neccessarily say that we
needed another f2f meeting to resolve that, but rather than this is
a very complex topic that takes a long time to resolve no matter
how we discuss it and the discussions had just happened to reach
a natural conclusion this time around. But lets see if this agreement
actually sticks this time

 I can see the argument that travel cost is an issue, but I think its
 also not a very strong argument. We have companies spending millions
 of dollars on OpenStack -- surely spending a relatively small amount
 on travel to keep the development team as efficient as possible isn't
 a big deal? I wouldn't be at all surprised if the financial costs of
 the v3 API debate (staff time mainly) were much higher than the travel
 costs of those involved in the summit and midcycle discussions which
 sorted it out.

I think the travel cost really is a big issue. Due to the number of
people who had to travel to the many mid-cycle meetups, a good number
of people I work with no longer have the ability to go to the Paris
design summit. This is going to make it harder for them to feel a
proper engaged part of our community. I can only see this situation
get worse over time if greater emphasis is placed on attending the
mid-cycle meetups.

 Travelling to places to talk to people isn't a great solution, but it
 is the most effective one we've found so far. We should continue to
 experiment with other options, but until we find something that works
 as well as meetups, I think we need to keep having them.
 
  IMHO, the reasons to cut back would be:
 
  - People leaving with a well, that was useless... feeling
  - Not enough people able to travel to make it worthwhile
 
  So far, neither of those have been outcomes of the midcycles we've had,
  so I think we're doing okay.
 
  The design summits are structured differently, where we see a lot more
  diverse attendance because of the colocation with the user summit. It
  doesn't lend itself well to long and in-depth discussions about specific
  things, but it's very useful for what it gives us in the way of
  exposure. We could try to have less of that at the summit and more
  midcycle-ish time, but I think it's unlikely to achieve the same level
  of usefulness in that environment.
 
  Specifically, the lack of colocation with too many other projects has
  been a benefit. This time, Mark and Maru where there from Neutron. Last
  time, Mark from Neutron and the other Mark from Glance were there. If
  they were having meetups in other rooms (like at summit) they wouldn't
  have been there exposed to discussions that didn't seem like they'd have
  a component for their participation, but did after all (re: nova and
  glance and who should own flavors).
 
 I agree. The ability to focus on the issues that were blocking nova
 was very important. That's hard to do at a design summit when there is
 so much happening at the same time.

Maybe we should change the way we structure the design summit to
improve that. If there are critical issues blocking nova, it feels
like it is better to be able to discuss and resolve as much as possible
at the start of the dev cycle rather than in the middle of the dev
cycle because I feel that means we are causing ourselves pain during
milestone 1/2.

  As I explain in the rest of my email below I'm not advocating
  getting rid of mid-cycle events entirely. I'm suggesting that
  we can attain a reasonable % of the benefits of f2f meetings
  by doing more formal virtual meetups and so be 

Re: [openstack-dev] [all] 3rd Party CI vs. Gerrit

2014-08-14 Thread Daniel P. Berrange
On Wed, Aug 13, 2014 at 12:05:27PM -0700, James E. Blair wrote:
 cor...@inaugust.com (James E. Blair) writes:
 
  Sean Dague s...@dague.net writes:
 
  This has all gone far enough that someone actually wrote a Grease Monkey
  script to purge all the 3rd Party CI content out of Jenkins UI. People
  are writing mail filters to dump all the notifications. Dan Berange
  filters all them out of his gerrit query tools.
 
  I should also mention that there is a pending change to do something
  similar via site-local Javascript in our Gerrit:
 
https://review.openstack.org/#/c/95743/
 
  I don't think it's an ideal long-term solution, but if it works, we may
  have some immediate relief without all having to install greasemonkey
  scripts.
 
 You may have noticed that this has merged, along with a further change
 that shows the latest results in a table format.  (You may need to
 force-reload in your browser to see the change.)

 Thanks again to Radoslav Gerganov for writing the original change.

These are both a great step forward in usability of the Gerrit web UI.
Thanks for the effort everyone put into these changes.

Regards,
Daniel
-- 
|: http://berrange.com  -o-http://www.flickr.com/photos/dberrange/ :|
|: http://libvirt.org  -o- http://virt-manager.org :|
|: http://autobuild.org   -o- http://search.cpan.org/~danberr/ :|
|: http://entangle-photo.org   -o-   http://live.gnome.org/gtk-vnc :|

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [nova] Retrospective veto revert policy

2014-08-14 Thread Mark McLoughlin
On Tue, 2014-08-12 at 15:56 +0100, Mark McLoughlin wrote:
 Hey
 
 (Terrible name for a policy, I know)
 
 From the version_cap saga here:
 
   https://review.openstack.org/110754
 
 I think we need a better understanding of how to approach situations
 like this.
 
 Here's my attempt at documenting what I think we're expecting the
 procedure to be:
 
   https://etherpad.openstack.org/p/nova-retrospective-veto-revert-policy
 
 If it sounds reasonably sane, I can propose its addition to the
 Development policies doc.

(In the spirit of we really need to step back and laugh at ourselves
sometimes ... )

Two years ago, we were worried about patches getting merged in less than
2 hours and had a discussion about imposing a minimum review time. How
times have changed! Is it even possible to land a patch in less than two
hours now? :)

Looking back over the thread, this part stopped me in my tracks:

  https://lists.launchpad.net/openstack/msg08625.html

On Tue, Mar 13, 2012, Mark McLoughlin markmc@xx wrote:

 Sometimes there can be a few folks working through an issue together and
 the patch gets pushed and approved so quickly that no-one else gets a
 chance to review.

Everyone has an opportunity to review even after a patch gets merged.

JE

It's not quite perfect, but if you squint you could conclude that
Johannes and I have both completely reversed our opinions in the
intervening two years :)

The lesson I take from that is to not get too caught up in the current
moment. We're growing and evolving rapidly. If we assume everyone is
acting in good faith, and allow each other to debate earnestly without
feelings getting hurt ... we should be able to work through anything.

Now, back on topic - digging through that thread, it doesn't seem we
settled on the idea of we can just revert it later if someone has an
objection in this thread. Does anyone recall when that idea first came
up?

Thanks,
Mark.


___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [nova] Retrospective veto revert policy

2014-08-14 Thread Daniel P. Berrange
On Thu, Aug 14, 2014 at 10:06:13AM +0100, Mark McLoughlin wrote:
 On Tue, 2014-08-12 at 15:56 +0100, Mark McLoughlin wrote:
  Hey
  
  (Terrible name for a policy, I know)
  
  From the version_cap saga here:
  
https://review.openstack.org/110754
  
  I think we need a better understanding of how to approach situations
  like this.
  
  Here's my attempt at documenting what I think we're expecting the
  procedure to be:
  
https://etherpad.openstack.org/p/nova-retrospective-veto-revert-policy
  
  If it sounds reasonably sane, I can propose its addition to the
  Development policies doc.
 
 (In the spirit of we really need to step back and laugh at ourselves
 sometimes ... )
 
 Two years ago, we were worried about patches getting merged in less than
 2 hours and had a discussion about imposing a minimum review time. How
 times have changed! Is it even possible to land a patch in less than two
 hours now? :)
 
 Looking back over the thread, this part stopped me in my tracks:
 
   https://lists.launchpad.net/openstack/msg08625.html
 
 On Tue, Mar 13, 2012, Mark McLoughlin markmc@xx wrote:
 
  Sometimes there can be a few folks working through an issue together and
  the patch gets pushed and approved so quickly that no-one else gets a
  chance to review.
 
 Everyone has an opportunity to review even after a patch gets merged.
 
 JE
 
 It's not quite perfect, but if you squint you could conclude that
 Johannes and I have both completely reversed our opinions in the
 intervening two years :)
 
 The lesson I take from that is to not get too caught up in the current
 moment. We're growing and evolving rapidly. If we assume everyone is
 acting in good faith, and allow each other to debate earnestly without
 feelings getting hurt ... we should be able to work through anything.
 
 Now, back on topic - digging through that thread, it doesn't seem we
 settled on the idea of we can just revert it later if someone has an
 objection in this thread. Does anyone recall when that idea first came
 up?

Probably lost in time - I've seen it said several times on Nova IRC
channel over the year(s) when we made a strategic decision to merge
something quickly.

Regards,
Daniel
-- 
|: http://berrange.com  -o-http://www.flickr.com/photos/dberrange/ :|
|: http://libvirt.org  -o- http://virt-manager.org :|
|: http://autobuild.org   -o- http://search.cpan.org/~danberr/ :|
|: http://entangle-photo.org   -o-   http://live.gnome.org/gtk-vnc :|

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [Neutron] Is network ordering of vNICs guaranteed?

2014-08-14 Thread Aaron Rosen
Most guests generate this on first boot. I haven't looked into how newer
versions of ubuntu handle this though  I noticed in 14.04 this file doesn't
exist anymore but I figure it's doing it somewhere else.

Aaron


On Wed, Aug 13, 2014 at 9:03 PM, Jian Wen wenjia...@gmail.com wrote:

 Ordering of vNICs is not 100% guaranteed for the cloud images which are
 not shipped with
 /etc/udev/rules.d/70-persistent-net.rules.
 e.g. attaching a port and deattaching another port, then reboot the
 instance.


 2014-08-12 15:52 GMT+08:00 Aaron Rosen aaronoro...@gmail.com:

 This bug was true in grizzly and older (and was reintroduced in icehouse
 for a few days but was fixed before the nova icehouse shipped).

 Aaron


 On Mon, Aug 11, 2014 at 7:10 AM, CARVER, PAUL pc2...@att.com wrote:

 Armando M. [mailto:arma...@gmail.com] wrote:



 On 9 August 2014 10:16, Jay Pipes jaypi...@gmail.com wrote:

 Paul, does this friend of a friend have a reproduceable test

 script for this?



 We would also need to know the OpenStack release where this issue
 manifest

 itself. A number of bugs have been raised in the past around this type
 of

 issue, and the last fix I recall is this one:

 

 https://bugs.launchpad.net/nova/+bug/1300325

 

 It's possible that this might have regressed, though.



 The reason I called it friend of a friend is because I think the info

 has filtered through a series of people and is not firsthand observation.

 I'll ask them to track back to who actually observed the behavior, how

 long ago, and with what version.



 It could be a regression, or it could just be old info that people have

 continued to assume is true without realizing it was considered a bug

 all along and has been fixed.



 Thanks! The moment I first heard it my first reaction was that it was

 almost certainly a bug and had probably already been fixed.



 ___
 OpenStack-dev mailing list
 OpenStack-dev@lists.openstack.org
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev



 ___
 OpenStack-dev mailing list
 OpenStack-dev@lists.openstack.org
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev




 --
 Best,

 Jian

 ___
 OpenStack-dev mailing list
 OpenStack-dev@lists.openstack.org
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [OpenStack-Dev] [Cinder] 3'rd party CI systems

2014-08-14 Thread David Pineau
Well, I could make it test every patchset, but then I have to look
into how to limit at one build per time + queue the other builds.
(didn't try that yet)

Anyways, I just got my new CI machine so I'll be able to finalize the
CI system relatively soon.

2014-08-14 1:01 GMT+02:00 Jeremy Stanley fu...@yuggoth.org:
 On 2014-08-13 16:30:23 + (+), Asselin, Ramy wrote:
 I remember infra team objected to the nightly builds. They wanted
 reports on every patch set in order to report to gerrit.
 [...]

 I can't imagine, nor do I recall, objecting to such an idea. The
 question is merely where you expect to publish the results, and how
 you deal with the fact that you're testing changes which have
 already merged rather than getting in front of the reviewers on
 proposed changes. I don't personally have any specific desire for
 third-party CI systems to report on changes in Gerrit, but
 individual projects supporting your drivers/features/whatever might.
 --
 Jeremy Stanley

 ___
 OpenStack-dev mailing list
 OpenStack-dev@lists.openstack.org
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev



-- 
David Pineau,
Developer RD at Scality

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [neutron][ml2] Mech driver as out-of-tree add-on

2014-08-14 Thread loy wolfe
On Thu, Aug 14, 2014 at 4:22 PM, Mathieu Rohon mathieu.ro...@gmail.com
wrote:

 Hi,

 I would like to add that it would be harder for the community to help
 maintaining drivers.
 such a work [1] wouldn't have occured with an out of tree ODL driver.


+1.
It's better to move all MD for none built-in backend out of tree,
maintaining these drivers shouldn't be the responsibility of community. Not
only MD, but also plugin, agent should all obey this rule



 [1] https://review.openstack.org/#/c/96459/

 On Wed, Aug 13, 2014 at 1:09 PM, Robert Kukura kuk...@noironetworks.com
 wrote:
  One thing to keep in mind is that the ML2 driver API does sometimes
 change,
  requiring updates to drivers. Drivers that are in-tree get updated along
  with the driver API change. Drivers that are out-of-tree must be updated
 by
  the owner.
 
  -Bob
 
 
  On 8/13/14, 6:59 AM, ZZelle wrote:
 
  Hi,
 
 
  The important thing to understand is how to integrate with neutron
 through
  stevedore/entrypoints:
 
 
 https://github.com/dave-tucker/odl-neutron-drivers/blob/master/setup.cfg#L32-L34
 
 
  Cedric
 
 
  On Wed, Aug 13, 2014 at 12:17 PM, Dave Tucker d...@dtucker.co.uk
 wrote:
 
  I've been working on this for OpenDaylight
  https://github.com/dave-tucker/odl-neutron-drivers
 
  This seems to work for me (tested Devstack w/ML2) but YMMV.
 
  -- Dave
 
  ___
  OpenStack-dev mailing list
  OpenStack-dev@lists.openstack.org
  http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
 
 
 
 
  ___
  OpenStack-dev mailing list
  OpenStack-dev@lists.openstack.org
  http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
 
 
 
  ___
  OpenStack-dev mailing list
  OpenStack-dev@lists.openstack.org
  http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
 

 ___
 OpenStack-dev mailing list
 OpenStack-dev@lists.openstack.org
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [neutron] Which changes need accompanying bugs?

2014-08-14 Thread Andreas Jaeger
On 08/14/2014 11:37 AM, Ihar Hrachyshka wrote:
 On 14/08/14 02:43, Angus Lees wrote:
 On Wed, 13 Aug 2014 11:11:51 AM Kevin Benton wrote:
 Is the pylint static analysis that caught that error prone to
 false positives? If not, I agree that it would be really nice if
 that were made part of the tox check so these don't have to be
 fixed after the fact.
 
 At the moment pylint on neutron is *very* noisy, and I've been
 looking through the reported issues by hand to get a feel for
 what's involved.  Enabling pylint is a separate discussion that I'd
 like to have - in some other thread.
 
 
 Just start with non-voting check. Once all the issues are fixed, we
 can enable it as voting.

Or blacklist all failures, make it voting - and then fix one issue after
the other - and with each patch, remove the blacklisted number.

That way no regressions come in...

Andreas
-- 
 Andreas Jaeger aj@{suse.com,opensuse.org} Twitter/Identica: jaegerandi
  SUSE LINUX Products GmbH, Maxfeldstr. 5, 90409 Nürnberg, Germany
   GF: Jeff Hawn,Jennifer Guild,Felix Imendörffer,HRB16746 (AG Nürnberg)
GPG fingerprint = 93A3 365E CE47 B889 DF7F  FED1 389A 563C C272 A126

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


[openstack-dev] [Congress] Help in understanding API Flow

2014-08-14 Thread Madhu Mohan
Hi,

Can some one help me understand how an API is exposed to the clients from
Congress server ?

I see that a cage service ('api-policy') is created in congress_server.py.
I believe this is implemented in policy_model. I tried to send a
json_request from my client on the server.

I tried sending list_members, get_items, PUT and  POST as methods
and all these give me NotImplemented error response.

Any help in this direction ?

I also want to add new APIs and hence understanding the API flow is crucial.

Thanks,
Madhu Mohan
___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [OpenStack-Infra] [infra] Third Party CI naming and contact (action required)

2014-08-14 Thread Franck Yelles
Hi James,

I have added the Nuage CI system to the list; I also took the liberty
to reorder alphabetically the list

Franck
Franck


On Wed, Aug 13, 2014 at 11:23 AM, James E. Blair cor...@inaugust.com wrote:
 Hi,

 We've updated the registration requirements for third-party CI systems
 here:

   http://ci.openstack.org/third_party.html

 We now have 86 third-party CI systems registered and have undertaken an
 effort to make things more user-friendly for the developers who interact
 with them.  There are two important changes to be aware of:

 1) We now generally name third-party systems in a descriptive manner
 including the company and product they are testing.  We have renamed
 currently-operating CI systems to match these standards to the best of
 our abilities.  Some of them ended up with particularly bad names (like
 Unknown Function...).  If your system is one of these, please join us
 in #openstack-infra on Freenode to establish a more descriptive name.

 2) We have established a standard wiki page template to supply a
 description of the system, what is tested, and contact information for
 each system.  See https://wiki.openstack.org/wiki/ThirdPartySystems for
 an index of such pages and instructions for creating them.  Each
 third-party CI system will have its own page in the wiki and it must
 include a link to that page in every comment that it leaves in Gerrit.

 If you operate a third-party CI system, please ensure that you register
 a wiki page and update your system to link to it in every new Gerrit
 comment by the end of August.  Beginning in September, we will disable
 systems that have not been updated.

 Thanks,

 Jim

 ___
 OpenStack-Infra mailing list
 openstack-in...@lists.openstack.org
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-infra

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [ceilometer] gate-ceilometer-python33 failed because of wrong setup.py in happybase

2014-08-14 Thread Dina Belova
Hisashi Osanai, np :)
You're welcome :)


On Thu, Aug 14, 2014 at 5:13 AM, Osanai, Hisashi 
osanai.hisa...@jp.fujitsu.com wrote:


 On Wed, Aug 13, 2014 at 2:35 PM, Julien Danjou  wrote:
  Means the py33 needs to execute on stable/icehouse. Here I
 misunderstand something...
  Not it does not, that line in tox.ini is not use by the gate.

  this is a problem in the infrastructure config.
  Means execfile function calls on python33 in happybase is a problem. If
 my understanding
  is correct, I agree with you and I think this is the direct cause of
 this problem.
 
  Your idea to solve this is creating a patch for the direct cause, right?
  My idea to solve this is to create a patch on
  http://git.openstack.org/cgit/openstack-infra/config/
  to exclude py33 on the stable/icehouse branch of Ceilometer in the gate.

 Sorry to use your time for explanation above again and thanks for it. I'm
 happy to have
 clear understanding your thought.

 On Wednesday, August 13, 2014 7:54 PM, Dina Belova wrote:
  Here it is: https://review.openstack.org/#/c/113842/
 Thank you for providing the fix. I surprised the speed for it. it's really
 fast...

 Thanks again!
 Hisashi Osanai




-- 

Best regards,

Dina Belova

Software Engineer

Mirantis Inc.
___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [neutron][ml2] Mech driver as out-of-tree add-on

2014-08-14 Thread Salvatore Orlando
I think there will soon be a discussion regarding what the appropriate
location for plugin and drivers should be.
My personal feeling is that Neutron has simply reached the tipping point
where the high number of drivers and plugins is causing unnecessary load
for the core team and frustration for the community.

There I would totally support Luke's initiative about maintaining an
out-of-tree ML2 driver. On the other hand, a plugin/driver diaspora might
also have negative consequences such as frequent breakages such as those
Bob was mentioning or confusion for users which might need to end up
fetching drivers from disparate sources.

As mentioned during the last Neutron IRC meeting this is another process
aspect which will be discussed soon, with the aim of defining a plan for:
- drastically reduce the number of plugins and drivers which must be
maintained in the main source tree
- enhance control of plugin/driver maintainers over their own code
- preserve the ability of doing CI checks on gerrit as we do today
- raise the CI bar (maybe finally set the smoketest as a minimum
requirement?)

Regards,
Salvatore



On 14 August 2014 11:47, loy wolfe loywo...@gmail.com wrote:




 On Thu, Aug 14, 2014 at 4:22 PM, Mathieu Rohon mathieu.ro...@gmail.com
 wrote:

 Hi,

 I would like to add that it would be harder for the community to help
 maintaining drivers.
 such a work [1] wouldn't have occured with an out of tree ODL driver.


 +1.
 It's better to move all MD for none built-in backend out of tree,
 maintaining these drivers shouldn't be the responsibility of community. Not
 only MD, but also plugin, agent should all obey this rule



 [1] https://review.openstack.org/#/c/96459/

 On Wed, Aug 13, 2014 at 1:09 PM, Robert Kukura kuk...@noironetworks.com
 wrote:
  One thing to keep in mind is that the ML2 driver API does sometimes
 change,
  requiring updates to drivers. Drivers that are in-tree get updated along
  with the driver API change. Drivers that are out-of-tree must be
 updated by
  the owner.
 
  -Bob
 
 
  On 8/13/14, 6:59 AM, ZZelle wrote:
 
  Hi,
 
 
  The important thing to understand is how to integrate with neutron
 through
  stevedore/entrypoints:
 
 
 https://github.com/dave-tucker/odl-neutron-drivers/blob/master/setup.cfg#L32-L34
 
 
  Cedric
 
 
  On Wed, Aug 13, 2014 at 12:17 PM, Dave Tucker d...@dtucker.co.uk
 wrote:
 
  I've been working on this for OpenDaylight
  https://github.com/dave-tucker/odl-neutron-drivers
 
  This seems to work for me (tested Devstack w/ML2) but YMMV.
 
  -- Dave
 
  ___
  OpenStack-dev mailing list
  OpenStack-dev@lists.openstack.org
  http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
 
 
 
 
  ___
  OpenStack-dev mailing list
  OpenStack-dev@lists.openstack.org
  http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
 
 
 
  ___
  OpenStack-dev mailing list
  OpenStack-dev@lists.openstack.org
  http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
 

 ___
 OpenStack-dev mailing list
 OpenStack-dev@lists.openstack.org
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev



 ___
 OpenStack-dev mailing list
 OpenStack-dev@lists.openstack.org
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [nova] Retrospective veto revert policy

2014-08-14 Thread Mark McLoughlin
On Tue, 2014-08-12 at 15:56 +0100, Mark McLoughlin wrote:
 Hey
 
 (Terrible name for a policy, I know)
 
 From the version_cap saga here:
 
   https://review.openstack.org/110754
 
 I think we need a better understanding of how to approach situations
 like this.
 
 Here's my attempt at documenting what I think we're expecting the
 procedure to be:
 
   https://etherpad.openstack.org/p/nova-retrospective-veto-revert-policy
 
 If it sounds reasonably sane, I can propose its addition to the
 Development policies doc.

Proposed here: https://review.openstack.org/114188

Thanks,
Mark.


___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [OpenStack-Infra] [infra] Third Party CI naming and contact (action required)

2014-08-14 Thread trinath.soman...@freescale.com
Hi Franck -

Thanks for the update. I too have that re-order in mind. :)

Here after CI owners may add their CI names in appropriate alphabetical order.

--
Trinath Somanchi - B39208
trinath.soman...@freescale.com | extn: 4048

-Original Message-
From: Franck Yelles [mailto:franck...@gmail.com] 
Sent: Thursday, August 14, 2014 3:33 PM
To: James E. Blair
Cc: OpenStack Development Mailing List (not for usage questions); 
openstack-in...@lists.openstack.org
Subject: Re: [OpenStack-Infra] [infra] Third Party CI naming and contact 
(action required)

Hi James,

I have added the Nuage CI system to the list; I also took the liberty to 
reorder alphabetically the list

Franck
Franck


On Wed, Aug 13, 2014 at 11:23 AM, James E. Blair cor...@inaugust.com wrote:
 Hi,

 We've updated the registration requirements for third-party CI systems
 here:

   http://ci.openstack.org/third_party.html

 We now have 86 third-party CI systems registered and have undertaken 
 an effort to make things more user-friendly for the developers who 
 interact with them.  There are two important changes to be aware of:

 1) We now generally name third-party systems in a descriptive manner 
 including the company and product they are testing.  We have renamed 
 currently-operating CI systems to match these standards to the best of 
 our abilities.  Some of them ended up with particularly bad names 
 (like Unknown Function...).  If your system is one of these, please 
 join us in #openstack-infra on Freenode to establish a more descriptive name.

 2) We have established a standard wiki page template to supply a 
 description of the system, what is tested, and contact information for 
 each system.  See https://wiki.openstack.org/wiki/ThirdPartySystems 
 for an index of such pages and instructions for creating them.  Each 
 third-party CI system will have its own page in the wiki and it must 
 include a link to that page in every comment that it leaves in Gerrit.

 If you operate a third-party CI system, please ensure that you 
 register a wiki page and update your system to link to it in every new 
 Gerrit comment by the end of August.  Beginning in September, we will 
 disable systems that have not been updated.

 Thanks,

 Jim

 ___
 OpenStack-Infra mailing list
 openstack-in...@lists.openstack.org
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-infra

___
OpenStack-Infra mailing list
openstack-in...@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-infra

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] Annoucing CloudKitty : an OpenSource Rating-as-a-Service project for OpenStack

2014-08-14 Thread Fei Long Wang
Glad to see there is another group interested in rating/billing. I'm 
thinking if I should announce our rating service to join the fun :-) You 
know, I'm kidding. The more groups join the rating service discussion, 
the better. But IMHO, we just need one rating service for OpenStack. So 
let's collaborate to work out a good rating service to benefit OpenStack 
and consumer. But before more duplicate work, it would be great if we 
can start some discussion. Cheers.


On 14/08/14 01:40, Christophe Sauthier wrote:
We are very pleased at Objectif Libre to intoduce CloudKitty, an 
effort to provide a fully OpenSource Rating-as-a-Service comptinent in 
OpenStack..


Following a first POC presented during the last summit in Atlanta to 
some Ceilometer devs (thanks again Julien Danjou for your great 
support !), we continued our effort to create a real service for 
rating. Today we are happy to share it with you all.



So what do we propose in CloudKitty?
 - a service for collecting metrics (using Ceilometer API)
 - a modular rating architecture to enable/disable modules and create 
your own rules on-the-fly, allowing you to use the rating patterns you 
like
 - an API to interact with the whole environment from core components 
to every rating module
 - a Horizon integration to allow configuration of the rating modules 
and display of pricing information in real time during instance 
creation
 - a CLI client to access this information and easily configure 
everything


Technically we are using all the elements that are used in the various 
OpenStack projects like olso, stevedore, pecan...
CloudKitty is highly modular and allows integration / developement of 
third party collection and rating modules and output formats.


A roadmap is available on the project wiki page (the link is at the 
end of this email), but we are clearly hoping to have some feedback 
and ideas on how to improve the project and reach a tighter 
integration with OpenStack.


The project source code is available at 
http://github.com/stackforge/cloudkitty
More stuff will be available on stackforge as soon as the reviews get 
validated like python-cloudkittyclient and cloudkitty-dashboard, so 
stay tuned.


The project's wiki page (https://wiki.openstack.org/wiki/CloudKitty) 
provides more information, and you can reach us via irc on freenode: 
#cloudkitty. Developper's documentation is on its way to readthedocs too.


We plan to present CloudKitty in detail during the Paris Summit, but 
we would love to hear from you sooner...


Cheers,

 Christophe and Objectif Libre


Christophe Sauthier   Mail : 
christophe.sauth...@objectif-libre.com

CEO  Fondateur   Mob : +33 (0) 6 16 98 63 96
Objectif LibreURL : 
www.objectif-libre.com

Infrastructure et Formations LinuxTwitter : @objectiflibre

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev



___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


[openstack-dev] [vmware] Canonical list of os types

2014-08-14 Thread Matthew Booth
I've just spent the best part of a day tracking down why instance
creation was failing on a particular setup. The error message from
CreateVM_Task was: 'A specified parameter was not correct'.

After discounting a great many possibilities, I finally discovered that
the problem was guestId, which was being set to 'CirrosGuest'.
Unusually, the vSphere API docs don't contain a list of valid values for
that field. Given the unhelpfulness of the error message, it might be
worthwhile validating that field (which we get from glance) and
displaying an appropriate warning.

Does anybody have a canonical list of valid values?

Thanks,

Matt
-- 
Matthew Booth
Red Hat Engineering, Virtualisation Team

Phone: +442070094448 (UK)
GPG ID:  D33C3490
GPG FPR: 3733 612D 2D05 5458 8A8A 1600 3441 EA19 D33C 3490

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [Devstack] [Cinder] Registering 3rd party driver as default

2014-08-14 Thread Amit Das
With further debugging, i find that none of the configuration options
present in /etc/cinder/cinder.conf are getting applied.

Regards,
Amit
*CloudByte Inc.* http://www.cloudbyte.com/


On Thu, Aug 14, 2014 at 11:40 AM, Amit Das amit@cloudbyte.com wrote:

 Hi folks,

 I have been trying to run devstack with my cinder driver as the default
 volume_driver but with no luck.

  Devstack seems to register the lvm driver as the default always.

 I have tried below approaches:

1. directly modifying the /etc/cinder/cinder.conf file
2. creating a driver file @ ./devstack/lib/cinder_plugins/driver-name
1. ref - https://review.openstack.org/#/c/68726/


 This is my localrc details:
 http://paste.openstack.org/show/94822/

 I run ./unstack.sh  then FORCE=yes ./stack.sh

 This is the cinder.conf that is generated after running above stack.sh. I
 comment out the [lvmdriver-1] section manually *(not sure if this section
 needs to be commented)*

 http://paste.openstack.org/show/94841/

 These are portions of c-sch  c-vol logs after restarting them in their
 respective screens.

 http://paste.openstack.org/show/94842/

 Regards,
 Amit
 *CloudByte Inc.* http://www.cloudbyte.com/

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [vmware] Canonical list of os types

2014-08-14 Thread Steve Gordon
- Original Message -
 From: Matthew Booth mbo...@redhat.com
 To: OpenStack Development Mailing List (not for usage questions) 
 openstack-dev@lists.openstack.org
 
 I've just spent the best part of a day tracking down why instance
 creation was failing on a particular setup. The error message from
 CreateVM_Task was: 'A specified parameter was not correct'.
 
 After discounting a great many possibilities, I finally discovered that
 the problem was guestId, which was being set to 'CirrosGuest'.
 Unusually, the vSphere API docs don't contain a list of valid values for
 that field. Given the unhelpfulness of the error message, it might be
 worthwhile validating that field (which we get from glance) and
 displaying an appropriate warning.
 
 Does anybody have a canonical list of valid values?
 
 Thanks,
 
 Matt

I found a page [1] linked from the Grizzly edition of the compute guide  [2] 
which has since been superseded. The content that would appear to have replaced 
it in more recent versions of the documentation suite [3] does not appear to 
contain such a link though. If a link to a more formal list is available it 
would be great to get this in the documentation.

Thanks,

Steve

[1] http://www.thinkvirt.com/?q=node/181
[2] 
http://docs.openstack.org/grizzly/openstack-compute/admin/content/image-metadata.html
[3] http://docs.openstack.org/icehouse/config-reference/content/vmware.html

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [neutron][cisco] Cisco Nexus requires patched ncclient

2014-08-14 Thread Ihar Hrachyshka
-BEGIN PGP SIGNED MESSAGE-
Hash: SHA512

FYI: I've uploaded a review for openstack/requirements to add the
upstream module into the list of potential dependencies [1]. Once it's
merged, I'm going to introduce this new requirement for Neutron.

[1]: https://review.openstack.org/114213

/Ihar

On 12/08/14 16:27, Ihar Hrachyshka wrote:
 Hey all,
 
 as per [1], Cisco Nexus ML2 plugin requires a patched version of 
 ncclient from github. I wonder:
 
 - whether this information is still current; - why don't we depend
 on ncclient thru our requirements.txt file.
 
 [1]: https://wiki.openstack.org/wiki/Neutron/ML2/MechCiscoNexus
 
 Cheers, /Ihar
 
 ___ OpenStack-dev
 mailing list OpenStack-dev@lists.openstack.org 
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
 
-BEGIN PGP SIGNATURE-
Version: GnuPG/MacGPG2 v2.0.22 (Darwin)

iQEcBAEBCgAGBQJT7KO6AAoJEC5aWaUY1u57ft0IAJzPC2o8fz9mwlC47tXGeHpJ
uP5HFIMaJaPXIwDnuz2uz8dJQZzBvTo6WrN+SRPb6PL3aGbIW4y2BA4n6NM276Xx
PwL8cnBdjQ9INwxn3g9jBceynbm2Yxx3I//2AIT1iu1Io/qjHppkUePxgH33PVMn
jw/n00mnCVJgxpHNXuFoe7Mn8UsduhB7xNCnW90t4rc9cfClGhW1T6/Pw2PWc07p
3Kw4OmmTS7q6r89iAmkBgbgdI2WBjdR902gwGxnuwmf7TJLo5Nd+jnvbe4bJ7V1d
DHzcFPNffKS2Fjbbxay0GjN+7/dAKHVRqkGQZvGzzEL8ZvqZGzyFojZ9BlnrM4Q=
=BZEc
-END PGP SIGNATURE-

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [nova][core] Expectations of core reviewers

2014-08-14 Thread Russell Bryant
On 08/13/2014 07:27 PM, Michael Still wrote:
 On Thu, Aug 14, 2014 at 3:44 AM, Russell Bryant rbry...@redhat.com wrote:
 On 08/13/2014 01:09 PM, Dan Smith wrote:
 Expecting cores to be at these sorts of things seems pretty reasonable
 to me, given the usefulness (and gravity) of the discussions we've been
 having so far. Companies with more cores will have to send more or make
 some hard decisions, but I don't want to cut back on the meetings until
 their value becomes unjustified.

 I disagree.  IMO, *expecting* people to travel, potentially across the
 globe, 4 times a year is an unreasonable expectation, and quite
 uncharacteristic of open source projects.  If we can't figure out a way
 to have the most important conversations in a way that is inclusive of
 everyone, we're failing with our processes.
 
 I am a bit confused by this stance to be honest. You yourself said
 when you were Icehouse PTL that you wanted cores to come to the
 summit. What changed?

Yes, I would love for core team members to come to the design summit
that's twice a year.  I still don't *expect* it for them to remain a
member of the team, and I certainly don't expect it 4 times a year.
It's a matter of frequency and requirement.

-- 
Russell Bryant

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [nova][core] Expectations of core reviewers

2014-08-14 Thread Russell Bryant
On 08/13/2014 11:31 PM, Michael Still wrote:
 On Thu, Aug 14, 2014 at 1:24 PM, Jay Pipes jaypi...@gmail.com wrote:
 
 Just wanted to quickly weigh in with my thoughts on this important topic. I
 very much valued the face-to-face interaction that came from the mid-cycle
 meetup in Beaverton (it was the only one I've ever been to).

 That said, I do not believe it should be a requirement that cores make it to
 the face-to-face meetings in-person. A number of folks have brought up very
 valid concerns about personal/family time, travel costs and burnout.
 
 I'm not proposing they be a requirement. I am proposing that they be
 strongly encouraged.

I'm not sure that's much different in reality.

 I believe that the issue raised about furthering the divide between core and
 non-core folks is actually the biggest reason I don't support a mandate to
 have cores at the face-to-face meetings, and I think we should make our best
 efforts to support quality virtual meetings that can be done on a more
 frequent basis than the face-to-face meetings that would be optional.
 
 I am all for online meetings, but we don't have a practical way to do
 them at the moment apart from IRC. Until someone has a concrete
 proposal that's been shown to work, I feel its a straw man argument.

Yes, IRC is one option which we already use on a regular basis.  We can
also switch to voice communication for higher bandwidth when needed.  We
even have a conferencing server set up in OpenStack's infrastructure:

https://wiki.openstack.org/wiki/Infrastructure/Conferencing

In theory it even supports basic video conferencing, though I haven't
tested it on this server yet.

-- 
Russell Bryant

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [nova][core] Expectations of core reviewers

2014-08-14 Thread Russell Bryant
On 08/13/2014 07:27 PM, Michael Still wrote:
 The etherpad for the meetup has extensive notes. Any summary I write
 will basically be those notes in prose. What are you looking for in a
 summary that isn't in the etherpad? There also wasn't a summary of the
 Icehouse midcycle produced that I can find. Whilst I am happy to do
 one for Juno, its a requirement that I hadn't planned for, and is
 therefore taking me some time to retrofit.
 
 I think we should chalk the request for summaries up experience and
 talk through how to better provide such things at future meetups.

The summary from the Icehouse meetup is here:

http://lists.openstack.org/pipermail/openstack-dev/2014-February/027370.html

-- 
Russell Bryant

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


[openstack-dev] [OpenStack][Docker] Run OpenStack Service in Docker Container

2014-08-14 Thread Jay Lau
I see a few mentions of OpenStack services themselves being containerized
in Docker. Is this a serious trend in the community?

http://allthingsopen.com/2014/02/12/why-containers-for-openstack-services/

-- 
Thanks,

Jay
___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] Annoucing CloudKitty : an OpenSource Rating-as-a-Service project for OpenStack

2014-08-14 Thread Sandy Walsh
Sounds very interesting. We're currently collecting detailed (and verified) 
usage information in StackTach and are keen to see what CloudKitty is able to 
offer. My one wish is that you keep the components as small pip 
redistributables with low coupling to promote reuse with other projects. Many 
tiny repos and clear API's (internal and external) are good for adoption and 
contribution. 

All the best!
-Sandy


From: Christophe Sauthier [christophe.sauth...@objectif-libre.com]
Sent: Wednesday, August 13, 2014 10:40 AM
To: openstack-dev@lists.openstack.org
Subject: [openstack-dev] Annoucing CloudKitty : an OpenSource 
Rating-as-a-Service project for OpenStack

We are very pleased at Objectif Libre to intoduce CloudKitty, an effort
to provide a fully OpenSource Rating-as-a-Service component in
OpenStack..

Following a first POC presented during the last summit in Atlanta to
some Ceilometer devs (thanks again Julien Danjou for your great support
!), we continued our effort to create a real service for rating. Today
we are happy to share it with you all.


So what do we propose in CloudKitty?
  - a service for collecting metrics (using Ceilometer API)
  - a modular rating architecture to enable/disable modules and create
your own rules on-the-fly, allowing you to use the rating patterns you
like
  - an API to interact with the whole environment from core components
to every rating module
  - a Horizon integration to allow configuration of the rating modules
and display of pricing information in real time during instance
creation
  - a CLI client to access this information and easily configure
everything

Technically we are using all the elements that are used in the various
OpenStack projects like olso, stevedore, pecan...
CloudKitty is highly modular and allows integration / developement of
third party collection and rating modules and output formats.

A roadmap is available on the project wiki page (the link is at the end
of this email), but we are clearly hoping to have some feedback and
ideas on how to improve the project and reach a tighter integration with
OpenStack.

The project source code is available at
http://github.com/stackforge/cloudkitty
More stuff will be available on stackforge as soon as the reviews get
validated like python-cloudkittyclient and cloudkitty-dashboard, so stay
tuned.

The project's wiki page (https://wiki.openstack.org/wiki/CloudKitty)
provides more information, and you can reach us via irc on freenode:
#cloudkitty. Developper's documentation is on its way to readthedocs
too.

We plan to present CloudKitty in detail during the Paris Summit, but we
would love to hear from you sooner...

Cheers,

  Christophe and Objectif Libre


Christophe Sauthier   Mail :
christophe.sauth...@objectif-libre.com
CEO  Fondateur   Mob : +33 (0) 6 16 98 63
96
Objectif LibreURL :
www.objectif-libre.com
Infrastructure et Formations LinuxTwitter : @objectiflibre

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [neutron][ml2] Mech driver as out-of-tree add-on

2014-08-14 Thread Kyle Mestery
I also feel like the drivers/plugins are currently BEYOND a tipping
point, and are in fact dragging down velocity of the core project in
many ways. I'm working on a proposal for Kilo where we move all
drivers/plugins out of the main Neutron tree and into a separate git
repository under the networking program. We have way too many drivers,
requiring way too man review cycles, for this to be a sustainable
model going forward. Since the main reason plugin/driver authors want
their code upstream is to be a part of the simultaneous release, and
thus be packaged by distributions, having a separate repository for
these will satisfy this requirement. I'm still working through the
details around reviews of this repository, etc.

Also, I feel as if the level of passion on the mailing list has died
down a bit, so I thought I'd send something out to try and liven
things up a bit. It's been somewhat non-emotional here for a day or
so. :)

Thanks,
Kyle

On Thu, Aug 14, 2014 at 5:09 AM, Salvatore Orlando sorla...@nicira.com wrote:
 I think there will soon be a discussion regarding what the appropriate
 location for plugin and drivers should be.
 My personal feeling is that Neutron has simply reached the tipping point
 where the high number of drivers and plugins is causing unnecessary load for
 the core team and frustration for the community.

 There I would totally support Luke's initiative about maintaining an
 out-of-tree ML2 driver. On the other hand, a plugin/driver diaspora might
 also have negative consequences such as frequent breakages such as those Bob
 was mentioning or confusion for users which might need to end up fetching
 drivers from disparate sources.

 As mentioned during the last Neutron IRC meeting this is another process
 aspect which will be discussed soon, with the aim of defining a plan for:
 - drastically reduce the number of plugins and drivers which must be
 maintained in the main source tree
 - enhance control of plugin/driver maintainers over their own code
 - preserve the ability of doing CI checks on gerrit as we do today
 - raise the CI bar (maybe finally set the smoketest as a minimum
 requirement?)

 Regards,
 Salvatore



 On 14 August 2014 11:47, loy wolfe loywo...@gmail.com wrote:




 On Thu, Aug 14, 2014 at 4:22 PM, Mathieu Rohon mathieu.ro...@gmail.com
 wrote:

 Hi,

 I would like to add that it would be harder for the community to help
 maintaining drivers.
 such a work [1] wouldn't have occured with an out of tree ODL driver.


 +1.
 It's better to move all MD for none built-in backend out of tree,
 maintaining these drivers shouldn't be the responsibility of community. Not
 only MD, but also plugin, agent should all obey this rule



 [1] https://review.openstack.org/#/c/96459/

 On Wed, Aug 13, 2014 at 1:09 PM, Robert Kukura kuk...@noironetworks.com
 wrote:
  One thing to keep in mind is that the ML2 driver API does sometimes
  change,
  requiring updates to drivers. Drivers that are in-tree get updated
  along
  with the driver API change. Drivers that are out-of-tree must be
  updated by
  the owner.
 
  -Bob
 
 
  On 8/13/14, 6:59 AM, ZZelle wrote:
 
  Hi,
 
 
  The important thing to understand is how to integrate with neutron
  through
  stevedore/entrypoints:
 
 
  https://github.com/dave-tucker/odl-neutron-drivers/blob/master/setup.cfg#L32-L34
 
 
  Cedric
 
 
  On Wed, Aug 13, 2014 at 12:17 PM, Dave Tucker d...@dtucker.co.uk
  wrote:
 
  I've been working on this for OpenDaylight
  https://github.com/dave-tucker/odl-neutron-drivers
 
  This seems to work for me (tested Devstack w/ML2) but YMMV.
 
  -- Dave
 
  ___
  OpenStack-dev mailing list
  OpenStack-dev@lists.openstack.org
  http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
 
 
 
 
  ___
  OpenStack-dev mailing list
  OpenStack-dev@lists.openstack.org
  http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
 
 
 
  ___
  OpenStack-dev mailing list
  OpenStack-dev@lists.openstack.org
  http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
 

 ___
 OpenStack-dev mailing list
 OpenStack-dev@lists.openstack.org
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev



 ___
 OpenStack-dev mailing list
 OpenStack-dev@lists.openstack.org
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev



 ___
 OpenStack-dev mailing list
 OpenStack-dev@lists.openstack.org
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [nova][core] Expectations of core reviewers

2014-08-14 Thread Daniel P. Berrange
On Thu, Aug 14, 2014 at 08:31:48AM -0400, Russell Bryant wrote:
 On 08/13/2014 11:31 PM, Michael Still wrote:
  On Thu, Aug 14, 2014 at 1:24 PM, Jay Pipes jaypi...@gmail.com wrote:
  
  Just wanted to quickly weigh in with my thoughts on this important topic. I
  very much valued the face-to-face interaction that came from the mid-cycle
  meetup in Beaverton (it was the only one I've ever been to).
 
  That said, I do not believe it should be a requirement that cores make it 
  to
  the face-to-face meetings in-person. A number of folks have brought up very
  valid concerns about personal/family time, travel costs and burnout.
  
  I'm not proposing they be a requirement. I am proposing that they be
  strongly encouraged.
 
 I'm not sure that's much different in reality.
 
  I believe that the issue raised about furthering the divide between core 
  and
  non-core folks is actually the biggest reason I don't support a mandate to
  have cores at the face-to-face meetings, and I think we should make our 
  best
  efforts to support quality virtual meetings that can be done on a more
  frequent basis than the face-to-face meetings that would be optional.
  
  I am all for online meetings, but we don't have a practical way to do
  them at the moment apart from IRC. Until someone has a concrete
  proposal that's been shown to work, I feel its a straw man argument.
 
 Yes, IRC is one option which we already use on a regular basis.  We can
 also switch to voice communication for higher bandwidth when needed.  We
 even have a conferencing server set up in OpenStack's infrastructure:
 
 https://wiki.openstack.org/wiki/Infrastructure/Conferencing
 
 In theory it even supports basic video conferencing, though I haven't
 tested it on this server yet.

Depending on the usage needs, I think Google hangouts is a quite useful
technology. For many-to-many session its limit of 10 participants can be
an issue, but for a few-to-many broadcast it could be practical. What I
find particularly appealing is the way it can live stream the session
over youtube which allows for unlimited number of viewers, as well as
being available offline for later catchup.

It could be useful in cases where one (or a handful) of people want to
present an idea / topic visually with slides / screencasts / etc and
then let the broader interactive discussion take place on IRC / mailing
list afterwards. I could see this being something that might let people
present proposals for new Nova features without having to wait (+try for
a limited) design summit slot in the 6 month cycle. One of the issues
I feel with the design summit is that it was the first time hearing about
many of the ideas, so there was not enough time to disgest the proposal
and so some of the thorny things to discuss only come to mind afterwards.
If we had to a approach for promoting features  knowledge in this way the
design summit could have more time to focus on areas of debate where the
f2f prescence is most valuable.

Regards,
Daniel
-- 
|: http://berrange.com  -o-http://www.flickr.com/photos/dberrange/ :|
|: http://libvirt.org  -o- http://virt-manager.org :|
|: http://autobuild.org   -o- http://search.cpan.org/~danberr/ :|
|: http://entangle-photo.org   -o-   http://live.gnome.org/gtk-vnc :|

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [all] The future of the integrated release

2014-08-14 Thread Devananda van der Veen
On Aug 14, 2014 2:04 AM, Eoghan Glynn egl...@redhat.com wrote:


   Letting the industry field-test a project and feed their experience
   back into the community is a slow process, but that is the best
   measure of a project's success. I seem to recall this being an
   implicit expectation a few years ago, but haven't seen it discussed
in
   a while.
  
   I think I recall us discussing a must have feedback that it's
   successfully deployed requirement in the last cycle, but we
recognized
   that deployers often wait until a project is integrated.
 
  In the early discussions about incubation, we respected the need to
  officially recognize a project as part of OpenStack just to create the
  uptick in adoption necessary to mature projects. Similarly, integration
is a
  recognition of the maturity of a project, but I think we have graduated
  several projects long before they actually reached that level of
maturity.
  Actually running a project at scale for a period of time is the only
way to
  know it is mature enough to run it in production at scale.
 
  I'm just going to toss this out there. What if we set the graduation
bar to
  is in production in at least two sizeable clouds (note that I'm not
saying
  public clouds). Trove is the only project that has, to my knowledge,
met
  that bar prior to graduation, and it's the only project that graduated
since
  Havana that I can, off hand, point at as clearly successful. Heat and
  Ceilometer both graduated prior to being in production; a few cycles
later,
  they're still having adoption problems and looking at large
architectural
  changes. I think the added cost to OpenStack when we integrate immature
or
  unstable projects is significant enough at this point to justify a more
  defensive posture.
 
  FWIW, Ironic currently doesn't meet that bar either - it's in
production in
  only one public cloud. I'm not aware of large private installations yet,
  though I suspect there are some large private deployments being spun up
  right now, planning to hit production with the Juno release.

 We have some hard data from the user survey presented at the Juno summit,
 with respectively 26  53 production deployments of Heat and Ceilometer
 reported.

 There's no cross-referencing of deployment size with services in
production
 in those data presented, though it may be possible to mine that out of the
 raw survey responses.

Indeed, and while that would be useful information, I was referring to the
deployment of those services at scale prior to graduation, not post
graduation.

Best,
Devananda
___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


[openstack-dev] [Neutron][DevStack] How to increase developer usage of Neutron

2014-08-14 Thread Mike Spreitzer
I'll bet I am not the only developer who is not highly competent with 
bridges and tunnels, Open VSwitch, Neutron configuration, and how DevStack 
transmutes all those.  My bet is that you would have more developers using 
Neutron if there were an easy-to-find and easy-to-follow recipe to use, to 
create a developer install of OpenStack with Neutron.  One that's a pretty 
basic and easy case.  Let's say a developer gets a recent image of Ubuntu 
14.04 from Canonical, and creates an instance in some undercloud, and that 
instance has just one NIC, at 10.9.8.7/16.  If there were a recipe for 
such a developer to follow from that point on, it would be great.

Regards,
Mike___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [Devstack] [Cinder] Registering 3rd party driver as default

2014-08-14 Thread Kerr, Andrew
You either need to comment out the enabled_backends line, or you'll want
to put something similar to this in your cinder.conf file:

[DEFAULT]
...
enabled_backends = cloudbyte
...
[cloudbyte]
volume_driver = volume_driver =
cinder.volume.drivers.cloudbyte.cloudbyte.ElasticenterISCSIDriver
SAN_IP=20.10.22.245
CB_APIKEY=masQwghrmPOVIqbjyyWKQdg4z4bP2sNZ13fRQyUMwm453PUiYB-xyRSMBDoZeMj6R
0-XU9DCscxMbe3AhleDyQ
CB_ACCOUNT_NAME=acc1
TSM_NAME=openstacktsm



If you have enabled_backends set then it will only use those driver specs
and ignore all driver related details in the DEFAULT section.

You also probably want to comment (or remove) the default_volume_type
line, unless you plan to create that volume type after the services come
up.

Andrew Kerr
OpenStack QA
Cloud Solutions Group
NetApp


From:  Amit Das amit@cloudbyte.com
Reply-To:  OpenStack Development Mailing List (not for usage questions)
openstack-dev@lists.openstack.org
Date:  Thursday, August 14, 2014 at 5:37 AM
To:  OpenStack Development Mailing List openstack-dev@lists.openstack.org
Subject:  Re: [openstack-dev] [Devstack] [Cinder] Registering 3rd
party   driver as default


With further debugging, i find that none of the configuration options
present in /etc/cinder/cinder.conf are getting applied.


Regards,
AmitCloudByte Inc. http://www.cloudbyte.com/




On Thu, Aug 14, 2014 at 11:40 AM, Amit Das
amit@cloudbyte.com wrote:

Hi folks,

I have been trying to run devstack with my cinder driver as the default
volume_driver but with no luck.

Devstack seems to register the lvm driver as the default always.

I have tried below approaches:

1. directly modifying the /etc/cinder/cinder.conf file
2. creating a driver file @ ./devstack/lib/cinder_plugins/driver-name

1. ref - https://review.openstack.org/#/c/68726/






This is my localrc details:
http://paste.openstack.org/show/94822/


I run ./unstack.sh  then FORCE=yes ./stack.sh

This is the cinder.conf that is generated after running above stack.sh. I
comment out the [lvmdriver-1] section manually
(not sure if this section needs to be commented)

http://paste.openstack.org/show/94841/


These are portions of c-sch  c-vol logs after restarting them in their
respective screens.

http://paste.openstack.org/show/94842/



Regards,
AmitCloudByte Inc. http://www.cloudbyte.com/








___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [Neutron][DevStack] How to increase developer usage of Neutron

2014-08-14 Thread CARVER, PAUL
Mike Spreitzer [mailto:mspre...@us.ibm.com] wrote:

I'll bet I am not the only developer who is not highly competent with
bridges and tunnels, Open VSwitch, Neutron configuration, and how DevStack
transmutes all those.  My bet is that you would have more developers using
Neutron if there were an easy-to-find and easy-to-follow recipe to use, to
create a developer install of OpenStack with Neutron.  One that's a pretty
basic and easy case.  Let's say a developer gets a recent image of Ubuntu
14.04 from Canonical, and creates an instance in some undercloud, and that
instance has just one NIC, at 10.9.8.7/16.  If there were a recipe for
such a developer to follow from that point on, it would be great.

https://wiki.openstack.org/wiki/NeutronDevstack worked for me.

However, I'm pretty sure it's only a single node all in one setup. At least,
I created only one VM to run it on and I don't think DevStack has created
multiple nested VMs inside of the one I create to run DevStack. I haven't
gotten around to figuring out how to setup a full multi-node DevStack
setup with separate compute nodes and network nodes and GRE/VXLAN tunnels.

There are multi-node instructions on that wiki page but I haven't tried
following them. If someone has a Vagrant file that creates a full multi-
node Neutron devstack complete with GRE/VXLAN tunnels it would be great
if they could add it to that wiki page.


___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


[openstack-dev] [neutron] requirements.txt: explicit vs. implicit

2014-08-14 Thread Ihar Hrachyshka
-BEGIN PGP SIGNED MESSAGE-
Hash: SHA512

Hi all,

some plugins depend on modules that are not mentioned in
requirements.txt. Among them, Cisco Nexus (ncclient), Brocade
(ncclient), Embrane (heleosapi)... Some other plugins put their
dependencies in requirements.txt though (like Arista depending on
jsonrpclib).

There are pros and cons in both cases. The obvious issue with not
putting those requirements in the file is that packagers are left
uninformed about those implicit requirements existing, meaning plugins
are shipped to users with broken dependencies. It also means we ship
code that depends on unknown modules grabbed from random places in the
internet instead of relying on what's available on pypi, which is a
bit scary.

With my packager hat on, I would like to suggest to make those
dependencies explicit by filling in requirements.txt. This will make
packaging a bit easier. Of course, runtime dependencies being set
correctly do not mean plugins are working and tested, but at least we
give them chance to be tested and used.

But, maybe there are valid concerns against doing so. In that case, I
would be glad to know how packagers are expected to track those
implicit dependencies.

I would like to ask community to decide what's the right way to handle
those cases.

Cheers,
/Ihar
-BEGIN PGP SIGNATURE-
Version: GnuPG/MacGPG2 v2.0.22 (Darwin)

iQEcBAEBCgAGBQJT7Lu2AAoJEC5aWaUY1u57tDkIAOrx1TWVjke8xxsJtz+tizmg
rDgoQyugU8bmaWUFzKi3yVLDFmkOH5iX9RFqj6pgXngydd+cO0Z8CB825uT7kimi
tTwTk2o1Ty4lIG38nwi/U8pn+nmzVApjOqtJmBmtZKBtoY7hRUs+QVTz5V5M1AmA
MQm0eYZXMQ531k4UTdaFxtZ2xPvnCEsFTWi0vosZLPvccVw33vUnQ0SnewQAgb4w
NZ7m302454S2INegqVYlZqQMQXxy6v/BAigyoLXBj8Pl3FsrNU0j3SMtzqSm71ty
GCz0qdWckUdgsDFnLyyNXjUV/G9xZ03pYZ5ID2WiVQl5MYbmkAHlJJkjCYIrv3c=
=tLTZ
-END PGP SIGNATURE-

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [heat] heat docker multi host scheduling support

2014-08-14 Thread Zane Bitter

On 14/08/14 03:21, Malawade, Abhijeet wrote:

Hi all,

I am trying to use heat to create docker containers. I have configured 
heat-docker plugin.
I am also able to create stack using heat successfully.

To start container on different host we need to provide 'docker_endpoint' in 
template. For this we have to provide host address where container will run in 
template.

Is there any way to schedule docker container on available hosts using 
heat-docker plugin without giving 'docker_endpoint' in template file.


Yes, there is a Nova driver on Stackforge that uses Docker as a 
back-end. Hopefully one day this will turn into a Nova-like container 
service, but for now the way to use Docker as a scheduler is through 
Nova with this driver.



Is heat-docker plugin supports managing docker hosts cluster with scheduling 
logic.


No.


Please let me know your suggestions on the same.

Thanks,
Abhijeet



___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [Devstack] [Cinder] Registering 3rd party driver as default

2014-08-14 Thread Amit Das
Thanks a lot.

This worked out like a charm.

Regards,
Amit
*CloudByte Inc.* http://www.cloudbyte.com/


On Thu, Aug 14, 2014 at 7:02 PM, Kerr, Andrew andrew.k...@netapp.com
wrote:

 You either need to comment out the enabled_backends line, or you'll want
 to put something similar to this in your cinder.conf file:

 [DEFAULT]
 ...
 enabled_backends = cloudbyte
 ...
 [cloudbyte]
 volume_driver = volume_driver =
 cinder.volume.drivers.cloudbyte.cloudbyte.ElasticenterISCSIDriver
 SAN_IP=20.10.22.245
 CB_APIKEY=masQwghrmPOVIqbjyyWKQdg4z4bP2sNZ13fRQyUMwm453PUiYB-xyRSMBDoZeMj6R
 0-XU9DCscxMbe3AhleDyQ
 CB_ACCOUNT_NAME=acc1
 TSM_NAME=openstacktsm



 If you have enabled_backends set then it will only use those driver specs
 and ignore all driver related details in the DEFAULT section.

 You also probably want to comment (or remove) the default_volume_type
 line, unless you plan to create that volume type after the services come
 up.

 Andrew Kerr
 OpenStack QA
 Cloud Solutions Group
 NetApp


 From:  Amit Das amit@cloudbyte.com
 Reply-To:  OpenStack Development Mailing List (not for usage questions)
 openstack-dev@lists.openstack.org
 Date:  Thursday, August 14, 2014 at 5:37 AM
 To:  OpenStack Development Mailing List openstack-dev@lists.openstack.org
 
 Subject:  Re: [openstack-dev] [Devstack] [Cinder] Registering 3rd
 party   driver as default


 With further debugging, i find that none of the configuration options
 present in /etc/cinder/cinder.conf are getting applied.


 Regards,
 AmitCloudByte Inc. http://www.cloudbyte.com/




 On Thu, Aug 14, 2014 at 11:40 AM, Amit Das
 amit@cloudbyte.com wrote:

 Hi folks,

 I have been trying to run devstack with my cinder driver as the default
 volume_driver but with no luck.

 Devstack seems to register the lvm driver as the default always.

 I have tried below approaches:

 1. directly modifying the /etc/cinder/cinder.conf file
 2. creating a driver file @ ./devstack/lib/cinder_plugins/driver-name

 1. ref - https://review.openstack.org/#/c/68726/






 This is my localrc details:
 http://paste.openstack.org/show/94822/


 I run ./unstack.sh  then FORCE=yes ./stack.sh

 This is the cinder.conf that is generated after running above stack.sh. I
 comment out the [lvmdriver-1] section manually
 (not sure if this section needs to be commented)

 http://paste.openstack.org/show/94841/


 These are portions of c-sch  c-vol logs after restarting them in their
 respective screens.

 http://paste.openstack.org/show/94842/



 Regards,
 AmitCloudByte Inc. http://www.cloudbyte.com/








 ___
 OpenStack-dev mailing list
 OpenStack-dev@lists.openstack.org
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [nova][core] Expectations of core reviewers

2014-08-14 Thread CARVER, PAUL
Daniel P. Berrange [mailto:berra...@redhat.com] wrote:

Depending on the usage needs, I think Google hangouts is a quite useful
technology. For many-to-many session its limit of 10 participants can be
an issue, but for a few-to-many broadcast it could be practical. What I
find particularly appealing is the way it can live stream the session
over youtube which allows for unlimited number of viewers, as well as
being available offline for later catchup.

I can't actually offer ATT resources without getting some level of
management approval first, but just for the sake of discussion here's
some info about the telepresence system we use.

-=-=-=-=-=-=-=-=-=-
ATS B2B Telepresence conferences can be conducted with an external company's
Telepresence room(s), which subscribe to the ATT Telepresence Solution,
or a limited number of other Telepresence service provider's networks.

Currently, the number of Telepresence rooms that can participate in a B2B
conference is limited to a combined total of 20 rooms (19 of which can be
ATT rooms, depending on the number of remote endpoints included).
-=-=-=-=-=-=-=-=-=-

We currently have B2B interconnect with over 100 companies and ATT has
telepresence rooms in many of our locations around the US and around
the world. If other large OpenStack companies also have telepresence
rooms that we could interconnect with I think it might be possible
to get management agreement to hold a couple OpenStack meetups per
year.

Most of our rooms are best suited for 6 people, but I know of at least
one 18 person telepresence room near me.

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [OpenStack-Dev][Cinder] Cinder Core nomination

2014-08-14 Thread Duncan Thomas
+1

On 14 August 2014 00:55, Boring, Walter walter.bor...@hp.com wrote:
 Hey guys,
I wanted to pose a nomination for Cinder core.

 Xing Yang.
 She has been active in the cinder community for many releases and has worked 
 on several drivers as well as other features for cinder itself.   She has 
 been doing an awesome job doing reviews and helping folks out in the 
 #openstack-cinder irc channel for a long time.   I think she would be a good 
 addition to the core team.


 Walt
 ___
 OpenStack-dev mailing list
 OpenStack-dev@lists.openstack.org
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev



-- 
Duncan Thomas

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [nova][core] Expectations of core reviewers

2014-08-14 Thread Russell Bryant
On 08/14/2014 10:04 AM, CARVER, PAUL wrote:
 Daniel P. Berrange [mailto:berra...@redhat.com] wrote:
 
 Depending on the usage needs, I think Google hangouts is a quite useful
 technology. For many-to-many session its limit of 10 participants can be
 an issue, but for a few-to-many broadcast it could be practical. What I
 find particularly appealing is the way it can live stream the session
 over youtube which allows for unlimited number of viewers, as well as
 being available offline for later catchup.
 
 I can't actually offer ATT resources without getting some level of
 management approval first, but just for the sake of discussion here's
 some info about the telepresence system we use.
 
 -=-=-=-=-=-=-=-=-=-
 ATS B2B Telepresence conferences can be conducted with an external company's
 Telepresence room(s), which subscribe to the ATT Telepresence Solution,
 or a limited number of other Telepresence service provider's networks.
 
 Currently, the number of Telepresence rooms that can participate in a B2B
 conference is limited to a combined total of 20 rooms (19 of which can be
 ATT rooms, depending on the number of remote endpoints included).
 -=-=-=-=-=-=-=-=-=-
 
 We currently have B2B interconnect with over 100 companies and ATT has
 telepresence rooms in many of our locations around the US and around
 the world. If other large OpenStack companies also have telepresence
 rooms that we could interconnect with I think it might be possible
 to get management agreement to hold a couple OpenStack meetups per
 year.
 
 Most of our rooms are best suited for 6 people, but I know of at least
 one 18 person telepresence room near me.

An ideal solution would allow attendees to join as individuals from
anywhere.  A lot of contributors work from home.  Is that sort of thing
compatible with your system?

-- 
Russell Bryant

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [nova][core] Expectations of core reviewers

2014-08-14 Thread Matt Riedemann



On 8/14/2014 3:47 AM, Daniel P. Berrange wrote:

On Thu, Aug 14, 2014 at 09:24:36AM +1000, Michael Still wrote:

On Thu, Aug 14, 2014 at 3:09 AM, Dan Smith d...@danplanet.com wrote:

I'm not questioning the value of f2f - I'm questioning the idea of
doing f2f meetings sooo many times a year. OpenStack is very much
the outlier here among open source projects - the vast majority of
projects get along very well with much less f2f time and a far
smaller % of their contributors attend those f2f meetings that do
happen. So I really do question what is missing from OpenStack's
community interaction that makes us believe that having 4 f2f
meetings a year is critical to our success.


How many is too many? So far, I have found the midcycles to be extremely
productive -- productive in a way that we don't see at the summits, and
I think other attendees agree. Obviously if budgets start limiting them,
then we'll have to deal with it, but I don't want to stop meeting
preemptively.


I agree they're very productive. Let's pick on the nova v3 API case as
an example... We had failed as a community to reach a consensus using
our existing discussion mechanisms (hundreds of emails, at least three
specs, phone calls between the various parties, et cetera), yet at the
summit and then a midcycle meetup we managed to nail down an agreement
on a very contentious and complicated topic.


We thought we had agreement on v3 API after Atlanta f2f summit and
after Hong Kong f2f too. So I wouldn't neccessarily say that we
needed another f2f meeting to resolve that, but rather than this is
a very complex topic that takes a long time to resolve no matter
how we discuss it and the discussions had just happened to reach
a natural conclusion this time around. But lets see if this agreement
actually sticks this time


I can see the argument that travel cost is an issue, but I think its
also not a very strong argument. We have companies spending millions
of dollars on OpenStack -- surely spending a relatively small amount
on travel to keep the development team as efficient as possible isn't
a big deal? I wouldn't be at all surprised if the financial costs of
the v3 API debate (staff time mainly) were much higher than the travel
costs of those involved in the summit and midcycle discussions which
sorted it out.


I think the travel cost really is a big issue. Due to the number of
people who had to travel to the many mid-cycle meetups, a good number
of people I work with no longer have the ability to go to the Paris
design summit. This is going to make it harder for them to feel a
proper engaged part of our community. I can only see this situation
get worse over time if greater emphasis is placed on attending the
mid-cycle meetups.


Travelling to places to talk to people isn't a great solution, but it
is the most effective one we've found so far. We should continue to
experiment with other options, but until we find something that works
as well as meetups, I think we need to keep having them.


IMHO, the reasons to cut back would be:

- People leaving with a well, that was useless... feeling
- Not enough people able to travel to make it worthwhile

So far, neither of those have been outcomes of the midcycles we've had,
so I think we're doing okay.

The design summits are structured differently, where we see a lot more
diverse attendance because of the colocation with the user summit. It
doesn't lend itself well to long and in-depth discussions about specific
things, but it's very useful for what it gives us in the way of
exposure. We could try to have less of that at the summit and more
midcycle-ish time, but I think it's unlikely to achieve the same level
of usefulness in that environment.

Specifically, the lack of colocation with too many other projects has
been a benefit. This time, Mark and Maru where there from Neutron. Last
time, Mark from Neutron and the other Mark from Glance were there. If
they were having meetups in other rooms (like at summit) they wouldn't
have been there exposed to discussions that didn't seem like they'd have
a component for their participation, but did after all (re: nova and
glance and who should own flavors).


I agree. The ability to focus on the issues that were blocking nova
was very important. That's hard to do at a design summit when there is
so much happening at the same time.


Maybe we should change the way we structure the design summit to
improve that. If there are critical issues blocking nova, it feels
like it is better to be able to discuss and resolve as much as possible
at the start of the dev cycle rather than in the middle of the dev
cycle because I feel that means we are causing ourselves pain during
milestone 1/2.


Just speaking from experience, I attended the Icehouse meetup before my 
first summit (Juno in ATL) and the design summit sessions for Juno were 
a big disappointment after the meetup sessions, basically because of the 
time constraints. The meetups are nice since there 

Re: [openstack-dev] use of compute node as a storage

2014-08-14 Thread Jyoti Ranjan
You run cinder-volume on compute node which you want to convert to storage.
Configure LVM driver of compute node as storage backend for cinder-volume.


On Fri, Aug 8, 2014 at 1:54 PM, shailendra acharya 
acharyashailend...@gmail.com wrote:

 i made 4 vm 1 controller, 1 network and 2 compute and i want 1 compute to
 run as a storage so plz help how can i do such ?

 ___
 OpenStack-dev mailing list
 OpenStack-dev@lists.openstack.org
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [nova][core] Expectations of core reviewers

2014-08-14 Thread Sandy Walsh
On 8/14/2014 11:28 AM, Russell Bryant wrote:
 On 08/14/2014 10:04 AM, CARVER, PAUL wrote:
 Daniel P. Berrange [mailto:berra...@redhat.com] wrote:

 Depending on the usage needs, I think Google hangouts is a quite useful
 technology. For many-to-many session its limit of 10 participants can be
 an issue, but for a few-to-many broadcast it could be practical. What I
 find particularly appealing is the way it can live stream the session
 over youtube which allows for unlimited number of viewers, as well as
 being available offline for later catchup.
 I can't actually offer ATT resources without getting some level of
 management approval first, but just for the sake of discussion here's
 some info about the telepresence system we use.

 -=-=-=-=-=-=-=-=-=-
 ATS B2B Telepresence conferences can be conducted with an external company's
 Telepresence room(s), which subscribe to the ATT Telepresence Solution,
 or a limited number of other Telepresence service provider's networks.

 Currently, the number of Telepresence rooms that can participate in a B2B
 conference is limited to a combined total of 20 rooms (19 of which can be
 ATT rooms, depending on the number of remote endpoints included).
 -=-=-=-=-=-=-=-=-=-

 We currently have B2B interconnect with over 100 companies and ATT has
 telepresence rooms in many of our locations around the US and around
 the world. If other large OpenStack companies also have telepresence
 rooms that we could interconnect with I think it might be possible
 to get management agreement to hold a couple OpenStack meetups per
 year.

 Most of our rooms are best suited for 6 people, but I know of at least
 one 18 person telepresence room near me.
 An ideal solution would allow attendees to join as individuals from
 anywhere.  A lot of contributors work from home.  Is that sort of thing
 compatible with your system?

http://bluejeans.com/ was a good experience.

What about Google Hangout OnAir for the PTL and core, while others are
view-only with chat/irc questions?

http://www.google.com/+/learnmore/hangouts/onair.html



___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [nova][core] Expectations of core reviewers

2014-08-14 Thread Anne Gentle
On Thu, Aug 14, 2014 at 10:13 AM, Sandy Walsh sandy.wa...@rackspace.com
wrote:

 On 8/14/2014 11:28 AM, Russell Bryant wrote:
  On 08/14/2014 10:04 AM, CARVER, PAUL wrote:
  Daniel P. Berrange [mailto:berra...@redhat.com] wrote:
 
  Depending on the usage needs, I think Google hangouts is a quite useful
  technology. For many-to-many session its limit of 10 participants can
 be
  an issue, but for a few-to-many broadcast it could be practical. What I
  find particularly appealing is the way it can live stream the session
  over youtube which allows for unlimited number of viewers, as well as
  being available offline for later catchup.
  I can't actually offer ATT resources without getting some level of
  management approval first, but just for the sake of discussion here's
  some info about the telepresence system we use.
 
  -=-=-=-=-=-=-=-=-=-
  ATS B2B Telepresence conferences can be conducted with an external
 company's
  Telepresence room(s), which subscribe to the ATT Telepresence Solution,
  or a limited number of other Telepresence service provider's networks.
 
  Currently, the number of Telepresence rooms that can participate in a
 B2B
  conference is limited to a combined total of 20 rooms (19 of which can
 be
  ATT rooms, depending on the number of remote endpoints included).
  -=-=-=-=-=-=-=-=-=-
 
  We currently have B2B interconnect with over 100 companies and ATT has
  telepresence rooms in many of our locations around the US and around
  the world. If other large OpenStack companies also have telepresence
  rooms that we could interconnect with I think it might be possible
  to get management agreement to hold a couple OpenStack meetups per
  year.
 
  Most of our rooms are best suited for 6 people, but I know of at least
  one 18 person telepresence room near me.
  An ideal solution would allow attendees to join as individuals from
  anywhere.  A lot of contributors work from home.  Is that sort of thing
  compatible with your system?
 
 http://bluejeans.com/ was a good experience.

 What about Google Hangout OnAir for the PTL and core, while others are
 view-only with chat/irc questions?

 http://www.google.com/+/learnmore/hangouts/onair.html


We've done Google Hangout OnAir for the docs team with varied success due
to the oddities with calendaring and having to know everyone's preferred
Google identity. It's nice for high-fidelity conversations without
requiring travel. It also lets people tune in then or watch a recording
later.
Anne






 ___
 OpenStack-dev mailing list
 OpenStack-dev@lists.openstack.org
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


[openstack-dev] [sahara] team meeting Aug 14 1800 UTC

2014-08-14 Thread Sergey Lukjanov
Hi folks,

We'll be having the Sahara team meeting as usual in
#openstack-meeting-alt channel.

Agenda: https://wiki.openstack.org/wiki/Meetings/SaharaAgenda#Next_meetings

http://www.timeanddate.com/worldclock/fixedtime.html?msg=Sahara+Meetingiso=20140814T18

-- 
Sincerely yours,
Sergey Lukjanov
Sahara Technical Lead
(OpenStack Data Processing)
Principal Software Engineer
Mirantis Inc.

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


[openstack-dev] [neutron] [third-party] What tests are required to be run

2014-08-14 Thread Kyle Mestery
Folks, I'm not sure if all CI accounts are running sufficient tests.
Per the requirements wiki page here [1], everyone needs to be running
more than just Tempest API tests, which I still see most neutron
third-party CI setups doing. I'd like to ask everyone who operates a
third-party CI account for Neutron to please look at the link below
and make sure you are running appropriate tests. If you have
questions, the weekly third-party meeting [2] is a great place to ask
questions.

Thanks,
Kyle

[1] https://wiki.openstack.org/wiki/NeutronThirdPartyTesting
[2] https://wiki.openstack.org/wiki/Meetings/ThirdParty

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [nova][core] Expectations of core reviewers

2014-08-14 Thread Kyle Mestery
On Thu, Aug 14, 2014 at 10:13 AM, Sandy Walsh sandy.wa...@rackspace.com wrote:
 On 8/14/2014 11:28 AM, Russell Bryant wrote:
 On 08/14/2014 10:04 AM, CARVER, PAUL wrote:
 Daniel P. Berrange [mailto:berra...@redhat.com] wrote:

 Depending on the usage needs, I think Google hangouts is a quite useful
 technology. For many-to-many session its limit of 10 participants can be
 an issue, but for a few-to-many broadcast it could be practical. What I
 find particularly appealing is the way it can live stream the session
 over youtube which allows for unlimited number of viewers, as well as
 being available offline for later catchup.
 I can't actually offer ATT resources without getting some level of
 management approval first, but just for the sake of discussion here's
 some info about the telepresence system we use.

 -=-=-=-=-=-=-=-=-=-
 ATS B2B Telepresence conferences can be conducted with an external company's
 Telepresence room(s), which subscribe to the ATT Telepresence Solution,
 or a limited number of other Telepresence service provider's networks.

 Currently, the number of Telepresence rooms that can participate in a B2B
 conference is limited to a combined total of 20 rooms (19 of which can be
 ATT rooms, depending on the number of remote endpoints included).
 -=-=-=-=-=-=-=-=-=-

 We currently have B2B interconnect with over 100 companies and ATT has
 telepresence rooms in many of our locations around the US and around
 the world. If other large OpenStack companies also have telepresence
 rooms that we could interconnect with I think it might be possible
 to get management agreement to hold a couple OpenStack meetups per
 year.

 Most of our rooms are best suited for 6 people, but I know of at least
 one 18 person telepresence room near me.
 An ideal solution would allow attendees to join as individuals from
 anywhere.  A lot of contributors work from home.  Is that sort of thing
 compatible with your system?

 http://bluejeans.com/ was a good experience.

 What about Google Hangout OnAir for the PTL and core, while others are
 view-only with chat/irc questions?

This is a terrible idea, as it perpetuates the core vs. non-core
argument. We need equal participation for cores and non-cores alike.

 http://www.google.com/+/learnmore/hangouts/onair.html



 ___
 OpenStack-dev mailing list
 OpenStack-dev@lists.openstack.org
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [OpenStack-Dev][Cinder] Cinder Core nomination

2014-08-14 Thread Martin, Kurt Frederick (ESSN Storage MSDU)
 +1, Xing  is very active in the community, provides valuable reviews and 
currently developing cinder core features.
~Kurt 

-Original Message-
From: Boring, Walter 
Sent: Wednesday, August 13, 2014 11:56 PM
To: openstack-dev@lists.openstack.org
Subject: [openstack-dev] [OpenStack-Dev][Cinder] Cinder Core nomination

Hey guys,
   I wanted to pose a nomination for Cinder core.

Xing Yang.
She has been active in the cinder community for many releases and has worked on 
several drivers as well as other features for cinder itself.   She has been 
doing an awesome job doing reviews and helping folks out in the 
#openstack-cinder irc channel for a long time.   I think she would be a good 
addition to the core team.


Walt
___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [nova][core] Expectations of core reviewers

2014-08-14 Thread CARVER, PAUL
Russell Bryant [mailto:rbry...@redhat.com] wrote:

An ideal solution would allow attendees to join as individuals from
anywhere.  A lot of contributors work from home.  Is that sort of thing
compatible with your system?

In principle, yes, but that loses the immersive telepresence aspect
which is the next best thing to an in-person meetup (which is where
this thread started.)

ATT Employees live and breathe on ATT Connect which is our
teleconferencing (not telepresence) service. It supports webcam
video as well as desktop sharing, but I'm on the verge of making
a sales pitch here which was NOT my intent.

I'm on ATT Connect meetings 5+ times a day but I'm biased so I
won't offer any opinion on how it compares to WebEx, GotoMeeting,
and other services. None of them are really equivalent to the
purpose built telepresence rooms.

My point was that there may well be a telepresence room within
reasonable driving distance for a large number of OpenStack
contributors if we were able to get a number of the large
OpenStack participant companies to open their doors to an
occasional meet-up. Instead of asking participants around
the globe to converge on a single physical location for
a meet-up, perhaps they could converge on the closest of
20 different locations that are linked via telepresence.

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [nova][core] Expectations of core reviewers

2014-08-14 Thread Kyle Mestery
On Thu, Aug 14, 2014 at 10:31 AM, CARVER, PAUL pc2...@att.com wrote:
 Russell Bryant [mailto:rbry...@redhat.com] wrote:

An ideal solution would allow attendees to join as individuals from
anywhere.  A lot of contributors work from home.  Is that sort of thing
compatible with your system?

 In principle, yes, but that loses the immersive telepresence aspect
 which is the next best thing to an in-person meetup (which is where
 this thread started.)

 ATT Employees live and breathe on ATT Connect which is our
 teleconferencing (not telepresence) service. It supports webcam
 video as well as desktop sharing, but I'm on the verge of making
 a sales pitch here which was NOT my intent.

 I'm on ATT Connect meetings 5+ times a day but I'm biased so I
 won't offer any opinion on how it compares to WebEx, GotoMeeting,
 and other services. None of them are really equivalent to the
 purpose built telepresence rooms.

 My point was that there may well be a telepresence room within
 reasonable driving distance for a large number of OpenStack
 contributors if we were able to get a number of the large
 OpenStack participant companies to open their doors to an
 occasional meet-up. Instead of asking participants around
 the globe to converge on a single physical location for
 a meet-up, perhaps they could converge on the closest of
 20 different locations that are linked via telepresence.

If we could make this work, I'd be up for it. It's a great balance to
the travel and face-to-face argument here.

 ___
 OpenStack-dev mailing list
 OpenStack-dev@lists.openstack.org
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [nova][core] Expectations of core reviewers

2014-08-14 Thread Russell Bryant
On 08/14/2014 11:40 AM, David Kranz wrote:
 On 08/14/2014 10:54 AM, Matt Riedemann wrote:


 On 8/14/2014 3:47 AM, Daniel P. Berrange wrote:
 On Thu, Aug 14, 2014 at 09:24:36AM +1000, Michael Still wrote:
 On Thu, Aug 14, 2014 at 3:09 AM, Dan Smith d...@danplanet.com wrote:
 I'm not questioning the value of f2f - I'm questioning the idea of
 doing f2f meetings sooo many times a year. OpenStack is very much
 the outlier here among open source projects - the vast majority of
 projects get along very well with much less f2f time and a far
 smaller % of their contributors attend those f2f meetings that do
 happen. So I really do question what is missing from OpenStack's
 community interaction that makes us believe that having 4 f2f
 meetings a year is critical to our success.

 How many is too many? So far, I have found the midcycles to be
 extremely
 productive -- productive in a way that we don't see at the summits,
 and
 I think other attendees agree. Obviously if budgets start limiting
 them,
 then we'll have to deal with it, but I don't want to stop meeting
 preemptively.

 I agree they're very productive. Let's pick on the nova v3 API case as
 an example... We had failed as a community to reach a consensus using
 our existing discussion mechanisms (hundreds of emails, at least three
 specs, phone calls between the various parties, et cetera), yet at the
 summit and then a midcycle meetup we managed to nail down an agreement
 on a very contentious and complicated topic.

 We thought we had agreement on v3 API after Atlanta f2f summit and
 after Hong Kong f2f too. So I wouldn't neccessarily say that we
 needed another f2f meeting to resolve that, but rather than this is
 a very complex topic that takes a long time to resolve no matter
 how we discuss it and the discussions had just happened to reach
 a natural conclusion this time around. But lets see if this agreement
 actually sticks this time

 I can see the argument that travel cost is an issue, but I think its
 also not a very strong argument. We have companies spending millions
 of dollars on OpenStack -- surely spending a relatively small amount
 on travel to keep the development team as efficient as possible isn't
 a big deal? I wouldn't be at all surprised if the financial costs of
 the v3 API debate (staff time mainly) were much higher than the travel
 costs of those involved in the summit and midcycle discussions which
 sorted it out.

 I think the travel cost really is a big issue. Due to the number of
 people who had to travel to the many mid-cycle meetups, a good number
 of people I work with no longer have the ability to go to the Paris
 design summit. This is going to make it harder for them to feel a
 proper engaged part of our community. I can only see this situation
 get worse over time if greater emphasis is placed on attending the
 mid-cycle meetups.

 Travelling to places to talk to people isn't a great solution, but it
 is the most effective one we've found so far. We should continue to
 experiment with other options, but until we find something that works
 as well as meetups, I think we need to keep having them.

 IMHO, the reasons to cut back would be:

 - People leaving with a well, that was useless... feeling
 - Not enough people able to travel to make it worthwhile

 So far, neither of those have been outcomes of the midcycles we've
 had,
 so I think we're doing okay.

 The design summits are structured differently, where we see a lot more
 diverse attendance because of the colocation with the user summit. It
 doesn't lend itself well to long and in-depth discussions about
 specific
 things, but it's very useful for what it gives us in the way of
 exposure. We could try to have less of that at the summit and more
 midcycle-ish time, but I think it's unlikely to achieve the same level
 of usefulness in that environment.

 Specifically, the lack of colocation with too many other projects has
 been a benefit. This time, Mark and Maru where there from Neutron.
 Last
 time, Mark from Neutron and the other Mark from Glance were there. If
 they were having meetups in other rooms (like at summit) they wouldn't
 have been there exposed to discussions that didn't seem like they'd
 have
 a component for their participation, but did after all (re: nova and
 glance and who should own flavors).

 I agree. The ability to focus on the issues that were blocking nova
 was very important. That's hard to do at a design summit when there is
 so much happening at the same time.

 Maybe we should change the way we structure the design summit to
 improve that. If there are critical issues blocking nova, it feels
 like it is better to be able to discuss and resolve as much as possible
 at the start of the dev cycle rather than in the middle of the dev
 cycle because I feel that means we are causing ourselves pain during
 milestone 1/2.

 Just speaking from experience, I attended the Icehouse meetup before
 my first summit (Juno in ATL) and the 

Re: [openstack-dev] [all] The future of the integrated release

2014-08-14 Thread Vishvananda Ishaya

On Aug 13, 2014, at 5:07 AM, Daniel P. Berrange berra...@redhat.com wrote:

 On Wed, Aug 13, 2014 at 12:55:48PM +0100, Steven Hardy wrote:
 On Wed, Aug 13, 2014 at 11:42:52AM +0100, Daniel P. Berrange wrote:
 On Thu, Aug 07, 2014 at 03:56:04AM -0700, Jay Pipes wrote:
 By ignoring stable branches, leaving it upto a
 small team to handle, I think we giving the wrong message about what
 our priorities as a team team are. I can't help thinking this filters
 through to impact the way people think about their work on master.
 
 Who is ignoring stable branches?  This sounds like a project specific
 failing to me, as all experienced core reviewers should consider offering
 their services to help with stable-maint activity.
 
 I don't personally see any reason why the *entire* project core team has to
 do this, but a subset of them should feel compelled to participate in the
 stable-maint process, if they have sufficient time, interest and historical
 context, it's not some other team IMO.
 
 I think that stable branch review should be a key responsibility for anyone
 on the core team, not solely those few who volunteer for stable team. As
 the number of projects in openstack grows I think the idea of having a
 single stable team with rights to approve across any project is ultimately
 flawed because it doesn't scale efficiently and they don't have the same
 level of domain knowledge as the respective project teams.

This side-thread is a bit off topic for the main discussion, but as a
stable-maint with not a lot of time, I would love more help from the core
teams here. That said, help is not just about aproving reviews. There are
three main steps in the process:
 1. Bugs get marked for backport
   I try to stay on top of this in nova by following the feed of merged patches
   and marking them icehouse-backport-potential[1] when they seem like they are
   appropriate but I’m sure I miss some.
 2. Patches get backprorted
   This is sometimes a very time-consuming process, especially late in the
   cycle or for patches that are being backported 2 releases.
 3. Patches get reviewed and merged
   The criteria for a stable backport are pretty straightforward and I think
   any core reviewer is capable of understanding and aply that criteria

While we have fallen behind in number 3. at times, we are much more often WAY
behind on 2. I also suspect that a whole bunch of patches get missed in some
of the other projects where someone isn’t specifically trying to mark them all
as they come in.

Vish

[1] https://bugs.launchpad.net/nova/+bugs?field.tag=icehouse-backport-potential



signature.asc
Description: Message signed with OpenPGP using GPGMail
___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [OpenStack-Dev][Cinder] Cinder Core nomination

2014-08-14 Thread Mike Perez
On 06:55 Thu 14 Aug , Boring, Walter wrote:
 Hey guys,
I wanted to pose a nomination for Cinder core.
 
 Xing Yang.
 She has been active in the cinder community for many releases and has worked 
 on several drivers as well as other features for cinder itself.   She has 
 been doing an awesome job doing reviews and helping folks out in the 
 #openstack-cinder irc channel for a long time.   I think she would be a good 
 addition to the core team.

+1

If you take a look at the last 3 months [1][2] she has been very active and
providing great feedback in reviews. She has also been setting great
expectations for other drivers with the recent third-party CI work. Thanks
Xing!

[1] - http://russellbryant.net/openstack-stats/cinder-reviewers-90.txt
[2] - http://stackalytics.com/?user_id=xing-yang

-- 
Mike Perez

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [neutron] requirements.txt: explicit vs. implicit

2014-08-14 Thread Ben Nemec
-BEGIN PGP SIGNED MESSAGE-
Hash: SHA1

On 08/14/2014 08:37 AM, Ihar Hrachyshka wrote:
 Hi all,
 
 some plugins depend on modules that are not mentioned in 
 requirements.txt. Among them, Cisco Nexus (ncclient), Brocade 
 (ncclient), Embrane (heleosapi)... Some other plugins put their 
 dependencies in requirements.txt though (like Arista depending on 
 jsonrpclib).
 
 There are pros and cons in both cases. The obvious issue with not 
 putting those requirements in the file is that packagers are left 
 uninformed about those implicit requirements existing, meaning
 plugins are shipped to users with broken dependencies. It also
 means we ship code that depends on unknown modules grabbed from
 random places in the internet instead of relying on what's
 available on pypi, which is a bit scary.
 
 With my packager hat on, I would like to suggest to make those 
 dependencies explicit by filling in requirements.txt. This will
 make packaging a bit easier. Of course, runtime dependencies being
 set correctly do not mean plugins are working and tested, but at
 least we give them chance to be tested and used.
 
 But, maybe there are valid concerns against doing so. In that case,
 I would be glad to know how packagers are expected to track those 
 implicit dependencies.
 
 I would like to ask community to decide what's the right way to
 handle those cases.

So I raised a similar issue about six months ago and completely failed
to follow up on the direction everyone seemed to be onboard with:
http://lists.openstack.org/pipermail/openstack-dev/2014-February/026976.html

I did add support to pbr for using nested requirements files, and I
had posted a PoC for oslo.messaging to allow requirements files for
different backends, but some of our CI jobs don't know how to handle
that and I never got around to addressing the limitation.

- From the packaging perspective, I think you could do a requirements
file that basically rolls up requirements.d/*.txt minus test.txt and
get all the runtime dependencies that the project knows about,
assuming we finished the implementation for this and started using it
in projects.

I don't really anticipate having time to pursue this in the near
future, so if you wanted to pick up the ball and run with it that
would be great! :-)

- -Ben

 
 Cheers, /Ihar
 
 ___ OpenStack-dev
 mailing list OpenStack-dev@lists.openstack.org 
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
 

-BEGIN PGP SIGNATURE-
Version: GnuPG v1
Comment: Using GnuPG with Thunderbird - http://www.enigmail.net/

iQEcBAEBAgAGBQJT7OS8AAoJEDehGd0Fy7uqtqsH/0kcIB+Q9iA5vR5evC+PDsFb
ek+cwvldbgpv0JwFwhgsLtsbRbI9xv1wDpu8L5i30yKzgcPQX5cqYe2WZeG5eCBJ
HUHb3t86rCanBU+kp7hpjHSoQbdwhY9gtu1LwiVha/2IeOHchZBzJccxEcACsv0q
Es8YkQy3qp9EfegumaL4OHvYFfB/j4NbewxZjAb3mkcpObb6NBM1v+qeubjTEg5I
nY8lJMLBXJOLNrR5cg8G7sObh3Cow51GtjwFaiFuZi/9o6whQFXipKnwXkaSRR5U
I3YV18sy3NLtoStZdr4/Oa9kUICw1MdDZAckoc5nP+AQeZCBFaPaCPpkLzIcMT8=
=CdPl
-END PGP SIGNATURE-

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [Ceilometer] Way to get wrapped method's name/class using Pecan secure decorators?

2014-08-14 Thread Pendergrass, Eric
Sure, Doug.  We want the ability to selectively apply policies to certain
Ceilometer
API methods based on user/tenant roles.

For example, we want to restrict the ability to execute Alarm deletes to
admins and user/tenants who have a special role, say domainadmin.

The policy file might look like this:
{
context_is_admin:  [[role:admin]],
admin_and_matching_project_domain_id:  [[role:domainadmin]],
admin_or_cloud_admin: [[rule:context_is_admin],
[rule:admin_and_matching_project_domain_id]],
telemetry:delete_alarms:  [[rule:admin_or_cloud_admin]]
}

The current acl.py and _query_to_kwargs access control setup either sets
project_id scope to None (do everything) or to the project_id in the request
header 'X-Project-Id'.  This allows for admin or project scope, but nothing
in
between.

We tried hooks.  Unfortunately we can't seem to turn the API controllers
into
HookControllers just by adding HookController to the Controller class
definition.  It causes infinite recursion on API startup.  For example, this
doesn't work because ceilometer-api will not start with it:
class MetersController(rest.RestController, HookController):

If there was a way to use hooks with the v2. API controllers that might work
really well.

So we are left using the @secure decorator and deriving the method name from
the request environ PATH_INFO and REQUEST_METHOD values.  This is how we
determine the wrapped method within the class (REQUEST_METHOD + PATH_INFO =
telemetry:delete_alarms with some munging).  We need the method name in
order to
selectively apply acces control to certain methods.

Deriving the method this way isn't ideal but it's the only thing we've
gotten working 
between hooks, @secure, and regular decorators.

I submitted a WIP BP here: https://review.openstack.org/#/c/112137/3.  It is
slightly out of date but should give you a beter idea of our goals.

Thanks

 Eric,

 If you can give us some more information about your end goal, independent
of the implementation, maybe we can propose an alternate technique to
achieve the same thing.

 Doug

 On Aug 12, 2014, at 6:21 PM, Ryan Petrello ryan.petre...@dreamhost.com
wrote:

  Yep, you're right, this doesn't seem to work.  The issue is that
  security is enforced at routing time (while the controller is still
  actually being discovered).  In order to do this sort of thing with
  the `check_permissions`, we'd probably need to add a feature to pecan.
 
  On 08/12/14 06:38 PM, Pendergrass, Eric wrote:
  Sure, here's the decorated method from v2.py:
 
 class MetersController(rest.RestController):
 Works on meters.
 
 @pecan.expose()
 def _lookup(self, meter_name, *remainder):
 return MeterController(meter_name), remainder
 
 @wsme_pecan.wsexpose([Meter], [Query])
 @secure(RBACController.check_permissions)
 def get_all(self, q=None):
 
  and here's the decorator called by the secure tag:
 
 class RBACController(object):
 global _ENFORCER
 if not _ENFORCER:
 _ENFORCER = policy.Enforcer()
 
 
 @classmethod
 def check_permissions(cls):
 # do some stuff
 
  In check_permissions I'd like to know the class and method with the
@secure tag that caused check_permissions to be invoked.  In this case, that
would be MetersController.get_all.
 
  Thanks
 
 
  Can you share some code?  What do you mean by, is there a way for the
decorator code to know it was called by MetersController.get_all
 
  On 08/12/14 04:46 PM, Pendergrass, Eric wrote:
  Thanks Ryan, but for some reason the controller attribute is None:
 
  (Pdb) from pecan.core import state
  (Pdb) state.__dict__
  {'hooks': [ceilometer.api.hooks.ConfigHook object at 0x31894d0,
  ceilometer.api.hooks.DBHook object at 0x3189650,
  ceilometer.api.hooks.PipelineHook object at 0x39871d0,
  ceilometer.api.hooks.TranslationHook object at 0x3aa5510], 'app':
  pecan.core.Pecan object at 0x2e76390, 'request': Request at
  0x3ed7390 GET http://localhost:8777/v2/meters, 'controller': None,
  'response': Response at 0x3ed74d0 200 OK}
 
  -Original Message-
  From: Ryan Petrello [mailto:ryan.petre...@dreamhost.com]
  Sent: Tuesday, August 12, 2014 10:34 AM
  To: OpenStack Development Mailing List (not for usage questions)
  Subject: Re: [openstack-dev] [Ceilometer] Way to get wrapped
method's name/class using Pecan secure decorators?
 
  This should give you what you need:
 
  from pecan.core import state
  state.controller
 
  On 08/12/14 04:08 PM, Pendergrass, Eric wrote:
  Hi, I'm trying to use the built in secure decorator in Pecan for
access control, and I'ld like to get the name of the method that is wrapped
from within the decorator.
 
  For instance, if I'm wrapping MetersController.get_all with an
@secure decorator, is there a way for the decorator code to know it was
called by MetersController.get_all?
 
  I don't see any global objects that provide this information.  I
can get the 

[openstack-dev] Using gerrymander, a client API and command line tool for gerrit

2014-08-14 Thread Daniel P. Berrange
Over the past couple of months I've worked on creating a python library
and command line tool for dealing with gerrit. I've blogged about this
a number of times on http://planet.openstack.org, but I figure there are
quite a few people who won't be reading the blogs there or easily miss
some posts hence this mail to alert people to its existance.

Gerrymander was born out of my frustration with the inefficiency of
handling reviews through the gerrit web UI. In particular (until
today - yay !) the review comments were splattered with noise from
CI bots, and it was hard to generate some kinds of reports for lists
of reviews you might want. The gerrit upgrade has improved the latter
to with better query language support, but it is still somewhat limited
in the way it can present the results and I prefer the CLI reports over
web UI anyway.

You might be familiar with tools like gerrit_view, reviewtodo and
reviewstats written by Josh Harlow and / or Russell Bryant in the
past. As a starting point, gerrymander aimed to offer all the
functionality present in those tools. More importantly though was
the idea to write it as a set of a python classes rather than just
put everything in a command line tool. This makes it easy for people
to customize  extend its functionality to pull our new interesting
types of data.

Personally, using gerrymander to query gerrit for reviews and being
able to quickly get a list of comments filtered for bots has had a
big positive impact on my efficiency as a core reviewer. I've been
able to review more code in less time and focus my attention better.
So if you are the kind of person who is more comfortable with having
command line tools instead of pointy-clicky web UIs, I'd encourage
you to try it out for yourself. Hopefully it will be of as much use
to other reviewers as it has been for me.

You can get it from pypi  (pip install gerrymander) and for a quick
start guide, check out the README file

   https://github.com/berrange/gerrymander/blob/master/README

Or my blog posts

   https://www.berrange.com/tags/gerrymander/

It is intended to work with any Gerrit instance, not just openstack,
so you'll need a config file that tailors it for openstack world,
like our team membership, bot accounts, etc. I've got a demo config
file here that should point you in the right direction at least:

   https://wiki.openstack.org/wiki/GerrymanderConfig

Feel free to update that wiki to add more project teams I've not
covered already, since there's many openstack projects I've not
familiar enough with.

NB, I've tested it to work on Linux systems. It would probably work
on OS-X since that resembles UNIX well enough, but Windows people
might be out of luck since it relies on forking stuff like ssh
and less with pipes connecting them up.


In terms of how I personally use it (assuming the above config file)...

At least once a day I look at all open changes in Nova which touch any
libvirt source files

  $ gerrymander -g nova --branch master libvirt

and will go through and review anything that I've not seen before.

In addition I'll process any open changes where I have commented on a
previous version of a patch, but not commented in the current version
which is a report generated by a special command:

  $ gerrymander -g nova todo-mine

If I have more time I'll look at any patches which have been submitted
where *no  one* has done any review, touching anything in nova source
tree

  $ gerrymander -g nova todo-noones

or any patches where I've not reviewed the current version

  $ gerrymander -g nova todo-others

That's pretty much all I need on a day-to-day basis to identify stuff
needing my review attention. For actual review I'll use the gerrit
web UI, or git fetch the patch locally depending on complexity of the
change in question. Also if I want to see comments fully expanded,
without bots, for a specific change number 104264 I would do

  $ gerrymander comments 104264

One day I'd love to be able to write some reports which pull priority
data on blueprints from nova-specs or launchpad, and correlate with 
reviews so important changes needing attention can be highlighted...

Regards,
Daniel
-- 
|: http://berrange.com  -o-http://www.flickr.com/photos/dberrange/ :|
|: http://libvirt.org  -o- http://virt-manager.org :|
|: http://autobuild.org   -o- http://search.cpan.org/~danberr/ :|
|: http://entangle-photo.org   -o-   http://live.gnome.org/gtk-vnc :|

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


[openstack-dev] [Cinder] 3'rd party CI systems: Not Whitelisted Volume type with name * could not be found.

2014-08-14 Thread Asselin, Ramy
Hi,

Does anyone know how to configure cinder ci tests to not have these errors?

12:32:11 *** Not Whitelisted *** 2014-08-14 12:23:56.179 18021 ERROR 
cinder.volume.volume_types [req-c9ec92ab-132b-4167-a467-15bd213659e8 
bd1f54a867ce47acb4097cd94149efa0 d15dcf4cb3c7495389799ec849e9036d - - -] 
Default volume type is not found, please check default_volume_type config: 
Volume type with name dot could not be found.

17:43:28 *** Not Whitelisted *** 2014-08-11 17:31:20.343 2097 ERROR 
cinder.volume.volume_types [req-01e539ad-5357-4a0f-8d58-464b970114f9 
f72cd499be584d9d9585bc26ca71c603 d845ad2d7e6a47dfb84bdf0754f60384 - - -] 
Default volume type is not found, please check default_volume_type config: 
Volume type with name cat_1 could not be found.

http://15.126.198.151/60/111460/18/check/dsvm-tempest-hp-lefthand/3ee1598/console.html
http://publiclogs.emc.com/vnx_ostack/EMC_VNX_FC/212/console.html

Thanks,
Ramy

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] Is there a way to let nova schedule plugin fetch glance image metadata

2014-08-14 Thread Dugger, Donald D
Indeed, no REST API yet, but that's one of the things we will be considering 
for Gantt.  I wonder if this is yet another argument for Gantt and, if you can 
wait until the Kilo cycle, maybe we can provide an appropriate service for Heat.

--
Don Dugger
Censeo Toto nos in Kansa esse decisse. - D. Gale
Ph: 303/443-3786

-Original Message-
From: Jay Pipes [mailto:jaypi...@gmail.com] 
Sent: Wednesday, August 13, 2014 10:12 PM
To: openstack-dev@lists.openstack.org
Subject: Re: [openstack-dev] Is there a way to let nova schedule plugin fetch 
glance image metadata

On 08/13/2014 11:06 PM, zhiwei wrote:
 Hi Jay.

 The case is: When heat create a stack, it will first call our 
 scheduler(will pass image_id), our scheduler will get image metadata 
 by image_id.

 Our scheduler will build a placement policy through image metadata, 
 then start booting VM.

How exactly is Heat calling your scheduler? The Nova scheduler does not have a 
public REST API, so I'm unsure how you are calling it.

More details needed, thanks!
  :)

-jay


___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [OpenStack-Infra] [infra] Third Party CI naming and contact (action required)

2014-08-14 Thread Lucas Eznarriaga
Hi,

Midokura CI should be up to date now too.
We have a problem with zuul-merger when the host id of review.openstack.org
changes. In this case, the new rsa key has to be added manually otherwise
the build will fail with a:

Merge Failed.

This change was unable to be automatically merged with the current state of
the repository. Please rebase your change and upload a new patchset

So if the review.openstack.org host changes soon it will start failing
again.
That said, taking into account that August is the usual vacation month in
Europe and probably in some other parts as well, end of August sounds like
a tight deadline. Also, I have been attending to third-party meetings
regularly but the last one (I have checked the logs) and I feel like an
announcement for these new requirements should have been done there.

Have a great summer!

Lucas




On Wed, Aug 13, 2014 at 8:23 PM, James E. Blair cor...@inaugust.com wrote:

 Hi,

 We've updated the registration requirements for third-party CI systems
 here:

   http://ci.openstack.org/third_party.html

 We now have 86 third-party CI systems registered and have undertaken an
 effort to make things more user-friendly for the developers who interact
 with them.  There are two important changes to be aware of:

 1) We now generally name third-party systems in a descriptive manner
 including the company and product they are testing.  We have renamed
 currently-operating CI systems to match these standards to the best of
 our abilities.  Some of them ended up with particularly bad names (like
 Unknown Function...).  If your system is one of these, please join us
 in #openstack-infra on Freenode to establish a more descriptive name.

 2) We have established a standard wiki page template to supply a
 description of the system, what is tested, and contact information for
 each system.  See https://wiki.openstack.org/wiki/ThirdPartySystems for
 an index of such pages and instructions for creating them.  Each
 third-party CI system will have its own page in the wiki and it must
 include a link to that page in every comment that it leaves in Gerrit.

 If you operate a third-party CI system, please ensure that you register
 a wiki page and update your system to link to it in every new Gerrit
 comment by the end of August.  Beginning in September, we will disable
 systems that have not been updated.

 Thanks,

 Jim

 ___
 OpenStack-Infra mailing list
 openstack-in...@lists.openstack.org
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-infra

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [OpenStack-Dev][Cinder] Cinder Core nomination

2014-08-14 Thread John Griffith
On Thu, Aug 14, 2014 at 10:29 AM, Mike Perez thin...@gmail.com wrote:

 On 06:55 Thu 14 Aug , Boring, Walter wrote:
  Hey guys,
 I wanted to pose a nomination for Cinder core.
 
  Xing Yang.
  She has been active in the cinder community for many releases and has
 worked on several drivers as well as other features for cinder itself.
 She has been doing an awesome job doing reviews and helping folks out in
 the #openstack-cinder irc channel for a long time.   I think she would be a
 good addition to the core team.

 +1

 If you take a look at the last 3 months [1][2] she has been very active and
 providing great feedback in reviews. She has also been setting great
 expectations for other drivers with the recent third-party CI work. Thanks
 Xing!

 [1] - http://russellbryant.net/openstack-stats/cinder-reviewers-90.txt
 [2] - http://stackalytics.com/?user_id=xing-yang

 --
 Mike Perez

 ___
 OpenStack-dev mailing list
 OpenStack-dev@lists.openstack.org
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


​+1 for sure!!​
___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [nova][core] Expectations of core reviewers

2014-08-14 Thread Joe Gordon
On Thu, Aug 14, 2014 at 1:47 AM, Daniel P. Berrange berra...@redhat.com
wrote:

 On Thu, Aug 14, 2014 at 09:24:36AM +1000, Michael Still wrote:
  On Thu, Aug 14, 2014 at 3:09 AM, Dan Smith d...@danplanet.com wrote:
   I'm not questioning the value of f2f - I'm questioning the idea of
   doing f2f meetings sooo many times a year. OpenStack is very much
   the outlier here among open source projects - the vast majority of
   projects get along very well with much less f2f time and a far
   smaller % of their contributors attend those f2f meetings that do
   happen. So I really do question what is missing from OpenStack's
   community interaction that makes us believe that having 4 f2f
   meetings a year is critical to our success.
  
   How many is too many? So far, I have found the midcycles to be
 extremely
   productive -- productive in a way that we don't see at the summits, and
   I think other attendees agree. Obviously if budgets start limiting
 them,
   then we'll have to deal with it, but I don't want to stop meeting
   preemptively.
 
  I agree they're very productive. Let's pick on the nova v3 API case as
  an example... We had failed as a community to reach a consensus using
  our existing discussion mechanisms (hundreds of emails, at least three
  specs, phone calls between the various parties, et cetera), yet at the
  summit and then a midcycle meetup we managed to nail down an agreement
  on a very contentious and complicated topic.

 We thought we had agreement on v3 API after Atlanta f2f summit and
 after Hong Kong f2f too. So I wouldn't neccessarily say that we
 needed another f2f meeting to resolve that, but rather than this is
 a very complex topic that takes a long time to resolve no matter
 how we discuss it and the discussions had just happened to reach
 a natural conclusion this time around. But lets see if this agreement
 actually sticks this time

  I can see the argument that travel cost is an issue, but I think its
  also not a very strong argument. We have companies spending millions
  of dollars on OpenStack -- surely spending a relatively small amount
  on travel to keep the development team as efficient as possible isn't
  a big deal? I wouldn't be at all surprised if the financial costs of
  the v3 API debate (staff time mainly) were much higher than the travel
  costs of those involved in the summit and midcycle discussions which
  sorted it out.

 I think the travel cost really is a big issue. Due to the number of
 people who had to travel to the many mid-cycle meetups, a good number
 of people I work with no longer have the ability to go to the Paris
 design summit. This is going to make it harder for them to feel a
 proper engaged part of our community. I can only see this situation
 get worse over time if greater emphasis is placed on attending the
 mid-cycle meetups.

  Travelling to places to talk to people isn't a great solution, but it
  is the most effective one we've found so far. We should continue to
  experiment with other options, but until we find something that works
  as well as meetups, I think we need to keep having them.
 
   IMHO, the reasons to cut back would be:
  
   - People leaving with a well, that was useless... feeling
   - Not enough people able to travel to make it worthwhile
  
   So far, neither of those have been outcomes of the midcycles we've had,
   so I think we're doing okay.
  
   The design summits are structured differently, where we see a lot more
   diverse attendance because of the colocation with the user summit. It
   doesn't lend itself well to long and in-depth discussions about
 specific
   things, but it's very useful for what it gives us in the way of
   exposure. We could try to have less of that at the summit and more
   midcycle-ish time, but I think it's unlikely to achieve the same level
   of usefulness in that environment.
  
   Specifically, the lack of colocation with too many other projects has
   been a benefit. This time, Mark and Maru where there from Neutron. Last
   time, Mark from Neutron and the other Mark from Glance were there. If
   they were having meetups in other rooms (like at summit) they wouldn't
   have been there exposed to discussions that didn't seem like they'd
 have
   a component for their participation, but did after all (re: nova and
   glance and who should own flavors).
 
  I agree. The ability to focus on the issues that were blocking nova
  was very important. That's hard to do at a design summit when there is
  so much happening at the same time.

 Maybe we should change the way we structure the design summit to
 improve that. If there are critical issues blocking nova, it feels
 like it is better to be able to discuss and resolve as much as possible
 at the start of the dev cycle rather than in the middle of the dev
 cycle because I feel that means we are causing ourselves pain during
 milestone 1/2.

   As I explain in the rest of my email below I'm not advocating
   getting 

Re: [openstack-dev] [Neutron][DevStack] How to increase developer usage of Neutron

2014-08-14 Thread Mike Spreitzer
CARVER, PAUL pc2...@att.com wrote on 08/14/2014 09:35:17 AM:

 Mike Spreitzer [mailto:mspre...@us.ibm.com] wrote:
 
 I'll bet I am not the only developer who is not highly competent with
 bridges and tunnels, Open VSwitch, Neutron configuration, and how 
DevStack
 transmutes all those.  My bet is that you would have more developers 
using
 Neutron if there were an easy-to-find and easy-to-follow recipe to use, 
to
 create a developer install of OpenStack with Neutron.  One that's a 
pretty
 basic and easy case.  Let's say a developer gets a recent image of 
Ubuntu
 14.04 from Canonical, and creates an instance in some undercloud, and 
that
 instance has just one NIC, at 10.9.8.7/16.  If there were a recipe for
 such a developer to follow from that point on, it would be great.
 
 https://wiki.openstack.org/wiki/NeutronDevstack worked for me.
 
 However, I'm pretty sure it's only a single node all in one setup. At 
least,
 I created only one VM to run it on and I don't think DevStack has 
created
 multiple nested VMs inside of the one I create to run DevStack. I 
haven't
 gotten around to figuring out how to setup a full multi-node DevStack
 setup with separate compute nodes and network nodes and GRE/VXLAN 
tunnels.
 
 There are multi-node instructions on that wiki page but I haven't tried
 following them. If someone has a Vagrant file that creates a full multi-
 node Neutron devstack complete with GRE/VXLAN tunnels it would be great
 if they could add it to that wiki page.

A working concrete recipe for a single-node install would be great.

https://wiki.openstack.org/wiki/NeutronDevstack is far from a concrete 
recipe, leaving many blanks to be filled in by the reader.
My problem is that as a non-expert in the relevant networking arcana, 
Neutron implementation,
and DevStack configuration options, it is not entirely obvious how to fill 
in the blanks.
For starters, 
http://docs.openstack.org/admin-guide-cloud/content/network-connectivity.html
speaks of four networks and, appropriately for a general page like that, 
does not relate them to NICs.
But at the start of the day, I need to know how many NICs to put on my 
host VM (the one in which I
will run DevStack to install OpenStack), how to configure them in the host 
VM's operating system,
and how to tell DevStack whatever details it needs and cannot figure out 
on its own
(I am not even clear on what that set is).
I need to know how to derive the fixed and floating IP address ranges from 
the networking context
of my host VM.
A recipe that requires more than one NIC on my host VM can be problematic 
in some situations,
which is why I suggested starting with a recipe for a host with a single 
NIC.

I am not using Xen in my undercloud.  I suspect many developers are not 
using Xen.
I did not even know it was possible to install OpenStack inside a Xen VM;
does that still work?

I was hoping for a working concrete recipe that does not depend on the 
undercloud,
rather something that works in a vanilla context that can easily be 
established with
whatever undercloud a given developer is using.

Thanks,
Mike
___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] Is there a way to let nova schedule plugin fetch glance image metadata

2014-08-14 Thread Sylvain Bauza
Mmm... I can understand that you perhaps need an external scheduler for
your own purposes, but I think you can't expect your plugin merged upstream
for two reasons :
- during Icehouse, there was an effort for not having the scheduler
proxying to the compute nodes
- Call for scheduler needs to go thru Nova-api and no endpoints are there
got just scheduling

That said, once Gantt will be lifted, discussing about possible endpoints
sounds reasonable to me.

-Sylvain

Le 14 août 2014 05:07, zhiwei zhiw...@gmail.com a écrit :

 Hi Jay.

 The case is: When heat create a stack, it will first call our
scheduler(will pass image_id), our scheduler will get image metadata by
image_id.

 Our scheduler will build a placement policy through image metadata, then
start booting VM.


 Thanks.


 On Thu, Aug 14, 2014 at 10:28 AM, Jay Pipes jaypi...@gmail.com wrote:

 On 08/13/2014 10:22 PM, zhiwei wrote:

 Thanks Jay.

 The scheduler plugin is not a scheduler filter.

 We implemented a scheduler instead of using nova native scheduler.


 OK. Any reason why you did this? Without any details on what your
scheduler does, it's tough to give advice on how to solve your problems.


 One of our scheduler component need to fetch image metadata by image_id(
 at this time, there is not instance ).


 Why? Again, the request_spec contains all the information you need about
the image...

 Best,
 -jay

 On Thu, Aug 14, 2014 at 9:29 AM, Jay Pipes jaypi...@gmail.com
 mailto:jaypi...@gmail.com wrote:

 On 08/13/2014 08:31 PM, zhiwei wrote:

 Hi all,

 We wrote a nova schedule plugin that need to fetch image
metadata by
 image_id, but encountered one thing, we did not have the glance
 context.

 Our solution is to configure OpenStack admin user and password
to
 nova.conf, as you know this is not good.

 So, I want to ask if there are any other ways to do this?


 You should not have to do a separate fetch of image metadata in a
 scheduler filter (which is what I believe you meant by plugin
above?).

 The filter object's host_passes() method has a filter_properties
 parameter that contains the request_spec, that in turn contains the
 image, which in turn contains the image metadata. You can access
 it like so:

   def host_passes(self, host_state, filter_properties):
   request_spec = filter_properties['request___spec']

   image_info = request_spec['image']
   # Certain image attributes are accessed via top-level keys,
like
   # size, disk_format, container_format and checksum
   image_size = image_info['size']
   # Other attributes can be accessed in the properties
collection
   # of key/value pairs
   image_props =  image.get('properties', {})
   for key, value in image_props.items():
   # do something...

 Best,
 -jay



 _
 OpenStack-dev mailing list
 OpenStack-dev@lists.openstack.__org
 mailto:OpenStack-dev@lists.openstack.org

http://lists.openstack.org/__cgi-bin/mailman/listinfo/__openstack-dev 
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev





 ___
 OpenStack-dev mailing list
 OpenStack-dev@lists.openstack.org
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev



 ___
 OpenStack-dev mailing list
 OpenStack-dev@lists.openstack.org
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev



 ___
 OpenStack-dev mailing list
 OpenStack-dev@lists.openstack.org
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [OpenStack-Infra] [infra] Third Party CI naming and contact (action required)

2014-08-14 Thread Jeremy Stanley
On 2014-08-14 20:50:21 +0200 (+0200), Lucas Eznarriaga wrote:
[...]
 We have a problem with zuul-merger when the host id of
 review.openstack.org changes. In this case, the new rsa key has to
 be added manually
[...]

If so, you were broken a while... the last (and effectively only in
recent history) time we replaced the Gerrit SSH API RSA host key on
review.openstack.org was in April as a safety precaution in the wake
of the Heartbleed Bug announcement. We definitely haven't touched it
in the four months since then.
-- 
Jeremy Stanley

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [Nova] Concerns around the Extensible Resource Tracker design - revert maybe?

2014-08-14 Thread Sylvain Bauza
Hi mikal,

Le 14 août 2014 01:49, Michael Still mi...@stillhq.com a écrit :

 So, there's been a lot of email in the last few days and I feel I am
 not keeping up.

 Sylvain, can you summarise for me what the plan is here? Can we roll
 forward or do we need to revert?

Well, as we agreed with Nikola, the problem is not with ERT but RT, as the
request data needs to be passed when claiming a resource.

I'm proposing to keep ERT and only consider plugins that are not needing
request_spec when claiming, but here there is no agreement yet.

Unfortunately, I'm on PTO till Tuesday, and Paul Murray this week as well.
So I propose to delay the discussion by these days as that's not impacted
by FPF.

In the meantime, I created a patch for discussing a workaround [1] for Juno
until we correctly figure out how to fix that issue, as it deserves a spec.

 Time is running out for Juno.


Indeed, I'm mostly concerned by the example exception spec that Nikola
mentioned [2] (isolate-scheduler-db) as it still needs a second +2 while
FPF is in 1 week...
I'm planning to deliver an alternative implementation without ERT wrt this
discussion.

-Sylvain

[1] https://review.openstack.org/#/c/113936/

[2] https://review.openstack.org/#/c/89893

 Thanks,
 Michael

 On Thu, Aug 14, 2014 at 3:40 AM, Sylvain Bauza sba...@redhat.com wrote:
 
  Le 13/08/2014 18:40, Brian Elliott a écrit :
 
  On Aug 12, 2014, at 5:21 AM, Nikola Đipanov ndipa...@redhat.com
wrote:
 
  Hey Nova-istas,
 
  While I was hacking on [1] I was considering how to approach the fact
  that we now need to track one more thing (NUMA node utilization) in
our
  resources. I went with - I'll add it to compute nodes table thinking
  it's a fundamental enough property of a compute host that it deserves
to
  be there, although I was considering  Extensible Resource Tracker at
one
  point (ERT from now on - see [2]) but looking at the code - it did not
  seem to provide anything I desperately needed, so I went with keeping
it
  simple.
 
  So fast-forward a few days, and I caught myself solving a problem
that I
  kept thinking ERT should have solved - but apparently hasn't, and I
  think it is fundamentally a broken design without it - so I'd really
  like to see it re-visited.
 
  The problem can be described by the following lemma (if you take
'lemma'
  to mean 'a sentence I came up with just now' :)):
 
  
  Due to the way scheduling works in Nova (roughly: pick a host based on
  stale(ish) data, rely on claims to trigger a re-schedule), _same
exact_
  information that scheduling service used when making a placement
  decision, needs to be available to the compute service when testing
the
  placement.
  “
 
  Correct
 
  This is not the case right now, and the ERT does not propose any way
to
  solve it - (see how I hacked around needing to be able to get
  extra_specs when making claims in [3], without hammering the DB). The
  result will be that any resource that we add and needs user supplied
  info for scheduling an instance against it, will need a buggy
  re-implementation of gathering all the bits from the request that
  scheduler sees, to be able to work properly.
 
  Agreed, ERT does not attempt to solve this problem of ensuring RT has
an
  identical set of information for testing claims.  I don’t think it was
  intended to.
 
  ERT does solve the issue of bloat in the RT with adding
  just-one-more-thing to test usage-wise.  It gives a nice hook for
inserting
  your claim logic for your specific use case.
 
 
  I think Nikola and I agreed on the fact that ERT is not responsible for
this
  design. That said I can talk on behalf of Nikola...
 
 
 
  This is obviously a bigger concern when we want to allow users to pass
  data (through image or flavor) that can affect scheduling, but still a
  huge concern IMHO.
 
  I think passing additional data through to compute just wasn’t a
problem
  that ERT aimed to solve.  (Paul Murray?)  That being said,
coordinating the
  passing of any extra data required to test a claim that is *not*
sourced
  from the host itself would be a very nice addition.  You are working
around
  it with some caching in your flavor db lookup use case, although one
could
  of course cook up a cleaner patch to pass such data through on the
“build
  this” request to the compute.
 
 
  Indeed, and that's why I think the problem can be resolved thanks to 2
  different things :
  1. Filters need to look at what ERT is giving them, that's what
  isolate-scheduler-db is trying to do (see my patches [2.3 and 2.4] on
the
  previous emails
  2. Some extra user request needs to be checked in the test() method of
ERT
  plugins (where claims are done), so I provided a WIP patch for
discussing it
  : https://review.openstack.org/#/c/113936/
 
 
 
  As I see that there are already BPs proposing to use this IMHO broken
  ERT ([4] for example), which will surely add to the proliferation of
  code that hacks around these design shortcomings in what is already a
  

Re: [openstack-dev] [nova] Review priorities as we approach juno-3

2014-08-14 Thread Dan Smith
 == Move Virt Drivers to use Objects (Juno Work) ==
 
 I couldn't actually find any code out for review for this one apart
 from https://review.openstack.org/#/c/94477/, is there more out there?

This was an umbrella one to cover a bunch of virt driver objects work
done early in the cycle. Much of that is done, I haven't gone looking
for anything to see if there are any obvious things to include under
this anymore, but I'll try to do that.

 == Add a virt driver for Ironic ==
 
 This one is in progress, but we need to keep going at it or we wont
 get it merged in time.
 
 * https://review.openstack.org/#/c/111223/ was approved, but a rebased
 ate it. Should be quick to re-approve.
 * https://review.openstack.org/#/c/111423/
 * https://review.openstack.org/#/c/111425/
 * ...there are more reviews in this series, but I'd be super happy to
 see even a few reviewed

I've been reviewing this pretty heavy and I think that it's just taking
a while to make changes given the roundabout way they're getting done
first in Ironic. I'm pretty confident that this one will be okay.

 == VMware: spawn refactor ==
 
 * https://review.openstack.org/#/c/104145/
 * https://review.openstack.org/#/c/104147/ (Dan Smith's -2 on this one
 seems procedural to me)

Yep, we're just trying to get MineSweeper votes on them before letting
them in. We already had one refactor go in without a minesweeper run
that ... broke minesweeper :)

--Dan



signature.asc
Description: OpenPGP digital signature
___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [neutron] [third-party] What tests are required to be run

2014-08-14 Thread trinath.soman...@freescale.com
Hi-

I'm hitting bugs for (basic/advanced)_server_ops testing. 

https://bugs.launchpad.net/nova/+bug/1232303

Kindly help me with a fix to this.

Thanking you.




--
Trinath Somanchi - B39208
trinath.soman...@freescale.com | extn: 4048

-Original Message-
From: Kyle Mestery [mailto:mest...@mestery.com] 
Sent: Thursday, August 14, 2014 9:00 PM
To: OpenStack Development Mailing List (not for usage questions)
Subject: [openstack-dev] [neutron] [third-party] What tests are required to be 
run

Folks, I'm not sure if all CI accounts are running sufficient tests.
Per the requirements wiki page here [1], everyone needs to be running more than 
just Tempest API tests, which I still see most neutron third-party CI setups 
doing. I'd like to ask everyone who operates a third-party CI account for 
Neutron to please look at the link below and make sure you are running 
appropriate tests. If you have questions, the weekly third-party meeting [2] is 
a great place to ask questions.

Thanks,
Kyle

[1] https://wiki.openstack.org/wiki/NeutronThirdPartyTesting
[2] https://wiki.openstack.org/wiki/Meetings/ThirdParty

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [all] The future of the integrated release

2014-08-14 Thread Joe Gordon
On Wed, Aug 13, 2014 at 12:24 PM, Doug Hellmann d...@doughellmann.com
wrote:


 On Aug 13, 2014, at 3:05 PM, Eoghan Glynn egl...@redhat.com wrote:

 
  At the end of the day, that's probably going to mean saying No to more
  things. Everytime I turn around everyone wants the TC to say No to
  things, just not to their particular thing. :) Which is human nature.
  But I think if we don't start saying No to more things we're going to
  end up with a pile of mud that no one is happy with.
 
  That we're being so abstract about all of this is frustrating. I get
  that no-one wants to start a flamewar, but can someone be concrete about
  what they feel we should say 'no' to but are likely to say 'yes' to?
 
 
  I'll bite, but please note this is a strawman.
 
  No:
  * Accepting any more projects into incubation until we are comfortable
 with
  the state of things again
  * Marconi
  * Ceilometer
 
  Well -1 to that, obviously, from me.
 
  Ceilometer is on track to fully execute on the gap analysis coverage
  plan agreed with the TC at the outset of this cycle, and has an active
  plan in progress to address architectural debt.

 Yes, there seems to be an attitude among several people in the community
 that the Ceilometer team denies that there are issues and refuses to work
 on them. Neither of those things is the case from our perspective.


Totally agree.



 Can you be more specific about the shortcomings you see in the project
 that aren’t being addressed?



Once again, this is just a strawman.

I'm just not sure OpenStack has 'blessed' the best solution out there.

https://wiki.openstack.org/wiki/Ceilometer/Graduation#Why_we_think_we.27re_ready



   - Successfully passed the challenge of being adopted by 3 related
   projects which have agreed to join or use ceilometer:
  - Synaps
  - Healthnmon
  - StackTach
  
https://wiki.openstack.org/w/index.php?title=StackTachaction=editredlink=1
  


Stacktach seems to still be under active development (
http://git.openstack.org/cgit/stackforge/stacktach/log/), is used by
rackspace in production and from everything I hear is more mature then
ceilometer.



 
  Divert all cross project efforts from the following projects so we can
 focus
  our cross project resources. Once we are in a bitter place we can
 expand our
  cross project resources to cover these again. This doesn't mean removing
  anything.
  * Sahara
  * Trove
  * Tripleo
 
  You write as if cross-project efforts are both of fixed size and
  amenable to centralized command  control.
 
  Neither of which is actually the case, IMO.
 
  Additional cross-project resources can be ponied up by the large
  contributor companies, and existing cross-project resources are not
  necessarily divertable on command.


Sure additional cross-project resources can and need to be ponied up, but I
am doubtful that will be enough.



 What “cross-project efforts” are we talking about? The liaison program in
 Oslo has been a qualified success so far. Would it make sense to extend
 that to other programs and say that each project needs at least one
 designated QA, Infra, Doc, etc. contact?

 Doug

 
  Yes:
  * All integrated projects that are not listed above
 
  And what of the other pending graduation request?
 
  Cheers,
  Eoghan
 
  ___
  OpenStack-dev mailing list
  OpenStack-dev@lists.openstack.org
  http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


 ___
 OpenStack-dev mailing list
 OpenStack-dev@lists.openstack.org
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [nova] Review priorities as we approach juno-3

2014-08-14 Thread Michael Still
On Fri, Aug 15, 2014 at 6:37 AM, Dan Smith d...@danplanet.com wrote:
 == Move Virt Drivers to use Objects (Juno Work) ==

 I couldn't actually find any code out for review for this one apart
 from https://review.openstack.org/#/c/94477/, is there more out there?

 This was an umbrella one to cover a bunch of virt driver objects work
 done early in the cycle. Much of that is done, I haven't gone looking
 for anything to see if there are any obvious things to include under
 this anymore, but I'll try to do that.

Thanks, I'd appreciate that. If its all done, we should mark it implemented.

 == Add a virt driver for Ironic ==

 This one is in progress, but we need to keep going at it or we wont
 get it merged in time.

 * https://review.openstack.org/#/c/111223/ was approved, but a rebased
 ate it. Should be quick to re-approve.
 * https://review.openstack.org/#/c/111423/
 * https://review.openstack.org/#/c/111425/
 * ...there are more reviews in this series, but I'd be super happy to
 see even a few reviewed

 I've been reviewing this pretty heavy and I think that it's just taking
 a while to make changes given the roundabout way they're getting done
 first in Ironic. I'm pretty confident that this one will be okay.

Yep, I appreciate your focus on this one -- as I am sure the ironic
people do too. If another core was available to pair up with you on
these we might be able to get them to land faster. I was doing that
for a while, but I haven't had time in the last week or so.

 == VMware: spawn refactor ==

 * https://review.openstack.org/#/c/104145/
 * https://review.openstack.org/#/c/104147/ (Dan Smith's -2 on this one
 seems procedural to me)

 Yep, we're just trying to get MineSweeper votes on them before letting
 them in. We already had one refactor go in without a minesweeper run
 that ... broke minesweeper :)

Is there a way for the minesweeper people to manually kick off runs
against specific reviews? Doing that might unblock this faster than
rechecking and hoping.

Michael

-- 
Rackspace Australia

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [nova][core] Expectations of core reviewers

2014-08-14 Thread Sylvain Bauza
Le 14 août 2014 21:56, Joe Gordon joe.gord...@gmail.com a écrit :




 On Thu, Aug 14, 2014 at 1:47 AM, Daniel P. Berrange berra...@redhat.com
wrote:

 On Thu, Aug 14, 2014 at 09:24:36AM +1000, Michael Still wrote:
  On Thu, Aug 14, 2014 at 3:09 AM, Dan Smith d...@danplanet.com wrote:
   I'm not questioning the value of f2f - I'm questioning the idea of
   doing f2f meetings sooo many times a year. OpenStack is very much
   the outlier here among open source projects - the vast majority of
   projects get along very well with much less f2f time and a far
   smaller % of their contributors attend those f2f meetings that do
   happen. So I really do question what is missing from OpenStack's
   community interaction that makes us believe that having 4 f2f
   meetings a year is critical to our success.
  
   How many is too many? So far, I have found the midcycles to be
extremely
   productive -- productive in a way that we don't see at the summits,
and
   I think other attendees agree. Obviously if budgets start limiting
them,
   then we'll have to deal with it, but I don't want to stop meeting
   preemptively.
 
  I agree they're very productive. Let's pick on the nova v3 API case as
  an example... We had failed as a community to reach a consensus using
  our existing discussion mechanisms (hundreds of emails, at least three
  specs, phone calls between the various parties, et cetera), yet at the
  summit and then a midcycle meetup we managed to nail down an agreement
  on a very contentious and complicated topic.

 We thought we had agreement on v3 API after Atlanta f2f summit and
 after Hong Kong f2f too. So I wouldn't neccessarily say that we
 needed another f2f meeting to resolve that, but rather than this is
 a very complex topic that takes a long time to resolve no matter
 how we discuss it and the discussions had just happened to reach
 a natural conclusion this time around. But lets see if this agreement
 actually sticks this time

  I can see the argument that travel cost is an issue, but I think its
  also not a very strong argument. We have companies spending millions
  of dollars on OpenStack -- surely spending a relatively small amount
  on travel to keep the development team as efficient as possible isn't
  a big deal? I wouldn't be at all surprised if the financial costs of
  the v3 API debate (staff time mainly) were much higher than the travel
  costs of those involved in the summit and midcycle discussions which
  sorted it out.

 I think the travel cost really is a big issue. Due to the number of
 people who had to travel to the many mid-cycle meetups, a good number
 of people I work with no longer have the ability to go to the Paris
 design summit. This is going to make it harder for them to feel a
 proper engaged part of our community. I can only see this situation
 get worse over time if greater emphasis is placed on attending the
 mid-cycle meetups.

  Travelling to places to talk to people isn't a great solution, but it
  is the most effective one we've found so far. We should continue to
  experiment with other options, but until we find something that works
  as well as meetups, I think we need to keep having them.
 
   IMHO, the reasons to cut back would be:
  
   - People leaving with a well, that was useless... feeling
   - Not enough people able to travel to make it worthwhile
  
   So far, neither of those have been outcomes of the midcycles we've
had,
   so I think we're doing okay.
  
   The design summits are structured differently, where we see a lot
more
   diverse attendance because of the colocation with the user summit. It
   doesn't lend itself well to long and in-depth discussions about
specific
   things, but it's very useful for what it gives us in the way of
   exposure. We could try to have less of that at the summit and more
   midcycle-ish time, but I think it's unlikely to achieve the same
level
   of usefulness in that environment.
  
   Specifically, the lack of colocation with too many other projects has
   been a benefit. This time, Mark and Maru where there from Neutron.
Last
   time, Mark from Neutron and the other Mark from Glance were there. If
   they were having meetups in other rooms (like at summit) they
wouldn't
   have been there exposed to discussions that didn't seem like they'd
have
   a component for their participation, but did after all (re: nova and
   glance and who should own flavors).
 
  I agree. The ability to focus on the issues that were blocking nova
  was very important. That's hard to do at a design summit when there is
  so much happening at the same time.

 Maybe we should change the way we structure the design summit to
 improve that. If there are critical issues blocking nova, it feels
 like it is better to be able to discuss and resolve as much as possible
 at the start of the dev cycle rather than in the middle of the dev
 cycle because I feel that means we are causing ourselves pain during
 milestone 1/2.

   As I 

Re: [openstack-dev] [Cinder] 3'rd party CI systems: Not Whitelisted Volume type with name * could not be found.

2014-08-14 Thread Martin, Kurt Frederick (ESSN Storage MSDU)
Cinder.conf needs to have a default_volume_type  entry set under the [Default] 
group and a volume type that is valid, for example, default_volume_type=bronze. 
This allows for a volume to be created when a volume type is not selected, the 
default 'None'  type. This feature has been available for some time in cinder 
but recently enabled in devstack.
~Kurt

-Original Message-
From: Asselin, Ramy 
Sent: Thursday, August 14, 2014 10:00 AM
To: OpenStack Development Mailing List (not for usage questions)
Subject: [openstack-dev] [Cinder] 3'rd party CI systems: Not Whitelisted Volume 
type with name * could not be found.

Hi,

Does anyone know how to configure cinder ci tests to not have these errors?

12:32:11 *** Not Whitelisted *** 2014-08-14 12:23:56.179 18021 ERROR 
cinder.volume.volume_types [req-c9ec92ab-132b-4167-a467-15bd213659e8 
bd1f54a867ce47acb4097cd94149efa0 d15dcf4cb3c7495389799ec849e9036d - - -] 
Default volume type is not found, please check default_volume_type config: 
Volume type with name dot could not be found.

17:43:28 *** Not Whitelisted *** 2014-08-11 17:31:20.343 2097 ERROR 
cinder.volume.volume_types [req-01e539ad-5357-4a0f-8d58-464b970114f9 
f72cd499be584d9d9585bc26ca71c603 d845ad2d7e6a47dfb84bdf0754f60384 - - -] 
Default volume type is not found, please check default_volume_type config: 
Volume type with name cat_1 could not be found.

http://15.126.198.151/60/111460/18/check/dsvm-tempest-hp-lefthand/3ee1598/console.html
http://publiclogs.emc.com/vnx_ostack/EMC_VNX_FC/212/console.html

Thanks,
Ramy

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [neutron] Which changes need accompanying bugs?

2014-08-14 Thread Maru Newby

On Aug 13, 2014, at 11:11 AM, Kevin Benton blak...@gmail.com wrote:

 Is the pylint static analysis that caught that error prone to false
 positives? If not, I agree that it would be really nice if that were made
 part of the tox check so these don't have to be fixed after the fact.
 
 To me that particular patch seems like one that should be accompanied with
 a unit test because it's a failure case with cleanup code that isn't being
 touched by the unit tests.

+1

As a general rule I would like to see test addition included with any fix 
targeting a bug that was merged due to a lack of coverage.


m.


 On Aug 13, 2014 8:34 AM, Armando M. arma...@gmail.com wrote:
 
 I am gonna add more color to this story by posting my replies on review
 [1]:
 
 Hi Angus,
 
 You touched on a number of points. Let me try to give you an answer to all
 of them.
 
 (I'll create a bug report too. I still haven't worked out which class
 of changes need an accompanying bug report and which don't.)
 
 The long story can be read below:
 
 https://wiki.openstack.org/wiki/BugFilingRecommendations
 
 https://wiki.openstack.org/wiki/GitCommitMessages
 
 IMO, there's a grey area for some of the issues you found, but when I am
 faced with a bug, I tend to answer myself? Would a bug report be useful to
 someone else? The author of the code? A consumer of the code? Not everyone
 follow the core review system all the time, whereas Launchpad is pretty
 much the tool everyone uses to stay abreast with the OpenStack release
 cycle. Obviously if you're fixing a grammar nit, or filing a cosmetic
 change that has no functional impact then I warrant the lack of a test, but
 in this case you're fixing a genuine error: let's say we want to backport
 this to icehouse, how else would we make the release manager of that?
 He/she is looking at Launchpad.
 
 I can add a unittest for this particular code path, but it would only
 check this particular short segment of code, would need to be maintained as
 the code changes, and wouldn't catch another occurrence somewhere else.
 This seems an unsatisfying return on the additional work :(
 
 You're looking at this from the wrong perspective. This is not about
 ensuring that other code paths are valid, but that this code path stays
 valid over time, ensuring that the code path is exercised and that no other
 regression of any kind creep in. The reason why this error was introduced
 in the first place is because the code wasn't tested when it should have.
 If you don't get that this mechanical effort of fixing errors by static
 analysis is kind of ineffective, which leads me to the last point
 
 I actually found this via static analysis using pylint - and my
 question is: should I create some sort of pylint unittest that tries to
 catch this class of problem across the entire codebase? [...]
 
 I value what you're doing, however I would see even more value if we
 prevented these types of errors from occurring in the first place via
 automation. You run pylint today, but what about tomorrow, or a week from
 now? Are you gonna be filing pylint fixes for ever? We might be better off
 automating the check and catch these types of errors before they land in
 the tree. This means that the work you are doing it two-pronged: a)
 automate the detection of some failures by hooking this into tox.ini via
 HACKING/pep8 or equivalent mechanism and b) file all the fixes that require
 these validation tests to pass; c) everyone is happy, or at least they
 should be.
 
 I'd welcome to explore a better strategy to ensure a better quality of the
 code base, without some degree of automation, nothing will stop these
 conversation from happening again.
 
 Cheers,
 
 Armando
 
 [1] https://review.openstack.org/#/c/113777/
 
 
 On 13 August 2014 03:02, Ihar Hrachyshka ihrac...@redhat.com wrote:
 
 -BEGIN PGP SIGNED MESSAGE-
 Hash: SHA512
 
 On 13/08/14 09:28, Angus Lees wrote:
 I'm doing various small cleanup changes as I explore the neutron
 codebase. Some of these cleanups are to fix actual bugs discovered
 in the code.  Almost all of them are tiny and obviously correct.
 
 A recurring reviewer comment is that the change should have had an
 accompanying bug report and that they would rather that change was
 not submitted without one (or at least, they've -1'ed my change).
 
 I often didn't discover these issues by encountering an actual
 production issue so I'm unsure what to include in the bug report
 other than basically a copy of the change description.  I also
 haven't worked out the pattern yet of which changes should have a
 bug and which don't need one.
 
 There's a section describing blueprints in NeutronDevelopment but
 nothing on bugs.  It would be great if someone who understands the
 nuances here could add some words on when to file bugs: Which type
 of changes should have accompanying bug reports? What is the
 purpose of that bug, and what should it contain?
 
 
 It was discussed before at:
 

Re: [openstack-dev] [Cinder] 3'rd party CI systems: Not Whitelisted Volume type with name * could not be found.

2014-08-14 Thread Asselin, Ramy
Both configurations have that set as you described. [1][2] 
Who actually creates that volume type? Is that supposed to be added manually to 
local.sh, or is this a bug in devstack?

[1] 
http://publiclogs.emc.com/vnx_ostack/EMC_VNX_FC/212/logs/etc/cinder/cinder.conf.txt.gz
[DEFAULT]
default_volume_type = cat_1

[2] 
http://15.126.198.151/60/111460/18/check/dsvm-tempest-hp-lefthand/3ee1598/logs/etc/cinder/cinder.conf.txt.gz
[DEFAULT]
default_volume_type = dot

-Original Message-
From: Martin, Kurt Frederick (ESSN Storage MSDU) 
Sent: Thursday, August 14, 2014 2:00 PM
To: OpenStack Development Mailing List (not for usage questions)
Subject: Re: [openstack-dev] [Cinder] 3'rd party CI systems: Not Whitelisted 
Volume type with name * could not be found.

Cinder.conf needs to have a default_volume_type  entry set under the [Default] 
group and a volume type that is valid, for example, default_volume_type=bronze. 
This allows for a volume to be created when a volume type is not selected, the 
default 'None'  type. This feature has been available for some time in cinder 
but recently enabled in devstack.
~Kurt

-Original Message-
From: Asselin, Ramy 
Sent: Thursday, August 14, 2014 10:00 AM
To: OpenStack Development Mailing List (not for usage questions)
Subject: [openstack-dev] [Cinder] 3'rd party CI systems: Not Whitelisted Volume 
type with name * could not be found.

Hi,

Does anyone know how to configure cinder ci tests to not have these errors?

12:32:11 *** Not Whitelisted *** 2014-08-14 12:23:56.179 18021 ERROR 
cinder.volume.volume_types [req-c9ec92ab-132b-4167-a467-15bd213659e8 
bd1f54a867ce47acb4097cd94149efa0 d15dcf4cb3c7495389799ec849e9036d - - -] 
Default volume type is not found, please check default_volume_type config: 
Volume type with name dot could not be found.

17:43:28 *** Not Whitelisted *** 2014-08-11 17:31:20.343 2097 ERROR 
cinder.volume.volume_types [req-01e539ad-5357-4a0f-8d58-464b970114f9 
f72cd499be584d9d9585bc26ca71c603 d845ad2d7e6a47dfb84bdf0754f60384 - - -] 
Default volume type is not found, please check default_volume_type config: 
Volume type with name cat_1 could not be found.

http://15.126.198.151/60/111460/18/check/dsvm-tempest-hp-lefthand/3ee1598/console.html
http://publiclogs.emc.com/vnx_ostack/EMC_VNX_FC/212/console.html

Thanks,
Ramy

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


[openstack-dev] [Nova] [Containers] Containers Service for OpenStack

2014-08-14 Thread Adrian Otto
OpenStack Devs,

In accordance with guidance from the Nova team at the Midcycle meeting, I have 
drafted the following proposal for a new OpenStack service to fit into the 
Compute program:

https://blueprints.launchpad.net/nova/+spec/containers-service
https://review.openstack.org/114044

Please take a look, and let me know your feedback so we can refine and resource 
this plan. Please let me know if you have an interest in working on this.

Thanks,

Adrian Otto


___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [Cinder] 3'rd party CI systems: Not Whitelisted Volume type with name * could not be found.

2014-08-14 Thread Martin, Kurt Frederick (ESSN Storage MSDU)
They need to be manually added. 

-Original Message-
From: Asselin, Ramy 
Sent: Thursday, August 14, 2014 2:17 PM
To: OpenStack Development Mailing List (not for usage questions)
Subject: Re: [openstack-dev] [Cinder] 3'rd party CI systems: Not Whitelisted 
Volume type with name * could not be found.

Both configurations have that set as you described. [1][2] Who actually creates 
that volume type? Is that supposed to be added manually to local.sh, or is this 
a bug in devstack?

[1] 
http://publiclogs.emc.com/vnx_ostack/EMC_VNX_FC/212/logs/etc/cinder/cinder.conf.txt.gz
[DEFAULT]
default_volume_type = cat_1

[2] 
http://15.126.198.151/60/111460/18/check/dsvm-tempest-hp-lefthand/3ee1598/logs/etc/cinder/cinder.conf.txt.gz
[DEFAULT]
default_volume_type = dot

-Original Message-
From: Martin, Kurt Frederick (ESSN Storage MSDU)
Sent: Thursday, August 14, 2014 2:00 PM
To: OpenStack Development Mailing List (not for usage questions)
Subject: Re: [openstack-dev] [Cinder] 3'rd party CI systems: Not Whitelisted 
Volume type with name * could not be found.

Cinder.conf needs to have a default_volume_type  entry set under the [Default] 
group and a volume type that is valid, for example, default_volume_type=bronze. 
This allows for a volume to be created when a volume type is not selected, the 
default 'None'  type. This feature has been available for some time in cinder 
but recently enabled in devstack.
~Kurt

-Original Message-
From: Asselin, Ramy
Sent: Thursday, August 14, 2014 10:00 AM
To: OpenStack Development Mailing List (not for usage questions)
Subject: [openstack-dev] [Cinder] 3'rd party CI systems: Not Whitelisted Volume 
type with name * could not be found.

Hi,

Does anyone know how to configure cinder ci tests to not have these errors?

12:32:11 *** Not Whitelisted *** 2014-08-14 12:23:56.179 18021 ERROR 
cinder.volume.volume_types [req-c9ec92ab-132b-4167-a467-15bd213659e8 
bd1f54a867ce47acb4097cd94149efa0 d15dcf4cb3c7495389799ec849e9036d - - -] 
Default volume type is not found, please check default_volume_type config: 
Volume type with name dot could not be found.

17:43:28 *** Not Whitelisted *** 2014-08-11 17:31:20.343 2097 ERROR 
cinder.volume.volume_types [req-01e539ad-5357-4a0f-8d58-464b970114f9 
f72cd499be584d9d9585bc26ca71c603 d845ad2d7e6a47dfb84bdf0754f60384 - - -] 
Default volume type is not found, please check default_volume_type config: 
Volume type with name cat_1 could not be found.

http://15.126.198.151/60/111460/18/check/dsvm-tempest-hp-lefthand/3ee1598/console.html
http://publiclogs.emc.com/vnx_ostack/EMC_VNX_FC/212/console.html

Thanks,
Ramy

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [nova] Review priorities as we approach juno-3

2014-08-14 Thread Sylvain Bauza
Le 14 août 2014 22:02, Michael Still mi...@stillhq.com a écrit :

 Hi.

 We're rapidly approaching j-3, so I want to remind people of the
 current reviews that are high priority. The definition of high
 priority I am using here is blueprints that are marked high priority
 in launchpad that have outstanding code for review -- I am sure there
 are other reviews that are important as well, but I want us to try to
 land more blueprints than we have so far. These are listed in the
 order they appear in launchpad.

 == Compute Manager uses Objects (Juno Work) ==


https://review.openstack.org/#/q/status:open+project:openstack/nova+branch:master+topic:bp/compute-manager-objects-juno,n,z

 This is ongoing work, but if you're after some quick code review
 points they're very easy to review and help push the project forward
 in an important manner.

 == Move Virt Drivers to use Objects (Juno Work) ==

 I couldn't actually find any code out for review for this one apart
 from https://review.openstack.org/#/c/94477/, is there more out there?

 == Add a virt driver for Ironic ==

 This one is in progress, but we need to keep going at it or we wont
 get it merged in time.

 * https://review.openstack.org/#/c/111223/ was approved, but a rebased
 ate it. Should be quick to re-approve.
 * https://review.openstack.org/#/c/111423/
 * https://review.openstack.org/#/c/111425/
 * ...there are more reviews in this series, but I'd be super happy to
 see even a few reviewed

 == Create Scheduler Python Library ==

 * https://review.openstack.org/#/c/82778/
 * https://review.openstack.org/#/c/104556/

 (There are a few abandoned patches in this series, I think those two
 are the active ones but please correct me if I am wrong).


Nope that's OK, those are the only one implementing the spec. We had a +2
on 82778 but it needed a rebase, FYI.

 == VMware: spawn refactor ==

 * https://review.openstack.org/#/c/104145/
 * https://review.openstack.org/#/c/104147/ (Dan Smith's -2 on this one
 seems procedural to me)
 * https://review.openstack.org/#/c/105738/
 * ...another chain with many more patches to review

 Thanks,
 Michael

 --
 Rackspace Australia

 ___
 OpenStack-dev mailing list
 OpenStack-dev@lists.openstack.org
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [all] The future of the integrated release

2014-08-14 Thread Doug Hellmann

On Aug 14, 2014, at 4:41 PM, Joe Gordon joe.gord...@gmail.com wrote:

 
 
 
 On Wed, Aug 13, 2014 at 12:24 PM, Doug Hellmann d...@doughellmann.com wrote:
 
 On Aug 13, 2014, at 3:05 PM, Eoghan Glynn egl...@redhat.com wrote:
 
 
  At the end of the day, that's probably going to mean saying No to more
  things. Everytime I turn around everyone wants the TC to say No to
  things, just not to their particular thing. :) Which is human nature.
  But I think if we don't start saying No to more things we're going to
  end up with a pile of mud that no one is happy with.
 
  That we're being so abstract about all of this is frustrating. I get
  that no-one wants to start a flamewar, but can someone be concrete about
  what they feel we should say 'no' to but are likely to say 'yes' to?
 
 
  I'll bite, but please note this is a strawman.
 
  No:
  * Accepting any more projects into incubation until we are comfortable with
  the state of things again
  * Marconi
  * Ceilometer
 
  Well -1 to that, obviously, from me.
 
  Ceilometer is on track to fully execute on the gap analysis coverage
  plan agreed with the TC at the outset of this cycle, and has an active
  plan in progress to address architectural debt.
 
 Yes, there seems to be an attitude among several people in the community that 
 the Ceilometer team denies that there are issues and refuses to work on them. 
 Neither of those things is the case from our perspective.
 
 Totally agree.
  
 
 Can you be more specific about the shortcomings you see in the project that 
 aren’t being addressed?
 
 
 Once again, this is just a straw man.

You’re not the first person to propose ceilometer as a project to kick out of 
the release, though, and so I would like to be talking about specific reasons 
rather than vague frustrations.

 
 I'm just not sure OpenStack has 'blessed' the best solution out there.
 
 https://wiki.openstack.org/wiki/Ceilometer/Graduation#Why_we_think_we.27re_ready
 
 
 Successfully passed the challenge of being adopted by 3 related projects 
 which have agreed to join or use ceilometer:
 Synaps
 Healthnmon
 StackTach
 
 Stacktach seems to still be under active development 
 (http://git.openstack.org/cgit/stackforge/stacktach/log/), is used by 
 rackspace in production and from everything I hear is more mature then 
 ceilometer.

Stacktach is older than ceilometer, but does not do all of the things 
ceilometer does now and aims to do in the future. It has been a while since I 
last looked at it, so the situation may have changed, but some of the reasons 
stacktach would not be a full replacement for ceilometer include: it only works 
with AMQP; it collects notification events, but doesn’t offer any metering 
ability per se (no tracking of values like CPU or bandwidth utilization); it 
only collects notifications from some projects, and doesn’t have a way to 
collect data from swift, which doesn’t emit notifications; and it does not 
integrate with Heat to trigger autoscaling alarms.

We did work with a few of the Stacktach developers on bringing event collection 
into ceilometer, and that work is allowing us to modify the way we store the 
meter data that causes a lot of the performance issues we’ve seen. That work is 
going on now and will be continued into Kilo, when we expect to be adding 
drivers for time-series databases more appropriate for that type of data.

We’ve just finished the TC meeting where some of these issues were discussed, 
but I want to reiterate my stance here more publicly. As a community, we need 
to be able to talk about the technical shortcomings of projects. We need to be 
able to say, for example, “ceilometer, your runtime performance isn’t good 
enough, you need to work on that rather than adding features.” But we must also 
be willing give the team in question the benefit of the doubt that they will 
work on the problem before we bring out the pitchforks of de-integration, 
because there are serious community implications around that level of rejection.

Doug

  
 
 
  Divert all cross project efforts from the following projects so we can 
  focus
  our cross project resources. Once we are in a bitter place we can expand 
  our
  cross project resources to cover these again. This doesn't mean removing
  anything.
  * Sahara
  * Trove
  * Tripleo
 
  You write as if cross-project efforts are both of fixed size and
  amenable to centralized command  control.
 
  Neither of which is actually the case, IMO.
 
  Additional cross-project resources can be ponied up by the large
  contributor companies, and existing cross-project resources are not
  necessarily divertable on command.
 
 Sure additional cross-project resources can and need to be ponied up, but I 
 am doubtful that will be enough.
  
 
 What “cross-project efforts” are we talking about? The liaison program in 
 Oslo has been a qualified success so far. Would it make sense to extend that 
 to other programs and say that each project needs at least one 

Re: [openstack-dev] [nova] Review priorities as we approach juno-3

2014-08-14 Thread Michael Still
I have also been reminded that http://54.201.139.117/nova-bugs.html
tracks bugs with outstanding code reviews (click on ready for
review). There are 179 at the moment, so it sure would be cool to
land some bug fixes.

Thanks,
Michael

On Fri, Aug 15, 2014 at 5:57 AM, Michael Still mi...@stillhq.com wrote:
 Hi.

 We're rapidly approaching j-3, so I want to remind people of the
 current reviews that are high priority. The definition of high
 priority I am using here is blueprints that are marked high priority
 in launchpad that have outstanding code for review -- I am sure there
 are other reviews that are important as well, but I want us to try to
 land more blueprints than we have so far. These are listed in the
 order they appear in launchpad.

 == Compute Manager uses Objects (Juno Work) ==

 https://review.openstack.org/#/q/status:open+project:openstack/nova+branch:master+topic:bp/compute-manager-objects-juno,n,z

 This is ongoing work, but if you're after some quick code review
 points they're very easy to review and help push the project forward
 in an important manner.

 == Move Virt Drivers to use Objects (Juno Work) ==

 I couldn't actually find any code out for review for this one apart
 from https://review.openstack.org/#/c/94477/, is there more out there?

 == Add a virt driver for Ironic ==

 This one is in progress, but we need to keep going at it or we wont
 get it merged in time.

 * https://review.openstack.org/#/c/111223/ was approved, but a rebased
 ate it. Should be quick to re-approve.
 * https://review.openstack.org/#/c/111423/
 * https://review.openstack.org/#/c/111425/
 * ...there are more reviews in this series, but I'd be super happy to
 see even a few reviewed

 == Create Scheduler Python Library ==

 * https://review.openstack.org/#/c/82778/
 * https://review.openstack.org/#/c/104556/

 (There are a few abandoned patches in this series, I think those two
 are the active ones but please correct me if I am wrong).

 == VMware: spawn refactor ==

 * https://review.openstack.org/#/c/104145/
 * https://review.openstack.org/#/c/104147/ (Dan Smith's -2 on this one
 seems procedural to me)
 * https://review.openstack.org/#/c/105738/
 * ...another chain with many more patches to review

 Thanks,
 Michael

 --
 Rackspace Australia



-- 
Rackspace Australia

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] Which program for Rally

2014-08-14 Thread Boris Pavlovic
Matt,

One thing did just occur to me while writing this though it's probably worth
 investigating splitting out the stress test framework as an external
 tool/project after we start work on the tempest library. [3]



I fully agree with the fact that stress testing doesn't belong to Tempest.

This current thread is all about this aspect and all my arguments, related
to splitting Rally and merging it to tempest are related to this.


Could you please elaborate, why instead of ready solution Rally  that has
community and is aligned with OpenStack and OpenStack processes you are
going to create from scratch similar solution?

I really don't see any reasons why we need to duplicate existing and
working solution and can't just work together on Rally?


Best regards,
Boris Pavlovic


On Fri, Aug 15, 2014 at 1:15 AM, Matthew Treinish mtrein...@kortar.org
wrote:

 On Wed, Aug 13, 2014 at 03:48:59PM -0600, Duncan Thomas wrote:
  On 13 August 2014 13:57, Matthew Treinish mtrein...@kortar.org wrote:
   On Tue, Aug 12, 2014 at 01:45:17AM +0400, Boris Pavlovic wrote:
   Keystone, Glance, Cinder, Neutron and Heat are running rally
 performance
   jobs, that can be used for performance testing, benchmarking,
 regression
   testing (already now). These jobs supports in-tree plugins for all
   components (scenarios, load generators, benchmark context) and they
 can use
   Rally fully without interaction with Rally team at all. More about
 these
   jobs:
  
 https://docs.google.com/a/mirantis.com/document/d/1s93IBuyx24dM3SmPcboBp7N47RQedT8u4AJPgOHp9-A/
   So I really don't see anything like this in tempest (even in observed
   future)
 
   So this is actually the communication problem I mentioned before.
 Singling out
   individual projects and getting them to add a rally job is not cross
 project
   communication. (this is part of what I meant by push using Rally)
 There was no
   larger discussion on the ML or a topic in the project meeting about
 adding these
   jobs. There was no discussion about the value vs risk of adding new
 jobs to the
   gate. Also, this is why less than half of the integrated projects have
 these
   jobs. Having asymmetry like this between gating workloads on projects
 helps no
   one.
 
  So the advantage of the approach, rather than having a massive
  cross-product discussion, is that interested projects (I've been very
  interested for a cinder core PoV) act as a test bed for other
  projects. 'Cross project' discussions rather come to other teams, they
  rely on people to find them, where as Boris came to us, said I've got
  this thing you might like, try it out, tell me what you want. He took
  feedback, iterated fast and investigated bugs. It has been a genuine
  pleasure to work with him, and I feel we made progress faster than we
  would have done if it was trying to please everybody.

 I'm not arguing whether Boris was great to work with or not. Or whether
 there
 isn't value in talking directly to the dev team when setting up a new job.
 That
 is definitely the fastest path to getting a new job up and running. But,
 for
 something like adding a new class of dsvm job which runs on every patch,
 that
 affects everyone, not just the project where the jobs are being added. A
 larger
 discussion is really necessary to weigh whether such a job should be
 added. It
 really only needs to happen once, just before the first one is added on an
 integrated project.

 
   That being said the reason I think osprofiler has been more accepted
 and it's
   adoption into oslo is not nearly as contentious is because it's an
 independent
   library that has value outside of itself. You don't need to pull in a
 monolithic
   stack to use it. Which is a design point more conducive with the rest
 of
   OpenStack.
 
  Sorry, are you suggesting tempest isn't a giant monolithic thing?
  Because I was able to comprehend the rally code very quickly, that
  isn't even slightly true of tempest. Having one simple tool that does
  one thing well is exactly what rally has tried to do - tempest seems
  to want to be five different things at once (CI, instalation tests,
  trademark, preformance, stress testing, ...)

 This is actually a common misconception about the purpose and role of
 Tempest.
 Tempest is strictly concerned with being the integration test suite for
 OpenStack, which just includes the actual tests and some methods of
 running the
 tests. This is attempted to be done in a manner which is independent of the
 environment in which tempest is run or run against. (for example, devstack
 vs a
 public cloud) Yes tempest is a large project and has a lot of tests which
 just
 adds to it's complexity, but it's scope is quite targeted. It's just that
 it
 grows at the same rate OpenStack scope grows because tempest has coverage
 for
 all the projects.

 Methods of running the tests does include the stress tests framework, but
 that
 is mostly just a method of leveraging the large quantity of tests we
 currently
 

Re: [openstack-dev] [nova][core] Expectations of core reviewers

2014-08-14 Thread Andrew Laski


On 08/14/2014 02:48 PM, Dugger, Donald D wrote:


My experience with mics, no matter how good, In conference rooms is 
not good.  You are always working hard to hear the soft spoken person 
who is drowned by the background noise. It’s always a strain and I 
don’t want to even think about doing it for a full 8 hrs.  It’s better 
than nothing so I agree, we need to provide some sort of remote 
conferencing at meetups but nothing compares with actually being there.


I’m not a core so I’m speaking from the outside but I think requiring 
attendance at the summits (absence requires a note from your parent 
Jwhile strongly encouraging attendance at mid-cycle meetups seems 
reasonable.




I think expecting summit attendance, rather than requiring, is a better 
way to say it, but otherwise I agree that this seems reasonable.



PS: In re companies spending millions on OpenStack so travel to f2f 
events should be a drop in the bucket.  It doesn’t always work that 
way with the finance department, money comes out of different buckets 
and the travel bucket `always` dries up first.


--

Don Dugger

Censeo Toto nos in Kansa esse decisse. - D. Gale

Ph: 303/443-3786

*From:*Joe Gordon [mailto:joe.gord...@gmail.com]
*Sent:* Wednesday, August 13, 2014 10:48 PM
*To:* OpenStack Development Mailing List (not for usage questions)
*Subject:* Re: [openstack-dev] [nova][core] Expectations of core reviewers

On Wed, Aug 13, 2014 at 8:31 PM, Michael Still mi...@stillhq.com 
mailto:mi...@stillhq.com wrote:


On Thu, Aug 14, 2014 at 1:24 PM, Jay Pipes jaypi...@gmail.com
mailto:jaypi...@gmail.com wrote:

 Just wanted to quickly weigh in with my thoughts on this
important topic. I
 very much valued the face-to-face interaction that came from the
mid-cycle
 meetup in Beaverton (it was the only one I've ever been to).

 That said, I do not believe it should be a requirement that
cores make it to
 the face-to-face meetings in-person. A number of folks have
brought up very
 valid concerns about personal/family time, travel costs and burnout.

I'm not proposing they be a requirement. I am proposing that they be
strongly encouraged.


 I believe that the issue raised about furthering the divide
between core and
 non-core folks is actually the biggest reason I don't support a
mandate to
 have cores at the face-to-face meetings, and I think we should
make our best
 efforts to support quality virtual meetings that can be done on
a more
 frequent basis than the face-to-face meetings that would be
optional.

I am all for online meetings, but we don't have a practical way to do
them at the moment apart from IRC. Until someone has a concrete
proposal that's been shown to work, I feel its a straw man argument.

What about making it easier for remote people to participate at the 
mid-cycle meetups? Set up some microphones and a Google hangout?  At 
least that way attending the mid-cycle is not all or nothing.


We did something like this last cycle (IIRC we didn't have enough 
mics) and it worked pretty well.



Michael

--
Rackspace Australia


___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
mailto:OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev



___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [all] The future of the integrated release

2014-08-14 Thread Eoghan Glynn

  Additional cross-project resources can be ponied up by the large
  contributor companies, and existing cross-project resources are not
  necessarily divertable on command.
 
 Sure additional cross-project resources can and need to be ponied up, but I
 am doubtful that will be enough.

OK, so what exactly do you suspect wouldn't be enough, for what
exactly?

Is it the likely number of such new resources, or the level of domain-
expertise that they can be realistically be expected bring to the
table, or the period of time to on-board them, or something else?

And which cross-project concern do you think is most strained by the
current set of projects in the integrated release? Is it:
 
 * QA
 * infra
 * release management
 * oslo
 * documentation
 * stable-maint
  
or something else?

Each of those teams has quite different prerequisite skill-sets, and
the on-ramp for someone jumping in seeking to make a positive impact
will vary from team to team.

Different approaches have been tried on different teams, ranging from
dedicated project-liaisons (Oslo) to shared cores (Sahara/Infra) to
newly assigned dedicated resources (QA/Infra). Which of these models
might work in your opinion? Which are doomed to failure, and why?

So can you be more specific here on why you think adding more cross-
project resources won't be enough to address an identified shortage
of cross-project resources, while de-integrating projects would be?

And, please, can we put the proverbial strawman back in its box on
this thread? It's all well and good as a polemic device, but doesn't
really move the discussion forward in a constructive way, IMO.

Thanks,
Eoghan

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [Congress] Help in understanding API Flow

2014-08-14 Thread Peter Balland
Hi Madhu,

If you are trying to expose data from a new service, you need to add a new 
data-source driver (examples from nova and neutron are in the tree.)  Once a 
data-source driver is registered, data from the data source will be exposed 
through the API automatically, and you will also be able to write policy based 
on the data.  (Note that using the current code in master your data source will 
not be activated until it is referenced by a policy - a commit is in flight to 
change that, however.)

If, instead, you are trying to expose non datasource data in the API, it would 
involve implementing the DataModel interface (see api/webservice.py) and 
binding it to a URI and handler in congress_server.py.  If this is what you 
want to do, please provide more details on what you are exposing and I can walk 
you through it.

- Peter

From: Madhu Mohan mmo...@mvista.commailto:mmo...@mvista.com
Reply-To: OpenStack Development Mailing List (not for usage questions) 
openstack-dev@lists.openstack.orgmailto:openstack-dev@lists.openstack.org
Date: Thursday, August 14, 2014 at 2:54 AM
To: openstack-dev 
openstack-dev@lists.openstack.orgmailto:openstack-dev@lists.openstack.org
Subject: [openstack-dev] [Congress] Help in understanding API Flow

Hi,

Can some one help me understand how an API is exposed to the clients from 
Congress server ?

I see that a cage service ('api-policy') is created in congress_server.py.
I believe this is implemented in policy_model. I tried to send a json_request 
from my client on the server.

I tried sending list_members, get_items, PUT and  POST as methods and 
all these give me NotImplemented error response.

Any help in this direction ?

I also want to add new APIs and hence understanding the API flow is crucial.

Thanks,
Madhu Mohan
___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [Neutron] Simple proposal for stabilizing new features in-tree

2014-08-14 Thread Wuhongning
Hi Sridar,

Yes I know this is only for phase 1, while I'm also thinking about how it 
should be in next phase. At least, zone concept should be introduced, we may 
use it to replace SG, to eliminate potential conflicts of defining ACL in two 
different places.


From: Sridar Kandaswamy (skandasw) [skand...@cisco.com]
Sent: Thursday, August 14, 2014 10:12 PM
To: OpenStack Development Mailing List (not for usage questions)
Subject: Re: [openstack-dev] [Neutron] Simple proposal for stabilizing new 
features in-tree

Hi Wuhongning:

Yes u are correct – this is phase 1 to at least get basic perimeter firewall 
support working with DVR before looking for an optimal way to address E – W 
traffic.

Thanks

Sridar

From: Wuhongning wuhongn...@huawei.commailto:wuhongn...@huawei.com
Reply-To: OpenStack List 
openstack-dev@lists.openstack.orgmailto:openstack-dev@lists.openstack.org
Date: Thursday, August 14, 2014 at 1:05 AM
To: OpenStack List 
openstack-dev@lists.openstack.orgmailto:openstack-dev@lists.openstack.org
Subject: Re: [openstack-dev] [Neutron] Simple proposal for stabilizing new 
features in-tree

FWaas can't seamlessly work with DVR yet. A BP [1] has been submitted, but it 
can only handle NS traffic, leaving W-E untouched. If we implement the WE 
firewall in DVR, the iptable might be applied at a per port basis, so there are 
some overlapping with SG (Can we image a packet run into iptable hook twice 
between VM and the wire, for both ingress and egress directions?).

Maybe the overall service plugins (including service extension in ML2) needs 
some cleaning up, It seems that Neutron is just built from separate single 
blocks.

[1]  
http://git.openstack.org/cgit/openstack/neutron-specs/tree/specs/juno/neutron-dvr-fwaas.rst

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


[openstack-dev] [Sahara] Swift authentication and backwards compatibility

2014-08-14 Thread Michael McCune
hello Sahara folks,

I am working to get the revamped spec[1] finalized and I'd like to know the 
group's thoughts on the idea of backward compatibility. It is possible to 
implement the new authentication method and remain backward compatible, but we 
will need to keep the username and password inputs in the Swift data forms.

Having the backward compatibility would also give Sahara a way to react in 
situations where the proxy domain is not available or the administrator doesn't 
wish to use it. I'm not sure this is the behavior we want, but I don't know if 
it is proper for Sahara to exit if no proxy domain can be found.

If we choose not to remain backward compatible then we are requiring Sahara 
operators to create the new proxy domain needed, and they must update all 
virtual machine images.

Thoughts?

regards,
mike

[1]: https://review.openstack.org/#/c/113591/

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


[openstack-dev] [TripleO] Python 3 support in os-*-config

2014-08-14 Thread Steve Kowalik
Hai,

I've been poking at Python 3 support in os-*-config recently, and
here's what is going on with it:

 * os-apply-config
   This one looks to be the hardest one to port. Multiple test failures,
hard coded use of '#!/usr/bin/env python' which is unhappy in a only
Python 3 environment, map() changes, byte changes ...

 * os-cloud-config
   I have a patch to fix current Python 3 issues up
(https://review.openstack.org/111606), and then my plan is to add Python
3 jobs to check *and* gate, so in-flight patches may need to watch out
for that soon.

 * os-collect-config
   This needs six sprinkled throughout it due to URL parsing, and I need
to work out what to do WRT absolute importing due to:
  File ./os_collect_config/collect.py, line 25, in module
from openstack.common import log
ImportError: No module named 'openstack'

 * os-net-config
   I know this isn't really a thing yet, but I also have yet to look
into it's Python 3 support.

 * os-refresh-config
   This has Python 3 jobs running in check and gate. Does anyone have a
small fix to test them out, the last code change that landed in
os-refresh-config was mid July.

Thanks,
-- 
Steve
I hate temporal mechanics!
 - Chief Miles O'Brien, Deep Space Nine

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [neutron][ml2] Mech driver as out-of-tree add-on

2014-08-14 Thread loy wolfe
On Thu, Aug 14, 2014 at 9:07 PM, Kyle Mestery mest...@mestery.com wrote:

 I also feel like the drivers/plugins are currently BEYOND a tipping
 point, and are in fact dragging down velocity of the core project in
 many ways. I'm working on a proposal for Kilo where we move all
 drivers/plugins out of the main Neutron tree and into a separate git


not all drivers/plugins, but most which are not built-in. For example,
ML2 with ovs/linux/sriov MD and l2pop MD should be kept in tree as the
default built-in backend, but all vendor specific MD and shim REST proxy
like all kinds of SDN controller MD should be removed out.


 repository under the networking program. We have way too many drivers,
 requiring way too man review cycles, for this to be a sustainable
 model going forward. Since the main reason plugin/driver authors want
 their code upstream is to be a part of the simultaneous release, and
 thus be packaged by distributions, having a separate repository for
 these will satisfy this requirement. I'm still working through the
 details around reviews of this repository, etc.

 Also, I feel as if the level of passion on the mailing list has died
 down a bit, so I thought I'd send something out to try and liven
 things up a bit. It's been somewhat non-emotional here for a day or
 so. :)

 Thanks,
 Kyle

 On Thu, Aug 14, 2014 at 5:09 AM, Salvatore Orlando sorla...@nicira.com
 wrote:
  I think there will soon be a discussion regarding what the appropriate
  location for plugin and drivers should be.
  My personal feeling is that Neutron has simply reached the tipping point
  where the high number of drivers and plugins is causing unnecessary load
 for
  the core team and frustration for the community.
 
  There I would totally support Luke's initiative about maintaining an
  out-of-tree ML2 driver. On the other hand, a plugin/driver diaspora
 might
  also have negative consequences such as frequent breakages such as those
 Bob
  was mentioning or confusion for users which might need to end up fetching
  drivers from disparate sources.
 
  As mentioned during the last Neutron IRC meeting this is another
 process
  aspect which will be discussed soon, with the aim of defining a plan for:
  - drastically reduce the number of plugins and drivers which must be
  maintained in the main source tree
  - enhance control of plugin/driver maintainers over their own code
  - preserve the ability of doing CI checks on gerrit as we do today
  - raise the CI bar (maybe finally set the smoketest as a minimum
  requirement?)
 
  Regards,
  Salvatore
 
 
 
  On 14 August 2014 11:47, loy wolfe loywo...@gmail.com wrote:
 
 
 
 
  On Thu, Aug 14, 2014 at 4:22 PM, Mathieu Rohon mathieu.ro...@gmail.com
 
  wrote:
 
  Hi,
 
  I would like to add that it would be harder for the community to help
  maintaining drivers.
  such a work [1] wouldn't have occured with an out of tree ODL driver.
 
 
  +1.
  It's better to move all MD for none built-in backend out of tree,
  maintaining these drivers shouldn't be the responsibility of community.
 Not
  only MD, but also plugin, agent should all obey this rule
 
 
 
  [1] https://review.openstack.org/#/c/96459/
 
  On Wed, Aug 13, 2014 at 1:09 PM, Robert Kukura 
 kuk...@noironetworks.com
  wrote:
   One thing to keep in mind is that the ML2 driver API does sometimes
   change,
   requiring updates to drivers. Drivers that are in-tree get updated
   along
   with the driver API change. Drivers that are out-of-tree must be
   updated by
   the owner.
  
   -Bob
  
  
   On 8/13/14, 6:59 AM, ZZelle wrote:
  
   Hi,
  
  
   The important thing to understand is how to integrate with neutron
   through
   stevedore/entrypoints:
  
  
  
 https://github.com/dave-tucker/odl-neutron-drivers/blob/master/setup.cfg#L32-L34
  
  
   Cedric
  
  
   On Wed, Aug 13, 2014 at 12:17 PM, Dave Tucker d...@dtucker.co.uk
   wrote:
  
   I've been working on this for OpenDaylight
   https://github.com/dave-tucker/odl-neutron-drivers
  
   This seems to work for me (tested Devstack w/ML2) but YMMV.
  
   -- Dave
  
   ___
   OpenStack-dev mailing list
   OpenStack-dev@lists.openstack.org
   http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
  
  
  
  
   ___
   OpenStack-dev mailing list
   OpenStack-dev@lists.openstack.org
   http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
  
  
  
   ___
   OpenStack-dev mailing list
   OpenStack-dev@lists.openstack.org
   http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
  
 
  ___
  OpenStack-dev mailing list
  OpenStack-dev@lists.openstack.org
  http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
 
 
 
  ___
  OpenStack-dev mailing list
  OpenStack-dev@lists.openstack.org
  

Re: [openstack-dev] [neutron][ml2] Mech driver as out-of-tree add-on

2014-08-14 Thread Kevin Benton
I also feel like the drivers/plugins are currently BEYOND a tipping
point, and are in fact dragging down velocity of the core project in
many ways.

Can you elaborate on the ways that they are slowing down the velocity? I
know they take up reviewer time, but are there other ways that you think
they slow down the project?


On Thu, Aug 14, 2014 at 6:07 AM, Kyle Mestery mest...@mestery.com wrote:

 I also feel like the drivers/plugins are currently BEYOND a tipping
 point, and are in fact dragging down velocity of the core project in
 many ways. I'm working on a proposal for Kilo where we move all
 drivers/plugins out of the main Neutron tree and into a separate git
 repository under the networking program. We have way too many drivers,
 requiring way too man review cycles, for this to be a sustainable
 model going forward. Since the main reason plugin/driver authors want
 their code upstream is to be a part of the simultaneous release, and
 thus be packaged by distributions, having a separate repository for
 these will satisfy this requirement. I'm still working through the
 details around reviews of this repository, etc.

 Also, I feel as if the level of passion on the mailing list has died
 down a bit, so I thought I'd send something out to try and liven
 things up a bit. It's been somewhat non-emotional here for a day or
 so. :)

 Thanks,
 Kyle

 On Thu, Aug 14, 2014 at 5:09 AM, Salvatore Orlando sorla...@nicira.com
 wrote:
  I think there will soon be a discussion regarding what the appropriate
  location for plugin and drivers should be.
  My personal feeling is that Neutron has simply reached the tipping point
  where the high number of drivers and plugins is causing unnecessary load
 for
  the core team and frustration for the community.
 
  There I would totally support Luke's initiative about maintaining an
  out-of-tree ML2 driver. On the other hand, a plugin/driver diaspora
 might
  also have negative consequences such as frequent breakages such as those
 Bob
  was mentioning or confusion for users which might need to end up fetching
  drivers from disparate sources.
 
  As mentioned during the last Neutron IRC meeting this is another
 process
  aspect which will be discussed soon, with the aim of defining a plan for:
  - drastically reduce the number of plugins and drivers which must be
  maintained in the main source tree
  - enhance control of plugin/driver maintainers over their own code
  - preserve the ability of doing CI checks on gerrit as we do today
  - raise the CI bar (maybe finally set the smoketest as a minimum
  requirement?)
 
  Regards,
  Salvatore
 
 
 
  On 14 August 2014 11:47, loy wolfe loywo...@gmail.com wrote:
 
 
 
 
  On Thu, Aug 14, 2014 at 4:22 PM, Mathieu Rohon mathieu.ro...@gmail.com
 
  wrote:
 
  Hi,
 
  I would like to add that it would be harder for the community to help
  maintaining drivers.
  such a work [1] wouldn't have occured with an out of tree ODL driver.
 
 
  +1.
  It's better to move all MD for none built-in backend out of tree,
  maintaining these drivers shouldn't be the responsibility of community.
 Not
  only MD, but also plugin, agent should all obey this rule
 
 
 
  [1] https://review.openstack.org/#/c/96459/
 
  On Wed, Aug 13, 2014 at 1:09 PM, Robert Kukura 
 kuk...@noironetworks.com
  wrote:
   One thing to keep in mind is that the ML2 driver API does sometimes
   change,
   requiring updates to drivers. Drivers that are in-tree get updated
   along
   with the driver API change. Drivers that are out-of-tree must be
   updated by
   the owner.
  
   -Bob
  
  
   On 8/13/14, 6:59 AM, ZZelle wrote:
  
   Hi,
  
  
   The important thing to understand is how to integrate with neutron
   through
   stevedore/entrypoints:
  
  
  
 https://github.com/dave-tucker/odl-neutron-drivers/blob/master/setup.cfg#L32-L34
  
  
   Cedric
  
  
   On Wed, Aug 13, 2014 at 12:17 PM, Dave Tucker d...@dtucker.co.uk
   wrote:
  
   I've been working on this for OpenDaylight
   https://github.com/dave-tucker/odl-neutron-drivers
  
   This seems to work for me (tested Devstack w/ML2) but YMMV.
  
   -- Dave
  
   ___
   OpenStack-dev mailing list
   OpenStack-dev@lists.openstack.org
   http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
  
  
  
  
   ___
   OpenStack-dev mailing list
   OpenStack-dev@lists.openstack.org
   http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
  
  
  
   ___
   OpenStack-dev mailing list
   OpenStack-dev@lists.openstack.org
   http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
  
 
  ___
  OpenStack-dev mailing list
  OpenStack-dev@lists.openstack.org
  http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
 
 
 
  ___
  OpenStack-dev mailing list
  

[openstack-dev] [OpenStack-Dev] [Cinder] Summer 2014 meetup summary

2014-08-14 Thread John Griffith
Hey Everyone,

I wanted to send out a quick message regarding the Cinder Mid-Cycle (kinda)
Meetup we just finished up at the Fort Collins HP Site.

 First off, I want to thank everyone who attended for a great and very
productive three days!!  Not only did we have a strong physical turn-out
(16 participants) but we also had a good number of folks attend via Google
HangOuts.  We covered a lot of topics and details as a group.  Agenda items
ranged from collective review of new Cinder features for Juno
(Pool-Scheduler, Consistency-Groups and Replication) to sharing approaches
to third-party CI deployment and lessons learned.  We also spent a good
deal of time discussing future plans and ideas for Cinder.

We had a great mix of folks, including long time core members, new
contributors and everything inbetween.  This even included representatives
from the Infra Team as well as a guest appearance from the Horizon team.
 It was a great format and everybody really pulled together in an inclusive
format to help accomplish quite a bit and set the tone for work to do in
the upcoming Kilo release.

Anyway, IMO it was a great effort and I really appreciate all those who
participated.  If you'd like more detail on the things that we discussed
you can check out the etherpad which includes some of the nitty gritty
details. Of course any questions or details for those that weren't present,
drop in to the Cinder channel with any questions or ideas that you might
have.

Thanks,
John

https://etherpad.openstack.org/p/cinder-meetup-summer-2014
___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


[openstack-dev] [Octavia] Object Model and DB Structure

2014-08-14 Thread Brandon Logan
So I've been assuming that the Octavia object model would be an exact
copy of the neutron lbaas one with additional information for Octavia.
However, after thinking about it I'm not sure this is the right way to
go because the object model in neutron lbaas may change in the future,
and Octavia can't just change it's object model when neutron
lbaas/openstack lbaas changes it's object model.  So if there are any
lessons learned we would like to apply to Octavia's object model now is
the time.

Entity name changes are also on the table if people don't really like
some of the names.  Even adding new entities or removing entities if
there are good reasons isn't out of the question.

Anyway here are a few of my suggestions.  Please add on to this if you
want.  Also, just flat out tell me I'm wrong on some of htese
suggestions if you feel as such.

A few improvements I'd suggest (using the current entity names):
-A real root object that is the only top level object (loadbalancer).
--This would be 1:M relationship with Listeners, but Listeners would
only be children of loadbalancers.
--Pools, Members, and Health Monitors would follow the same workflow.
--Basically no shareable entities.

-Provisioning status only on the root object (loadbalancer).
--PENDING_CREATE, PENDING_UPDATE, PENDING_DELETE, ACTIVE (No need for a
DEFEERRED status! YAY!)
--Also maybe a DELETED status.

-Operating status on other entities
--ACTIVE or ONLINE, DEGRADED, INACTIVE or OFFLINE
--Pools and Members
--Listeners have been mentioned but I'd like to hear more details on
that.

-Adding status_description field in, or something similar.  Would only
eixst on loadbalancer entity if loadbalancer is the only top level
object.

Thanks,
Brandon
___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev