Re: [Openstack] Four compute-node, everytime the 1st and 2nd are choosen

2011-11-10 Thread Jorge Luiz Correa
Is there a flag in nova.conf that permits we configure that? In
documentation we can see that exists some algorithms used by scheduler.
But, I don't know how to choose that one best fit our requirements.

Thanks!
:)

On Wed, Nov 9, 2011 at 1:36 PM, Ed Leafe ed.le...@rackspace.com wrote:

 On Nov 9, 2011, at 7:51 AM, Razique Mahroua wrote:

  I use the default scheduler, in fact, I've never tunned it really.
  The hypervisors all run KVM

This is where the flag is defined in nova.scheduler.least_cost.py:

  32 FLAGS = flags.FLAGS
  33 flags.DEFINE_list('least_cost_functions',
  34 ['nova.scheduler.least_cost.compute_fill_first_cost_fn'],
  35 'Which cost functions the LeastCostScheduler should use.')

Since the default weighting function is
 'compute_fill_first_cost_fn', which, as its name suggests, chooses hosts so
 as to fill up one host as much as possible before selecting another, the
 pattern you're seeing is expected. If you change that flag to
 'nova.scheduler.noop_cost_fn', you should see the hosts selected randomly.
 The idea is that you can create your own weighting functions that will
 select potential hosts in a way that best fits your needs.


 -- Ed Leafe


 ___
 Mailing list: https://launchpad.net/~openstack
 Post to : openstack@lists.launchpad.net
 Unsubscribe : https://launchpad.net/~openstack
 More help   : https://help.launchpad.net/ListHelp




-- 
- MSc. Correa, J.L.
___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Four compute-node, everytime the 1st and 2nd are choosen

2011-11-10 Thread Razique Mahroua
+1 :)
Razique Mahrouarazique.mahr...@gmail.com

Le 10 nov. 2011 à 11:27, Jorge Luiz Correa a écrit :Is there a flag in nova.conf that permits we configure that? In documentation we can see that exists some algorithms used by scheduler. But, I don't know how to choose that one best fit our requirements.
Thanks!:)On Wed, Nov 9, 2011 at 1:36 PM, Ed Leafe ed.le...@rackspace.com wrote:
On Nov 9, 2011, at 7:51 AM, Razique Mahroua wrote:

 I use the default scheduler, in fact, I've never tunned it really.
 The hypervisors all run KVM

This is where the flag is defined in nova.scheduler.least_cost.py:

32 FLAGS = flags.FLAGS
33 flags.DEFINE_list('least_cost_functions',
34 ['nova.scheduler.least_cost.compute_fill_first_cost_fn'],
35 'Which cost functions the LeastCostScheduler should use.')

Since the default weighting function is 'compute_fill_first_cost_fn', which, as its name suggests, chooses hosts so as to fill up one host as much as possible before selecting another, the pattern you're seeing is expected. If you change that flag to 'nova.scheduler.noop_cost_fn', you should see the hosts selected randomly. The idea is that you can create your own weighting functions that will select potential hosts in a way that best fits your needs.



-- Ed Leafe


___
Mailing list: https://launchpad.net/~openstack
Post to   : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help  : https://help.launchpad.net/ListHelp
-- - MSc. Correa, J.L.

___Mailing list: https://launchpad.net/~openstackPost to : openstack@lists.launchpad.netUnsubscribe : https://launchpad.net/~openstackMore help : https://help.launchpad.net/ListHelp___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Proposal to add Johannes Erdfelt to nova-core

2011-11-10 Thread Soren Hansen
We now have a sufficient amount of +1 to go ahead with this. If noone
objects before Tuesday next week, I'll make it so.

-- 
Soren Hansen        | http://linux2go.dk/
Ubuntu Developer    | http://www.ubuntu.com/
OpenStack Developer | http://www.openstack.org/

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Proposal to add Kevin Mitchell to nova-core

2011-11-10 Thread Soren Hansen
That brings us to +5. If noone objects by Tuesday next week, I'll make it so.

-- 
Soren Hansen        | http://linux2go.dk/
Ubuntu Developer    | http://www.ubuntu.com/
OpenStack Developer | http://www.openstack.org/

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Host Aggregates ...

2011-11-10 Thread Armando Migliaccio
Hi Sandy,

Thanks for taking the time to read this.

My understanding is that a typical Nova deployment would span across multiple 
zones, that zones may have subzones, and that child zones will have a number of 
availability zones in them; please do correct me if I am wrong :)

That stated, it was assumed that an aggregate will be a grouping of servers 
within an availability zone (hence the introduction of the extra concept), and 
would be used to manage hypervisor pools when and if required. This introduces 
benefits like VM live migration, VM HA and zero-downtime host upgrades. The 
introduction of hypervisor pools is just the easy way to get these benefits in 
the short term. 

Going back to your point, it is possible to match host-aggregates with 
single-zone that uses capabilities on the implementation level (assumed that 
it is okay to be unable to represent aggregates as children of availability 
zones). Nevertheless, I still see zones and aggregates as being different on 
the conceptual level. 

What is your view if we went with the approach of implementing an aggregate as 
a special single-zone that uses capabilities? Would there be a risk of 
tangling the zone management API a bit?

Thanks for feedback!

Cheers,
Armando

 -Original Message-
 From: Sandy Walsh [mailto:sandy.wa...@rackspace.com]
 Sent: 09 November 2011 21:10
 To: Armando Migliaccio
 Cc: openstack@lists.launchpad.net
 Subject: Host Aggregates ...
 
 Hi Armando,
 
 I finally got around to reading
 https://blueprints.launchpad.net/nova/+spec/host-aggregates.
 
 Perhaps you could elaborate a little on how this differs from host
 capabilities (key-value pairs associated with a service) that the scheduler
 can use when making decisions?
 
 The distributed scheduler doesn't need zones to operate, but will use them if
 available. Would host-aggregates simply be a single-zone that uses
 capabilities?
 
 Cheers,
 Sandy

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Four compute-node, everytime the 1st and 2nd are choosen

2011-11-10 Thread Sandy Walsh
Are you using Diablo or Trunk?

If you're using trunk the default scheduler is MultiScheduler, which uses 
Chance scheduler. I think Diablo uses Chance by default?

--scheduler_driver

Unless you've explicitly selected the LeastCostScheduler (which only exists in 
Diablo now) I wouldn't worry about those settings.

Did you explicitly define a scheduler to use?


From: openstack-bounces+sandy.walsh=rackspace@lists.launchpad.net 
[openstack-bounces+sandy.walsh=rackspace@lists.launchpad.net] on behalf of 
Jorge Luiz Correa [corre...@gmail.com]
Sent: Thursday, November 10, 2011 6:27 AM
To: openstack@lists.launchpad.net
Subject: Re: [Openstack] Four compute-node, everytime the 1st and 2nd are 
choosen

Is there a flag in nova.conf that permits we configure that? In documentation 
we can see that exists some algorithms used by scheduler. But, I don't know how 
to choose that one best fit our requirements.

Thanks!
:)

On Wed, Nov 9, 2011 at 1:36 PM, Ed Leafe 
ed.le...@rackspace.commailto:ed.le...@rackspace.com wrote:
On Nov 9, 2011, at 7:51 AM, Razique Mahroua wrote:

 I use the default scheduler, in fact, I've never tunned it really.
 The hypervisors all run KVM

   This is where the flag is defined in 
nova.scheduler.least_cost.pyhttp://nova.scheduler.least_cost.py:

 32 FLAGS = flags.FLAGS
 33 flags.DEFINE_list('least_cost_functions',
 34 ['nova.scheduler.least_cost.compute_fill_first_cost_fn'],
 35 'Which cost functions the LeastCostScheduler should use.')

   Since the default weighting function is 'compute_fill_first_cost_fn', 
which, as its name suggests, chooses hosts so as to fill up one host as much as 
possible before selecting another, the pattern you're seeing is expected. If 
you change that flag to 'nova.scheduler.noop_cost_fn', you should see the hosts 
selected randomly. The idea is that you can create your own weighting functions 
that will select potential hosts in a way that best fits your needs.


-- Ed Leafe


___
Mailing list: https://launchpad.net/~openstack
Post to : 
openstack@lists.launchpad.netmailto:openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp



--
- MSc. Correa, J.L.

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Host Aggregates ...

2011-11-10 Thread Sandy Walsh
Ok, that helps ... now I see the abstraction your going for (a new layer under 
availability zones).

Personally I prefer a tagging approach to a modeled hierarchy. It was something 
we debated at great length with Zones. In this case, the tag would be in the 
capabilities assigned to the host.

I think both availability zones (and host aggregates) should be modeled using 
tags/capabilities without having to explicitly model it as a tree or in the db 
... which is how I see this evolving. At the scheduler level we should be able 
to make decisions using simple tag collections.

WestCoast, HasGPU, GeneratorBackup, PriorityNetwork

Are we saying the same thing?

Are there use cases that this approach couldn't handle?

-S


From: Armando Migliaccio [armando.migliac...@eu.citrix.com]
Sent: Thursday, November 10, 2011 8:50 AM
To: Sandy Walsh
Cc: openstack@lists.launchpad.net
Subject: RE: Host Aggregates ...

Hi Sandy,

Thanks for taking the time to read this.

My understanding is that a typical Nova deployment would span across multiple 
zones, that zones may have subzones, and that child zones will have a number of 
availability zones in them; please do correct me if I am wrong :)

That stated, it was assumed that an aggregate will be a grouping of servers 
within an availability zone (hence the introduction of the extra concept), and 
would be used to manage hypervisor pools when and if required. This introduces 
benefits like VM live migration, VM HA and zero-downtime host upgrades. The 
introduction of hypervisor pools is just the easy way to get these benefits in 
the short term.

Going back to your point, it is possible to match host-aggregates with 
single-zone that uses capabilities on the implementation level (assumed that 
it is okay to be unable to represent aggregates as children of availability 
zones). Nevertheless, I still see zones and aggregates as being different on 
the conceptual level.

What is your view if we went with the approach of implementing an aggregate as 
a special single-zone that uses capabilities? Would there be a risk of 
tangling the zone management API a bit?

Thanks for feedback!

Cheers,
Armando

 -Original Message-
 From: Sandy Walsh [mailto:sandy.wa...@rackspace.com]
 Sent: 09 November 2011 21:10
 To: Armando Migliaccio
 Cc: openstack@lists.launchpad.net
 Subject: Host Aggregates ...

 Hi Armando,

 I finally got around to reading
 https://blueprints.launchpad.net/nova/+spec/host-aggregates.

 Perhaps you could elaborate a little on how this differs from host
 capabilities (key-value pairs associated with a service) that the scheduler
 can use when making decisions?

 The distributed scheduler doesn't need zones to operate, but will use them if
 available. Would host-aggregates simply be a single-zone that uses
 capabilities?

 Cheers,
 Sandy

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Four compute-node, everytime the 1st and 2nd are choosen

2011-11-10 Thread Sateesh Chodapuneedi
The flag (of type list) you can use is --least_cost_functions
Multiple algorithms can be specified, the default one is 
compute_fill_first_cost_fn which gives priority to the compute host 
(XenServer/ESX etc.) with more free RAM.

Regards,
Sateesh


This e-mail message is for the sole use of the intended recipient(s) and may 
contain confidential and/or privileged information. Any unauthorized review, 
use, disclosure, or distribution is prohibited. If you are not the intended 
recipient, please contact the sender by reply e-mail and destroy all copies of 
the original message.


From: openstack-bounces+sateesh.chodapuneedi=citrix@lists.launchpad.net 
[mailto:openstack-bounces+sateesh.chodapuneedi=citrix@lists.launchpad.net] 
On Behalf Of Razique Mahroua
Sent: Thursday, November 10, 2011 4:12 PM
To: Jorge Luiz Correa
Cc: openstack@lists.launchpad.net
Subject: Re: [Openstack] Four compute-node, everytime the 1st and 2nd are 
choosen

+1 :)

Razique Mahroua
razique.mahr...@gmail.commailto:razique.mahr...@gmail.com

[cid:image002.jpg@01CC9FD5.EC50CAF0]

Le 10 nov. 2011 à 11:27, Jorge Luiz Correa a écrit :


Is there a flag in nova.conf that permits we configure that? In documentation 
we can see that exists some algorithms used by scheduler. But, I don't know how 
to choose that one best fit our requirements.

Thanks!
:)
On Wed, Nov 9, 2011 at 1:36 PM, Ed Leafe 
ed.le...@rackspace.commailto:ed.le...@rackspace.com wrote:
On Nov 9, 2011, at 7:51 AM, Razique Mahroua wrote:

 I use the default scheduler, in fact, I've never tunned it really.
 The hypervisors all run KVM

   This is where the flag is defined in 
nova.scheduler.least_cost.pyhttp://nova.scheduler.least_cost.py/:

 32 FLAGS = flags.FLAGS
 33 flags.DEFINE_list('least_cost_functions',
 34 ['nova.scheduler.least_cost.compute_fill_first_cost_fn'],
 35 'Which cost functions the LeastCostScheduler should use.')

   Since the default weighting function is 'compute_fill_first_cost_fn', 
which, as its name suggests, chooses hosts so as to fill up one host as much as 
possible before selecting another, the pattern you're seeing is expected. If 
you change that flag to 'nova.scheduler.noop_cost_fn', you should see the hosts 
selected randomly. The idea is that you can create your own weighting functions 
that will select potential hosts in a way that best fits your needs.


-- Ed Leafe


___
Mailing list: https://launchpad.net/~openstack
Post to : 
openstack@lists.launchpad.netmailto:openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp



--
- MSc. Correa, J.L.
___
Mailing list: https://launchpad.net/~openstack
Post to : 
openstack@lists.launchpad.netmailto:openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp

inline: image001.gifinline: image002.jpg___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Stable branch reviews

2011-11-10 Thread Mark McLoughlin
Hey,

On Wed, 2011-11-09 at 16:50 +0100, Thierry Carrez wrote:
 Hi everyone,
 
 Since there seems to be some confusion around master vs. stable/diablo
 vs. core reviewers, I think it warrants a small thread.
 
 When at the Design Summit we discussed setting up stable branches, I
 warned about the risks that setting them up brings for trunk development:
 
 1) Reduce resources affected to trunk development
 2) Reduce quality of trunk

The it must be in trunk first policy is the best way to mitigate
against that.

 To mitigate that, we decided that the group doing stable branch
 maintenance would be a separate group (i.e. *not* core developers), and
 we decided that whatever ends up in the stable branch must first land in
 the master branch.

Well, I recall it a little differently - that both the stable branch
maint group and the core team members would have +2 privileges on the
stable branch.

Maybe I just misread James here:

  https://lists.launchpad.net/openstack/msg04751.html

  only members of the maintainers team or core devs can +/-2

I also seem to have imagined the core teams being members of the stable
branch maint team:

  https://launchpad.net/~openstack-stable-maint/+members

But, whatever :)

We have a separate team with me, Chuck and Dave on it. Only one of us
can +2.

But wait! Vish +2ed a stable branch patch yesterday:

  https://review.openstack.org/328

James, help a poor confused soul out here, would you? :)

 So a change goes like this:
 * Change is proposed to trunk
 * Change is reviewed by core (is it appropriate, well-written, etc)
 * Change lands in trunk
 * Change is proposed to stable/diablo
 * Change is reviewed by stable team (is it relevant for a stable update,
 did it land in trunk first)
 * Change lands in stable/diablo
 
 This avoids the aforementioned risks, avoids duplicating review efforts
 (the two reviews actually check for different things), and keep the
 teams separate (so trunk reviews are not slowed down by stable reviews).
 
 Note that this does not prevent core developers that have an interest in
 stable/diablo from being in the two teams.
 
 Apparently people in core can easily mistake master for stable/diablo,
 and can also +2 stable/diablo changes. In order to avoid mistakes, I
 think +2 powers on stable/diablo should be limited to members of the
 stable maintenance team (who know their stable review policy).
 
 That should help avoid mistakes (like landing a fix in stable/diablo
 that never made it to master), while not preventing individual core devs
 from helping in stable reviews.

Right, that makes sense. Only folks that understand the stable branch
policy[1] should be allowed to +2 on the stable branch.

Basically, a stable branch reviewer should only +2 if:

  - It fixes a significant issue, seen, or potentially seen, by someone
during real life use

  - The fix, or equivalent, must be in master already

  - The fix was either a fairly trivial cherry-pick that looks 
equally correct for the stable branch, or that the fix has 
sufficient technical review (e.g. a +1 from another stable 
reviewer if it's fairly straightforward, or one or more +1s from 
folks on core it it's really gnarly)

  - If this reviewer proposed the patch originally, another stable
branch reviewer should have +1ed it 

All we need is an understanding of the policy and reasonable judgement,
it's not rocket science. I'd encourage folks to apply to the team for
membership after reviewing a few patches.

Cheers,
Mark.

[1] - http://wiki.openstack.org/StableBranch


___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


[Openstack] Hardware HA

2011-11-10 Thread Viacheslav Biriukov
Hi all.

What are the best practices for HA of the hardware compute-node, and
virtual machines.
After googling I found matahari, pacemaker-cloud, but nothing about
build-in fiches  openstack.

1) How do you create such environments?
2) Does it is right way to use pacemaker-cloud with openstack? Is it stable?

Tnx
-- 
Viacheslav Biriukov
BR
http://biriukov.com
___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Hardware HA

2011-11-10 Thread Pádraig Brady
On 11/10/2011 02:38 PM, Viacheslav Biriukov wrote:
 Hi all.
 
 What are the best practices for HA of the hardware compute-node, and virtual 
 machines.
 After googling I found matahari, pacemaker-cloud, but nothing about build-in 
 fiches  openstack. 
 
 1) How do you create such environments?  
 2) Does it is right way to use pacemaker-cloud with openstack? Is it stable?

About pacemaker-cloud, one can currently use it with openstack to some extent:
http://ahsalkeld.wordpress.com/a-mash-up-of-openstack-and-pacemaker-cloud/
However there are lots of manual steps and it's still in development.
It would be cool if you wanted to give that a shot and reported any issues or 
thoughts.

cheers,
Pádraig.

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Hardware HA

2011-11-10 Thread Soren Hansen
2011/11/10 Viacheslav Biriukov v.v.biriu...@gmail.com:
 Hi all.
 What are the best practices for HA of the hardware compute-node, and virtual
 machines.
 After googling I found matahari, pacemaker-cloud, but nothing about
 build-in fiches  openstack.
 1) How do you create such environments?
 2) Does it is right way to use pacemaker-cloud with openstack? Is it stable?

I'd avoid depending on anything like that altogether. Try to design
your application so that it doesn't depend on any one instance being
up. It'll work out better in the long run.

-- 
Soren Hansen        | http://linux2go.dk/
Ubuntu Developer    | http://www.ubuntu.com/
OpenStack Developer | http://www.openstack.org/

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Hardware HA

2011-11-10 Thread Armando Migliaccio
There is a blueprint that touches these aspects:

https://blueprints.launchpad.net/nova/+spec/guest-ha

This is tailored at use cases where you cannot redesign an existing app. 

The work is at the early stages, but you are more than welcome to join the 
effort!

Cheers,
Armando

 -Original Message-
 From: openstack-bounces+armando.migliaccio=eu.citrix@lists.launchpad.net
 [mailto:openstack-
 bounces+armando.migliaccio=eu.citrix@lists.launchpad.net] On Behalf Of
 Soren Hansen
 Sent: 10 November 2011 15:51
 To: Viacheslav Biriukov
 Cc: openstack@lists.launchpad.net
 Subject: Re: [Openstack] Hardware HA
 
 2011/11/10 Viacheslav Biriukov v.v.biriu...@gmail.com:
  Hi all.
  What are the best practices for HA of the hardware compute-node, and virtual
  machines.
  After googling I found matahari, pacemaker-cloud, but nothing about
  build-in fiches  openstack.
  1) How do you create such environments?
  2) Does it is right way to use pacemaker-cloud with openstack? Is it stable?
 
 I'd avoid depending on anything like that altogether. Try to design
 your application so that it doesn't depend on any one instance being
 up. It'll work out better in the long run.
 
 --
 Soren Hansen        | http://linux2go.dk/
 Ubuntu Developer    | http://www.ubuntu.com/
 OpenStack Developer | http://www.openstack.org/
 
 ___
 Mailing list: https://launchpad.net/~openstack
 Post to : openstack@lists.launchpad.net
 Unsubscribe : https://launchpad.net/~openstack
 More help   : https://help.launchpad.net/ListHelp
___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Stable branch reviews

2011-11-10 Thread James E. Blair
Mark McLoughlin mar...@redhat.com writes:

 To mitigate that, we decided that the group doing stable branch
 maintenance would be a separate group (i.e. *not* core developers), and
 we decided that whatever ends up in the stable branch must first land in
 the master branch.

 Well, I recall it a little differently - that both the stable branch
 maint group and the core team members would have +2 privileges on the
 stable branch.

If we discussed that explicitly, I'm afraid it didn't make it into my
notes from the summit.  It wouldn't surprise me if we left the answer to
that question in the bike shed.

 Maybe I just misread James here:

   https://lists.launchpad.net/openstack/msg04751.html

   only members of the maintainers team or core devs can +/-2

You read correctly; I took your proposal, translated it into slightly
more formal gerrit terms, and implemented that.  So currently gerrit is
configured so that _either_ core devs or the maintainers team can
approve changes to stable/ branches.

That can be changed, and maintainers can be given exclusive approval
authority over the stable/ branches.

 I also seem to have imagined the core teams being members of the stable
 branch maint team:

   https://launchpad.net/~openstack-stable-maint/+members

Right now it doesn't matter because in gerrit core devs have the same
authority as maintainers.  However, if we make maintainers exclusive
approvers, it might be better for individuals with interests in both to
be individually members of both since the stable branch has extra
behavioral rules.

 But wait! Vish +2ed a stable branch patch yesterday:

   https://review.openstack.org/328

 James, help a poor confused soul out here, would you? :)

 Right, that makes sense. Only folks that understand the stable branch
 policy[1] should be allowed to +2 on the stable branch.

 Basically, a stable branch reviewer should only +2 if:

   - It fixes a significant issue, seen, or potentially seen, by someone
 during real life use

   - The fix, or equivalent, must be in master already

   - The fix was either a fairly trivial cherry-pick that looks 
 equally correct for the stable branch, or that the fix has 
 sufficient technical review (e.g. a +1 from another stable 
 reviewer if it's fairly straightforward, or one or more +1s from 
 folks on core it it's really gnarly)

   - If this reviewer proposed the patch originally, another stable
 branch reviewer should have +1ed it 

 All we need is an understanding of the policy and reasonable judgement,
 it's not rocket science. I'd encourage folks to apply to the team for
 membership after reviewing a few patches.

It sounds like the best way to implement this policy is to give
openstack-stable-maint exclusive approval authority on stable branches,
and then make sure people understand those rules when adding them to
that team.  If that's the consensus, I can make the change.

-Jim

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Hardware HA

2011-11-10 Thread Viacheslav Biriukov
Hm
If we planning vm hosting we work on the other level. So if hw node fails
we need fast automatic migration to other node.

2011/11/10 Soren Hansen so...@linux2go.dk

 2011/11/10 Viacheslav Biriukov v.v.biriu...@gmail.com:
  Hi all.
  What are the best practices for HA of the hardware compute-node, and
 virtual
  machines.
  After googling I found matahari, pacemaker-cloud, but nothing about
  build-in fiches  openstack.
  1) How do you create such environments?
  2) Does it is right way to use pacemaker-cloud with openstack? Is it
 stable?

 I'd avoid depending on anything like that altogether. Try to design
 your application so that it doesn't depend on any one instance being
 up. It'll work out better in the long run.

 --
 Soren Hansen| http://linux2go.dk/
 Ubuntu Developer| http://www.ubuntu.com/
 OpenStack Developer | http://www.openstack.org/




-- 
Viacheslav Biriukov
BR
http://biriukov.com
___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Hardware HA

2011-11-10 Thread Razique Mahroua
Funny you bring that up today;I spent the day working on that. I've implemented Gluster FS on my openstack running installation and written a script along that.Here is the implementationnode1- 1 instance runningthe node 1 crashes (could be anything atm)the script detect the node is gone (to be defined : heartbeat, monitoring, etc...) and relaunch the instance in the specified node.I've tested it and successfully works. Just need to write the final touchesRegards
Razique Mahrouarazique.mahr...@gmail.com

Le 10 nov. 2011 à 17:15, Viacheslav Biriukov a écrit :HmIf we planning vm hosting we work on the other level. So if hw node fails we need fast automatic migration to other node.2011/11/10 Soren Hansen so...@linux2go.dk
2011/11/10 Viacheslav Biriukov v.v.biriu...@gmail.com:
 Hi all.
 What are the bestpractices for HA of the hardware compute-node, and virtual
 machines.
 After googling I found matahari, pacemaker-cloud, but nothing about
 build-infiches openstack.
 1) How do you create suchenvironments?
 2) Does it is right way to use pacemaker-cloud with openstack? Is it stable?

I'd avoid depending on anything like that altogether. Try to design
your application so that it doesn't depend on any one instance being
up. It'll work out better in the long run.

--
Soren Hansen| http://linux2go.dk/
Ubuntu Developer  | http://www.ubuntu.com/
OpenStack Developer | http://www.openstack.org/
-- Viacheslav BiriukovBRhttp://biriukov.com

___Mailing list: https://launchpad.net/~openstackPost to : openstack@lists.launchpad.netUnsubscribe : https://launchpad.net/~openstackMore help : https://help.launchpad.net/ListHelp___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Tutorials of how to install openstack swift into centos 6

2011-11-10 Thread Stefano Maffulli
Hello pf

thanks for sending this. Did you write this down on a wiki page
somewhere, too? I fear that in the mailing list archive this won't get
the visibility it deserves.

Cheers,
stef

On Wed, 2011-11-09 at 17:00 +1100, pf shineyear wrote:
 openstack swift install on centos 6
 
 1. proxy install
 
 1) check your python version must = 2.6
 
 2) yum install libvirt
 
 3) yum install memcached
 
 4) yum install xfsprogs
 
 5) yum install python-setuptools python-devel python-simplejson
 python-config
 
 6) easy_install webob
 
 7) easy_install eventlet
 
 8) install xattr-0.6.2.tar.gz, python setup.py build, python setup.py
 install
 
 9) install coverage-3.5.1.tar.gz, python setup.py build, python
 setup.py install
 
 10) wget http://www.openstack.org/projects/storage/latest-release/;
  python setup.py build
  python setup.py install
 
 11) wget
 https://github.com/downloads/gholt/swauth/swauth-lucid-build-1.0.2-1.tgz;
 python setup.py build
 python setup.py install
 
 12) mkdir /etc/swift
 
 13) yum install openssh-server
 
 14) yum install git-core
 
 15) vi /etc/swift/swift.conf
 
 [swift-hash]
 # random unique string that can never change (DO NOT LOSE)
 swift_hash_path_suffix = `od -t x8 -N 8 -A n /dev/random`
 
 
 16) goto /etc/swift/
 
 17) openssl req -new -x509 -nodes -out cert.crt -keyout cert.key
 
 18) service memcached restart, ps -aux | grep mem
 
 495  16954  0.0  0.1 330756   816 ?Ssl  18:19   0:00
 memcached -d -p 11211 -u memcached -m 64 -c 1024
 -P /var/run/memcached/memcached.pid
 
 19) easy_install netifaces
 
 20) vi /etc/swift/proxy-server.conf
 
 [DEFAULT]
 cert_file = /etc/swift/cert.crt
 key_file = /etc/swift/cert.key
 bind_port = 8080
 workers = 8
 user = swift
 log_facility = LOG_LOCAL0
 allow_account_management = true
 
 [pipeline:main]
 pipeline = healthcheck cache swauth proxy-server
 
 [app:proxy-server]
 use = egg:swift#proxy
 allow_account_management = true
 account_autocreate = true
 log_facility = LOG_LOCAL0
 log_headers = true
 log_level =DEBUG
 
 [filter:swauth]
 use = egg:swauth#swauth
 #use = egg:swift#swauth
 default_swift_cluster = local#https://10.38.10.127:8080/v1
 # Highly recommended to change this key to something else!
 super_admin_key = swauthkey
 log_facility = LOG_LOCAL1
 log_headers = true
 log_level =DEBUG
 allow_account_management = true
 
 [filter:healthcheck]
 use = egg:swift#healthcheck
 
 [filter:cache]
 use = egg:swift#memcache
 memcache_servers = 10.38.10.127:11211
 
 
 21) config /etc/rsyslog.conf
 
 local0.*/var/log/swift/proxy.log
 local1.*/var/log/swift/swauth.log
 
 
 
 
 
 21) build the ring, i have 3 node, 1 proxy
 
  swift-ring-builder account.builder create 18 3 1
  swift-ring-builder account.builder add z1-10.38.10.109:6002/sdb1 1
  swift-ring-builder account.builder add z2-10.38.10.119:6002/sdb1 1
  swift-ring-builder account.builder add z3-10.38.10.114:6002/sdb1 1
  swift-ring-builder object.builder rebalance
 
  swift-ring-builder account.builder rebalance
  swift-ring-builder object.builder create 18 3 1
  swift-ring-builder object.builder add z1-10.38.10.109:6000/sdb1 1
  swift-ring-builder object.builder add z2-10.38.10.119:6000/sdb1 1
  swift-ring-builder object.builder add z3-10.38.10.114:6000/sdb1 1
  swift-ring-builder object.builder rebalance
 
  swift-ring-builder container.builder create 18 3 1
  swift-ring-builder container.builder add z1-10.38.10.109:6001/sdb1 1
  swift-ring-builder container.builder add z2-10.38.10.119:6001/sdb1 1
  swift-ring-builder container.builder add z3-10.38.10.114:6001/sdb1 1
  swift-ring-builder container.builder rebalance
 
 
 22) easy_install configobj
 
 23) easy_install nose
 
 24) easy_install simplejson
 
 25) easy_install xattr
 
 26) easy_install eventlet
 
 27) easy_install greenlet
 
 28) easy_install pastedeploy
 
 29) groupadd swift
 
 30) useradd -g swift swift
 
 31) chown -R swift:swift /etc/swift/
 
 32) service rsyslog restart
 
 33) swift-init proxy start
 
 
 2. storage node install
 
 1) yum install python-setuptools python-devel python-simplejson
 python-configobj python-nose
 
 2) yum install openssh-server
 
 3) easy_install webob
 
 4) yum install curl gcc memcached sqlite xfsprogs
 
 
 
 5) easy_install eventlet
 
 
 6) wget
 http://pypi.python.org/packages/source/x/xattr/xattr-0.6.2.tar.gz#md5=5fc899150d03c082558455483fc0f89f;
 
 
  python setup.py build
 
  python setup.py install
 
 
 
 7)  wget
 http://pypi.python.org/packages/source/c/coverage/coverage-3.5.1.tar.gz#md5=410d4c8155a4dab222f2bc51212d4a24;
 
 
  python setup.py build
 
  python setup.py install
 
 
 8) yum install libvirt
 
 
 9) groupadd swift
 
 
 10) useradd -g swift swift
 
 
 11) mkdir -p /etc/swift
 
 
 12) chown -R swift:swift /etc/swift/
 
 
 13) cp swift.conf account.ring.gz container.ring.gz
 object.ring.gz /etc/swift/  (scp from proxy server)
 
 
 14) yum install xfsprogs
 
 
 15) wget 

Re: [Openstack] Host Aggregates ...

2011-11-10 Thread Vishvananda Ishaya
The main thing that the idea of host-aggregates provides is the ability to 
specify metadata at a group of hosts level.  If you put 10 hosts into a single 
aggregate, you can specify characteristics for those hosts as a group (i.e. 
which san they are backing files onto, whether they support live migration, 
etc.)  It is also the way that xen (and esx) model things internally, so if we 
don't do it we have to map our custom tags  to resource pools in the hypervisor 
anyway. Finally, it also simplifies some scheduler logic.  For example if you 
are trying to find a valid host to migrate instances onto, you can just ask for 
hosts in the same group and it is the cloud administrator's responsibility to 
make sure that all of the systems in that group have the required 
functionality.  These groups could be implemented with tags, but I think 
conceptually the idea of a tag implies a more tenuous relationship than 
aggregate group.

I look at it as a grouping of hosts that is smaller than a zone (say a 
cluster).  Large enough for shared metadata but small enough so that splitting 
the db and api would be a pain.

Vish

On Nov 10, 2011, at 5:01 AM, Sandy Walsh wrote:

 Ok, that helps ... now I see the abstraction your going for (a new layer 
 under availability zones).
 
 Personally I prefer a tagging approach to a modeled hierarchy. It was 
 something we debated at great length with Zones. In this case, the tag 
 would be in the capabilities assigned to the host.
 
 I think both availability zones (and host aggregates) should be modeled using 
 tags/capabilities without having to explicitly model it as a tree or in the 
 db ... which is how I see this evolving. At the scheduler level we should be 
 able to make decisions using simple tag collections.
 
 WestCoast, HasGPU, GeneratorBackup, PriorityNetwork
 
 Are we saying the same thing?
 
 Are there use cases that this approach couldn't handle?
 
 -S
 
 
 From: Armando Migliaccio [armando.migliac...@eu.citrix.com]
 Sent: Thursday, November 10, 2011 8:50 AM
 To: Sandy Walsh
 Cc: openstack@lists.launchpad.net
 Subject: RE: Host Aggregates ...
 
 Hi Sandy,
 
 Thanks for taking the time to read this.
 
 My understanding is that a typical Nova deployment would span across multiple 
 zones, that zones may have subzones, and that child zones will have a number 
 of availability zones in them; please do correct me if I am wrong :)
 
 That stated, it was assumed that an aggregate will be a grouping of servers 
 within an availability zone (hence the introduction of the extra concept), 
 and would be used to manage hypervisor pools when and if required. This 
 introduces benefits like VM live migration, VM HA and zero-downtime host 
 upgrades. The introduction of hypervisor pools is just the easy way to get 
 these benefits in the short term.
 
 Going back to your point, it is possible to match host-aggregates with 
 single-zone that uses capabilities on the implementation level (assumed 
 that it is okay to be unable to represent aggregates as children of 
 availability zones). Nevertheless, I still see zones and aggregates as being 
 different on the conceptual level.
 
 What is your view if we went with the approach of implementing an aggregate 
 as a special single-zone that uses capabilities? Would there be a risk of 
 tangling the zone management API a bit?
 
 Thanks for feedback!
 
 Cheers,
 Armando
 
 -Original Message-
 From: Sandy Walsh [mailto:sandy.wa...@rackspace.com]
 Sent: 09 November 2011 21:10
 To: Armando Migliaccio
 Cc: openstack@lists.launchpad.net
 Subject: Host Aggregates ...
 
 Hi Armando,
 
 I finally got around to reading
 https://blueprints.launchpad.net/nova/+spec/host-aggregates.
 
 Perhaps you could elaborate a little on how this differs from host
 capabilities (key-value pairs associated with a service) that the scheduler
 can use when making decisions?
 
 The distributed scheduler doesn't need zones to operate, but will use them if
 available. Would host-aggregates simply be a single-zone that uses
 capabilities?
 
 Cheers,
 Sandy
 
 ___
 Mailing list: https://launchpad.net/~openstack
 Post to : openstack@lists.launchpad.net
 Unsubscribe : https://launchpad.net/~openstack
 More help   : https://help.launchpad.net/ListHelp


___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Stable branch reviews

2011-11-10 Thread Vishvananda Ishaya

On Nov 10, 2011, at 6:22 AM, Mark McLoughlin wrote:

 But wait! Vish +2ed a stable branch patch yesterday:
 
  https://review.openstack.org/328


I don't mind losing my powers over stable/diablo.

On a related note, is there a way we can change the color scheme in gerrit (to 
red??) for stable branches?  I think there are a number of cases with core 
members reviewing stable/diablo patches thinking they were for trunk.

Vish

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Hardware HA

2011-11-10 Thread Ilya Alekseyev
Hi Armando!

It is very interesting feature. Are you already have specs for this
blueprints or may be etherpad?

Regards,
Ilya

10 ноября 2011 г. 20:16 пользователь Armando Migliaccio 
armando.migliac...@eu.citrix.com написал:

 There is a blueprint that touches these aspects:

 https://blueprints.launchpad.net/nova/+spec/guest-ha

 This is tailored at use cases where you cannot redesign an existing app.

 The work is at the early stages, but you are more than welcome to join the
 effort!

 Cheers,
 Armando

  -Original Message-
  From: openstack-bounces+armando.migliaccio=
 eu.citrix@lists.launchpad.net
  [mailto:openstack-
  bounces+armando.migliaccio=eu.citrix@lists.launchpad.net] On Behalf
 Of
  Soren Hansen
  Sent: 10 November 2011 15:51
  To: Viacheslav Biriukov
  Cc: openstack@lists.launchpad.net
  Subject: Re: [Openstack] Hardware HA
 
  2011/11/10 Viacheslav Biriukov v.v.biriu...@gmail.com:
   Hi all.
   What are the best practices for HA of the hardware compute-node, and
 virtual
   machines.
   After googling I found matahari, pacemaker-cloud, but nothing about
   build-in fiches  openstack.
   1) How do you create such environments?
   2) Does it is right way to use pacemaker-cloud with openstack? Is it
 stable?
 
  I'd avoid depending on anything like that altogether. Try to design
  your application so that it doesn't depend on any one instance being
  up. It'll work out better in the long run.
 
  --
  Soren Hansen| http://linux2go.dk/
  Ubuntu Developer| http://www.ubuntu.com/
  OpenStack Developer | http://www.openstack.org/
 
  ___
  Mailing list: https://launchpad.net/~openstack
  Post to : openstack@lists.launchpad.net
  Unsubscribe : https://launchpad.net/~openstack
  More help   : https://help.launchpad.net/ListHelp
 ___
 Mailing list: https://launchpad.net/~openstack
 Post to : openstack@lists.launchpad.net
 Unsubscribe : https://launchpad.net/~openstack
 More help   : https://help.launchpad.net/ListHelp

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Stable branch reviews

2011-11-10 Thread Dave Walker
On Thu, Nov 10, 2011 at 08:02:23AM -0800, James E. Blair wrote:
SNIP
 
  But wait! Vish +2ed a stable branch patch yesterday:
 
https://review.openstack.org/328
 
  James, help a poor confused soul out here, would you? :)
 
  Right, that makes sense. Only folks that understand the stable branch
  policy[1] should be allowed to +2 on the stable branch.
 
  Basically, a stable branch reviewer should only +2 if:
 
- It fixes a significant issue, seen, or potentially seen, by someone
  during real life use
 
- The fix, or equivalent, must be in master already
 
- The fix was either a fairly trivial cherry-pick that looks 
  equally correct for the stable branch, or that the fix has 
  sufficient technical review (e.g. a +1 from another stable 
  reviewer if it's fairly straightforward, or one or more +1s from 
  folks on core it it's really gnarly)
 
- If this reviewer proposed the patch originally, another stable
  branch reviewer should have +1ed it 
 
  All we need is an understanding of the policy and reasonable judgement,
  it's not rocket science. I'd encourage folks to apply to the team for
  membership after reviewing a few patches.
 
 It sounds like the best way to implement this policy is to give
 openstack-stable-maint exclusive approval authority on stable branches,
 and then make sure people understand those rules when adding them to
 that team.  If that's the consensus, I can make the change.

Hi,

Thanks for helping to add clarification to this.  From our
perspective, I have confidence that ~*-core members know the
difference between trunk and stable policy.  Therefore for the short
term, it makes sense to have more eyes - especially those which are
likely to have good knowledge of the internals.

Therefore, I am happy for ~*-core to still have +2 access; especially
if it helps seed the maint team.

Going forward, it probably will make sense to have a distinction, but
I feel it might be quite early for that to be a requirement.

Thanks.

Kind Regards,
Dave Walker


signature.asc
Description: Digital signature
___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Bug fixes and test cases submitted against stable/diablo

2011-11-10 Thread Jay Pipes
On Wed, 2011-11-09 at 14:57 +0100, Thierry Carrez wrote:
 Soren Hansen wrote:
  2011/11/9 Nachi Ueno ueno.na...@nttdata-agilenet.com:
  I understand your point. Stop QAing stable/diablo and focus on Essex.
  
  Oh, no no. That's not the point. I'm thrilled to have you work on
  QAing Diablo. The only issue is that the fixes you come up with should
  be pushed to Essex first. There are two reasons for this:
  
   * If we don't push the fixes to Essex, the problems will still be
  present in Essex and every release after that.
  
   * Having them in Essex lets us try them out, vet them and validate
  them more thoroughly before we let them into the stable branch. When a
  patch lands in the stable branch it has to be well tested already
  (unless of course Essex has deviated too much, in which case we'll
  have to accept the risk of getting it into Diablo directly).
 
 +1
 
 You should submit patches to master and then backport them to
 stable/diablo, rather than proposing them for stable/diablo directly.
 That ensures your work benefits both branches: making diablo better
 without making essex worse than diablo.
 
 If that's just too much work, maybe you should raise the issue at the
 next QA meeting to try to get some outside help ?

At the QA meeting yesterday, I offered my help to Nati. I will handle
proposing his patches to Essex up to a future date where Nati and his
team will switch to code against Essex, not Diablo/stable and propose
first to master, then others will backport to diablo/stable.

Nati and I will decide on that future date for his team to switch their
focus to Essex trunk and not have to have someone manually
forward-port these patches to trunk.

Cheers,
-jay


___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Which nova scheduler for different hardware sizes?

2011-11-10 Thread Christian Wittwer
Thanks a lot for your answers. Unfortunately I can't follow the trunk
and I have to use the Diablo release. Is it possible to backport that
new scheduler to Diablo?
Anyway I gave the least cost scheduler a try, it loads but never
schedules a vm correctly.

--compute_scheduler_driver=nova.scheduler.least_cost.LeastCostScheduler

But then log fires that message when I try to launch an instance

2011-11-10 19:31:15,588 DEBUG nova.scheduler.least_cost [-] Weighted
Costs = [] from (pid=1697) weigh_hosts
/usr/lib/pymodules/python2.7/nova/scheduler/least_cost.py:170
2011-11-10 19:31:15,591 ERROR nova.rpc [-] Exception during message handling
(nova.rpc): TRACE: Traceback (most recent call last):
(nova.rpc): TRACE:   File
/usr/lib/pymodules/python2.7/nova/rpc/impl_kombu.py, line 620, in
_process_data
(nova.rpc): TRACE: rval = node_func(context=ctxt, **node_args)
(nova.rpc): TRACE:   File
/usr/lib/pymodules/python2.7/nova/scheduler/manager.py, line 103, in
_schedule
(nova.rpc): TRACE: host = real_meth(*args, **kwargs)
(nova.rpc): TRACE:   File
/usr/lib/pymodules/python2.7/nova/scheduler/abstract_scheduler.py,
line 231, in schedule_run_instance
(nova.rpc): TRACE: raise driver.NoValidHost(_('No hosts were available'))
(nova.rpc): TRACE: NoValidHost: No hosts were available
(nova.rpc): TRACE:

My compute hosts are up and running, and there are no other instances running.

root@unic-dev-os-controller:~# nova-manage service list
Binary   Host Zone
Status State Updated_At
nova-compute unic-dev-os-controller   nova
enabled:-)   2011-11-10 18:35:06
nova-scheduler   unic-dev-os-controller   nova
enabled:-)   2011-11-10 18:35:06
nova-network unic-dev-os-controller   nova
enabled:-)   2011-11-10 18:35:06
nova-compute unic-dev-os-compute1 nova
enabled:-)   2011-11-10 18:35:06

Any ideas why the scheduler does not find a valid host?

Christian

2011/11/1 Lorin Hochstein lo...@isi.edu:
 Christian:

 Sandy's branch just landed in the repository. You should be able to use the 
 distributed scheduler with the least cost functionality by specifying the 
 following flag in nova.conf for the nova-scheduler service:

 --compute_scheduler_driver=nova.scheduler.distributed_scheduler.DistributedScheduler

 By default, this uses the 
 nova.scheduler.least_cost.compute_fill_first_cost_fn weighting function.

 Note, however, that this function will favor scheduling instances to nodes 
 that have the smallest amount of RAM available that can still fit the 
 instance. If you're looking for the opposite effect (deploy to the node that 
 has the most amount of RAM free), then you'll have to write your own cost 
 function.  One way would be to add the following method to least_cost.py:


 def compute_least_loaded_cost_fn(host_info):
    return -compute_fill_first_cost_fn(host_info)


 Then add the following flag to your nova.conf

 --least_cost_functions=nova.scheduler.least_cost.compute_least_loaded_cost_fn


 Lorin
 --
 Lorin Hochstein, Computer Scientist
 USC Information Sciences Institute
 703.812.3710
 http://www.east.isi.edu/~lorin




 On Nov 1, 2011, at 11:37 AM, Sandy Walsh wrote:

 I'm hoping to land this branch asap.
 https://review.openstack.org/#change,1192

 It replaces all the kind of alike schedulers with a single 
 DistributedScheduler.

 -S

 
 From: openstack-bounces+sandy.walsh=rackspace@lists.launchpad.net 
 [openstack-bounces+sandy.walsh=rackspace@lists.launchpad.net] on behalf 
 of Christian Wittwer [wittwe...@gmail.com]
 Sent: Tuesday, November 01, 2011 5:38 AM
 To: Lorin Hochstein
 Cc: openstack@lists.launchpad.net
 Subject: Re: [Openstack] Which nova scheduler for different hardware sizes?

 Lorin,
 Thanks for your reply. Well the least cost scheduler with these cost
 functions looks interesting.
 Unfortunately there is not much documenation about it. Can somebody
 give me an example how to switch to that scheduler using the memory
 cost function which already exist?

 Cheers,
 Christian

 2011/10/24 Lorin Hochstein lo...@isi.edu:
 Christian:
 You could use the least cost scheduler, but I think you'd have to write your
 own cost function to take into account the different number of cores.
 Looking at the source, the only cost function it comes with only takes into
 account the amount of memory that's free, not loading in terms of total
 physical cores and allocated virtual cores. (We use a custom scheduler at
 our site, so I don't have any firsthand experience with the least-cost
 scheduler).
 Lorin
 --
 Lorin Hochstein, Computer Scientist
 USC Information Sciences Institute
 703.812.3710
 http://www.east.isi.edu/~lorin



 On Oct 22, 2011, at 3:17 AM, Christian Wittwer wrote:

 I'm planning to build a openstack nova installation with older
 hardware. These servers obviously doesn't have the same hardware
 configuration like memory and 

Re: [Openstack] Hardware HA

2011-11-10 Thread Soren Hansen
2011/11/10 Viacheslav Biriukov v.v.biriu...@gmail.com:
 Hm
 If we planning vm hosting we work on the other level. So if hw node fails we
 need fast automatic migration to other node.

That's the whole point. For most interesting applications, fast
automatic migration isn't anywhere near fast enough. Don't try to
avoid failure. Expect it and design around it.

-- 
Soren Hansen        | http://linux2go.dk/
Ubuntu Developer    | http://www.ubuntu.com/
OpenStack Developer | http://www.openstack.org/

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


[Openstack] Essex-1 milestone available for Keystone, Glance, Nova and Horizon

2011-11-10 Thread Thierry Carrez
Hi everyone,

It's my great pleasure to announce the immediate availability for
Keystone, Glance, Nova and Horizon of the first milestone of the Essex
development cycle, called essex-1.

This milestone picks up all development made on trunk since Diablo
release branches were cut in early September. You can see the full list
of new features and fixed bugs, as well as tarball downloads, at:

https://launchpad.net/keystone/essex/essex-1
https://launchpad.net/glance/essex/essex-1
https://launchpad.net/nova/essex/essex-1
https://launchpad.net/horizon/essex/essex-1

Note that you can also test the Glance  Nova milestones on Ubuntu by
enabling the following PPAs:
 ppa:nova-core/milestone
 ppa:glance-core/milestone

The next milestone, essex-2, is scheduled for release on December 15th.
Enjoy !

-- 
Thierry Carrez (ttx)
Release Manager, OpenStack

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Bug fixes and test cases submitted against stable/diablo

2011-11-10 Thread Nachi Ueno
Hi folks

Thank you for your help  Mark and Jay and Reviewers
I removed all review request for diablo/stable from Gerrit.
And, We will follow community policy.

Current our test code and bug fix is based on stable/diablo.
For each branch. forward-porting is needed.

12 bug patch branch is in progress ( they are almost done)
34 bug patch branch is on github(*)
30 test code branch is on github.
(*)https://github.com/ntt-pf-lab/nova/branches

From next work alter these branches, We will follow the policy (Essex first).

However, for now, we have not enough man-power.
So please help us.
I wrote a script which shows bug description and conflict files and
merge command.
(See https://gist.github.com/1355816)

Each branch is linked to bug report.

If you guys help forward-porting work, would you please assign the bug
for yourself.
(Thanks Jay!)

Naming rule in our repository is like this.
https://github.com/ntt-pf-lab/nova/tree/openstack-qa-nova-(bugID)

For now, There are bugs which is not fixed yet, so test code fails.
So I think we should start from bug fix.

Cheers
Nati

2011/11/10 Jay Pipes jaypi...@gmail.com:
 On Wed, 2011-11-09 at 14:57 +0100, Thierry Carrez wrote:
 Soren Hansen wrote:
  2011/11/9 Nachi Ueno ueno.na...@nttdata-agilenet.com:
  I understand your point. Stop QAing stable/diablo and focus on Essex.
 
  Oh, no no. That's not the point. I'm thrilled to have you work on
  QAing Diablo. The only issue is that the fixes you come up with should
  be pushed to Essex first. There are two reasons for this:
 
   * If we don't push the fixes to Essex, the problems will still be
  present in Essex and every release after that.
 
   * Having them in Essex lets us try them out, vet them and validate
  them more thoroughly before we let them into the stable branch. When a
  patch lands in the stable branch it has to be well tested already
  (unless of course Essex has deviated too much, in which case we'll
  have to accept the risk of getting it into Diablo directly).

 +1

 You should submit patches to master and then backport them to
 stable/diablo, rather than proposing them for stable/diablo directly.
 That ensures your work benefits both branches: making diablo better
 without making essex worse than diablo.

 If that's just too much work, maybe you should raise the issue at the
 next QA meeting to try to get some outside help ?

 At the QA meeting yesterday, I offered my help to Nati. I will handle
 proposing his patches to Essex up to a future date where Nati and his
 team will switch to code against Essex, not Diablo/stable and propose
 first to master, then others will backport to diablo/stable.

 Nati and I will decide on that future date for his team to switch their
 focus to Essex trunk and not have to have someone manually
 forward-port these patches to trunk.

 Cheers,
 -jay


 ___
 Mailing list: https://launchpad.net/~openstack
 Post to     : openstack@lists.launchpad.net
 Unsubscribe : https://launchpad.net/~openstack
 More help   : https://help.launchpad.net/ListHelp


___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Hardware HA

2011-11-10 Thread Soren Hansen
2011/11/10 Ryan Lane rl...@wikimedia.org:
 That's the whole point. For most interesting applications, fast
 automatic migration isn't anywhere near fast enough. Don't try to
 avoid failure. Expect it and design around it.
 This assumes all application designers are doing this. Most web
 applications do this fairly well, but most enterprise applications do
 this very poorly.

I know. That's what makes them a poor fit for the cloud.

 Hardware HA is useful for more than just poorly designed applications
 though. I have a cloud instance that runs my personal website. I don't
 want to pay for two (or more, realistically) instances just to ensure
 that if my host dies that my site will continue to run. My provider
 should automatically detect the hardware failure and re-launch my
 instance on another piece of hardware; it should also notify me that
 it happened, but that's a different story ;).

I'm not sure I count that as High Availability. It's more like
Eventual Availability. :)

-- 
Soren Hansen        | http://linux2go.dk/
Ubuntu Developer    | http://www.ubuntu.com/
OpenStack Developer | http://www.openstack.org/

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Draft API specifications

2011-11-10 Thread Jay Pipes
On Thu, 2011-11-10 at 13:22 -0600, Anne Gentle wrote:
 Thanks, Jay. I usually try to be more careful with the API names, so
 thanks for clarifying.

Sorry if I sounded patronizing there.. didn't mean to be.

 I think the landing page containing Draft API docs looks something
 like the attached mockup, let me know your feedback.

Looks great to me.

 Jay, you're the only PTL using Google Docs for feedback, others have
 used the doc comments system, Disqus. I can set up doc comments feeds
 specifically for RFC periods on particular specs, though your Google
 Docs approach works fine also. It would be nice to standardize but I
 also like that Google docs lets you click Resolved on a comment. 

Yeah, I like the fact that you can comment on a specific block of the
proposal as well, instead of all in one list of comments at the bottom
of a page. Also, yes, the Resolved comment feature is very nice for this
kind of iterative feedback.

That said, however, if it is your wish to do all draft API proposals
using a single system, so that it is easier for you to maintain, I will
bend to your will :)

 I have the ability to make DRAFT in big red letters on the output. I
 could also put RFC as a watermark on the page during the RFC period in
 addition to the DRAFT in each page. 
 
 I mostly want to ensure easy access so that the draft APIs get plenty
 of comments and reviews.

Yep, me too. :)

-jay


___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Hardware HA

2011-11-10 Thread Ryan Lane
 That's the whole point. For most interesting applications, fast
 automatic migration isn't anywhere near fast enough. Don't try to
 avoid failure. Expect it and design around it.


This assumes all application designers are doing this. Most web
applications do this fairly well, but most enterprise applications do
this very poorly.

Hardware HA is useful for more than just poorly designed applications
though. I have a cloud instance that runs my personal website. I don't
want to pay for two (or more, realistically) instances just to ensure
that if my host dies that my site will continue to run. My provider
should automatically detect the hardware failure and re-launch my
instance on another piece of hardware; it should also notify me that
it happened, but that's a different story ;).

- Ryan Lane

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Hardware HA

2011-11-10 Thread Ryan Lane
 I know. That's what makes them a poor fit for the cloud.


Meh. Private clouds will still use applications like this. I think
the cloud is great for cloud providers, but why limit nova's
usefulness to just cloud providers?

The cloud way of doing things pushes the responsibility of keeping
applications alive to the client. There's a lot of clients that don't
have this level of sophistication.

 Hardware HA is useful for more than just poorly designed applications
 though. I have a cloud instance that runs my personal website. I don't
 want to pay for two (or more, realistically) instances just to ensure
 that if my host dies that my site will continue to run. My provider
 should automatically detect the hardware failure and re-launch my
 instance on another piece of hardware; it should also notify me that
 it happened, but that's a different story ;).

 I'm not sure I count that as High Availability. It's more like
 Eventual Availability. :)


So, this is one HA mode for VMware. There is also a newer HA mode that
is much more expensive (from the resources perspective) that keeps a
shadow copy of a virtual machine on another piece of hardware, and if
the primary instance's hardware dies, it automatically switches over
to the shadow copy.

Both modes are really useful. There's a huge level of automation
needed for doing things the cloud way that is completely
unnecessary. I don't want to have to monitor my instances to see if
one died due to a hardware failure, then start new ones, then pool
them, then depool the dead ones. I want my provider to handle hardware
deaths for me. If I have 200 web servers instances, and 40 of them die
because they are on nodes that die, I want them to restart somewhere
else. It removes all the bullshit automation I'd need to do otherwise.

- Ryan Lane

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


[Openstack] Problem with Gerrit Workflow (probably with ssh public key)

2011-11-10 Thread Dugger, Donald D
I'm having problems trying to follow the steps in Gerrit Workflow Quick 
Reference (wiki.openstack.org/GerritWorkflow), in fact I'm failing on the first 
step.  When I try and get the list of available projects I'm getting the 
failure:

Permission denied (publickey).

I've uploaded my public key and this key does work when I `ssh' into other 
machines so I think the key itself is correct but `review.openstack.org' seems 
to be unhappy with it.  Note that the machine I'm using is behind a NAT server 
but that hasn't caused problems with other `ssh' connections.  I've pasted the 
output from a `ssh -v ...' commenad but it doesn't really show anything.

If anyone has any ideas on what I'm doing wrong I would greatly appreciate it.

--
Don Dugger
Censeo Toto nos in Kansa esse decisse. - D. Gale
Ph: 303/443-3786

--- cut here for log ---
OpenSSH_5.1p1 Debian-5, OpenSSL 0.9.8g 19 Oct 2007
debug1: Reading configuration data /home/n0ano/.ssh/config
debug1: Reading configuration data /etc/ssh/ssh_config
debug1: Applying options for *
debug1: Connecting to review.openstack.org [173.203.103.119] port 29418.
debug1: Connection established.
debug1: identity file /home/n0ano/.ssh/identity type -1
debug1: identity file /home/n0ano/.ssh/id_rsa type 1
debug1: Checking blacklist file /usr/share/ssh/blacklist.RSA-1024
debug1: Checking blacklist file /etc/ssh/blacklist.RSA-1024
debug1: identity file /home/n0ano/.ssh/id_dsa type 2
debug1: Checking blacklist file /usr/share/ssh/blacklist.DSA-1024
debug1: Checking blacklist file /etc/ssh/blacklist.DSA-1024
debug1: Remote protocol version 2.0, remote software version 
GerritCodeReview_2.2.1-4-g4b9d4ed (SSHD-CORE-0.5.1-R1095809)
debug1: no match: GerritCodeReview_2.2.1-4-g4b9d4ed (SSHD-CORE-0.5.1-R1095809)
debug1: Enabling compatibility mode for protocol 2.0
debug1: Local version string SSH-2.0-OpenSSH_5.1p1 Debian-5
debug1: SSH2_MSG_KEXINIT sent
debug1: SSH2_MSG_KEXINIT received
debug1: kex: server-client aes128-cbc hmac-md5 none
debug1: kex: client-server aes128-cbc hmac-md5 none
debug1: sending SSH2_MSG_KEXDH_INIT
debug1: expecting SSH2_MSG_KEXDH_REPLY
debug1: Host '[review.openstack.org]:29418' is known and matches the RSA host 
key.
debug1: Found key in /home/n0ano/.ssh/known_hosts:17
debug1: ssh_rsa_verify: signature correct
debug1: SSH2_MSG_NEWKEYS sent
debug1: expecting SSH2_MSG_NEWKEYS
debug1: SSH2_MSG_NEWKEYS received
debug1: SSH2_MSG_SERVICE_REQUEST sent
debug1: SSH2_MSG_SERVICE_ACCEPT received
debug1: Authentications that can continue: publickey
debug1: Next authentication method: publickey
debug1: Trying private key: /home/n0ano/.ssh/identity
debug1: Offering public key: /home/n0ano/.ssh/id_rsa
debug1: Authentications that can continue: publickey
debug1: Offering public key: /home/n0ano/.ssh/id_dsa
debug1: Authentications that can continue: publickey
debug1: No more authentication methods to try.
Permission denied (publickey).



___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Problem with Gerrit Workflow (probably with ssh public key)

2011-11-10 Thread Dugger, Donald D
Nevermind.  I'm not sure how I did it but by some magic incantation of 
resetting login name/public key it's now working.

Sorry for the noise.

--
Don Dugger
Censeo Toto nos in Kansa esse decisse. - D. Gale
Ph: 303/443-3786


-Original Message-
From: Dugger, Donald D 
Sent: Thursday, November 10, 2011 6:14 PM
To: openstack@lists.launchpad.net
Subject: Problem with Gerrit Workflow (probably with ssh public key)

I'm having problems trying to follow the steps in Gerrit Workflow Quick 
Reference (wiki.openstack.org/GerritWorkflow), in fact I'm failing on the first 
step.  When I try and get the list of available projects I'm getting the 
failure:

Permission denied (publickey).

I've uploaded my public key and this key does work when I `ssh' into other 
machines so I think the key itself is correct but `review.openstack.org' seems 
to be unhappy with it.  Note that the machine I'm using is behind a NAT server 
but that hasn't caused problems with other `ssh' connections.  I've pasted the 
output from a `ssh -v ...' commenad but it doesn't really show anything.

If anyone has any ideas on what I'm doing wrong I would greatly appreciate it.

--
Don Dugger
Censeo Toto nos in Kansa esse decisse. - D. Gale
Ph: 303/443-3786

--- cut here for log ---
OpenSSH_5.1p1 Debian-5, OpenSSL 0.9.8g 19 Oct 2007
debug1: Reading configuration data /home/n0ano/.ssh/config
debug1: Reading configuration data /etc/ssh/ssh_config
debug1: Applying options for *
debug1: Connecting to review.openstack.org [173.203.103.119] port 29418.
debug1: Connection established.
debug1: identity file /home/n0ano/.ssh/identity type -1
debug1: identity file /home/n0ano/.ssh/id_rsa type 1
debug1: Checking blacklist file /usr/share/ssh/blacklist.RSA-1024
debug1: Checking blacklist file /etc/ssh/blacklist.RSA-1024
debug1: identity file /home/n0ano/.ssh/id_dsa type 2
debug1: Checking blacklist file /usr/share/ssh/blacklist.DSA-1024
debug1: Checking blacklist file /etc/ssh/blacklist.DSA-1024
debug1: Remote protocol version 2.0, remote software version 
GerritCodeReview_2.2.1-4-g4b9d4ed (SSHD-CORE-0.5.1-R1095809)
debug1: no match: GerritCodeReview_2.2.1-4-g4b9d4ed (SSHD-CORE-0.5.1-R1095809)
debug1: Enabling compatibility mode for protocol 2.0
debug1: Local version string SSH-2.0-OpenSSH_5.1p1 Debian-5
debug1: SSH2_MSG_KEXINIT sent
debug1: SSH2_MSG_KEXINIT received
debug1: kex: server-client aes128-cbc hmac-md5 none
debug1: kex: client-server aes128-cbc hmac-md5 none
debug1: sending SSH2_MSG_KEXDH_INIT
debug1: expecting SSH2_MSG_KEXDH_REPLY
debug1: Host '[review.openstack.org]:29418' is known and matches the RSA host 
key.
debug1: Found key in /home/n0ano/.ssh/known_hosts:17
debug1: ssh_rsa_verify: signature correct
debug1: SSH2_MSG_NEWKEYS sent
debug1: expecting SSH2_MSG_NEWKEYS
debug1: SSH2_MSG_NEWKEYS received
debug1: SSH2_MSG_SERVICE_REQUEST sent
debug1: SSH2_MSG_SERVICE_ACCEPT received
debug1: Authentications that can continue: publickey
debug1: Next authentication method: publickey
debug1: Trying private key: /home/n0ano/.ssh/identity
debug1: Offering public key: /home/n0ano/.ssh/id_rsa
debug1: Authentications that can continue: publickey
debug1: Offering public key: /home/n0ano/.ssh/id_dsa
debug1: Authentications that can continue: publickey
debug1: No more authentication methods to try.
Permission denied (publickey).



___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Stable branch reviews

2011-11-10 Thread Mark McLoughlin
On Thu, 2011-11-10 at 09:02 -0800, Vishvananda Ishaya wrote:
 On Nov 10, 2011, at 6:22 AM, Mark McLoughlin wrote:
 
  But wait! Vish +2ed a stable branch patch yesterday:
  
   https://review.openstack.org/328
 
 
 I don't mind losing my powers over stable/diablo.
 
 On a related note, is there a way we can change the color scheme in
 gerrit (to red??) for stable branches?  I think there are a number of
 cases with core members reviewing stable/diablo patches thinking they
 were for trunk.

No doubt gerrit could be improved, but what works for me is to just look
at reviews for the master branch with e.g.

https://review.openstack.org/#q,status:open+project:openstack/nova+branch:master,n,z

Gerrit's query syntax is actually quite useful, e.g.

https://review.openstack.org/#q,status:open+project:openstack/nova+branch:master+owner:vish,n,z

Docs on it here:

http://review.coreboot.org/Documentation/user-search.html

Cheers,
Mark.


___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Stable branch reviews

2011-11-10 Thread Mark McLoughlin
On Thu, 2011-11-10 at 08:02 -0800, James E. Blair wrote:
 Mark McLoughlin mar...@redhat.com writes:
  Only folks that understand the stable branch policy[1] should be 
  allowed to +2 on the stable branch.
 
  Basically, a stable branch reviewer should only +2 if:
 
- It fixes a significant issue, seen, or potentially seen, by someone
  during real life use
 
- The fix, or equivalent, must be in master already
 
- The fix was either a fairly trivial cherry-pick that looks 
  equally correct for the stable branch, or that the fix has 
  sufficient technical review (e.g. a +1 from another stable 
  reviewer if it's fairly straightforward, or one or more +1s from 
  folks on core it it's really gnarly)
 
- If this reviewer proposed the patch originally, another stable
  branch reviewer should have +1ed it 
 
  All we need is an understanding of the policy and reasonable judgement,
  it's not rocket science. I'd encourage folks to apply to the team for
  membership after reviewing a few patches.
 
 It sounds like the best way to implement this policy is to give
 openstack-stable-maint exclusive approval authority on stable branches,
 and then make sure people understand those rules when adding them to
 that team.  If that's the consensus, I can make the change.

Yes, that's what Thierry initially suggested and I'm persuaded now
too :)

Cheers,
Mark.


___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Stable branch reviews

2011-11-10 Thread Mark McLoughlin
Hi Dave,

On Thu, 2011-11-10 at 17:33 +, Dave Walker wrote:
 On Thu, Nov 10, 2011 at 08:02:23AM -0800, James E. Blair wrote:
 SNIP
  
   But wait! Vish +2ed a stable branch patch yesterday:
  
 https://review.openstack.org/328
  
   James, help a poor confused soul out here, would you? :)
  
   Right, that makes sense. Only folks that understand the stable branch
   policy[1] should be allowed to +2 on the stable branch.
  
   Basically, a stable branch reviewer should only +2 if:
  
 - It fixes a significant issue, seen, or potentially seen, by someone
   during real life use
  
 - The fix, or equivalent, must be in master already
  
 - The fix was either a fairly trivial cherry-pick that looks 
   equally correct for the stable branch, or that the fix has 
   sufficient technical review (e.g. a +1 from another stable 
   reviewer if it's fairly straightforward, or one or more +1s from 
   folks on core it it's really gnarly)
  
 - If this reviewer proposed the patch originally, another stable
   branch reviewer should have +1ed it 
  
   All we need is an understanding of the policy and reasonable judgement,
   it's not rocket science. I'd encourage folks to apply to the team for
   membership after reviewing a few patches.
  
  It sounds like the best way to implement this policy is to give
  openstack-stable-maint exclusive approval authority on stable branches,
  and then make sure people understand those rules when adding them to
  that team.  If that's the consensus, I can make the change.
 
 Hi,
 
 Thanks for helping to add clarification to this.  From our
 perspective, I have confidence that ~*-core members know the
 difference between trunk and stable policy.  Therefore for the short
 term, it makes sense to have more eyes - especially those which are
 likely to have good knowledge of the internals.
 
 Therefore, I am happy for ~*-core to still have +2 access; especially
 if it helps seed the maint team.
 
 Going forward, it probably will make sense to have a distinction, but
 I feel it might be quite early for that to be a requirement.

I basically said the same thing initially to Thierry on irc, but he
turned me around.

I'm not actually sure all folks on core do grok (or even want to grok)
the subtleties of the stable branch policy and the tradeoffs you need to
make when deciding whether to +2 some on the stable branch. Thierry has
had some similar experience with the milestone-proposed branch, I guess.

Also, I'm not even sure all folks on core always notice that a patch is
being submitted against stable, not master :)

But, of course, if anyone in core wanted to help with +2ing on the
stable branch, we'd add them to stable-maint in a flash.

Cheers,
Mark.


___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp