Mike,

You mention  "We are now extending that example to include storage, and we are 
also working examples with Hadoop. "

In the context of your examples / scenarios, do these placement decisions 
consider storage performance and capacity on a physical node?

For example: Based on application needs, and IOPS, latency requirements - 
carving out a SSD storage or a traditional spinning disk block volume?  Or say 
for cost-efficiency reasons using SSD caching on Hadoop name nodes? 

I'm investigating  a) Per node PCIe SSD deployment need in Openstack 
environment /  Hadoop environment and ,b) selected node SSD caching, 
specifically for OpenStack Cinder.  Hope this is the right forum to ask this 
question.

rgds,
S

On Sep 12, 2013, at 12:29 AM, Mike Spreitzer <mspre...@us.ibm.com> wrote:

> Yes, I've seen that material.  In my group we have worked larger and more 
> complex examples.  I have a proposed breakout session at the Hong Kong summit 
> to talk about one, you might want to vote for it.  The URL is 
> http://www.openstack.org/summit/openstack-summit-hong-kong-2013/become-a-speaker/TalkDetails/109
>  and the title is "Continuous Delivery of Lotus Connections on OpenStack".  
> We used our own technology to do the scheduling (make placement decisions) 
> and orchestration, calling Nova and Quantum to carry out the decisions our 
> software made.  Above the OpenStack infrastructure we used two layers of our 
> own software, one focused on infrastructure and one adding concerns for the 
> software running on that infrastructure.  Each used its own language for a 
> whole topology AKA pattern AKA application AKA cluster.  For example, our 
> pattern has 16 VMs running the WebSphere application server, organized into 
> four homogenous groups (members are interchangeable) of four each.  For each 
> group, we asked that it both (a) be spread across at least two racks, with no 
> more than half the VMs on any one rack and (b) have no two VMs on the same 
> hypervisor.  You can imagine how this would involve multiple levels of 
> grouping and relationships between groups (and you will probably be surprised 
> by the particulars).  We also included information on licensed products, so 
> that the placement decision can optimize license cost (for the IBM 
> "sub-capacity" licenses, placement of VMs can make a cost difference).  Thus, 
> multiple policies per thing.  We are now extending that example to include 
> storage, and we are also working examples with Hadoop. 
> 
> Regards, 
> Mike 
> 
> 
> 
> From:        Gary Kotton <gkot...@vmware.com> 
> To:        OpenStack Development Mailing List 
> <openstack-dev@lists.openstack.org>, 
> Date:        09/11/2013 06:06 AM 
> Subject:        Re: [openstack-dev] [heat] Comments/questions on the 
> instance-group-api-extension blueprint 
> 
> 
> 
> 
> 
> From: Mike Spreitzer <mspre...@us.ibm.com>
> Reply-To: OpenStack Development Mailing List 
> <openstack-dev@lists.openstack.org>
> Date: Tuesday, September 10, 2013 11:58 PM
> To: OpenStack Development Mailing List <openstack-dev@lists.openstack.org>
> Subject: [openstack-dev] [heat] Comments/questions on the 
> instance-group-api-extension blueprint 
> 
> First, I'm a newbie here, wondering: is this the right place for 
> comments/questions on blueprints?  Supposing it is... 
> 
> [Gary Kotton] Yeah, as Russel said this is the correct place 
> 
> I am referring to 
> https://blueprints.launchpad.net/nova/+spec/instance-group-api-extension
> 
> In my own research group we have experience with a few systems that do 
> something like that, and more (as, indeed, that blueprint explicitly states 
> that it is only the start of a longer roadmap).  I would like to highlight a 
> couple of differences that alarm me.  One is the general overlap between 
> groups.  I am not saying this is wrong, but as a matter of natural 
> conservatism we have shied away from unnecessary complexities.  The only 
> overlap we have done so far is hierarchical nesting.  As the 
> instance-group-api-extension explicitly contemplates groups of groups as a 
> later development, this would cover the overlap that we have needed.  On the 
> other hand, we already have multiple "policies" attached to a single group.  
> We have policies for a variety of concerns, so some can combine completely or 
> somewhat independently.  We also have relationships (of various sorts) 
> between groups (as well as between individuals, and between individuals and 
> groups).  The policies and relationships, in general, are not simply names 
> but also have parameters. 
> 
> [Gary Kotton] The instance groups was meant to be the first step towards what 
> we had presented in Portland. Please look at the presentation that we gave an 
> this may highlight what the aims were: 
> https://docs.google.com/presentation/d/1oDXEab2mjxtY-cvufQ8f4cOHM0vIp4iMyfvZPqg8Ivc/edit?usp=sharing.
>  Sadly for this release we did not manage to get the instance groups through 
> (it was an issue of timing and bad luck). We will hopefully get this though 
> in the first stages of the I cycle and then carry on building on it as it has 
> a huge amount of value for OpenStack. It will be great if you can also 
> participate in the discussions. 
> 
> Thanks, 
> Mike_______________________________________________
> OpenStack-dev mailing list
> OpenStack-dev@lists.openstack.org
> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
> 
> _______________________________________________
> OpenStack-dev mailing list
> OpenStack-dev@lists.openstack.org
> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev

_______________________________________________
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev

Reply via email to