Re: [openstack-dev] [nova][PCI] one use case make the flavor/extra-info based solution to be right choice
于 2014年03月21日 03:18, Jay Pipes 写道: On Thu, 2014-03-20 at 13:50 +, Robert Li (baoli) wrote: Hi Yongli, I'm very glad that you bring this up and relive our discussion on PCI passthrough and its application on networking. The use case you brought up is: user wants a FASTER NIC from INTEL to join a virtual networking. By FASTER, I guess that you mean that the user is allowed to select a particular vNIC card. Therefore, the above statement can be translated into the following requests for a PCI device: . Intel vNIC . 1G or 10G or ? . network to join First of all, I'm not sure in a cloud environment, a user would care about the vendor or card type. Correct. Nor would/should a user of the cloud know what vendor or card type is in use on a particular compute node. At most, all a user of the cloud would be able to select from is an instance type (flavor) that listed some capability like high_io_networking or something like that, and the mapping of what high_io_networking meant on the back end of Nova would need to be done by the operator (i.e. if the tag high_io_networking is on a flavor a user has asked to launch a server with, then that tag should be translated into a set of capabilities that is passed to the scheduler and used to determine where the instance can be scheduled by looking at which compute nodes support that set of capabilities. This is what I've been babbling about with regards to leaking implementation through the API. What happens if, say, the operator decides to use IBM cards (instead of or in addition to Intel ones)? If you couple the implementation with the API, like the example above shows (user wants a FASTER NIC from INTEL), then you have to add more complexity to the front-end API that a user deals with, instead of just adding a capabilities mapping for new compute nodes that says high_io_networking tag can match to these new compute nodes with IBM cards. Jay thank you, sorry for later reply in this use case, use might so not care about the vendor id/product id. but for a specific image , the product's model(which related to the vendor id/product id) might cared by user. cause the image might could not support new device which possibly use vendor_id and product id to eliminate the unsupported device. anyway, even without the product/vendor id, the multiple extra tag still needed. and consideration this case, some accelerate card for encryption and decryption/hash there are many supported feature, and most likely different pci card might support different feature set,like : md5, DES,3DES, AES, RSA, SHA-x, IDEA, RC4,5,6 the way to select such a device is use it's feature set instead of one or 2 of group, so the extra information about a pci card is need, in a flexible way. Yongli He Best, -jay ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [OpenStack-Dev] [Cinder FFE] Request for HDS FFE
On 21:29 Mon 24 Mar , Tom Goulding wrote: I'd like to request an FFE for Cinder drivers for Hitachi HNAS storage. We showed this driver running in the demo theater at the Hong Kong summit and we've had end-users using it since January. The driver was initially submitted on Feb 18, 2014 some 45 days ago with regular work since then to address all review comments (a total of 8 revisions). The only real functional change that has occurred since Hong Kong was making CHAP protocol support optional. (The blueprint for this driver was submitted back on Aug. 29, 2013.) We've been working through the new Cinder requirements for certification testing which slowed things down as has obtaining a stable devstack environment. Early documentation on the new certification tests was not clear and has since been improved. We are convinced that the driver poses no risk to the stability and quality of the Icehouse release, so given that it's in the community's best interest to encourage a broad ecosystem of device support, and that we've been working in good faith to get this driver through the process, please help us get this into Icehouse. Hi Tom, Thanks for submitting your driver. Unfortunately I would have to give a -1 for an exception to be made for a new driver this late in the release. I'm sorry that there were problems with getting the cert tests working. However, on March 13th [1], your cert test did reveal that it was not passing for ISCSI and NFS [2][3], which would've gone out with the release. I think you would agree it's far more important to make sure things actually work than making a release. There were other issues with the unit tests not using mock and other standards that Cinder reviewers hold to every review, that this late in the release just didn't make the time frame doable. Regardless of risk to Cinder core, this is a time that contributors should be spending on getting targeted core issues stable and focusing review time on those issues. I would encourage you to propose a review for the Juno release. [1] - https://review.openstack.org/#/c/74371/ [2] - https://gist.github.com/sombrafam/9493308 [3] - https://gist.github.com/sombrafam/9416255 -- Mike Perez ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [Murano] Code review
Hi Murano team, can we review this commit asap: https://bugs.launchpad.net/murano/+bug/1291968? This is fix for the critical issue in release 0.5: #1291968https://bugs.launchpad.net/murano/+bug/1291968 . Thanks! -- Timur, QA Engineer OpenStack Projects Mirantis Inc ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] Separating our Murano PL core in own package
Joshua, I was talking about simple python sub-package inside existing repository, in existing package. I am suggesting to add muranoapi.engine.name sub-package, and nothing more. On Mon, Mar 24, 2014 at 10:29 PM, Ruslan Kamaldinov rkamaldi...@mirantis.com wrote: On Mon, Mar 24, 2014 at 10:08 PM, Joshua Harlow harlo...@yahoo-inc.com wrote: Seeing that the following repos already exist, maybe there is some need for cleanup? - https://github.com/stackforge/murano-agent - https://github.com/stackforge/murano-api - https://github.com/stackforge/murano-common - https://github.com/stackforge/murano-conductor - https://github.com/stackforge/murano-dashboard - https://github.com/stackforge/murano-deployment - https://github.com/stackforge/murano-docs - https://github.com/stackforge/murano-metadataclient - https://github.com/stackforge/murano-repository - https://github.com/stackforge/murano-tests ...(did I miss others?) Can we maybe not have more git repositories and instead figure out a way to have 1 repository (perhaps with submodules?) ;-) It appears like murano is already exploding all over stackforge which makes it hard to understand why yet another repo is needed. I understand why from a code point of view, but it doesn't seem right from a code organization point of view to continue adding repos. It seems like murano (https://github.com/stackforge/murano) should just have 1 repo, with sub-repos (tests, docs, api, agent...) for its own organizational usage instead of X repos that expose others to murano's internal organizational details. -Josh Joshua, I agree that this huge number of repositories is confusing for newcomers. I've spent some time to understand mission of each of these repos. That's why we already did the cleanup :) [0] And I personally will do everything to prevent creation of new repo for Murano. Here is the list of repositories targeted for the next Murano release (Apr 17): * murano-api * murano-agent * python-muranoclient * murano-dashboard * murano-docs The rest of these repos will be deprecated right after the release. Also we will rename murano-api to just murano. murano-api will include all the Murano services, functionaltests for Tempest, Devstack scripts, developer docs. I guess we already can update README files in deprecated repos to avoid further confusion. I wouldn't agree that there should be just one repo. Almost every OpenStack project has it's own repo for python client. All the user docs are kept in a separate repo. Guest agent code should live in it's own repository to keep number of dependencies as low as possible. I'd say there should be required/comfortable minimum of repositories per project. And one more nit correction: OpenStack has it's own git repository [1]. We shoul avoid referring to github since it's just a convinient mirror, while [1] is an official OpenStack repository. [0] https://blueprints.launchpad.net/murano/+spec/repository-reorganization [1] http://git.openstack.org/cgit/ Thanks, Ruslan ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev -- Serg Melikyan, Senior Software Engineer at Mirantis, Inc. http://mirantis.com | smelik...@mirantis.com +7 (495) 640-4904, 0261 +7 (903) 156-0836 ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] Separating our Murano PL core in own package
Ruslan, What about murano-deployment repo? The most important part of it are PowerSheel scripts, Windows Image Builder, package manifests, and some other scripts that better to keep somewhere. Where do we plan to move them? On Mon, Mar 24, 2014 at 10:29 PM, Ruslan Kamaldinov rkamaldi...@mirantis.com wrote: On Mon, Mar 24, 2014 at 10:08 PM, Joshua Harlow harlo...@yahoo-inc.com wrote: Seeing that the following repos already exist, maybe there is some need for cleanup? - https://github.com/stackforge/murano-agent - https://github.com/stackforge/murano-api - https://github.com/stackforge/murano-common - https://github.com/stackforge/murano-conductor - https://github.com/stackforge/murano-dashboard - https://github.com/stackforge/murano-deployment - https://github.com/stackforge/murano-docs - https://github.com/stackforge/murano-metadataclient - https://github.com/stackforge/murano-repository - https://github.com/stackforge/murano-tests ...(did I miss others?) Can we maybe not have more git repositories and instead figure out a way to have 1 repository (perhaps with submodules?) ;-) It appears like murano is already exploding all over stackforge which makes it hard to understand why yet another repo is needed. I understand why from a code point of view, but it doesn't seem right from a code organization point of view to continue adding repos. It seems like murano (https://github.com/stackforge/murano) should just have 1 repo, with sub-repos (tests, docs, api, agent...) for its own organizational usage instead of X repos that expose others to murano's internal organizational details. -Josh Joshua, I agree that this huge number of repositories is confusing for newcomers. I've spent some time to understand mission of each of these repos. That's why we already did the cleanup :) [0] And I personally will do everything to prevent creation of new repo for Murano. Here is the list of repositories targeted for the next Murano release (Apr 17): * murano-api * murano-agent * python-muranoclient * murano-dashboard * murano-docs The rest of these repos will be deprecated right after the release. Also we will rename murano-api to just murano. murano-api will include all the Murano services, functionaltests for Tempest, Devstack scripts, developer docs. I guess we already can update README files in deprecated repos to avoid further confusion. I wouldn't agree that there should be just one repo. Almost every OpenStack project has it's own repo for python client. All the user docs are kept in a separate repo. Guest agent code should live in it's own repository to keep number of dependencies as low as possible. I'd say there should be required/comfortable minimum of repositories per project. And one more nit correction: OpenStack has it's own git repository [1]. We shoul avoid referring to github since it's just a convinient mirror, while [1] is an official OpenStack repository. [0] https://blueprints.launchpad.net/murano/+spec/repository-reorganization [1] http://git.openstack.org/cgit/ Thanks, Ruslan ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev -- Thanks, Dmitry Teselkin Deployment Engineer Mirantis http://www.mirantis.com ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Neutron][LBaaS]
Hi Vijay, Currently Neutron LBaaS supports only namespace based implementation for HAProxy. You can however run LBaaS agent on the host other than network controller node - in that case HAProxy processes will be running on that host but still in namespaces. Also there is an effort in Neutron regarding adding support of advanced services in VMs [1]. After it is completed I hope it will be possible to adopt it in LBaaS and run HAProxy in such a service VM. [1] https://blueprints.launchpad.net/neutron/+spec/adv-services-in-vms Thanks, Oleg On Tue, Mar 25, 2014 at 1:39 AM, Vijay B os.v...@gmail.com wrote: Hi Eugene, Thanks for the reply! How/where is the agent configuration done for HAProxy? If I don't want to go with a network namespace based HAProxy process, but want to deploy my own HAProxy instance on a host outside of the network controller node, and make neutron deploy pools/VIPs on that HAProxy instance, does neutron currently support this scenario? If so, what are the configuration steps I will need to carry out to deploy HAProxy on a separate host (for example, where do I specify the ip address of the haproxy host, etc)? Regards, Vijay On Mon, Mar 24, 2014 at 2:04 PM, Eugene Nikanorov enikano...@mirantis.com wrote: Hi, HAProxy driver has not removed from the trunk, instead it became a base for agent-based driver, so the only haproxy-specific thing in the plugin driver is device driver name. Namespace driver is a device driver on the agent side and it was there from the beginning. The reason for the change is mere refactoring: it seems that solutions that employ agents could share the same code with only device driver being specific. So, everything is in place, HAProxy continues to be the default implementation of Neutron LBaaS service. It supports spawning haproxy processes on any host that runs lbaas agent. Thanks, Eugene. On Tue, Mar 25, 2014 at 12:33 AM, Vijay B os.v...@gmail.com wrote: Hi, I'm looking at HAProxy support in Neutron, and I observe that the drivers/haproxy/plugin_driver.py file in the stable/havana release has been effectively removed from trunk (master), in that the plugin driver in the master simply points to the namespace driver. What was the reason to do this? Was the plugin driver in havana tested and documented? I can't seem to get hold of any relevant documentation that describes how to configure HAProxy LBs installed on separate boxes (and not brought up in network namespaces) - can anyone please point me to the same? Also, are there any plans to bring back the HAProxy plugin driver to talk to remote HAProxy instances? Thanks, Regards, Vijay ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [qa] [neutron] Neutron Full Parallel job - Last 4 days failures
Inline Salvatore On 24 March 2014 23:01, Matthew Treinish mtrein...@kortar.org wrote: On Mon, Mar 24, 2014 at 09:56:09PM +0100, Salvatore Orlando wrote: Thanks a lot! We now need to get on these bugs, and define with QA an acceptable failure rate criterion for switching the full job to voting. It would be good to have a chance to only run the tests against code which is already in master. To this aim we might push a dummy patch, and keep it spinning in the check queue. Honestly, there isn't really a number. I had a thread trying to get consensus on that back when I first made tempest run in parallel. What I ended up doing back then and what we've done since for this kind of change is to just pick a slower week for the gate and just green light it, of course after checking to make sure if it blows up we're not blocking anything critical. Then I guess the ideal period would be after RC2s are cut. Also, we'd need to run also a postgres flavour of the job at least. Meaning that when calculating the probability of a patch passing in the gate is actually the combined probability of two jobs completing successfully. On another note, we noticed that the duplicated jobs currently executed for redundancy in neutron actually seem to point all to the same build id. I'm not sure then if we're actually executing each job twice or just duplicating lines in the jenkins report. If it looks like it's passing at roughly the same rate as everything else and you guys think it's ready. 25% is definitely too high, for comparison when I looked at a couple of min. ago at the numbers for the past 4 days on the equivalent job with nova-network it only failed 4% of the time. (12 out of 300) But that number does fluctuate quite a bit for example looking at the past week the number grows to 11.6%. (171 out of 1480) Even with 11.6% I would not enable it. Running mysql and pg jobs this will give us a combined success rate of 78.1%, which pretty much means the chances of clearing successfully a 5-deep queue in the gate will be a mere 29%. My gut metric is that we should achieve a degree of pass rate which allows us to clear a 10-deep gate queue with a 50% success rate. This translates to a 3.5% failure rate per job, which is indeed inline with what's currently observed for nova-network. Doing it this way doesn't seem like the best, but until it's gating things really don't get the attention they deserve and more bugs will just slip in while you wait. There will most likely be initial pain after it merges, but it's the only real way to lock it down and make forward progress. -Matt Treinish On 24 March 2014 21:45, Rossella Sblendido rsblend...@suse.com wrote: Hello all, here is an update regarding the Neutron full parallel job. I used the following Logstash query [1] that checks the failures of the last 4 days (the last bug fix related with the full job was merged 4 days ago). These are the results: 123 failure (25% of the total) I took a sample of 50 failures and I obtained the following: 22% legitimate failures (they are due to the code change introduced by the patch) 22% infra issues 12% https://bugs.launchpad.net/openstack-ci/+bug/1291611 12% https://bugs.launchpad.net/tempest/+bug/1281969 8% https://bugs.launchpad.net/tempest/+bug/1294603 3% https://bugs.launchpad.net/neutron/+bug/1283522 3% https://bugs.launchpad.net/neutron/+bug/1291920 3% https://bugs.launchpad.net/nova/+bug/1290642 3% https://bugs.launchpad.net/tempest/+bug/1252971 3% https://bugs.launchpad.net/horizon/+bug/1257885 3% https://bugs.launchpad.net/tempest/+bug/1292242 3% https://bugs.launchpad.net/neutron/+bug/1277439 3% https://bugs.launchpad.net/neutron/+bug/1283599 cheers, Rossella [1] http://logstash.openstack.org/#eyJzZWFyY2giOiJidWlsZF9uYW1lOi BcImNoZWNrLXRlbXBlc3QtZHN2bS1uZXV0cm9uLWZ1bGxcIiBBTkQgbWVzc2 FnZTpcIkZpbmlzaGVkOiBGQUlMVVJFXCIgQU5EIHRhZ3M6Y29uc29sZSIsIm ZpZWxkcyI6W10sIm9mZnNldCI6MCwidGltZWZyYW1lIjoiY3VzdG9tIiwiZ3 JhcGhtb2RlIjoiY291bnQiLCJ0aW1lIjp7ImZyb20iOiIyMDE0LTAzLTIwVD EzOjU0OjI1KzAwOjAwIiwidG8iOiIyMDE0LTAzLTI0VDEzOjU0OjI1KzAwOj AwIiwidXNlcl9pbnRlcnZhbCI6IjAifSwibW9kZSI6IiIsImFuYWx5emVfZm llbGQiOiIiLCJzdGFtcCI6MTM5NTY3MDY2ODc0OX0= ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list
Re: [openstack-dev] [Nova] Updates to Juno blueprint review process
On 03/25/2014 12:01 AM, Stefano Maffulli wrote: On 03/24/2014 09:20 AM, Russell Bryant wrote: Another critical point of clarification ... we are *not* moving out of blueprints at all. We're still using them for tracking, just as before. We are *adding* the use of gerrit for reviewing the design. That changes things, thank you for the clarification. If I understand correctly, pages like https://wiki.openstack.org/wiki/HowToContribute#Feature_development will still be valid at pointing at Launchpad Blueprints as the tool we use for the design, roadmap, tracking of progress. What changes is that for Nova all blueprints will have to have a URL added to the specifications, filed in a gerrit repository. Correct. I can live with that, even though I personally think this is a horrible hack and a major tech debt we need to solve properly in the shortest amount of time. I really don't see why this is so bad. We're using a tool specifically designed for reviewing things that is already working *really* well, not just for code, but also for TC governance documents. Unless Storyboard plans to re-implement what gerrit does (I sure hope it doesn't), I expect we would keep working like this. Do you expect storyboard to have an interface for iterating on text documents, where you can provide inline comments, review differences between revisions, etc? What am I missing? -- Russell Bryant ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [openstack-qa] Graduation Requirements + Scope of Tempest
On Mar 21, 2014, at 9:01 AM, David Kranz dkr...@redhat.com wrote: On 03/20/2014 04:19 PM, Rochelle.RochelleGrober wrote: -Original Message- From: Malini Kamalambal [mailto:malini.kamalam...@rackspace.com] Sent: Thursday, March 20, 2014 12:13 PM 'project specific functional testing' in the Marconi context is treating Marconi as a complete system, making Marconi API calls verifying the response - just like an end user would, but without keystone. If one of these tests fail, it is because there is a bug in the Marconi code , and not because its interaction with Keystone caused it to fail. That being said there are certain cases where having a project specific functional test makes sense. For example swift has a functional test job that starts swift in devstack. But, those things are normally handled on a per case basis. In general if the project is meant to be part of the larger OpenStack ecosystem then Tempest is the place to put functional testing. That way you know it works with all of the other components. The thing is in openstack what seems like a project isolated functional test almost always involves another project in real use cases. (for example keystone auth with api requests) One of the concerns we heard in the review was 'having the functional tests elsewhere (I.e within the project itself) does not count and they have to be in Tempest'. This has made us as a team wonder if we should migrate all our functional tests to Tempest. But from Matt's response, I think it is reasonable to continue in our current path have the functional tests in Marconi coexist along with the tests in Tempest. I think that what is being asked, really is that the functional tests could be a single set of tests that would become a part of the tempest repository and that these tests would have an ENV variable as part of the configuration that would allow either no Keystone or Keystone or some such, if that is the only configuration issue that separates running the tests isolated vs. integrated. The functional tests need to be as much as possible a single set of tests to reduce duplication and remove the likelihood of two sets getting out of sync with each other/development. If they only run in the integrated environment, that's ok, but if you want to run them isolated to make debugging easier, then it should be a configuration option and a separate test job. So, if my assumptions are correct, QA only requires functional tests for integrated runs, but if the project QAs/Devs want to run isolated for dev and devtest purposes, more power to them. Just keep it a single set of functional tests and put them in the Tempest repository so that if a failure happens, anyone can find the test and do the debug work without digging into a separate project repository. Hopefully, the tests as designed could easily take a new configuration directive and a short bit of work with OS QA will get the integrated FTs working as well as the isolated ones. --Rocky This issue has been much debated. There are some active members of our community who believe that all the functional tests should live outside of tempest in the projects, albeit with the same idea that such tests could be run either as part of today's real tempest runs or mocked in various ways to allow component isolation or better performance. Maru Newby posted a patch with an example of one way to do this but I think it expired and I don't have a pointer. I think the best place for functional api tests to be maintained is in the projects themselves. The domain expertise required to write api tests is likely to be greater among project resources, and they should be tasked with writing api tests pre-merge. The current 'merge-first, test-later' procedure of maintaining api tests in the Tempest repo makes that impossible. Worse, the cost of developing functional api tests is higher in the integration environment that is the Tempest default. The patch in question [1] proposes allowing pre-merge functional api test maintenance and test reuse in an integration environment. m. 1: https://review.openstack.org/#/c/72585/ IMO there are valid arguments on both sides, but I hope every one could agree that functional tests should not be arbitrarily split between projects and tempest as they are now. The Tempest README states a desire for complete coverage of the OpenStack API but Tempest is not close to that. We have been discussing and then ignoring this issue for some time but I think the recent action to say that Tempest will be used to determine if something can use the OpenStack trademark will force more completeness on tempest (more tests, that is). I think we need to resolve this issue but it won't be easy and modifying existing api tests to be more flexible will be a lot of work. But at least new projects could get on the right
[openstack-dev] [Trove] Meeting Wednesday, 26 March @ 1800 UTC
Just a quick reminder for the weekly Trove meeting. https://wiki.openstack.org/wiki/Meetings#Trove_.28DBaaS.29_meeting Date/Time: Wednesday 26 March - 1800 UTC / 1100 PDT / 1300 CDT IRC channel: #openstack-meeting-alt Meeting Agenda (https://wiki.openstack.org/wiki/Meetings/TroveMeeting): 1. Data Store abstraction start/stop/status/control https://blueprints.launchpad.net/trove/+spec/trove-guest-agent-datastore-control 2. Point in time recovery https://wiki.openstack.org/wiki/Trove/PointInTimeRecovery 3. Data volume snapshot https://wiki.openstack.org/wiki/Trove/volume-data-snapshot-design Cheers, Nikhil ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [depfreeze] Dependency freeze coming up (EOD Tuesday March 25)
Sergey Lukjanov wrote: RE Sahara, we'll need one more version bump to remove all backward compat code added for smooth transition. What's the deadline for doing it? Personally, I'd like to do it next week. Is it ok? If you are only *removing* dependencies I think next week is fine :) -- Thierry Carrez (ttx) ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] Separating our Murano PL core in own package
Hi, I suggest to move all needed Powershell scripts and etc. to the main repository 'murano' in the separate folder. +1 on this. The scripts will not go inside the PyPi package, they will be just grouped in a subfolder. Completely agree on the repo-reorganization topic in general. However And I personally will do everything to prevent creation of new repo for Murano. Well, this may be unavoidable :) We may face a need to create a murano-contrib repository where Murano users will be able to contribute sources of their own murano packages, improve the core library etc. Given that we get rid of murano-conductor, murano-repository, murano-metadataclient, murano-common, murano-tests and, probably, murano-deployment, we are probably ok with having one more. Technically, we may reuse murano-repository for this. But this can be discussed right after there 0.5 release. -- Regards, Alexander Tivelkov On Tue, Mar 25, 2014 at 12:09 PM, Timur Nurlygayanov tnurlygaya...@mirantis.com wrote: Dmitry, I suggest to move all needed Powershell scripts and etc. to the main repository 'murano' in the separate folder. On Tue, Mar 25, 2014 at 11:38 AM, Dmitry Teselkin dtesel...@mirantis.comwrote: Ruslan, What about murano-deployment repo? The most important part of it are PowerSheel scripts, Windows Image Builder, package manifests, and some other scripts that better to keep somewhere. Where do we plan to move them? On Mon, Mar 24, 2014 at 10:29 PM, Ruslan Kamaldinov rkamaldi...@mirantis.com wrote: On Mon, Mar 24, 2014 at 10:08 PM, Joshua Harlow harlo...@yahoo-inc.com wrote: Seeing that the following repos already exist, maybe there is some need for cleanup? - https://github.com/stackforge/murano-agent - https://github.com/stackforge/murano-api - https://github.com/stackforge/murano-common - https://github.com/stackforge/murano-conductor - https://github.com/stackforge/murano-dashboard - https://github.com/stackforge/murano-deployment - https://github.com/stackforge/murano-docs - https://github.com/stackforge/murano-metadataclient - https://github.com/stackforge/murano-repository - https://github.com/stackforge/murano-tests ...(did I miss others?) Can we maybe not have more git repositories and instead figure out a way to have 1 repository (perhaps with submodules?) ;-) It appears like murano is already exploding all over stackforge which makes it hard to understand why yet another repo is needed. I understand why from a code point of view, but it doesn't seem right from a code organization point of view to continue adding repos. It seems like murano (https://github.com/stackforge/murano) should just have 1 repo, with sub-repos (tests, docs, api, agent...) for its own organizational usage instead of X repos that expose others to murano's internal organizational details. -Josh Joshua, I agree that this huge number of repositories is confusing for newcomers. I've spent some time to understand mission of each of these repos. That's why we already did the cleanup :) [0] And I personally will do everything to prevent creation of new repo for Murano. Here is the list of repositories targeted for the next Murano release (Apr 17): * murano-api * murano-agent * python-muranoclient * murano-dashboard * murano-docs The rest of these repos will be deprecated right after the release. Also we will rename murano-api to just murano. murano-api will include all the Murano services, functionaltests for Tempest, Devstack scripts, developer docs. I guess we already can update README files in deprecated repos to avoid further confusion. I wouldn't agree that there should be just one repo. Almost every OpenStack project has it's own repo for python client. All the user docs are kept in a separate repo. Guest agent code should live in it's own repository to keep number of dependencies as low as possible. I'd say there should be required/comfortable minimum of repositories per project. And one more nit correction: OpenStack has it's own git repository [1]. We shoul avoid referring to github since it's just a convinient mirror, while [1] is an official OpenStack repository. [0] https://blueprints.launchpad.net/murano/+spec/repository-reorganization [1] http://git.openstack.org/cgit/ Thanks, Ruslan ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev -- Thanks, Dmitry Teselkin Deployment Engineer Mirantis http://www.mirantis.com ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev -- Timur, QA Engineer OpenStack Projects Mirantis Inc ___ OpenStack-dev mailing list
Re: [openstack-dev] [neutron][rootwrap] Performance considerations, sudo?
On 03/24/2014 07:23 PM, Yuriy Taraday wrote: On Mon, Mar 24, 2014 at 9:51 PM, Carl Baldwin c...@ecbaldwin.net mailto:c...@ecbaldwin.net wrote: Don't discard the first number so quickly. For example, say we use a timeout mechanism for the daemon running inside namespaces to avoid using too much memory with a daemon in every namespace. That means we'll pay the startup cost repeatedly but in a way that amortizes it down. Even if it is really a one time cost, then if you collect enough samples then the outlier won't have much affect on the mean anyway. It actually affects all numbers but mean (e.g. deviation is gross). Carl is right, I thought of it later in the evening, when the timeout mechanism is in place we must consider the number. I'd say keep it in there. +1 I agree. Carl On Mon, Mar 24, 2014 at 2:04 AM, Miguel Angel Ajo majop...@redhat.com mailto:majop...@redhat.com wrote: It's the first call starting the daemon / loading config files, etc?, May be that first sample should be discarded from the mean for all processes (it's an outlier value). I thought about cutting max from counting deviation and/or showing second-max value. But I don't think it matters much and there's not much people here who're analyzing deviation. It's pretty clear what happens with the longest run with this case and I think we can let it be as is. It's mean value that matters most here. Yes, I agree, but as Carl said, having timeouts in place, in a practical environment, the mean will be shifted too. Timeouts are needed within namespaces, to avoid excessive memory consumption. But it could be OK as we'd be cutting out the ip netns delay. Or , if we find a simpler setns mechanism enough for our needs, may be we don't need to care about short-timeouts in ip netns at all... Best, Miguel Ángel. -- Kind regards, Yuriy. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Murano][Heat] MuranoPL questions?
Hi Thomas, I think we went to the second loop of the discussion about generic language concepts. Murano does not use a new language for the sole purpose of having parameters, constraints and polymorphism. These are generic concepts which are common for different languages, so keeping arguing about these generic concepts is just like a holy war like Python vs. C. Keeping these arguments is just like to say that we don't need Python as functions and parameters already exists in C which is used under the hood in Python. Yes Murano DSL have some generic concepts similar to HOT. I think this is a benefit as user will see the familiar syntax constructions and it will be a lower threshold for him to start using Murano DSL. In a simplified view Murano uses DSL for application definition to solve several particular problems: a) control UI rendering of Application Catalog b) control HOT template generation These aspects are not covered in HOT and probably should not be covered. I don't like an idea of expressing HOT template generation in HOT as it sounds like a creation another Lisp like language :-) I don't think that your statement that most of the people in the community are against new DSL is a right summary. There are some disagreements how it should look like and what are the goals. You will be probably surprised but we are not the first who use DSL for HOT templates generation. Here is an e-mail thread right about Ruby based DSL used in IBM for the same purpose: http://lists.openstack.org/pipermail/openstack-dev/2014-February/026606.html The term Orchestration is quite generic. Saying that orchestration should be Heat job sounds like a well know Henry Ford's phrase You can have any colour as long as it's black.. I think this is again a lack of understanding of the difference between Orchestration program and Heat project. There are many aspects of Orchestration and OpenStack has the Orchestration program for the projects which are focused on some aspects of orchestration. Heat is one of the project inside Orchestration program but it does not mean that Heat should cover everything. That is why we discussed in this thread how workflows aspects should be aligned and how they should be placed into this Orchestration program. Thanks Georgy On Mon, Mar 24, 2014 at 8:28 AM, Dmitry mey...@gmail.com wrote: MuranoPL supposed to provide a solution for the real needs to manage services in the centralized manner and to allow cloud provider customers to create their own services. The application catalog similar to AppDirect (www.appdirect.com) natively supported by OpenStack is a huge step forward. Think about Amazon which provides different services for the different needs: Amazon Cloud Formation, Amazon OpsWorks and Amazon Beanstalk. Following the similar logic (which is fully makes sense for me), OpenStack should provide resource reservation and orchestration (Heat and Climate), Application Catalog (Murano) and PaaS (Solum). Every project can live in harmony with other and contribute for the cloud service provider service completeness. This is my opinion and i would happy to use Murano in our internal solution. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev -- Georgy Okrokvertskhov Architect, OpenStack Platform Products, Mirantis http://www.mirantis.com Tel. +1 650 963 9828 Mob. +1 650 996 3284 On Mon, Mar 24, 2014 at 5:13 AM, Thomas Herve thomas.he...@enovance.comwrote: Hi Stan, Comments inline. Zane, I appreciate your explanations on Heat/HOT. This really makes sense. I didn't mean to say that MuranoPL is better for Heat. Actually HOT is good for Heat's mission. I completely acknowledge it. I've tried to avoid comparison between languages and I'm sorry if it felt that way. This is not productive as I don't offer you to replace HOT with MuranoPL (although I believe that certain elements of MuranoPL syntax can be contributed to HOT and be valuable addition there). Also people tend to protect what they have developed and invested into and to be fair this is what we did in this thread to great extent. What I'm trying to achieve is that you and the rest of Heat team understand why it was designed the way it is. I don't feel that Murano can become full-fledged member of OpenStack ecosystem without a bless from Heat team. And it would be even better if we agree on certain design, join our efforts and contribute to each other for sake of Orchestration program. Note that I feel that most people outside of the Murano project are against the idea of using a DSL. You should feel that it could block the integration in OpenStack. I'm sorry for long mail texts written in not-so-good English and appreciate you patience reading and answering them. Having said that let me step backward and explain our design
Re: [openstack-dev] [Nova] Updates to Juno blueprint review process
Russell Bryant wrote: On 03/24/2014 11:42 AM, Stefano Maffulli wrote: At this point I'd like to get a fair assessment of storyboard's status and timeline: it's clear that Launchpad blueprints need to be abandoned lightspeed fast and I (and others) have been sold the idea that Storyboard is a *thing* that will happen *soon*. I also was told that spec reviews are an integral part of Storyboard use case scenarios, not just defects. Another critical point of clarification ... we are *not* moving out of blueprints at all. We're still using them for tracking, just as before. We are *adding* the use of gerrit for reviewing the design. Yes, there is a clear misunderstanding here. There never was any kind of spec review system in Launchpad. Launchpad tracks feature completion, not design specs. It supported a link and behind that link was a set of tools (wiki pages, etherpads, google docs) where the spec review would hopefully happen. This proposal is *just* about replacing all those design documents with a clear change and using Gerrit to iterate and track approvals on it. This is not moving off Launchpad blueprints nor is it bypassing StoryBoard. It's adding a bit of formal process around something that randomly lived out of our tooling up to now. About StoryBoard progress now: People got very excited with the Django proof-of-concept because it looked like it was almost ready to replace Launchpad. But I always made clear this was just a proof-of-concept and it would take a *lot* of time and effort to make it real. And my usual amount of free time would probably not allow me to help that much. It also took us more time than we expected to set up a basic team to work on it (many thanks to HP and Mirantis for jumping on it with significant resources), to align that team on clear goals, to rewrite the main data server as an OpenStack-style API server, to write from scratch a JavaScript webclient and get it properly tested at the gate, etc. We now have the base infrastructure in place, we continuously deploy, and the minimal viable product is almost completed. We expect the infrastructure team to start dogfooding it in Juno and hopefully we'll iterate faster next cycle... to make it a viable Launchpad alternative for adventurous projects in the K cycle. So it's not vaporware, it exists for real. But there is still a lot of work needed on it to be generally usable, so it shouldn't be used as an argument to stall everything else. -- Thierry Carrez (ttx) ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Murano][Heat] MuranoPL questions?
What I can say is that I'm not convinced. The only use-case for a DSL would be if you have to upload user-written code, but what you mentioned is a Web interface, where the user doesn't use the DSL, and the cloud provider is the developer. There is no reason in this case to have a secure environment for the code. I didn't say that. There are at least 2 different roles application developers/publishers and application users. Application developer is not necessary cloud provider. The whole point of AppCatalog is to support scenario when anyone can create and package some application and that package can be uploaded by user alone. Think Apple AppStore or Google Play. Some cloud providers may configure ACLs so that user be allowed to consume applications they decided while others may permit to upload applications to some configurable scope (e.g. apps that would be visible to all cloud users, to particular tenant or be private to the user). We also think to have some of peer relations so that it would be possible to have application upload in one catalog to become automatically available in all connected catalogs. This is similar to how Linux software repos work - AppCatalog is repo, Murano package is what DEB/RPMs are to repo and DSL is what DEB/RPMs manifests are to packages. Just that is run on cloud and designed to handle complex multi-node apps as well as trivial ones in which case this may be narrowed to actual installation of DEB/RPM I'm glad that you bring packages up. This is a really good example of why you don't need a new programming language. Packages uses whatever technology they prefer to handle their scripting needs. They then have an declarative interface which hides the imperative parts behind. You trust OpenStack developers with their code, you trust package developers with their code, why not trust catalog developers? -- Thomas ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] Multiple patches in one review
On Mon, 2014-03-24 at 10:49 -0400, Russell Bryant wrote: Gerrit support for a patch series could certainly be better. There has long been talking about gerrit getting topic review functionality, whereby you could e.g. approve a whole series of patches from a topic view. See: https://code.google.com/p/gerrit/issues/detail?id=51 https://groups.google.com/d/msg/repo-discuss/5oRra_tLKMA/rxwU7pPAQE8J My understanding is there's a fork of gerrit out there with this functionality that some projects are using successfully. Mark. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Murano][Heat] MuranoPL questions?
Hi Thomas, I think we went to the second loop of the discussion about generic language concepts. Murano does not use a new language for the sole purpose of having parameters, constraints and polymorphism. These are generic concepts which are common for different languages, so keeping arguing about these generic concepts is just like a holy war like Python vs. C. Keeping these arguments is just like to say that we don't need Python as functions and parameters already exists in C which is used under the hood in Python. Yes Murano DSL have some generic concepts similar to HOT. I think this is a benefit as user will see the familiar syntax constructions and it will be a lower threshold for him to start using Murano DSL. In a simplified view Murano uses DSL for application definition to solve several particular problems: a) control UI rendering of Application Catalog b) control HOT template generation These aspects are not covered in HOT and probably should not be covered. I don't like an idea of expressing HOT template generation in HOT as it sounds like a creation another Lisp like language :-) I'm not saying that HOT will cover all your needs. I think it will cover a really good portion. And I'm saying that for the remaining part, you can use an existing language and not create a new one. I don't think that your statement that most of the people in the community are against new DSL is a right summary. There are some disagreements how it should look like and what are the goals. You will be probably surprised but we are not the first who use DSL for HOT templates generation. Here is an e-mail thread right about Ruby based DSL used in IBM for the same purpose: http://lists.openstack.org/pipermail/openstack-dev/2014-February/026606.html The term Orchestration is quite generic. Saying that orchestration should be Heat job sounds like a well know Henry Ford's phrase You can have any colour as long as it's black.. That worked okay for him :). I think this is again a lack of understanding of the difference between Orchestration program and Heat project. There are many aspects of Orchestration and OpenStack has the Orchestration program for the projects which are focused on some aspects of orchestration. Heat is one of the project inside Orchestration program but it does not mean that Heat should cover everything. That is why we discussed in this thread how workflows aspects should be aligned and how they should be placed into this Orchestration program. Well, today Heat is the one and only program in the Orchestration program. If and when you have orchestration needs not covered, we are there to make sure Heat is not the best place to handle them. The answer will probably not Heat forever, but we need good use cases to delegate those needs to another project. -- Thomas ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] OpenStack vs. SQLA 0.9
FYI, allowing 0.9 recently merged into openstack/requirements: https://review.openstack.org/79817 This is a good example of how we should be linking gerrit and mailing list discussions together more. I don't think the gerrit review was linked in this thread nor was the mailing list discussion linked in the gerrit review. Mark. On Thu, 2014-03-13 at 22:45 -0700, Roman Podoliaka wrote: Hi all, I think it's actually not that hard to fix the errors we have when using SQLAlchemy 0.9.x releases. I uploaded two changes two Nova to fix unit tests: - https://review.openstack.org/#/c/80431/ (this one should also fix the Tempest test run error) - https://review.openstack.org/#/c/80432/ Thanks, Roman On Thu, Mar 13, 2014 at 7:41 PM, Thomas Goirand z...@debian.org wrote: On 03/14/2014 02:06 AM, Sean Dague wrote: On 03/13/2014 12:31 PM, Thomas Goirand wrote: On 03/12/2014 07:07 PM, Sean Dague wrote: Because of where we are in the freeze, I think this should wait until Juno opens to fix. Icehouse will only be compatible with SQLA 0.8, which I think is fine. I expect the rest of the issues can be addressed during Juno 1. -Sean Sean, No, it's not fine for me. I'd like things to be fixed so we can move forward. Debian Sid has SQLA 0.9, and Jessie (the next Debian stable) will be released SQLA 0.9 and with Icehouse, not Juno. We're past freeze, and this requires deep changes in Nova DB to work. So it's not going to happen. Nova provably does not work with SQLA 0.9, as seen in Tempest tests. -Sean I'd be nice if we considered more the fact that OpenStack, at some point, gets deployed on top of distributions... :/ Anyway, if we can't do it because of the freeze, then I will have to carry the patch in the Debian package. Never the less, someone will have to work and fix it. If you know how to help, it'd be very nice if you proposed a patch, even if we don't accept it before Juno opens. Thomas Goirand (zigo) ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] OpenStack vs. SQLA 0.9
In fairness, there were also copious IRC conversations on the topic as well. I think because there were so few people in this thread, and all those people were participating in the review and irc conversations, updating the thread just fell off the list. My bad. When conversations jump media it's some times hard to remember that each one might have had different people watching passively. On 03/25/2014 06:36 AM, Mark McLoughlin wrote: FYI, allowing 0.9 recently merged into openstack/requirements: https://review.openstack.org/79817 This is a good example of how we should be linking gerrit and mailing list discussions together more. I don't think the gerrit review was linked in this thread nor was the mailing list discussion linked in the gerrit review. Mark. On Thu, 2014-03-13 at 22:45 -0700, Roman Podoliaka wrote: Hi all, I think it's actually not that hard to fix the errors we have when using SQLAlchemy 0.9.x releases. I uploaded two changes two Nova to fix unit tests: - https://review.openstack.org/#/c/80431/ (this one should also fix the Tempest test run error) - https://review.openstack.org/#/c/80432/ Thanks, Roman On Thu, Mar 13, 2014 at 7:41 PM, Thomas Goirand z...@debian.org wrote: On 03/14/2014 02:06 AM, Sean Dague wrote: On 03/13/2014 12:31 PM, Thomas Goirand wrote: On 03/12/2014 07:07 PM, Sean Dague wrote: Because of where we are in the freeze, I think this should wait until Juno opens to fix. Icehouse will only be compatible with SQLA 0.8, which I think is fine. I expect the rest of the issues can be addressed during Juno 1. -Sean Sean, No, it's not fine for me. I'd like things to be fixed so we can move forward. Debian Sid has SQLA 0.9, and Jessie (the next Debian stable) will be released SQLA 0.9 and with Icehouse, not Juno. We're past freeze, and this requires deep changes in Nova DB to work. So it's not going to happen. Nova provably does not work with SQLA 0.9, as seen in Tempest tests. -Sean I'd be nice if we considered more the fact that OpenStack, at some point, gets deployed on top of distributions... :/ Anyway, if we can't do it because of the freeze, then I will have to carry the patch in the Debian package. Never the less, someone will have to work and fix it. If you know how to help, it'd be very nice if you proposed a patch, even if we don't accept it before Juno opens. Thomas Goirand (zigo) ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev -- Sean Dague Samsung Research America s...@dague.net / sean.da...@samsung.com http://dague.net signature.asc Description: OpenPGP digital signature ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] Spec repos for blueprint development and review
On Mar 24, 2014, at 2:23 PM, Sean Dague s...@dague.netmailto:s...@dague.net wrote: On 03/24/2014 02:05 PM, James E. Blair wrote: Russell Bryant rbry...@redhat.commailto:rbry...@redhat.com writes: On 03/24/2014 12:34 PM, James E. Blair wrote: Hi, So recently we started this experiment with the compute and qa programs to try using Gerrit to review blueprints. Launchpad is deficient in this area, and while we hope Storyboard will deal with it much better, but it's not ready yet. This seems to be a point of confusion. My view is that Storyboard isn't intended to implement what gerrit provides. Given that, it seems like we'd still be using this whether the tracker is launchpad or storyboard. I don't think it's intended to implement what Gerrit provides, however, I'm not sure what Gerrit provides is _exactly_ what's needed here. I do agree that Gerrit is a much better tool than launchpad for collaborating on some kinds of blueprints. Agreed, but the current blueprint system is broken in a way that a half step with an imperfect tool is better than keeping the status quo. However, one of the reasons we're creating StoryBoard is so that we have a tool that is compatible with our workflow and meets our requirements. It's not just about tracking work items, it should be a tool for creating, evaluating, and progressing changes to projects (stories), across all stages. I don't envision the end-state for storyboard to be that we end up copying data back and forth between it and Gerrit. Since we're designing a system from scratch, we might as well design it to do what we want. One of our early decisions was to say that UX and code stories have equally important use cases in StoryBoard. Collaboration around UX style blueprints (especially those with graphical mock-ups) sets a fairly high bar for the kind of interaction we will support. Gerrit is a great tool for reviewing code and other text media. But somehow it is even worse than launchpad for collaborating when visual media are involved. Quite a number of blueprints could benefit from better support for that (not just UI mockups but network diagrams, etc). We can learn a lot from the experiment of using Gerrit for blueprint review, and I think it's going to help make StoryBoard a lot better for all of our use cases. Diagram handling was one of the first questions I asked Russell when I saw the repo creation proposal. Diagrams are very helpful and while gerrit is not ideal for handling diagrams, Sphinx should allow us to incorporate them in basic way for now. I view this as an improvement over the 5 different formats blueprints are submitted in now. I think that's fine if long term this whole thing is optimized. I just do very much worry that StoryBoard keeps going under progressive scope creep before we've managed to ship the base case. That's a dangerous situation to be in, as it means it's evolving without a feedback loop. I'd much rather see Storyboard get us off launchpad ASAP across all the projects, and then work on solving the things launchpad doesn't do. +1000 I don’t want to see Storyboard changing scope to the point where it doesn’t adequately deliver because it is trying to solve too many problems in the first iteration. mark ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Murano][Heat] MuranoPL questions?
On Tue, Mar 25, 2014 at 3:32 AM, Thomas Herve thomas.he...@enovance.comwrote: Hi Thomas, I think we went to the second loop of the discussion about generic language concepts. Murano does not use a new language for the sole purpose of having parameters, constraints and polymorphism. These are generic concepts which are common for different languages, so keeping arguing about these generic concepts is just like a holy war like Python vs. C. Keeping these arguments is just like to say that we don't need Python as functions and parameters already exists in C which is used under the hood in Python. Yes Murano DSL have some generic concepts similar to HOT. I think this is a benefit as user will see the familiar syntax constructions and it will be a lower threshold for him to start using Murano DSL. In a simplified view Murano uses DSL for application definition to solve several particular problems: a) control UI rendering of Application Catalog b) control HOT template generation These aspects are not covered in HOT and probably should not be covered. I don't like an idea of expressing HOT template generation in HOT as it sounds like a creation another Lisp like language :-) I'm not saying that HOT will cover all your needs. I think it will cover a really good portion. And I'm saying that for the remaining part, you can use an existing language and not create a new one. As a user can't run arbitrary python code in openstack we used Python language to create a new API for the remaining parts. This API service accepts a yaml based description of what should be done. There is no intention to create a new generic programming language. We used OpenStack approach and created a service for specific functions around Application Catalog features. Due to dynamic nature of applications we had to add a bit of dynamics to the service input just because of the same reason why Heat uses templates. I don't think that your statement that most of the people in the community are against new DSL is a right summary. There are some disagreements how it should look like and what are the goals. You will be probably surprised but we are not the first who use DSL for HOT templates generation. Here is an e-mail thread right about Ruby based DSL used in IBM for the same purpose: http://lists.openstack.org/pipermail/openstack-dev/2014-February/026606.html The term Orchestration is quite generic. Saying that orchestration should be Heat job sounds like a well know Henry Ford's phrase You can have any colour as long as it's black.. That worked okay for him :). Not really. The world acknowledged his inventions and new approaches. Other manufacturers adopted his ideas and moved forward, providing more variety, while Ford stuck with his model-T, which was very successful though. The history shows that variety won the battle over single approach and now we have different colors, shapes, engines :-) I think this is again a lack of understanding of the difference between Orchestration program and Heat project. There are many aspects of Orchestration and OpenStack has the Orchestration program for the projects which are focused on some aspects of orchestration. Heat is one of the project inside Orchestration program but it does not mean that Heat should cover everything. That is why we discussed in this thread how workflows aspects should be aligned and how they should be placed into this Orchestration program. Well, today Heat is the one and only program in the Orchestration program. If and when you have orchestration needs not covered, we are there to make sure Heat is not the best place to handle them. The answer will probably not Heat forever, but we need good use cases to delegate those needs to another project. That is exactly the reason why we have these discussions :-) We have the use cases for new functionality and we are trying to find a place for it. -- Thomas ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev -- Georgy Okrokvertskhov Architect, OpenStack Platform Products, Mirantis http://www.mirantis.com Tel. +1 650 963 9828 Mob. +1 650 996 3284 ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Murano][Heat] MuranoPL questions?
On Tue, Mar 25, 2014 at 2:27 PM, Thomas Herve thomas.he...@enovance.comwrote: What I can say is that I'm not convinced. The only use-case for a DSL would be if you have to upload user-written code, but what you mentioned is a Web interface, where the user doesn't use the DSL, and the cloud provider is the developer. There is no reason in this case to have a secure environment for the code. I didn't say that. There are at least 2 different roles application developers/publishers and application users. Application developer is not necessary cloud provider. The whole point of AppCatalog is to support scenario when anyone can create and package some application and that package can be uploaded by user alone. Think Apple AppStore or Google Play. Some cloud providers may configure ACLs so that user be allowed to consume applications they decided while others may permit to upload applications to some configurable scope (e.g. apps that would be visible to all cloud users, to particular tenant or be private to the user). We also think to have some of peer relations so that it would be possible to have application upload in one catalog to become automatically available in all connected catalogs. This is similar to how Linux software repos work - AppCatalog is repo, Murano package is what DEB/RPMs are to repo and DSL is what DEB/RPMs manifests are to packages. Just that is run on cloud and designed to handle complex multi-node apps as well as trivial ones in which case this may be narrowed to actual installation of DEB/RPM I'm glad that you bring packages up. This is a really good example of why you don't need a new programming language. Packages uses whatever technology they prefer to handle their scripting needs. They then have an declarative interface which hides the imperative parts behind. The same is true for Murano. MuranoPL is not used to express what should be deployed. In Murano there is object model that describes view of the word. It serves for the same purpose as HOT in Heat but it is simpler because it says just what need to be deployed but not how it should be accomplished as this information is already contained in application definitions. There is REST API to edit/submit Object Model which is again has nothing to do with MuranoPL. UI dashboard talks to AppCatalog to see what applications/classes are available and AppCatalog also knows what are the properties of those classes. This is needed for UI so that it can ask user for appropriate input. This is similar to how Horizon asks user to input parameters that are declared in HOT template. But all the imperative stuff is hidden inside Murano packages and is not used for anything outside Murano engine. MuranoPL is not a replacement for scripting languages. You still use bash/puppet/PowerShell/whatever you like for actual deployment. No MuranoPL code is executed on VM side. So the analogy with RPMs manifests is valid. You trust OpenStack developers with their code, you trust package developers with their code, why not trust catalog developers? They do trust catalog developers (hopefully). But catalog developers have nothing to do with catalog contents. Anyone can create and upload application to App Catalog the same way how anyone can upload his application to Google Play. The fact that I trust Google doesn't mean that I trust all applications in Google Play. The fact that I trust catalog developers doesn't mean that I (as a cloud operator) is going to allow execution of untrusted code in that catalog. Unless that code is sandboxed by design. Similar to that I can navigate to any web site google points me out and let it execute any JavaScript it wants unless it it JavaScript and not browser plugin or desktop application. -- Thomas ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev -- Sincerely yours Stanislav (Stan) Lagun Senior Developer Mirantis 35b/3, Vorontsovskaya St. Moscow, Russia Skype: stanlagun www.mirantis.com sla...@mirantis.com ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [openstack-qa] Graduation Requirements + Scope of Tempest
We are talking about different levels of testing, 1. Unit tests - which everybody agrees should be in the individual project itself 2. System Tests - 'System' referring to ( limited to), all the components that make up the project. These are also the functional tests for the project. 3. Integration Tests - This is to verify that the OS components interact well and don't break other components -Keystone being the most obvious example. This is where I see getting the maximum mileage out of Tempest. Its not easy to detect what the integration points with other projects are, any project can use any stable API from any other project. Because of this all OpenStack APIs should fit into this category. Any project can use any stable API –but that does not make all API tests , Integration Tests. A test becomes Integration test when it has two or more projects interacting in the test. Individual projects should be held accountable to make sure that their API's work – no matter who consumes them. We should be able to treat the project as a complete system, make API calls and validate that the response matches the API definitions. Identifying issues earlier in the pipeline reduces the Total Cost of Quality. I agree that Integration Testing is hard. It is complicated because it requires knowledge of how systems interact with each other – and knowledge comes from a lot of time spent on analysis. It requires people with project expertise to talk to each other identify possible test cases. openstack-qa is the ideal forum to do this. Holding projects responsible for their functionality will help the QA team focus on complicated integration tests. Having a second group writing tests to Nova's public APIs has been really helpful in keeping us honest as well. Sounds like a testimonial for more project level testing :) I see value in projects taking ownership of the System Tests - because if the project is not 'functionally ready', it is not ready to integrate with other components of Openstack. What do you mean by not ready? 'Functionally Ready' - The units that make up a project can work together as a system, all API's have been exercised with positive negative test cases by treating the project as a complete system. There are no known critical bugs. The point here being identify as many issues as possible, earlier in the game. But for this approach to be successful, projects should have diversity in the team composition - we need more testers who focus on creating these tests. This will keep the teams honest in their quality standards. As long as individual projects cannot guarantee functional test coverage, we will need more tests in Tempest. But that will shift focus away from Integration Testing, which can be done ONLY in Tempest. Regardless of whatever we end up deciding, it will be good to have these discussions sooner than later. This will help at least the new projects to move in the right direction. -Malini ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.orgmailto:OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Neutron][LBaaS] Neutron LBaaS, Libra and managed services
John, Brandon, I agree that we cannot have a multitude of drivers doing the same thing or close to because then we end-up in the same situation as we are today where we have duplicate effort and technical debt. The goal would be here to be able to built a framework around the drivers that would allow for resiliency, failover, etc... If the differentiators are in higher level APIs then we can have one a single driver (in the best case) for each software LB e.g. HA proxy, nginx, etc. Thoughts? Susanne On Mon, Mar 24, 2014 at 11:26 PM, John Dewey j...@dewey.ws wrote: I have a similar concern. The underlying driver may support different functionality, but the differentiators need exposed through the top level API. I see the SSL work is well underway, and I am in the process of defining L7 scripting requirements. However, I will definitely need L7 scripting prior to the API being defined. Is this where vendor extensions come into play? I kinda like the route the Ironic guy safe taking with a vendor passthru API. John On Monday, March 24, 2014 at 3:17 PM, Brandon Logan wrote: Creating a separate driver for every new need brings up a concern I have had. If we are to implement a separate driver for every need then the permutations are endless and may cause a lot drivers and technical debt. If someone wants an ha-haproxy driver then great. What if they want it to be scalable and/or HA, is there supposed to be scalable-ha-haproxy, scalable-haproxy, and ha-haproxy drivers? Then what if instead of doing spinning up processes on the host machine we want a nova VM or a container to house it? As you can see the permutations will begin to grow exponentially. I'm not sure there is an easy answer for this. Maybe I'm worrying too much about it because hopefully most cloud operators will use the same driver that addresses those basic needs, but worst case scenarios we have a ton of drivers that do a lot of similar things but are just different enough to warrant a separate driver. -- *From:* Susanne Balle [sleipnir...@gmail.com] *Sent:* Monday, March 24, 2014 4:59 PM *To:* OpenStack Development Mailing List (not for usage questions) *Subject:* Re: [openstack-dev] [Neutron][LBaaS] Neutron LBaaS, Libra and managed services Eugene, Thanks for your comments, See inline: Susanne On Mon, Mar 24, 2014 at 4:01 PM, Eugene Nikanorov enikano...@mirantis.com wrote: Hi Susanne, a couple of comments inline: We would like to discuss adding the concept of managed services to the Neutron LBaaS either directly or via a Neutron LBaaS plug-in to Libra/HA proxy. The latter could be a second approach for some of the software load-balancers e.g. HA proxy since I am not sure that it makes sense to deploy Libra within Devstack on a single VM. Currently users would have to deal with HA, resiliency, monitoring and managing their load-balancers themselves. As a service provider we are taking a more managed service approach allowing our customers to consider the LB as a black box and the service manages the resiliency, HA, monitoring, etc. for them. As far as I understand these two abstracts, you're talking about making LBaaS API more high-level than it is right now. I think that was not on our roadmap because another project (Heat) is taking care of more abstracted service. The LBaaS goal is to provide vendor-agnostic management of load balancing capabilities and quite fine-grained level. Any higher level APIs/tools can be built on top of that, but are out of LBaaS scope. [Susanne] Yes. Libra currently has some internal APIs that get triggered when an action needs to happen. We would like similar functionality in Neutron LBaaS so the user doesn't have to manage the load-balancers but can consider them as black-boxes. Would it make sense to maybe consider integrating Neutron LBaaS with heat to support some of these use cases? We like where Neutron LBaaS is going with regards to L7 policies and SSL termination support which Libra is not currently supporting and want to take advantage of the best in each project. We have a draft on how we could make Neutron LBaaS take advantage of Libra in the back-end. The details are available at: https://wiki.openstack.org/wiki/Neutron/LBaaS/LBaaS%2Band%2BLibra%2Bintegration%2BDraft I looked at the proposal briefly, it makes sense to me. Also it seems to be the simplest way of integrating LBaaS and Libra - create a Libra driver for LBaaS. [Susanne] Yes that would be the short team solution to get us where we need to be. But We do not want to continue to enhance Libra. We would like move to Neutron LBaaS and not have duplicate efforts. While this would allow us to fill a gap short term we would like to discuss the longer term strategy since we believe that everybody would benefit from having such managed services artifacts built directly into
[openstack-dev] [Ironic][Keystone] Move drivers credentials to Keystone
Hi, Right now Ironic is being responsible for storing the credentials for the IPMI and SSH drivers (and potentially other drivers in the future), I wonder if we should delegate this task to Keystone. The Keystone V3 API now has a /credentials endpoint which would allow us to specify arbitrary types (not only ec2 anymore) and use it as a credential store[1]. That would avoid further fragmentation of credentials being stored in different places in OpenStack, and make the management of the credentials easier (Think about a situation where many nodes share the same IPMI username/password and we need to update it, if this is stored in Keystone it only needs to be updated there once cause nodes will only hold a reference to it) It also was pointed to me that setting a hard dependency on Keystone V3 might significantly raises the bar for integration with existing clouds*. So perhaps we should make it optional? In the same way we can specify a username/password or key_filename for the ssh driver we could have a reference to a credential in Keystone V3? What you guys think about the idea? What are the cloud operators/sysadmins view on that? * There's also some ongoing thoughts about using v3 for other things in Ironic (e.g signed url's) but that's kinda out of the topic. [1] https://github.com/openstack/identity-api/blob/master/openstack-identity-api/v3/src/markdown/identity-api-v3.md#create-credential-post-credentials Ironic bp (discussion): https://blueprints.launchpad.net/ironic/+spec/credentials-keystone-v3 ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [depfreeze] Dependency freeze coming up (EOD Tuesday March 25)
We'd like to bump python-saharaclient min version to the next major version (=0.6.0 - = 0.7.0) and remove python-savannaclient. On Tue, Mar 25, 2014 at 1:47 PM, Thierry Carrez thie...@openstack.org wrote: Sergey Lukjanov wrote: RE Sahara, we'll need one more version bump to remove all backward compat code added for smooth transition. What's the deadline for doing it? Personally, I'd like to do it next week. Is it ok? If you are only *removing* dependencies I think next week is fine :) -- Thierry Carrez (ttx) ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev -- Sincerely yours, Sergey Lukjanov Sahara Technical Lead (OpenStack Data Processing) Mirantis Inc. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Ironic][Keystone] Move drivers credentials to Keystone
Hi, Right now Ironic is being responsible for storing the credentials for the IPMI and SSH drivers (and potentially other drivers in the future), I wonder if we should delegate this task to Keystone. The Keystone V3 API now has a /credentials endpoint which would allow us to specify arbitrary types (not only ec2 anymore) and use it as a credential store[1]. That would avoid further fragmentation of credentials being stored in different places in OpenStack, and make the management of the credentials easier (Think about a situation where many nodes share the same IPMI username/password and we need to update it, if this is stored in Keystone it only needs to be updated there once cause nodes will only hold a reference to it) It also was pointed to me that setting a hard dependency on Keystone V3 might significantly raises the bar for integration with existing clouds*. So perhaps we should make it optional? In the same way we can specify a username/password or key_filename for the ssh driver we could have a reference to a credential in Keystone V3? What you guys think about the idea? Hi Lucas, At a high level, this sounds like an excellent idea to me. IIUC the major blocker to ceilometer taking point on controlling the IPMI polling cycle has been secure access to these credentials. If these were available to ceilometer in a controlled way via keystone, then the IPMI polling cycle could be managed in a very similar way to the ceilo polling activity on the hypervisor and SMNP daemons. However, I'm a little fuzzy on the detail of enabling this via keystone v3, so it would be great to drill down into the detail either on the ML or at summit. For example, would it be in the guise of a trust that delegates limited privilege to allow the ceilometer user call GET /credentials to retrieve the IPMI user/pass? Or would the project_id parameter to POST /credentials suffice to limit access to IPMI credentials to the ceilometer tenant only? (as opposed to allowing any other openstack service access these creds) In that case, would we need to also decouple the ceilometer user from the generic service tenant? Cheers, Eoghan What are the cloud operators/sysadmins view on that? * There's also some ongoing thoughts about using v3 for other things in Ironic (e.g signed url's) but that's kinda out of the topic. [1] https://github.com/openstack/identity-api/blob/master/openstack-identity-api/v3/src/markdown/identity-api-v3.md#create-credential-post-credentials Ironic bp (discussion): https://blueprints.launchpad.net/ironic/+spec/credentials-keystone-v3 ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Neutron][LBaaS] Neutron LBaaS, Libra and managed services
Hi Brandon, On Tue, Mar 25, 2014 at 2:17 AM, Brandon Logan brandon.lo...@rackspace.comwrote: Creating a separate driver for every new need brings up a concern I have had. If we are to implement a separate driver for every need then the permutations are endless and may cause a lot drivers and technical debt. If someone wants an ha-haproxy driver then great. What if they want it to be scalable and/or HA, is there supposed to be scalable-ha-haproxy, scalable-haproxy, and ha-haproxy drivers? Then what if instead of doing spinning up processes on the host machine we want a nova VM or a container to house it? As you can see the permutations will begin to grow exponentially. I'm not sure there is an easy answer for this. Maybe I'm worrying too much about it because hopefully most cloud operators will use the same driver that addresses those basic needs, but worst case scenarios we have a ton of drivers that do a lot of similar things but are just different enough to warrant a separate driver. The driver is what implements communicating to a particular device/appliance and translating logical service configuration to a backend-specific configuration. I never said the driver is per feature. But different drivers may implement different features in their own way, the general requirement is that user expectations should be properly satisfied. Thanks, Eugene. -- *From:* Susanne Balle [sleipnir...@gmail.com] *Sent:* Monday, March 24, 2014 4:59 PM *To:* OpenStack Development Mailing List (not for usage questions) *Subject:* Re: [openstack-dev] [Neutron][LBaaS] Neutron LBaaS, Libra and managed services Eugene, Thanks for your comments, See inline: Susanne On Mon, Mar 24, 2014 at 4:01 PM, Eugene Nikanorov enikano...@mirantis.com wrote: Hi Susanne, a couple of comments inline: We would like to discuss adding the concept of managed services to the Neutron LBaaS either directly or via a Neutron LBaaS plug-in to Libra/HA proxy. The latter could be a second approach for some of the software load-balancers e.g. HA proxy since I am not sure that it makes sense to deploy Libra within Devstack on a single VM. Currently users would have to deal with HA, resiliency, monitoring and managing their load-balancers themselves. As a service provider we are taking a more managed service approach allowing our customers to consider the LB as a black box and the service manages the resiliency, HA, monitoring, etc. for them. As far as I understand these two abstracts, you're talking about making LBaaS API more high-level than it is right now. I think that was not on our roadmap because another project (Heat) is taking care of more abstracted service. The LBaaS goal is to provide vendor-agnostic management of load balancing capabilities and quite fine-grained level. Any higher level APIs/tools can be built on top of that, but are out of LBaaS scope. [Susanne] Yes. Libra currently has some internal APIs that get triggered when an action needs to happen. We would like similar functionality in Neutron LBaaS so the user doesn't have to manage the load-balancers but can consider them as black-boxes. Would it make sense to maybe consider integrating Neutron LBaaS with heat to support some of these use cases? We like where Neutron LBaaS is going with regards to L7 policies and SSL termination support which Libra is not currently supporting and want to take advantage of the best in each project. We have a draft on how we could make Neutron LBaaS take advantage of Libra in the back-end. The details are available at: https://wiki.openstack.org/wiki/Neutron/LBaaS/LBaaS%2Band%2BLibra%2Bintegration%2BDraft I looked at the proposal briefly, it makes sense to me. Also it seems to be the simplest way of integrating LBaaS and Libra - create a Libra driver for LBaaS. [Susanne] Yes that would be the short team solution to get us where we need to be. But We do not want to continue to enhance Libra. We would like move to Neutron LBaaS and not have duplicate efforts. While this would allow us to fill a gap short term we would like to discuss the longer term strategy since we believe that everybody would benefit from having such managed services artifacts built directly into Neutron LBaaS. I'm not sure about building it directly into LBaaS, although we can discuss it. [Susanne] The idea behind the managed services aspect/extensions would be reusable for other software LB. For instance, HA is definitely on roadmap and everybody seems to agree that HA should not require user/tenant to do any specific configuration other than choosing HA capability of LBaaS service. So as far as I see it, requirements for HA in LBaaS look very similar to requirements for HA in Libra. [Susanne] Yes. Libra works well for us in the public cloud but we would like to move to Neutron LBaaS and not have duplicate
Re: [openstack-dev] [Neutron][LBaaS] Neutron LBaaS, Libra and managed services
On Tue, Mar 25, 2014 at 9:24 AM, Eugene Nikanorov enikano...@mirantis.comwrote: That for sure can be implemented. I only would recommend to implement such kind of management system out of Neutron/LBaaS tree, e.g. to only have client within Libra driver that will communicate with the management backend. [Susanne] Again this would only be a short term solution since as we move forward and want to contribute new features it would result in duplication of efforts because the features might need to be done in Libra and not Neutron LBaaS. That seems to be a way other vendors are taking right now. Regarding the features, could you point to description of those? Our end goal is to be able to move to just use Neutron LBaaS. For example SSL termination is not in Libra and we don't want to have to implement it when it is already in Neutron LBaaS. the same with L7 policies. Having the service be resilient beyond just a pair of HA proxies is a biggy for us. We cannot expect our customers to manage the LB themselves. Susanne Thanks, Eugene. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Neutron][LBaaS] Neutron LBaaS, Libra and managed services
Hi John, On Tue, Mar 25, 2014 at 7:26 AM, John Dewey j...@dewey.ws wrote: I have a similar concern. The underlying driver may support different functionality, but the differentiators need exposed through the top level API. Not really, whole point of the service is to abstract the user from specifics of backend implementation. So for any feature there is a common API, not specific to any implementation. There probably could be some exception to this guide line that lays in the area of admin API, but that's yet to be discussed. I see the SSL work is well underway, and I am in the process of defining L7 scripting requirements. However, I will definitely need L7 scripting prior to the API being defined. Is this where vendor extensions come into play? I kinda like the route the Ironic guy safe taking with a vendor passthru API. I may say that core team has rejected 'vendor extensions' idea due to potential non-uniform user API experience. That becomes even worse with flavors introduced, because users don't know what vendor is backing up the service they have created. Thanks, Eugene. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [nova] Devstack. Fail to boot an instance if more than 1 network is defined
Anyone else is facing this bug? https://bugs.launchpad.net/nova/+bug/1296808 Thanks Avishay ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Neutron][LBaaS] Neutron LBaaS, Libra and managed services
On Tue, Mar 25, 2014 at 9:36 AM, Eugene Nikanorov enikano...@mirantis.comwrote: Hi John, On Tue, Mar 25, 2014 at 7:26 AM, John Dewey j...@dewey.ws wrote: I have a similar concern. The underlying driver may support different functionality, but the differentiators need exposed through the top level API. Not really, whole point of the service is to abstract the user from specifics of backend implementation. So for any feature there is a common API, not specific to any implementation. There probably could be some exception to this guide line that lays in the area of admin API, but that's yet to be discussed. Admin APIs would make sense. I see the SSL work is well underway, and I am in the process of defining L7 scripting requirements. However, I will definitely need L7 scripting prior to the API being defined. Is this where vendor extensions come into play? I kinda like the route the Ironic guy safe taking with a vendor passthru API. I may say that core team has rejected 'vendor extensions' idea due to potential non-uniform user API experience. That becomes even worse with flavors introduced, because users don't know what vendor is backing up the service they have created. Thanks, Eugene. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [TripleO] Moving tripleo-ci towards the gate
On 24/03/14 22:58, Joe Gordon wrote: On Fri, Mar 21, 2014 at 6:29 AM, Derek Higgins der...@redhat.com mailto:der...@redhat.com wrote: Hi All, I'm trying to get a handle on what needs to happen before getting tripleo-ci(toci) into the gate, I realize this may take some time but I'm trying to map out how to get to the end goal of putting multi node tripleo based deployments in the gate which should cover a lot of uses cases that devstact-gate doesn't. Here are some of the stages I think we need to achieve before being in the gate along with some questions where people may be able to fill in the blanks. Stage 1: check - tripleo projects This is what we currently have running, 5 separate jobs running non voting checks against tripleo projects Stage 2 (a). reliability Obviously keeping the reliability of both the results and the ci system is a must and we should always aim towards 0% false test results, but is there an acceptable number of false negatives for example that would be acceptable to infa, what are the numbers on the gate at the moment? should we aim to match those at the very least (Maybe we already have). And for how long do we need to maintain those levels before considering the system proven? I cannot come up with a specific number for this, perhaps someone else can. I see the results and CI system reliability as two very different things, for the CI system it should ideally never go down for very long (although this is less critical while tripleo is non-voting check only, like all other 3rd party systems). As for false negatives in the results, they should be on par with devstack-gate jobs especially once you start running tempest. Yup, that would seem like a reasonable/fair target. Stage 2 (b). speedup How long can the longest jobs take? We have plans in place to speed up our current jobs but what should the target be? Gate jobs currently take up to a little over an hour [0][1] [0] https://jenkins01.openstack.org/job/check-tempest-dsvm-postgres-full/buildTimeTrend [1] https://jenkins02.openstack.org/job/check-tempest-dsvm-postgres-full/buildTimeTrend Our overcloud job is currently just under 90 minutes, I'm confident we can get below an hour (of course then we have to run tempest and whatever else we add which will bring us back up) 3. More Capacity If you wanted to run tripleo-check everwhere a ''check-tempest-dsvm-full' job is run that is over 600 jobs in a 24 hour period. Looks like I was a little short in my guesstimation and presumably it wont be 600 this time next year [3] graphite http://graphite.openstack.org/render/?from=00%3A00_20140203fgcolor=00title=Check%20Hit%20Count_t=0.2247244759928435height=308bgcolor=ffwidth=586hideGrid=falseuntil=23%3A59_20140324showTarget=color(alias(hitcount(sum(stats.zuul.pipeline.gate.job.gate-tempest-dsvm-neutron-heat-slow.%7BSUCCESS%2CFAILURE%7D)%2C%275hours%27)%2C%20%27gate-tempest-dsvm-neutron-heat-slow%27)%2C%27green%27)_salt=1395701365.817lineMode=staircasetarget=color(alias(hitcount(sum(stats.zuul.pipeline.check.job.check-tempest-dsvm-full.%7BSUCCESS%2CFAILURE%7D)%2C%2724hours%27)%2C%20%27check-tempest-dsvm-full%20hits%20over%2024%20hours%27)%2C%27orange%27) color(alias(hitcount(sum(stats.zuul.pipeline.check.job.check-tempest-dsvm-full.{SUCCESS,FAILURE}),'24hours'), 'check-tempest-dsvm-full hits over 24 hours'),'orange') I'm going to talk about RAM here as its probably the resource where we will hit our infrastructure limits first. Each time a suite of toci jobs is kicked off we currently kick off 5 jobs (which will double once Fedora is added[1]) In total these jobs spawn 15 vm's consuming 80G of RAM (its actually 120G to workaround a bug we will should soon have fixed[2]), we also have plans that will reduce this 80G further but lets stick with it for the moment. Some of these jobs complete after about 30 minutes but lets say our target is an overall average of 45 minutes. With Fedora that means each run will tie up 160G for 45 minutes. Or 160G can provide us with 32 runs (each including 10 jobs) per day So to kick off 500 (I made this number up) runs per day, we would need (500 / 32.0) * 160G = 2500G of RAM We then need to double this number to allow for redundancy, so thats 5000G of RAM We probably have about 3/4 of this available to us at the moment but its not evenly balanced between the 2 clouds so we're not covered from a redundancy point of view. So we need more hardware (either by expanding the clouds we have or added new clouds), I'd like for us to start a separate effort to map out exactly what our medium term goals should be, including o jobs we want to run
Re: [openstack-dev] [Neutron][LBaaS] Neutron LBaaS, Libra and managed services
There are (at least) two ways of expressing differentiation: - through an API extension, visible to the tenant - though an internal policy mechanism, with specific policies inferred from tenant or network characteristics Both have their place. Please don't fall into the trap of thinking that differentiation requires API extension. Sent from my iPhone - please excuse any typos or creative spelling corrections! On Mar 25, 2014, at 1:36 PM, Eugene Nikanorov enikano...@mirantis.com wrote: Hi John, On Tue, Mar 25, 2014 at 7:26 AM, John Dewey j...@dewey.ws wrote: I have a similar concern. The underlying driver may support different functionality, but the differentiators need exposed through the top level API. Not really, whole point of the service is to abstract the user from specifics of backend implementation. So for any feature there is a common API, not specific to any implementation. There probably could be some exception to this guide line that lays in the area of admin API, but that's yet to be discussed. I see the SSL work is well underway, and I am in the process of defining L7 scripting requirements. However, I will definitely need L7 scripting prior to the API being defined. Is this where vendor extensions come into play? I kinda like the route the Ironic guy safe taking with a “vendor passthru” API. I may say that core team has rejected 'vendor extensions' idea due to potential non-uniform user API experience. That becomes even worse with flavors introduced, because users don't know what vendor is backing up the service they have created. Thanks, Eugene. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [nova][PCI] one use case make the flavor/extra-info based solution to be right choice
On Tue, 2014-03-25 at 14:21 +0800, yongli he wrote: 于 2014年03月21日 03:18, Jay Pipes 写道: On Thu, 2014-03-20 at 13:50 +, Robert Li (baoli) wrote: Hi Yongli, I'm very glad that you bring this up and relive our discussion on PCI passthrough and its application on networking. The use case you brought up is: user wants a FASTER NIC from INTEL to join a virtual networking. By FASTER, I guess that you mean that the user is allowed to select a particular vNIC card. Therefore, the above statement can be translated into the following requests for a PCI device: . Intel vNIC . 1G or 10G or ? . network to join First of all, I'm not sure in a cloud environment, a user would care about the vendor or card type. Correct. Nor would/should a user of the cloud know what vendor or card type is in use on a particular compute node. At most, all a user of the cloud would be able to select from is an instance type (flavor) that listed some capability like high_io_networking or something like that, and the mapping of what high_io_networking meant on the back end of Nova would need to be done by the operator (i.e. if the tag high_io_networking is on a flavor a user has asked to launch a server with, then that tag should be translated into a set of capabilities that is passed to the scheduler and used to determine where the instance can be scheduled by looking at which compute nodes support that set of capabilities. This is what I've been babbling about with regards to leaking implementation through the API. What happens if, say, the operator decides to use IBM cards (instead of or in addition to Intel ones)? If you couple the implementation with the API, like the example above shows (user wants a FASTER NIC from INTEL), then you have to add more complexity to the front-end API that a user deals with, instead of just adding a capabilities mapping for new compute nodes that says high_io_networking tag can match to these new compute nodes with IBM cards. Jay thank you, sorry for later reply in this use case, use might so not care about the vendor id/product id. but for a specific image , the product's model(which related to the vendor id/product id) might cared by user. cause the image might could not support new device which possibly use vendor_id and product id to eliminate the unsupported device. anyway, even without the product/vendor id, the multiple extra tag still needed. and consideration this case, some accelerate card for encryption and decryption/hash there are many supported feature, and most likely different pci card might support different feature set,like : md5, DES,3DES, AES, RSA, SHA-x, IDEA, RC4,5,6 the way to select such a device is use it's feature set instead of one or 2 of group, so the extra information about a pci card is need, in a flexible way. Hi Yongli, I'm all for enabling users to take advantage of technology. I just will push to make sure that the public user-facing API hides things like vendor or product information as much as possible. Best, -jay ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [nova][cinder] Refactor ISCSIDriver to support other iSCSI transports besides TCP
Hi, I want to share with the community the following challenge: Currently, Vendors who have their iSCSI driver, and want to add RDMA transport (iSER), cannot leverage their existing plug-in which inherit from iSCSI And must modify their driver or create an additional plug-in driver which inherit from iSER, and copy the exact same code. Instead I believe a simpler approach is to add a new attribute to ISCSIDriver to support other iSCSI transports besides TCP, which will allow minimal changes to support iSER. The existing ISERDriver code will be removed, this will eliminate significant code and class duplication, and will work with all the iSCSI vendors who supports both TCP and RDMA without the need to modify their plug-in drivers. To achieve that both cinder nova requires slight changes: For cinder, I wish to add a parameter called transport (default to iscsi) to distinguish between the transports and use the existing iscsi_ip_address parameter for any transport type connection. For nova, I wish to add a parameter called default_rdma (default to false) to enable initiator side. The outcome will avoid code duplication and the need to add more classes. I am not sure what will be the right approach to handle this, I already have the code, should I open a bug or blueprint to track this issue? Best Regards, Shlomi ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [Neutron][ML2][Ml2Plugin] Setting _original_network in NetworkContext:
Hi All, In the current Ml2Plugin code when 'create_network' is called, as shown below: def create_network(self, context, network) net_data = network['network'] ... session = context.session with session.begin(subtransactions=True): self._ensure_default_security_group(context, tenant_id) result = super(Ml2Plugin, self).create_network(context, network) ... mech_context = driver_context.NetworkContext(self, context, result) self.mechanism_manager.create_network_precommit(mech_context) ... the original_network parameter is not set (the default is None) when instantiating NetworkContext, and as a result the mech_context has only the value of network object returned from super(Ml2Plugin, self).create_network(). This causes issue when a mechanism driver needs to use the original network parameters (given to the create_network), specially when extension is used for the network resources. (The 'result' only has the network attributes without extension which is used to set the '_network' in the NetwrokContext object). Even using extension function registration using db_base_plugin_v2.NeutronDbPluginV2.register_dict_extend_funcs(...) won't help as the network object that is passed to the registered function does not include the extension parameters. Is there any reason that the original_network is not set when initializing the NetworkContext? Would that cause any issue to set it to 'net_data' so that any mechanism driver can use original network parameters as they are available when create_network is called? Appreciate your comments. Thanks, Nader. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Ironic] Nodeless Vendor Passthru API
Hi Russell, Ironic allows drivers to expose a vendor passthru API on a Node. This basically serves two purposes: 1. Allows drivers to expose functionality that hasn't yet been standardized in the Ironic API. For example, the Seamicro driver exposes attach_volume, set_boot_device and set_node_vlan_id passthru methods. 2. Vendor passthru is also used by the PXE deploy driver as an internal RPC callback mechanism. The deploy ramdisk makes calls to the passthru API to signal for a deployment to continue once a server has booted. For the purposes of this discussion I want to focus on case #2. Case #1 is certainly worth a separate discussion - we started this in #openstack-ironic on Friday. In the new agent we are working on, we want to be able to look up what node the agent is running on, and eventually to be able to register a new node automatically. We will perform an inventory of the server and submit that to Ironic, where it can be used to map the agent to an existing Node or to create a new one. Once the agent knows what node it is on, it will check in with a passthru API much like that used by the PXE driver - in some configurations this might trigger an immediate continue of an ongoing deploy, in others it might simply register the agent as available for new deploys in the future. Maybe another way to look up what node the agent is running on would be by looking at the MAC address of that node, having it on hand you could then GET /ports/detail and find which port has that MAC associated with it, after you find the port you can look at the node_uuid field which holds the UUID of the node which that port belongs to (All ports have a node_uuid, it's mandatory). So, right now you would need to list all the ports and find that MAC address there, but I got a review up that might help you with it by allowing you to get a port using its address as input: https://review.openstack.org/#/c/82773/ (GET /ports/detail?address=mac). What you think? The point here is that we need a way for the agent driver to expose a top-level lookup API, which doesn't require a Node UUID in the URL. I've got a review (https://review.openstack.org/#/c/81919/) up which explores one possible implementation of this. It basically routes POSTs to /drivers/driver_name/vendor_passthru/method_name to a new method on the vendor interface. Importantly, I don't believe that this is a useful way for vendors to implement new consumer-facing functionality. If we decide to take this approach, we should reject drivers try to do so. It is intended *only* for internal communications with deploy agents. Another possibility is that we could create a new API service intended explicitly to serve use case #2 described above, which doesn't include most of the existing public paths. In our environment I expect us to allow agents whitelisted access to only two specific paths (lookup and checkin), but this might be a better way to achieve that. Thoughts? Thanks, Russell ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] Hyper-V Meeting
Hi All, We have numerous people travelling today, and therefore we need to cancel the meeting for today. We will resume next week. p We will resume next week. Peter J. Pouliot CISSP Sr. SDET OpenStack Microsoft New England Research Development Center 1 Memorial Drive Cambridge, MA 02142 P: 1.(857).4536436 E: ppoul...@microsoft.commailto:ppoul...@microsoft.com ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [nova][cinder] Refactor ISCSIDriver to support other iSCSI transports besides TCP
Hi Shlomi, Another solution to consider is to create a subclass per transport (iSCSI, iSER) which reference the same shared common code. This is the solution used for the 3PAR iSCSI FC transports. See these for reference: cinder/volume/drivers/san/hp/hp_3par_common.py cinder/volume/drivers/san/hp/hp_3par_fc.py cinder/volume/drivers/san/hp/hp_3par_iscsi.py Hope this helps. Ramy From: Shlomi Sasson [mailto:shlo...@mellanox.com] Sent: Tuesday, March 25, 2014 8:07 AM To: openstack-dev@lists.openstack.org Subject: [openstack-dev] [nova][cinder] Refactor ISCSIDriver to support other iSCSI transports besides TCP Hi, I want to share with the community the following challenge: Currently, Vendors who have their iSCSI driver, and want to add RDMA transport (iSER), cannot leverage their existing plug-in which inherit from iSCSI And must modify their driver or create an additional plug-in driver which inherit from iSER, and copy the exact same code. Instead I believe a simpler approach is to add a new attribute to ISCSIDriver to support other iSCSI transports besides TCP, which will allow minimal changes to support iSER. The existing ISERDriver code will be removed, this will eliminate significant code and class duplication, and will work with all the iSCSI vendors who supports both TCP and RDMA without the need to modify their plug-in drivers. To achieve that both cinder nova requires slight changes: For cinder, I wish to add a parameter called transport (default to iscsi) to distinguish between the transports and use the existing iscsi_ip_address parameter for any transport type connection. For nova, I wish to add a parameter called default_rdma (default to false) to enable initiator side. The outcome will avoid code duplication and the need to add more classes. I am not sure what will be the right approach to handle this, I already have the code, should I open a bug or blueprint to track this issue? Best Regards, Shlomi ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Neutron] OVS 2.1.0 is available but not the Neutron ARP responder
Hi all, As promise, the blog post [1] about running devstack into containers. [1] http://dev.cloudwatt.com/en/blog/running-devstack-into-linux-containers.html Regards, Edouard. Le 21 mars 2014 14:12, Kyle Mestery mest...@noironetworks.com a écrit : Getting this type of functional testing into the gate would be pretty phenomenal. Thanks for your continued efforts here Mathieu! If there is anything I can do to help here, let me know. One other concern here is that the infra team may have issues running a version of OVS which isn't packaged into Ubuntu/CentOS. Keep that in mind as well. Edourard, I look forward to your blog, please share it here once you've written it! Thanks, Kyle On Fri, Mar 21, 2014 at 6:15 AM, Édouard Thuleau thul...@gmail.comwrote: Thanks Mathieu for your support and work onto CI to enable multi-node. I wrote a blog post about how to run devstack development environment with LXC. I hope it will be publish next week. Just add a pointer about OVS support network namespaces since 2 years ago now [1]. [1] http://git.openvswitch.org/cgi-bin/gitweb.cgi?p=openvswitch;a=commitdiff;h=2a4999f3f33467f4fa22ed6e5b06350615fb2dac Regards, Édouard. On Fri, Mar 21, 2014 at 11:31 AM, Mathieu Rohon mathieu.ro...@gmail.comwrote: Hi edouard, thanks for the information. I would love to see your patch getting merged to have l2-population MD fully functional with an OVS based deployment. Moreover, this patch has a minimal impact on neutron, since the code is used only if l2-population MD is used in the ML2 plugin. markmcclain was concerned that no functional testing is done, but L2-population MD needs mutlinode deployment to be tested. A deployment based on a single VM won't create overlay tunnels, which is a mandatory technology to have l2-population activated. The Opensatck-CI is not able, for the moment, to run job based on multi-node deployment. We proposed an evolution of devstack to have a multinode deployment based on a single VM which launch compute nodes in LXC containers [1], but this evolution has been refused by Opensatck-CI since there is other ways to run multinode setup with devstack, and LXC container is not compatible with iscsi and probably ovs [2][3]. One way to have functional test for this feature would be to deploy 3rd party testing environment, but it would be a pity to have to maintain a 3rd party to test some functionalities which are not based on 3rd party equipments. So we are currently learning about the Openstack-CI tools to propose some evolutions to have mutinode setup inside the gate [4]. There are a lot of way to implement it (node-pools evolution, usage of tripleO, of Heat [5]), and we don't know which one would be the easiest, and so the one we have to work on to have the multinode feature available ASAP. This feature looks very important for Neutron, at least to test overlay tunneling. I thinks it's very important for nova too, to test live-migration. [1]https://blueprints.launchpad.net/devstack/+spec/lxc-computes [2]https://bugs.launchpad.net/ubuntu/+source/lxc/+bug/1226855 [3] http://eavesdrop.openstack.org/meetings/infra/2014/infra.2014-02-18-19.01.log.html [4] https://www.mail-archive.com/openstack-infra@lists.openstack.org/msg00968.html [5] http://lists.openstack.org/pipermail/openstack-infra/2013-July/000128.html On Fri, Mar 21, 2014 at 10:08 AM, Édouard Thuleau thul...@gmail.com wrote: Hi, Just to inform you that the new OVS release 2.1.0 was done yesterday [1]. This release contains new features and significant performance improvements [2]. And in that new features, one [3] was use to add local ARP responder with OVS agent and the plugin ML2 with the MD l2-pop [4]. Perhaps, it's time to reconsider that review? [1] https://www.mail-archive.com/discuss@openvswitch.org/msg09251.html [2] http://openvswitch.org/releases/NEWS-2.1.0 [3] http://git.openvswitch.org/cgi-bin/gitweb.cgi?p=openvswitch;a=commitdiff;h=f6c8a6b163af343c66aea54953553d84863835f7 [4] https://review.openstack.org/#/c/49227/ Regards, Édouard. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org
Re: [openstack-dev] [Mistral][TaskFlow] Long running actions
On 03/21/2014 12:33 AM, W Chan wrote: Can the long running task be handled by putting the target task in the workflow in a persisted state until either an event triggers it or timeout occurs? An event (human approval or trigger from an external system) sent to the transport will rejuvenate the task. The timeout is configurable by the end user up to a certain time limit set by the mistral admin. Based on the TaskFlow examples, it seems like the engine instance managing the workflow will be in memory until the flow is completed. Unless there's other options to schedule tasks in TaskFlow, if we have too many of these workflows with long running tasks, seems like it'll become a memory issue for mistral... Look into the Trusts capability of Keystone for Authorization support on long running tasks. On Thu, Mar 20, 2014 at 3:07 PM, Dmitri Zimine d...@stackstorm.com mailto:d...@stackstorm.com wrote: For the 'asynchronous manner' discussion see http://tinyurl.com/n3v9lt8; I'm still not sure why u would want to make is_sync/is_async a primitive concept in a workflow system, shouldn't this be only up to the entity running the workflow to decide? Why is a task allowed to be sync/async, that has major side-effects for state-persistence, resumption (and to me is a incorrect abstraction to provide) and general workflow execution control, I'd be very careful with this (which is why I am hesitant to add it without much much more discussion). Let's remove the confusion caused by async. All tasks [may] run async from the engine standpoint, agreed. Long running tasks - that's it. Examples: wait_5_days, run_hadoop_job, take_human_input. The Task doesn't do the job: it delegates to an external system. The flow execution needs to wait (5 days passed, hadoob job finished with data x, user inputs y), and than continue with the received results. The requirement is to survive a restart of any WF component without loosing the state of the long running operation. Does TaskFlow already have a way to do it? Or ongoing ideas, considerations? If yes let's review. Else let's brainstorm together. I agree, that has major side-effects for state-persistence, resumption (and to me is a incorrect abstraction to provide) and general workflow execution control, I'd be very careful with this But these requirement comes from customers' use cases: wait_5_day - lifecycle management workflow, long running external system - Murano requirements, user input - workflow for operation automations with control gate checks, provisions which require 'approval' steps, etc. DZ ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org mailto:OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [OpenStack-Dev] [Cinder FFE] Request for HDS FFE
On 03/25/2014 10:42 AM, Steven Sonnenberg wrote: I just want to point out, there were no changes required to pass the tests. We were running those tests in Brazil and tunneling NFS and iSCSI across the Internet which explain timeout issues. Those are the same tests that passed a month earlier before we went into the cycle of review/fix/format etc. I think the key point is the current timing. We're aiming to do RC1 for projects this week if possible. FFEs were really only allowed weeks ago. -- Russell Bryant ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [ceilometer][all] persisting dump tables after migration
in ceilometer we have a bug regarding residual dump tables left after migration: https://bugs.launchpad.net/ceilometer/+bug/1259724 basically, in a few prior migrations, when adding missing constraints, a dump table was create to backup values which didn't fit into the new constraints. i raised the initial bug because i believe there is very little value to these tables as i would expect any administrator capturing data of some importance would backup their data before any migration to begin with. i noticed in Nova, they also clean up their dump tables but i wanted to raise this on mailing list so that everyone is aware of the issue before i add a patch which blows away these dump tables. :) i'd be interested if anyone actually finds value in having these dump tables persist just so we can see if your use case can be handle without the tables. for reference, the dump tables are created in: https://github.com/openstack/ceilometer/blob/master/ceilometer/storage/sqlalchemy/migrate_repo/versions/012_add_missing_foreign_keys.py https://github.com/openstack/ceilometer/blob/master/ceilometer/storage/sqlalchemy/migrate_repo/versions/027_remove_alarm_fk_constraints.py cheers, gordon chung openstack, ibm software standards___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Ironic][Keystone] Move drivers credentials to Keystone
On Tue, 2014-03-25 at 12:23 +, Lucas Alvares Gomes wrote: Hi, Right now Ironic is being responsible for storing the credentials for the IPMI and SSH drivers (and potentially other drivers in the future), I wonder if we should delegate this task to Keystone. The Keystone V3 API now has a /credentials endpoint which would allow us to specify arbitrary types (not only ec2 anymore) and use it as a credential store[1]. That would avoid further fragmentation of credentials being stored in different places in OpenStack, and make the management of the credentials easier (Think about a situation where many nodes share the same IPMI username/password and we need to update it, if this is stored in Keystone it only needs to be updated there once cause nodes will only hold a reference to it) It also was pointed to me that setting a hard dependency on Keystone V3 might significantly raises the bar for integration with existing clouds*. So perhaps we should make it optional? In the same way we can specify a username/password or key_filename for the ssh driver we could have a reference to a credential in Keystone V3? I think the idea of using Keystone for keypair management in Nova is a good one. There is already precedent in Nova for doing this kind of thing ... it's already been done for images, volumes, and network. One problem with the Keystone v3 credentials API, though, is that it does not have support for unique names of keypairs per project, as that is how the Nova API /keypairs resource endpoint works. What you guys think about the idea? What are the cloud operators/sysadmins view on that? As long as the functionality was enabled using the standard driver-based setup (as was done for glance, nova, cinder, and neutron integration), I don't see any issues for deployers. Of course, you'd need a migration script, but that's not a huge deal. Best, -jay ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Murano][Heat] MuranoPL questions?
Murano folks, I guess we should stop our debates with the Heat team about MuranoPL. What we should do instead is to carefully read this thread couple of times and collect and process all the feedback we've got. Steve and Zane did a very good job helping us to find a way to align with Heat and Orhcestration program. Let me outline the most important (imho) quotes: # START quotes Steven Dake: I see no issue with HOT remaining simple and tidy focused entirely on orchestration (taking a desired state and converting that into reality) with some other imperative language layered on top to handle workflow and ALM. I believe this separation of concerns is best for OpenStack and should be the preferred development path. Zane Bitter: I do think we should implement the hooks I mentioned at the start of this thread to allow tighter integration between Heat and a workflow engine (i.e. Mistral). So building a system on top of Heat is good. Building it on top of Mistral as well would also be good, and that was part of the feedback from the TC. To me, building on top means building on top of the languages (which users will have to invest a lot of work in learning) as well, rather than having a completely different language and only using the underlying implementation(s). To me that implies that Murano should be a relatively thin wrapper that ties together HOT and Mistral's DSL. Steve Dake: --- I don't think HOT can do these things and I don't think we want HOT to do these things. I am ok with that, since I don't see the pushback on having two languages for two different things in OpenStack. I got from gokrove on iRC today that the rationale for the pushback was the TC wanted Murano folks to explore how to integrate better with Heat and possibly the orchestration program. I don't see HOT as a place where there is an opportunity for scope expansion. I see instead Murano creating HOT blobs and feeding them to Heat. Zane Bitter: Because there have to be some base types that you can use as building blocks. In the case of Heat, those base types are the set of things that you can create by talking to OpenStack APIs authenticated with the user's token. In the case of Mistral, I would expect it to be the set of actions that you can take by talking to OpenStack APIs authenticated with the user's token. And in the case of Murano, I would expect it to be the union of those two. Everything is a combination of existing resources, because the set of existing resources is the set of things which the operator provides as-a-Service. The set of things that the operator provides as a service plus the set of things that you can implement yourself on your own server (virtual or not) covers the entire universe of things. What you appear to be suggesting is that OpenStack must provide *Everything*-as-a-Service by allowing users to write their own services and have the operator execute them as-a-Service. This would be a breathtakingly ambitious undertaking, and I don't mean that in a good way. When the TC said Murano is slightly too far up the stack at this point to meet the measured progression of openstack as a whole requirement, IMO one of the major things they meant was that you're inventing your own workflow thing, leading to duplication of effort between this and Workflow as a Service. (And Mistral folks are in turn doing the same thing by not using the same workflow library, taskflow, as the rest of OpenStack.) Candidly, I'm surprised that this is not the #1 thing on your priority list because IMO it's the #1 thing that will delay getting the project incubated. # END quotes Also I should quote what Georgy Okrokvertkhov said: Having this aligned I see a Murano package as an archive with all necessary definitions and resources and Murano service will just properly pass them to related services like Heat and Mistral. I think then Murano DSL will be much more simple and probably will be closer to declarative format with some specific data operations. Summary: I would like to propose further path for Murano evolution where it evolves as an OpenStack project, aligned with others and developed in agreement between all the interested sides. Here is the plan: * Step forward and implement hooks (for workflow customization) in Heat. We will register a blueprint, discuss it with the Heat team and implement it * Use our cross-project session on ATL summit to set clear goals and expectations * Build pilot in Murano which: a. Uses new HOT Software components to do VM side software deployments b. Uses Mistral DSL to describe workflow. It'll require more focused discussion with Mistral team c. Bundles all that into an application package using new Murano DSL d. Allows users to combine applications in a single environenment * Continuously align with the Heat team * Continuously re-evaluate current state of pilot
Re: [openstack-dev] [Ironic][Keystone] Move drivers credentials to Keystone
Why not use Barbican? It stores credentials after encrypting them. -Original Message- From: Jay Pipes [mailto:jaypi...@gmail.com] Sent: Tuesday, March 25, 2014 9:50 AM To: openstack-dev@lists.openstack.org Subject: Re: [openstack-dev] [Ironic][Keystone] Move drivers credentials to Keystone On Tue, 2014-03-25 at 12:23 +, Lucas Alvares Gomes wrote: Hi, Right now Ironic is being responsible for storing the credentials for the IPMI and SSH drivers (and potentially other drivers in the future), I wonder if we should delegate this task to Keystone. The Keystone V3 API now has a /credentials endpoint which would allow us to specify arbitrary types (not only ec2 anymore) and use it as a credential store[1]. That would avoid further fragmentation of credentials being stored in different places in OpenStack, and make the management of the credentials easier (Think about a situation where many nodes share the same IPMI username/password and we need to update it, if this is stored in Keystone it only needs to be updated there once cause nodes will only hold a reference to it) It also was pointed to me that setting a hard dependency on Keystone V3 might significantly raises the bar for integration with existing clouds*. So perhaps we should make it optional? In the same way we can specify a username/password or key_filename for the ssh driver we could have a reference to a credential in Keystone V3? I think the idea of using Keystone for keypair management in Nova is a good one. There is already precedent in Nova for doing this kind of thing ... it's already been done for images, volumes, and network. One problem with the Keystone v3 credentials API, though, is that it does not have support for unique names of keypairs per project, as that is how the Nova API /keypairs resource endpoint works. What you guys think about the idea? What are the cloud operators/sysadmins view on that? As long as the functionality was enabled using the standard driver-based setup (as was done for glance, nova, cinder, and neutron integration), I don't see any issues for deployers. Of course, you'd need a migration script, but that's not a huge deal. Best, -jay ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Ironic][Keystone] Move drivers credentials to Keystone
On Tue, 2014-03-25 at 17:39 +, Miller, Mark M (EB SW Cloud - RD - Corvallis) wrote: Why not use Barbican? It stores credentials after encrypting them. No reason not to add a Barbican driver as well. Best, -jay ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Murano][Heat][TC] MuranoPL questions?
On 03/25/2014 09:55 AM, Ruslan Kamaldinov wrote: Murano folks, I guess we should stop our debates with the Heat team about MuranoPL. What we should do instead is to carefully read this thread couple of times and collect and process all the feedback we've got. Steve and Zane did a very good job helping us to find a way to align with Heat and Orhcestration program. Let me outline the most important (imho) quotes: Ruslan, +2! The plan forward at the end of your email seem reasonable as well, but Zane is probably in a bit better position to clarify if any suggested changes would be recommended. As a heat-core member, I would be supportive of ALM scope expansion happening within the Orchestration program. I also think from a technical perspective ALM fits nicely within the scope of OpenStack. Just to head off possible questions about ALM and Solum, I do not see them as competitive. I am only one voice among many, and think we need some broad general agreement with heat-core and the TC that suggested scope expansion makes sense and we won't get pushback 6-12 months down the road. Perhaps the TC may think it belongs in a new program. I honestly don't know. I'm not quite sure how to get that conversation started, so I added [TC] to the subject tags in an attempt to get the ball rolling. Regards, -steve # START quotes Steven Dake: I see no issue with HOT remaining simple and tidy focused entirely on orchestration (taking a desired state and converting that into reality) with some other imperative language layered on top to handle workflow and ALM. I believe this separation of concerns is best for OpenStack and should be the preferred development path. Zane Bitter: I do think we should implement the hooks I mentioned at the start of this thread to allow tighter integration between Heat and a workflow engine (i.e. Mistral). So building a system on top of Heat is good. Building it on top of Mistral as well would also be good, and that was part of the feedback from the TC. To me, building on top means building on top of the languages (which users will have to invest a lot of work in learning) as well, rather than having a completely different language and only using the underlying implementation(s). To me that implies that Murano should be a relatively thin wrapper that ties together HOT and Mistral's DSL. Steve Dake: --- I don't think HOT can do these things and I don't think we want HOT to do these things. I am ok with that, since I don't see the pushback on having two languages for two different things in OpenStack. I got from gokrove on iRC today that the rationale for the pushback was the TC wanted Murano folks to explore how to integrate better with Heat and possibly the orchestration program. I don't see HOT as a place where there is an opportunity for scope expansion. I see instead Murano creating HOT blobs and feeding them to Heat. Zane Bitter: Because there have to be some base types that you can use as building blocks. In the case of Heat, those base types are the set of things that you can create by talking to OpenStack APIs authenticated with the user's token. In the case of Mistral, I would expect it to be the set of actions that you can take by talking to OpenStack APIs authenticated with the user's token. And in the case of Murano, I would expect it to be the union of those two. Everything is a combination of existing resources, because the set of existing resources is the set of things which the operator provides as-a-Service. The set of things that the operator provides as a service plus the set of things that you can implement yourself on your own server (virtual or not) covers the entire universe of things. What you appear to be suggesting is that OpenStack must provide *Everything*-as-a-Service by allowing users to write their own services and have the operator execute them as-a-Service. This would be a breathtakingly ambitious undertaking, and I don't mean that in a good way. When the TC said Murano is slightly too far up the stack at this point to meet the measured progression of openstack as a whole requirement, IMO one of the major things they meant was that you're inventing your own workflow thing, leading to duplication of effort between this and Workflow as a Service. (And Mistral folks are in turn doing the same thing by not using the same workflow library, taskflow, as the rest of OpenStack.) Candidly, I'm surprised that this is not the #1 thing on your priority list because IMO it's the #1 thing that will delay getting the project incubated. # END quotes Also I should quote what Georgy Okrokvertkhov said: Having this aligned I see a Murano package as an archive with all necessary definitions and resources and Murano service will just properly pass them to related services like Heat and Mistral. I think then Murano DSL will be much more simple and probably will be closer to declarative
Re: [openstack-dev] [Neutron] Using Python-Neutronclient from Python - docstrings needed?
On Fri, Mar 21, 2014 at 08:35:05PM EDT, Rajdeep Dua wrote: Sean, If you can point me to the project file in github which needs to be modified , i will include these docs Thanks Rajdeep I imagine inside the openstack-manuals git repo https://github.com/openstack/openstack-manuals Possibly inside the doc/user-guide tree. Although others may have better suggestions. -- Sean M. Collins ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Neutron][LBaaS]
Hi Oleg! Thanks for the confirmation, and for the link to the services in service VM blueprint link. I went through it and while the concept of using service VMs is one way to go (cloudstack uses the approach, for example), I am not inclined to take that approach for some reasons. Having a service image essentially means forcing customers/deployers to either go with a specific OS version or to maintain multiple formats (qcow/ova/vhd etc) for different hypervisors. It will mean putting in packaging code into openstack to build these VMs with every build. Upgradation from one Openstack release to the next will sometime eventually force upgradation of these service VMs as well. We will need to track service VM versions in glance or elsewhere and prevent deployment across versions if required. Since we're talking about VMs and not just modules, downtime also increases during upgrades. It adds a whole level of complexity to handle all these scenarios, and we will need to do that in Openstack code. I am very averse to doing that. Also as an aside, when service VMs come into picture, deployment (depending on the hypervisor) can take quite a long time (and scalability issues can be hit). Take VMWare for example. We can create full or linked clones on it. Full clone creation is quite costly in it. Storage migration is another area. The issues will add up with time. From a developer's pov, debugging can get tricky at times with service VMs. The issues brought up in the blueprint are definitely valid and must be addressed, but we'll need to have a discussion on what the optimal and cleanest approach would be. There is another interesting discussion going on regarding LBaaS APIs for Libra and HAProxy between Susanne Balle/Eugene and others - I'll chip in with my 2 cents on that.. Thanks a lot again! Regards, Vijay On Tue, Mar 25, 2014 at 1:02 AM, Oleg Bondarev obonda...@mirantis.comwrote: Hi Vijay, Currently Neutron LBaaS supports only namespace based implementation for HAProxy. You can however run LBaaS agent on the host other than network controller node - in that case HAProxy processes will be running on that host but still in namespaces. Also there is an effort in Neutron regarding adding support of advanced services in VMs [1]. After it is completed I hope it will be possible to adopt it in LBaaS and run HAProxy in such a service VM. [1] https://blueprints.launchpad.net/neutron/+spec/adv-services-in-vms Thanks, Oleg On Tue, Mar 25, 2014 at 1:39 AM, Vijay B os.v...@gmail.com wrote: Hi Eugene, Thanks for the reply! How/where is the agent configuration done for HAProxy? If I don't want to go with a network namespace based HAProxy process, but want to deploy my own HAProxy instance on a host outside of the network controller node, and make neutron deploy pools/VIPs on that HAProxy instance, does neutron currently support this scenario? If so, what are the configuration steps I will need to carry out to deploy HAProxy on a separate host (for example, where do I specify the ip address of the haproxy host, etc)? Regards, Vijay On Mon, Mar 24, 2014 at 2:04 PM, Eugene Nikanorov enikano...@mirantis.com wrote: Hi, HAProxy driver has not removed from the trunk, instead it became a base for agent-based driver, so the only haproxy-specific thing in the plugin driver is device driver name. Namespace driver is a device driver on the agent side and it was there from the beginning. The reason for the change is mere refactoring: it seems that solutions that employ agents could share the same code with only device driver being specific. So, everything is in place, HAProxy continues to be the default implementation of Neutron LBaaS service. It supports spawning haproxy processes on any host that runs lbaas agent. Thanks, Eugene. On Tue, Mar 25, 2014 at 12:33 AM, Vijay B os.v...@gmail.com wrote: Hi, I'm looking at HAProxy support in Neutron, and I observe that the drivers/haproxy/plugin_driver.py file in the stable/havana release has been effectively removed from trunk (master), in that the plugin driver in the master simply points to the namespace driver. What was the reason to do this? Was the plugin driver in havana tested and documented? I can't seem to get hold of any relevant documentation that describes how to configure HAProxy LBs installed on separate boxes (and not brought up in network namespaces) - can anyone please point me to the same? Also, are there any plans to bring back the HAProxy plugin driver to talk to remote HAProxy instances? Thanks, Regards, Vijay ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Ironic][Keystone] Move drivers credentials to Keystone
Yes, this is exactly the use case we’re trying to address with Barbican. I think this is something that definitely belongs in Barbican, especially now that we are an incubated project. We’d love to help out with any integration questions you may have. -Doug Mendizabal On 3/25/14, 12:49 PM, Jay Pipes jaypi...@gmail.com wrote: On Tue, 2014-03-25 at 17:39 +, Miller, Mark M (EB SW Cloud - RD - Corvallis) wrote: Why not use Barbican? It stores credentials after encrypting them. No reason not to add a Barbican driver as well. Best, -jay ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev smime.p7s Description: S/MIME cryptographic signature ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] Requirements for dropping / skipping a Tempest test
Because this has come up a few times during freeze, it's probably worth actively advertising the policy we've developed on dropping / skipping a Tempest test. Tempest tests encode the behavior of the system (to some degree), which means that once we know the behavior of the system, if code in a core project can only land if we skip or drop a tempest test, that's clearly a behavior change. We want to be really deliberate about this, so in these situations we require: * an extremely clear commit message about why this is required (and why backwards compatible behavior is not an option) * a link in the commit message to the dependent review (you can just put the idempotent change id in there) * a +2 on the dependent review in the project by a core member The 3rd part is important, because incompatible behavior changes are something we want to make the core team for any given project is sure they want to do before we drop or skip a test and let those changes come into the project. This system isn't perfect, but it's trying to strike a balance. My hope is in the future we could implement cross project dependencies in zuul so that you could gather test results for a project change - assuming the tempest change was applied, that would prevent all these changes from being -1ed until the tempest change is landed. -Sean -- Sean Dague Samsung Research America s...@dague.net / sean.da...@samsung.com http://dague.net signature.asc Description: OpenPGP digital signature ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [Neutron][IPv6] Neutron Routers and LLAs
During the review[0] of the patch that only allows RAs from known addresses, Robert Li brought up a bug in Neutron, where a IPv6 Subnet could be created, with a link local address for the gateway, that would fail to create the Neutron router because the IP address that the router's port would be assigned, was a link local address that was not on the subnet. This may or may have been run before force_gateway_on_subnet flag was introduced. Robert - if you can give us what version of Neutron you were running that would be helpful. Here's the full text of what Robert posted in the review, which shows the bug, which was later filed[1]. This is what I've tried, creating a subnet with a LLA gateway address: neutron subnet-create --ip-version 6 --name myipv6sub --gateway fe80::2001:1 mynet :::/64 Created a new subnet: +--++ | Field | Value | +--++ | allocation_pools | {start: :::1, end: ::::::fffe} | | cidr | :::/64 | | dns_nameservers | | | enable_dhcp | True | | gateway_ip | fe80::2001:1 | | host_routes | | | id | a1513aa7-fb19-4b87-9ce6-25fd238ce2fb | | ip_version | 6 | | name | myipv6sub | | network_id | 9c25c905-da45-4f97-b394-7299ec586cff | | tenant_id | fa96d90f267b4a93a5198c46fc13abd9 | +--++ openstack@devstack-16:~/devstack$ neutron router-list +--+-+-+ | id | name | external_gateway_info | +--+-+-+ | 7cf084b4-fafd-4da2-9b15-0d25a3e27e67 | router1 | {network_id: 02673c3c-35c3-40a9-a5c2-9e5c093aca48, enable_snat: true} | +--+-+-+ openstack@devstack-16:~/devstack$ neutron router-interface-add 7cf084b4-fafd-4da2-9b15-0d25a3e27e67 myipv6sub 400-{u'NeutronError': {u'message': u'Invalid input for operation: IP address fe80::2001:1 is not a valid IP for the defined subnet.', u'type': u'InvalidInput', u'detail': u''}} During last week's meeting, we had a bit of confusion near the end of the meeting[2] about the following bug, and the fix[3]. If I am not mistaken - the fix is so that when you create a v6 Subnet with a link local address, then create a Neutron router to serve as the gateway for that subnet - the operation will successfully complete and a router will be created. We may need to take a look at the code that create a router - to ensure that only one gateway port is created, and that the link local address from the subnet's 'gateway' attribute is used as the address. This is at least my understanding of the problem as it stands today - and that this bug and fix does not involve any external gateways or physical devices that Neutron does not control - this is exclusively about Neutron routers. [0]: https://review.openstack.org/#/c/72252/ [1]: https://bugs.launchpad.net/neutron/+bug/1284518 [2]: http://eavesdrop.openstack.org/meetings/neutron_ipv6/2014/neutron_ipv6.2014-03-18-14.02.log.html [3]: https://review.openstack.org/#/c/76125/ -- Sean M. Collins ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [OpenStack-Dev] [Cinder FFE] Request for HDS FFE
On Tue, Mar 25, 2014 at 10:19 AM, Russell Bryant rbry...@redhat.com wrote: On 03/25/2014 10:42 AM, Steven Sonnenberg wrote: I just want to point out, there were no changes required to pass the tests. We were running those tests in Brazil and tunneling NFS and iSCSI across the Internet which explain timeout issues. Those are the same tests that passed a month earlier before we went into the cycle of review/fix/format etc. I think the key point is the current timing. We're aiming to do RC1 for projects this week if possible. FFEs were really only allowed weeks ago. -- Russell Bryant ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev Hey Steven, As we discussed last night, my issue is that it's just very late at this point. Yes, your patch has been in review for 45 days, however even then it was pushing it. By the way 45 days at this stage of the release is not a long time. My question is if people have been running this since HongKong why did you wait until February before submitting it? It is minimal risk to the core code, no doubt. However, I've taking a pretty hard stance with other 3'rd party drivers that have missed the cut off dates and I don't see a compelling reason to make an exception here based just on principal. I'd also like to point out that contributions help when it comes to FFE's. In other words I don't see any activity other than this driver for the last 4 months (reviews or otherwise). When somebody comes in with a patch past a date and asks for an exception the first thing I consider is whether they've been active for the cycle or if they're just racing the clock to get a driver in for the next release. Something else I consider is if current code is maintained, in other words you have a driver in the code base currently and it hasn't been maintained since August (again last minute fixes before RC). Now during RC you have another driver that you want added. If there was active involvement and maintenance of the code and I didn't see a pattern here (pattern of late/last minute submission) I likely would have a different opinion. I'm still a -1 on the exception, even if it inherently doesn't introduce significant risk. It's not the code or the driver itself at this point but the point regarding dates, process etc. [1] History of maintenance for existing HDS code in Cinder [2] Commit history for Erlon (author of the current patch/driver) [3] Commit history for the author of the previous driver including the last minute updates Thanks, John [1]: https://github.com/openstack/cinder/commits/master/cinder/volume/drivers/hds [2]: https://review.openstack.org/#/dashboard/10058 [3]: https://review.openstack.org/#/dashboard/7447 ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Swift] Swift storage policies in Icehouse
As a quick review, storage policies allow objects to be stored across a particular subset of hardware...and with a particular storage algorithm Having worked on backup software in the past, this sounds interesting. :D What is the scope of these policies? Are they per-object, per-container, and/or per-project? Or do they not work like that? --- Kurt G. | @kgriffs ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Swift] Swift storage policies in Icehouse
On Mar 25, 2014, at 12:11 PM, Kurt Griffiths kurt.griffi...@rackspace.com wrote: As a quick review, storage policies allow objects to be stored across a particular subset of hardware...and with a particular storage algorithm Having worked on backup software in the past, this sounds interesting. :D What is the scope of these policies? Are they per-object, per-container, and/or per-project? Or do they not work like that? A storage policy is set on a container when it is created. So, for example, create your photos container with a global 3-replica scheme and also a thumbnails-west with 2 replicas in your West Coast region and thumbnails-east with 2 replicas in your East Coast region. Then make a container for server-backups that is erasure coded and stored in the EU. And all of that is stored and managed in the same logical Swift cluster. So you can see that this feature set gives deployers and users a ton of flexibility. How will storage policies be exposed? I'm glad you asked... A deployer (ie the cluster operator) will configure the storage policies (including which is the default). At that point, an end-user can create containers with a particular storage policy and start saving objects there. What about automatically moving data between storage policies? This is something that is explicitly not in scope for this set of work. Maybe someday, but in the meantime, I fully expect the Swift ecosystem to create and support tools to help manage data lifecycle management. For now, that doesn't belong in Swift. --John signature.asc Description: Message signed with OpenPGP using GPGMail ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Ironic] Nodeless Vendor Passthru API
On Tue, Mar 25, 2014 at 6:56 AM, Lucas Alvares Gomes lucasago...@gmail.comwrote: Hi Russell, Ironic allows drivers to expose a vendor passthru API on a Node. This basically serves two purposes: 1. Allows drivers to expose functionality that hasn't yet been standardized in the Ironic API. For example, the Seamicro driver exposes attach_volume, set_boot_device and set_node_vlan_id passthru methods. 2. Vendor passthru is also used by the PXE deploy driver as an internal RPC callback mechanism. The deploy ramdisk makes calls to the passthru API to signal for a deployment to continue once a server has booted. For the purposes of this discussion I want to focus on case #2. Case #1 is certainly worth a separate discussion - we started this in #openstack-ironic on Friday. In the new agent we are working on, we want to be able to look up what node the agent is running on, and eventually to be able to register a new node automatically. We will perform an inventory of the server and submit that to Ironic, where it can be used to map the agent to an existing Node or to create a new one. Once the agent knows what node it is on, it will check in with a passthru API much like that used by the PXE driver - in some configurations this might trigger an immediate continue of an ongoing deploy, in others it might simply register the agent as available for new deploys in the future. Maybe another way to look up what node the agent is running on would be by looking at the MAC address of that node, having it on hand you could then GET /ports/detail and find which port has that MAC associated with it, after you find the port you can look at the node_uuid field which holds the UUID of the node which that port belongs to (All ports have a node_uuid, it's mandatory). So, right now you would need to list all the ports and find that MAC address there, but I got a review up that might help you with it by allowing you to get a port using its address as input: https://review.openstack.org/#/c/82773/ (GET /ports/detail?address=mac). What you think? Right, we discussed this possibility as well. Internally that's actually how the first iteration of our lookup call will work (mapping MAC addresses to ports to nodes). This can definitely be made to work, but in my mind it has a few limitations: 1. It limits how we can do lookups. In the future I'd like to be able to consider serial numbers, hardware profiles, etc when trying to map an agent to a node. Needing to expose an API for each of these is going to be infeasible. 2. It limits how we do enrollment. 3. It forces us to expose more of the API to agents. Items 1 and 2 seem like things that someone deploying Ironic is especially likely to want to customize. For example, I might have some business process around replacing NICs where I want to customize how a server is uniquely identified (or even hit an external source to do it), or I might want want to override enrollment to hook into Fuel. While we can probably find a way to solve our current problem, how can we generically solve the need for an agent to talk to a driver (either in order to centralize orchestration, or because access to the database is needed) without needing a Node UUID? Thanks, Russell ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [nova] should there be an audit to clear the REBOOTING task_state?
I've reported a bug (https://bugs.launchpad.net/nova/+bug/1296967) where we got stuck with a task_state of REBOOTING due to what seem to be RPC issues. Regardless of how we got there, currently there is no audit that will clear the task_state if it gets stuck. Because of this, once we got into this state it required manual action to get out of it. This doesn't seem like a good design...is there a way to do some sort of audit to ensure that the task_state is actually valid? Chris ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Neutron][LBaaS] addition to requirement wiki
Thanks Itsuro, Good requirement since Neutron LBaaS is an asynchronous API. Cheers, --Jorge On 3/24/14 7:27 PM, Itsuro ODA o...@valinux.co.jp wrote: Hi LBaaS developpers, I added 'Status Indication' to requirement Wiki. It may be independent from object model discussion but I think this is an item which should not be forgotten. Thanks. -- Itsuro ODA o...@valinux.co.jp ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [python-openstacksdk] Meeting Minutes - Tuesday 25 March
Thanks for another good meeting. More code coming along, and a few reviews out there to take a look at. Minutes: http://eavesdrop.openstack.org/meetings/python_openstacksdk/2014/python_openstacksdk.2014-03-25-19.01.html Minutes (text): http://eavesdrop.openstack.org/meetings/python_openstacksdk/2014/python_openstacksdk.2014-03-25-19.01.txt Log: http://eavesdrop.openstack.org/meetings/python_openstacksdk/2014/python_openstacksdk.2014-03-25-19.01.log.html We're all in #openstack-sdks in Freenode ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [marconi] Performance numbers
?Hello, I wonder if any performance measurements have been done with Marconi? Are there results available somewhere? I am generally trying to set my expectations in terms of latency and throughput as a function of the size of the deployment, number of queues, number of producers/consumers, type of backend, size of backend cluster, guarantees, message size etc. Trying to understand how Marconi would do in a large multi-tenant deployment.? Any and all data points would be helpful. Thanks, Tomasz Janczuk ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [TripleO][reviews] We're falling behind
TripleO has just seen an influx of new contributors. \o/. Flip side - we're now slipping on reviews /o\. In the meeting today we had basically two answers: more cores, and more work by cores. We're slipping by 2 reviews a day, which given 16 cores is a small amount. I'm going to propose some changes to core in the next couple of days - I need to go and re-read a bunch of reviews first - but, right now we don't have a hard lower bound on the number of reviews we request cores commit to (on average). We're seeing 39/day from the 16 cores - which isn't enough as we're falling behind. Thats 2.5 or so. So - I'd like to ask all cores to commit to doing 3 reviews a day, across all of tripleo (e.g. if your favourite stuff is all reviewed, find two new things to review even if outside comfort zone :)). And we always need more cores - so if you're not a core, this proposal implies that we'll be asking that you a) demonstrate you can sustain 3 reviews a day on average as part of stepping up, and b) be willing to commit to that. Obviously if we have enough cores we can lower the minimum commitment - so I don't think this figure should be fixed in stone. And now - time for a loose vote - who (who is a tripleo core today) supports / disagrees with this proposal - lets get some consensus here. I'm in favour, obviously :), though it is hard to put reviews ahead of direct itch scratching, its the only way to scale the project. -Rob -- Robert Collins rbtcoll...@hp.com Distinguished Technologist HP Converged Cloud ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [nova][scheduler] Availability Zones and Host aggregates..
Hi, The availability Zones filter states that theoretically a compute node can be part of multiple availability zones. I have a requirement where I need to make a compute node part to 2 AZ. When I try to create a host aggregates with AZ I can not add the node in two host aggregates that have AZ defined. However if I create a host aggregate without associating an AZ then I can add the compute nodes to it. After doing that I can update the host-aggregate an associate an AZ. This looks like a bug. I can see the compute node to be listed in the 2 AZ with the availability-zone-list command. The problem that I have is that I can still not boot a VM on the compute node when I do not specify the AZ in the command though I have set the default availability zone and the default schedule zone in nova.conf. I get the error “ERROR: The requested availability zone is not available” What I am trying to achieve is have two AZ that the user can select during the boot but then have a default AZ which has the HV from both AZ1 AND AZ2 so that when the user does not specify any AZ in the boot command I scatter my VM on both the AZ in a balanced way. Any pointers. Thanks, Sangeeta ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [TripleO][reviews] We're falling behind
Hey +1 3 a day seems pretty sustainable. Cheers, -- Chris Jones On 25 Mar 2014, at 20:17, Robert Collins robe...@robertcollins.net wrote: TripleO has just seen an influx of new contributors. \o/. Flip side - we're now slipping on reviews /o\. In the meeting today we had basically two answers: more cores, and more work by cores. We're slipping by 2 reviews a day, which given 16 cores is a small amount. I'm going to propose some changes to core in the next couple of days - I need to go and re-read a bunch of reviews first - but, right now we don't have a hard lower bound on the number of reviews we request cores commit to (on average). We're seeing 39/day from the 16 cores - which isn't enough as we're falling behind. Thats 2.5 or so. So - I'd like to ask all cores to commit to doing 3 reviews a day, across all of tripleo (e.g. if your favourite stuff is all reviewed, find two new things to review even if outside comfort zone :)). And we always need more cores - so if you're not a core, this proposal implies that we'll be asking that you a) demonstrate you can sustain 3 reviews a day on average as part of stepping up, and b) be willing to commit to that. Obviously if we have enough cores we can lower the minimum commitment - so I don't think this figure should be fixed in stone. And now - time for a loose vote - who (who is a tripleo core today) supports / disagrees with this proposal - lets get some consensus here. I'm in favour, obviously :), though it is hard to put reviews ahead of direct itch scratching, its the only way to scale the project. -Rob -- Robert Collins rbtcoll...@hp.com Distinguished Technologist HP Converged Cloud ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] auto-delete in amqp reply_* queues in OpenStack
Ok, assuming that you've run that query against the 5 stuck queues I would expect the following results: * if an active listener for a queue lives on one of compute hosts: that queue was created by compute service initiating rpc command. Since you didn't restart them during switchover, the compute services still use the same queues. * if queue does not have a listener: the queue was created by the controller which was active before the switchover. That queue could have become stuck not exactly at previous switchover, but as well at some other switchover occurred in the past. 2014-03-25 0:33 GMT+04:00 Chris Friesen chris.frie...@windriver.com: On 03/24/2014 01:27 PM, Dmitry Mescheryakov wrote: I see two possible explanations for these 5 remaining queues: * They were indeed recreated by 'compute' services. I.e. controller service send some command over rpc and then it was shut down. Its reply queue was automatically deleted, since its only consumer was disconnected. The compute services replied after that and so recreated the queue. According to Rabbit MQ docs, such queue will be stuck alive indefinitely, since it will never have a consumer. * Possibly there are services on compute nodes which initiate RPC calls themselves. I don't know OpenStack architecture enough to say if services running on compute nodes do so. In that case these 5 queues are still used by compute services. Do Rabbit MQ management tools (web or cli) allow to view active consumers for queues? If yes, then you can find out which of the cases above you encountered. Or it maybe be some third case I didn't account for :-) It appears that the cli tools do not provide a way to print the info, but if you query a single queue via the web API it will give the IP address and port of the consumers for the queue. The vhost needs to be URL-encoded, so the query looks something like this: curl -i -u guest:guest http://192.168.204.2:15672/api/queues/%2f/reply_08e35acffe2c4078ae4603f08e9d0997 Chris ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Ironic][Keystone] Move drivers credentials to Keystone
On Tue, Mar 25, 2014 at 12:49 PM, Jay Pipes jaypi...@gmail.com wrote: On Tue, 2014-03-25 at 17:39 +, Miller, Mark M (EB SW Cloud - RD - Corvallis) wrote: Why not use Barbican? It stores credentials after encrypting them. No reason not to add a Barbican driver as well. If Keystone's /v3/credentials API has any legs, it should be backed by barbican, if not superseded by it completely. Best, -jay ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Ironic][Keystone] Move drivers credentials to Keystone
On 25/03/14 12:23 +, Lucas Alvares Gomes wrote: Hi, Right now Ironic is being responsible for storing the credentials for the IPMI and SSH drivers (and potentially other drivers in the future), I wonder if we should delegate this task to Keystone. The Keystone V3 API now has a /credentials endpoint which would allow us to specify arbitrary types (not only ec2 anymore) and use it as a credential store[1]. That would avoid further fragmentation of credentials being stored in different places in OpenStack, and make the management of the credentials easier (Think about a situation where many nodes share the same IPMI username/password and we need to update it, if this is stored in Keystone it only needs to be updated there once cause nodes will only hold a reference to it) It also was pointed to me that setting a hard dependency on Keystone V3 might significantly raises the bar for integration with existing clouds*. So perhaps we should make it optional? In the same way we can specify a username/password or key_filename for the ssh driver we could have a reference to a credential in Keystone V3? As others seem to be saying, I think it might make sense to make this pluggable. Store it in driver metadata, or store it in Keystone, or store it in Barbican. Of course, that's 3x the effort. As a relative newcomer -- is it problematic for us to leverage an incubated service? I suppose that a pluggable approach with Barbican as one option would alleviate any problems that might exist. This would argue to me that the easiest thing for Ceilometer might be to query us for IPMI stats, if the credential store is pluggable. Fetch these bare metal statistics doesn't seem too off-course for Ironic to me. The alternative is that Ceilometer and Ironic would both have to be configured for the same pluggable credential store. Or am I crazy? -- Matt ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] Jenkins test logs and their retention period
On Mon, Mar 24, 2014 at 5:49 AM, Sean Dague s...@dague.net wrote: ... Part of the challenge is turning off DEBUG is currently embedded in code in oslo log, which makes it kind of awkward to set sane log levels for included libraries because it requires an oslo round trip with code to all the projects to do it. Here's how it's done in Keystone: https://review.openstack.org/#/c/62068/10/keystone/config.py It's definitely awkward. - Brant ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Nova] Updates to Juno blueprint review process
On 03/25/2014 01:30 AM, Russell Bryant wrote: I really don't see why this is so bad. We're using a tool specifically designed for reviewing things that is already working *really* well, right, for code reviews. not just for code, but also for TC governance documents. I disagree: that's another hack, needed because there is nothing better around. The only decent tool that I remember had a decent UX to discuss text documents was the tool used to discuss online the GPLv3 drafts. Google Docs has features similar to that... Unfortunately free software tools somehow got stuck even when they pioneered many features. Unless Storyboard plans to re-implement what gerrit does (I sure hope it doesn't), I expect we would keep working like this. Do you expect storyboard to have an interface for iterating on text documents, where you can provide inline comments, review differences between revisions, etc? What am I missing? Mainly I think you're missing things related to usability, discoverability, search, reporting and all sorts of drawbacks related to having two separate tools to do one thing, loosely coupled. Gerrit's search is obscure for non-daily gerrit users, indexing by web search engines is non existent... grep on local filesystem may work for some, but that's not all the people interested in knowing why a feature was implemented that way (for example), git is not unfriendly only picky about who his friends are... and I can go on but I'll stop here: I understand we have no better solution in the given timeframe, it's pointless to debate it further. /stef -- Ask and answer questions on https://ask.openstack.org ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Neutron][LBaaS] Neutron LBaaS, Libra and managed services
Hey Susanne, I think it makes sense to group drivers by each LB software. For example, there would be a driver for HAProxy, one for Citrix's Netscalar, one for Riverbed's Stingray, etc. One important aspect about Openstack that I don't want us to forget though is that a tenant should be able to move between cloud providers at their own will (no vendor lock-in). The API contract is what allows this. The challenging aspect is ensuring different drivers support the API contract in the same way. What components should drivers share is also and interesting conversation to be had. Cheers, --Jorge From: Susanne Balle sleipnir...@gmail.commailto:sleipnir...@gmail.com Reply-To: OpenStack Development Mailing List (not for usage questions) openstack-dev@lists.openstack.orgmailto:openstack-dev@lists.openstack.org Date: Tuesday, March 25, 2014 6:59 AM To: OpenStack Development Mailing List (not for usage questions) openstack-dev@lists.openstack.orgmailto:openstack-dev@lists.openstack.org Subject: Re: [openstack-dev] [Neutron][LBaaS] Neutron LBaaS, Libra and managed services John, Brandon, I agree that we cannot have a multitude of drivers doing the same thing or close to because then we end-up in the same situation as we are today where we have duplicate effort and technical debt. The goal would be here to be able to built a framework around the drivers that would allow for resiliency, failover, etc... If the differentiators are in higher level APIs then we can have one a single driver (in the best case) for each software LB e.g. HA proxy, nginx, etc. Thoughts? Susanne On Mon, Mar 24, 2014 at 11:26 PM, John Dewey j...@dewey.wsmailto:j...@dewey.ws wrote: I have a similar concern. The underlying driver may support different functionality, but the differentiators need exposed through the top level API. I see the SSL work is well underway, and I am in the process of defining L7 scripting requirements. However, I will definitely need L7 scripting prior to the API being defined. Is this where vendor extensions come into play? I kinda like the route the Ironic guy safe taking with a “vendor passthru” API. John On Monday, March 24, 2014 at 3:17 PM, Brandon Logan wrote: Creating a separate driver for every new need brings up a concern I have had. If we are to implement a separate driver for every need then the permutations are endless and may cause a lot drivers and technical debt. If someone wants an ha-haproxy driver then great. What if they want it to be scalable and/or HA, is there supposed to be scalable-ha-haproxy, scalable-haproxy, and ha-haproxy drivers? Then what if instead of doing spinning up processes on the host machine we want a nova VM or a container to house it? As you can see the permutations will begin to grow exponentially. I'm not sure there is an easy answer for this. Maybe I'm worrying too much about it because hopefully most cloud operators will use the same driver that addresses those basic needs, but worst case scenarios we have a ton of drivers that do a lot of similar things but are just different enough to warrant a separate driver. From: Susanne Balle [sleipnir...@gmail.commailto:sleipnir...@gmail.com] Sent: Monday, March 24, 2014 4:59 PM To: OpenStack Development Mailing List (not for usage questions) Subject: Re: [openstack-dev] [Neutron][LBaaS] Neutron LBaaS, Libra and managed services Eugene, Thanks for your comments, See inline: Susanne On Mon, Mar 24, 2014 at 4:01 PM, Eugene Nikanorov enikano...@mirantis.commailto:enikano...@mirantis.com wrote: Hi Susanne, a couple of comments inline: We would like to discuss adding the concept of “managed services” to the Neutron LBaaS either directly or via a Neutron LBaaS plug-in to Libra/HA proxy. The latter could be a second approach for some of the software load-balancers e.g. HA proxy since I am not sure that it makes sense to deploy Libra within Devstack on a single VM. Currently users would have to deal with HA, resiliency, monitoring and managing their load-balancers themselves. As a service provider we are taking a more managed service approach allowing our customers to consider the LB as a black box and the service manages the resiliency, HA, monitoring, etc. for them. As far as I understand these two abstracts, you're talking about making LBaaS API more high-level than it is right now. I think that was not on our roadmap because another project (Heat) is taking care of more abstracted service. The LBaaS goal is to provide vendor-agnostic management of load balancing capabilities and quite fine-grained level. Any higher level APIs/tools can be built on top of that, but are out of LBaaS scope. [Susanne] Yes. Libra currently has some internal APIs that get triggered when an action needs to happen. We would like similar functionality in Neutron LBaaS so the user doesn’t have to manage the load-balancers but can consider them as
[openstack-dev] [Nova][Neutron] Neutron + Nova + OVS security group fix
Hi Nova, Neturon Team I would like to discuss issue of Neutron + Nova + OVS security group fix. We have a discussion in IRC today, but the issue is complicated so we will have a conf call tomorrow 17:00 UST (10AM PDT). #openstack-neutron (I'll put conf call information in IRC) -- Please let me know if this time won't work with you. Bug Report https://bugs.launchpad.net/neutron/+bug/1297469 Background of this issue: ML2 + OVSDriver + IptablesBasedFirewall combination is a default plugin setting in the Neutron. In this case, we need a special handing in VIF. Because OpenVSwitch don't support iptables, we are using linuxbride + openvswitch bridge. We are calling this as hybrid driver. On the other discussion, we generalized the Nova side VIF plugging to the Libvirt GenericVIFDriver. The idea is let neturon tell the VIF plugging configration details to the GenericDriver, and GerericDriver takes care of it. Unfortunatly, HybridDriver is removed before GenericDriver is ready for security group. This makes ML2 + OVSDriver + IptablesBasedFirewall combination unfunctional. We were working on realfix, but we can't make it until Icehouse release due to design discussions [1]. # Even if neturon side patch isn't merged yet. So we are proposing a workaround fix to Nova side. In this fix, we are adding special version of the GenericVIFDriver which can work with the combination. There is two point on this new Driver. (1) It prevent set conf.filtername. Because we should use NoopFirewallDriver, we need conf.filtername should be None when we use it. (2) use plug_ovs_hybrid and unplug_ovs_hybrid by enforcing get_require_firewall as True. Here is patchs with UT. Workaournd fix: Nova https://review.openstack.org/#/c/82904/ Devstack patch for ML2 (Tested with 82904) https://review.openstack.org/#/c/82937/ We have tested the patch 82904 with following test, and this works. - Launch VM - Assign floating ip - make sure ping to the floating ip is failing from GW - modify security group rule to allow ping from anywhere - make sure ping is working [1] Real fix: (defered to Juno) Improve vif attributes related with firewalling https://review.openstack.org/#/c/21946/ Support binding:vif_security parameter in neutron https://review.openstack.org/#/c/44596/ -- I'll put latest update on here https://etherpad.openstack.org/p/neturon_security_group_fix_workaround_icehouse Best Nachi ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [neutron][chaining][policy] Port-oriented Network service chaining
Hi, Most of the advanced services and group policy sub-team members who attended last week’s meeting should remember I promised to start a drafting proposal regarding network service chaining. This week I got to start writing a document which is accessible here: https://docs.google.com/document/d/1Bk1e8-diE1VnzlbM8l479Mjx2vKliqdqC_3l5S56ITU It should not be considered a formal blueprint as it yet requires large discussion from the community wrt the validation (or sanity if you will) of the proposed idea. I will be joining the advanced service IRC meeting tomorrow, and the group policy IRC meeting thursday, making myself available to answer any questions you may have. In the meantime you can also start discussing in this email thread or commenting in the document. Thanks, Carlos Goncalves ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [All][Keystone] Deprecation of the v2 API
On 03/25/2014 04:50 PM, Russell Bryant wrote: We discussed the deprecation of the v2 keystone API in the cross-project meeting today [1]. This thread is to recap and bring that discussion to some consensus. snip In summary, until we have completed v3 support within OpenStack itself, it's premature to mark the API deprecated since that's a signal to end users and deployers that says action is required. Thoughts? Makes sense to me. Chris ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [TripleO][reviews] We're falling behind
Excerpts from Robert Collins's message of 2014-03-25 13:17:29 -0700: TripleO has just seen an influx of new contributors. \o/. Flip side - we're now slipping on reviews /o\. In the meeting today we had basically two answers: more cores, and more work by cores. We're slipping by 2 reviews a day, which given 16 cores is a small amount. I'm going to propose some changes to core in the next couple of days - I need to go and re-read a bunch of reviews first - but, right now we don't have a hard lower bound on the number of reviews we request cores commit to (on average). We're seeing 39/day from the 16 cores - which isn't enough as we're falling behind. Thats 2.5 or so. So - I'd like to ask all cores to commit to doing 3 reviews a day, across all of tripleo (e.g. if your favourite stuff is all reviewed, find two new things to review even if outside comfort zone :)). And we always need more cores - so if you're not a core, this proposal implies that we'll be asking that you a) demonstrate you can sustain 3 reviews a day on average as part of stepping up, and b) be willing to commit to that. Obviously if we have enough cores we can lower the minimum commitment - so I don't think this figure should be fixed in stone. And now - time for a loose vote - who (who is a tripleo core today) supports / disagrees with this proposal - lets get some consensus here. I'm in favour, obviously :), though it is hard to put reviews ahead of direct itch scratching, its the only way to scale the project. +1 FWIW I think we just haven't made reviews as much of a priority at the same time that we got an influx of contribution. We also started getting serious on CI, which I think has slowed the pace through the review machine, with the bonus that we break ourselves less often. :) -Rob ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] User mailing lists for OpenStack projects
Hi Shaunak, the lists for the OpenStack project are hosted on http://lists.openstack.org. You can see the full roster on that page. If you need a special list for the users of OpenStack PHP SDK you can propose a change to the lists manifest on openstack-infra/config/modules/openstack_project/manifests/lists.pp. Follow the instructions on this wiki page: https://wiki.openstack.org/wiki/Community#Mailing_lists_in_local_languages Have people interested in the new list vote the review, I think it would be a good discussion to have. Cheers, stef On 03/20/2014 09:34 PM, Shaunak Kashyap wrote: Hi folks, I am relatively new to OpenStack development as one of the developers on the unified PHP SDK for OpenStack [1]. We were recently discussing about a mailing list for the users of this SDK (as opposed to it’s contributors who will use openstack-dev@). The purpose of such as mailing list would be for users of the SDK to communicate with the contributors as well as each other. Of course, there would be other avenues for such communication as well (IRC, for instance). Specifically, we would like to know whether existing OpenStack projects have mailing lists for their users and, if so, where they are being hosted. Thanks, Shaunak [1] https://wiki.openstack.org/wiki/OpenStack-SDK-PHP ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev -- Ask and answer questions on https://ask.openstack.org ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Nova][Neutron] Neutron + Nova + OVS security group fix
I hope we can sort this out on the mailing list IRC, without having to schedule emergency meetings. Salvatore On 25 March 2014 22:58, Nachi Ueno na...@ntti3.com wrote: Hi Nova, Neturon Team I would like to discuss issue of Neutron + Nova + OVS security group fix. We have a discussion in IRC today, but the issue is complicated so we will have a conf call tomorrow 17:00 UST (10AM PDT). #openstack-neutron (I'll put conf call information in IRC) thanks, but I'd prefer you discuss the matter on IRC. I won't be available at that time and having IRC logs on eavesdrop will allow me to catch up without having to ask people or waiting for minutes on the mailing list. -- Please let me know if this time won't work with you. Bug Report https://bugs.launchpad.net/neutron/+bug/1297469 Background of this issue: ML2 + OVSDriver + IptablesBasedFirewall combination is a default plugin setting in the Neutron. In this case, we need a special handing in VIF. Because OpenVSwitch don't support iptables, we are using linuxbride + openvswitch bridge. We are calling this as hybrid driver. The hybrid solution in Neutron has been around for such a long time that I would hardly call it a special handling. To summarize, the VIF is plugged into a linux bridge, which has another leg plugged in the OVS integration bridge. On the other discussion, we generalized the Nova side VIF plugging to the Libvirt GenericVIFDriver. The idea is let neturon tell the VIF plugging configration details to the GenericDriver, and GerericDriver takes care of it. The downside of the generic driver is that so far it's assuming local configuration values are sufficient to correctly determine VIF plugging. The generic VIF driver will use the hybrid driver if get_firewall_required is true. And this will happen if the firewall driver is anything different from the NoOp driver. This was uncovered by a devstack commit (1143f7e). When I previously discussed with the people involved this issue, I was under the impression that the devstack patch introduced the problem. Apparently the Generic VIF driver is not taking at the moments hints from neutron regarding the driver to use, and therefore, from what I gather, makes a decision based on nova conf flags only. So a quick fix would be to tell the Generic VIF driver to always use hybrid plugging when neutron is enabled (which can be gathered by nova conf flags). This will fix the issue for ML2, but will either break or insert an unnecessary extra hop for other plugins. Unfortunatly, HybridDriver is removed before GenericDriver is ready for security group. The drivers were marked for deprecation in Havana, and if we thought the GenericDriver was not good for neutron security groups we had enough time to scream. This makes ML2 + OVSDriver + IptablesBasedFirewall combination unfunctional. We were working on realfix, but we can't make it until Icehouse release due to design discussions [1]. # Even if neturon side patch isn't merged yet. So we are proposing a workaround fix to Nova side. In this fix, we are adding special version of the GenericVIFDriver which can work with the combination. There is two point on this new Driver. (1) It prevent set conf.filtername. Because we should use NoopFirewallDriver, we need conf.filtername should be None when we use it. (2) use plug_ovs_hybrid and unplug_ovs_hybrid by enforcing get_require_firewall as True. Here is patchs with UT. Workaournd fix: Nova https://review.openstack.org/#/c/82904/ Devstack patch for ML2 (Tested with 82904) https://review.openstack.org/#/c/82937/ Are there other plugins which need the hybrid driver for sec groups to work? I think so. And also - the patch does not seem to work according to Jenkins. The failures look genuine to me. We have tested the patch 82904 with following test, and this works. - Launch VM - Assign floating ip - make sure ping to the floating ip is failing from GW - modify security group rule to allow ping from anywhere - make sure ping is working You can actually run your devstack patch with your patch under review in the check queue. Check what Aaron did here: https://review.openstack.org/#/c/78694/11 I wonder if instead of putting this bit of tape, we could just leverage the VIF_TYPE attribute of the port binding extension. Most plugins use VIF_TYPE_OVS already. It's a pity the ml2 plugin with the OVS mech driver did not specify VIF_TYPE_OVS_HYBRID. But, in my opinion if we fix the relevant constants in the plugins which require hybrid plugging, and we slightly change the generic VIF driver logic to make a decision according to the VIF_TYPE binding attribute we should fine, as we'll end up with two small, contained patches, which, IMHO, are not even much ugly. But again, I'm far from being a subject matter expert when it comes to nova/neutron integration and the ML2 plugin. [1] Real fix: (defered to Juno) Improve vif attributes related with firewalling
Re: [openstack-dev] [All][Keystone] Deprecation of the v2 API
On Tue, Mar 25, 2014 at 5:50 PM, Russell Bryant rbry...@redhat.com wrote: We discussed the deprecation of the v2 keystone API in the cross-project meeting today [1]. This thread is to recap and bring that discussion to some consensus. The issue is that Keystone has marked the v2 API as deprecated in Icehouse: https://blueprints.launchpad.net/keystone/+spec/deprecate-v2-api If you use the API, deployments will get this in their logs: WARNING keystone.openstack.common.versionutils [-] Deprecated: v2 API is deprecated as of Icehouse in favor of v3 API and may be removed in K. The deprecation status is reflected in the API for end users, as well. For example, from the CLI: $ keystone discover Keystone found at http://172.16.12.38:5000/v2.0 - supports version v2.0 (deprecated) here http://172.16.12.38:5000/v2.0/ My proposal is that this deprecation be reverted. Here's why: First, it seems there isn't a common use of deprecated. To me, marking something deprecated means that the deprecated feature: - has been completely replaced by something else - end users / deployers should take action to migrate to the new thing immediately. - The project has provided a documented migration path - the old thing will be removed at a specific date/release The problem with the above is that most OpenStack projects do not support the v3 API yet. From talking to Dolph in the meeting, it sounds like the intention is: - fully support v2, just don't add features - signal to other projects that they should be migrating to v3 Given that intention, I believe the proper thing to do is to actually leave the API marked as fully supported / stable. Keystone should be working with other OpenStack projects to migrate them to v3. Once that is complete, deprecation can be re-visited. Sounds great, thanks for the discussion. Anne In summary, until we have completed v3 support within OpenStack itself, it's premature to mark the API deprecated since that's a signal to end users and deployers that says action is required. Thoughts? [1] http://eavesdrop.openstack.org/meetings/project/2014/project.2014-03-25-21.01.log.html#l-103 -- Russell Bryant ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Murano][Heat] MuranoPL questions?
On 21/03/14 18:58, Stan Lagun wrote: Zane, I appreciate your explanations on Heat/HOT. This really makes sense. I didn't mean to say that MuranoPL is better for Heat. Actually HOT is good for Heat's mission. I completely acknowledge it. I've tried to avoid comparison between languages and I'm sorry if it felt that way. This is not productive as I don't offer you to replace No problem, I didn't interpret it that way. I just wanted to use it as an opportunity to document a bit more information about Heat works that may be helpful to you. HOT with MuranoPL (although I believe that certain elements of MuranoPL syntax can be contributed to HOT and be valuable addition there). Also people tend to protect what they have developed and invested into and to be fair this is what we did in this thread to great extent. Yes, we are all human and this is an unavoidable part of life :) What I'm trying to achieve is that you and the rest of Heat team understand why it was designed the way it is. I don't feel that Murano can become full-fledged member of OpenStack ecosystem without a bless from Heat team. And it would be even better if we agree on certain design, join our efforts and contribute to each other for sake of Orchestration program. Thanks. To be clear, this is what I am looking for. At the beginning of this thread I proposed a very simple possible implementation, and I'm wanting folks to tell me what is missing from that model that would justify a more complicated approach. I'm sorry for long mail texts written in not-so-good English and appreciate you patience reading and answering them. Having said that let me step backward and explain our design decisions. Cloud administrators are usually technical guys that are capable of learning HOT and writing YAML templates. They know exact configuration of their cloud (what services are available, what is the version of OpenStack cloud is running) and generally understands how OpenStack works. They also know about software they intent to install. If such guy wants to install Drupal he knows exactly that he needs HOT template describing Fedora VM with Apache + PHP + MySQL + Drupal itself. It is not a problem for him to write such HOT template. I'm aware that TOSCA has these types of constraints, and in fact I suggested to the TOSCA TC that maybe this is where we should draw the line between Heat and some TOSCA-compatible service: HOT should be a concrete description of exactly what you're going to get, whereas some other service (in this case Murano) would act as the constraints solver. e.g. something like an image name would not be hardcoded in a Murano template, you have some constraints about which operating system and what versions should be allowed, and it would pick one and pass it to Heat. So I am interested in this approach. The worst outcome here would be to end up with something that was equivalent to TOSCA but not easily translatable to the TOSCA Simple Profile YAML format (currently a Working Draft). Where 'easily translatable' preferably means 'by just changing some names'. I can't comment on whether this is the case as things stand. Note that such template would be designed for very particular configuration. There are hundreds of combinations that may be used to install that Drupal - use RHEL/Windows/etc instead of Fedora, use ngnix/IIS/etc instead of Apache, use FastCGI instead of mod_php, PostgreSQL instead of MySQL. You may choose to have all software on single VM or have one VM for database and another for Drupal. There are also constraints to those combinations. For example you cannot have Fedora + IIS on the same VM. You cannot have Apache and Drupal on different VMs. So the HOT template represent fixed combination of those software components. HOT may have input parameters like username or dbImageName but the overall structure of template is fixed. You cannot have template that choses whether to use Windows or Linux based on parameter value. As Thomas mentioned, CloudFormation now has conditionals. If we were to add this feature in HOT (which seems likely), you would actually be able to do that. http://docs.aws.amazon.com/AWSCloudFormation/latest/UserGuide/conditions-section-structure.html You cannot write HOT that accepts number of instances it allowed to create and then decide what would be installed on each of them. This is just not needed for Heat users. The path we've gone down in Heat is to introduce software component resources so that the definition of the software component to install is separated from the definition of the server it runs on, and may even be in a separate template file. So you can write a template that launches a single server with (e.g.) both Wordpress and MySQL installed, or you can have a separate template (though with the conditional stuff above it could theoretically be the same template even) that launches two servers with Wordpress on one and MySQL on the
Re: [openstack-dev] [All][Keystone] Deprecation of the v2 API
On Tue, Mar 25, 2014 at 5:50 PM, Russell Bryant rbry...@redhat.com wrote: We discussed the deprecation of the v2 keystone API in the cross-project meeting today [1]. This thread is to recap and bring that discussion to some consensus. The issue is that Keystone has marked the v2 API as deprecated in Icehouse: https://blueprints.launchpad.net/keystone/+spec/deprecate-v2-api If you use the API, deployments will get this in their logs: WARNING keystone.openstack.common.versionutils [-] Deprecated: v2 API is deprecated as of Icehouse in favor of v3 API and may be removed in K. The deprecation status is reflected in the API for end users, as well. For example, from the CLI: $ keystone discover Keystone found at http://172.16.12.38:5000/v2.0 - supports version v2.0 (deprecated) here http://172.16.12.38:5000/v2.0/ My proposal is that this deprecation be reverted. Here's why: First, it seems there isn't a common use of deprecated. To me, marking something deprecated means that the deprecated feature: - has been completely replaced by something else - end users / deployers should take action to migrate to the new thing immediately. - The project has provided a documented migration path - the old thing will be removed at a specific date/release Agree on all points. Unfortunately, we have yet to succeed on the documentation front: https://blueprints.launchpad.net/keystone/+spec/document-v2-to-v3-transition The problem with the above is that most OpenStack projects do not support the v3 API yet. From talking to Dolph in the meeting, it sounds like the intention is: - fully support v2, just don't add features - signal to other projects that they should be migrating to v3 Above all else, this was our primary goal: to raise awareness about our path forward, and to identify the non-obvious stakeholders that we needed to work with in order to drop support for v2. With today's discussion as evidence, I think we've succeeded in that regard :) Given that intention, I believe the proper thing to do is to actually leave the API marked as fully supported / stable. Keystone should be working with other OpenStack projects to migrate them to v3. Once that is complete, deprecation can be re-visited. Happy to! Revert deprecation of the v2 API: https://review.openstack.org/#/c/82963/ Although I'd prefer to apply this patch directly to milestone-proposed, so we can continue into Juno with the deprecation in master. In summary, until we have completed v3 support within OpenStack itself, it's premature to mark the API deprecated since that's a signal to end users and deployers that says action is required. Thoughts? [1] http://eavesdrop.openstack.org/meetings/project/2014/project.2014-03-25-21.01.log.html#l-103 -- Russell Bryant ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Nova][Neutron] Neutron + Nova + OVS security group fix
Hi Salvatore 2014-03-25 17:57 GMT-07:00 Salvatore Orlando sorla...@nicira.com: I hope we can sort this out on the mailing list IRC, without having to schedule emergency meetings. Russel requested to have a conf call on this, so let him decide it. Salvatore On 25 March 2014 22:58, Nachi Ueno na...@ntti3.com wrote: Hi Nova, Neturon Team I would like to discuss issue of Neutron + Nova + OVS security group fix. We have a discussion in IRC today, but the issue is complicated so we will have a conf call tomorrow 17:00 UST (10AM PDT). #openstack-neutron (I'll put conf call information in IRC) thanks, but I'd prefer you discuss the matter on IRC. I won't be available at that time and having IRC logs on eavesdrop will allow me to catch up without having to ask people or waiting for minutes on the mailing list. -- Please let me know if this time won't work with you. Bug Report https://bugs.launchpad.net/neutron/+bug/1297469 Background of this issue: ML2 + OVSDriver + IptablesBasedFirewall combination is a default plugin setting in the Neutron. In this case, we need a special handing in VIF. Because OpenVSwitch don't support iptables, we are using linuxbride + openvswitch bridge. We are calling this as hybrid driver. The hybrid solution in Neutron has been around for such a long time that I would hardly call it a special handling. To summarize, the VIF is plugged into a linux bridge, which has another leg plugged in the OVS integration bridge. On the other discussion, we generalized the Nova side VIF plugging to the Libvirt GenericVIFDriver. The idea is let neturon tell the VIF plugging configration details to the GenericDriver, and GerericDriver takes care of it. The downside of the generic driver is that so far it's assuming local configuration values are sufficient to correctly determine VIF plugging. The generic VIF driver will use the hybrid driver if get_firewall_required is true. And this will happen if the firewall driver is anything different from the NoOp driver. This was uncovered by a devstack commit (1143f7e). When I previously discussed with the people involved this issue, I was under the impression that the devstack patch introduced the problem. Apparently the Generic VIF driver is not taking at the moments hints from neutron regarding the driver to use, and therefore, from what I gather, makes a decision based on nova conf flags only. So a quick fix would be to tell the Generic VIF driver to always use hybrid plugging when neutron is enabled (which can be gathered by nova conf flags). This will fix the issue for ML2, but will either break or insert an unnecessary extra hop for other plugins. get_firewall_required = True won't fix this issue. We need to make sure we won't set config.filtername in this case. Unfortunatly, HybridDriver is removed before GenericDriver is ready for security group. The drivers were marked for deprecation in Havana, and if we thought the GenericDriver was not good for neutron security groups we had enough time to scream. The reason we missed this issue is we lack the negative test for security groups.. so we couldn't realize this. This makes ML2 + OVSDriver + IptablesBasedFirewall combination unfunctional. We were working on realfix, but we can't make it until Icehouse release due to design discussions [1]. # Even if neturon side patch isn't merged yet. So we are proposing a workaround fix to Nova side. In this fix, we are adding special version of the GenericVIFDriver which can work with the combination. There is two point on this new Driver. (1) It prevent set conf.filtername. Because we should use NoopFirewallDriver, we need conf.filtername should be None when we use it. (2) use plug_ovs_hybrid and unplug_ovs_hybrid by enforcing get_require_firewall as True. Here is patchs with UT. Workaournd fix: Nova https://review.openstack.org/#/c/82904/ Devstack patch for ML2 (Tested with 82904) https://review.openstack.org/#/c/82937/ Are there other plugins which need the hybrid driver for sec groups to work? I think so. And also - the patch does not seem to work according to Jenkins. The failures look genuine to me. I agere with you, however we should start with minimal fix for this. We can work for another plugin in another patch. This patch will fail in Jenkins because it needs 82904. We have tested the patch 82904 with following test, and this works. - Launch VM - Assign floating ip - make sure ping to the floating ip is failing from GW - modify security group rule to allow ping from anywhere - make sure ping is working You can actually run your devstack patch with your patch under review in the check queue. Check what Aaron did here: https://review.openstack.org/#/c/78694/11 Nice hack. let me try it I wonder if instead of putting this bit of tape, we could just leverage the VIF_TYPE attribute of the port binding extension. Most plugins use
Re: [openstack-dev] [Ironic][Keystone] Move drivers credentials to Keystone
On 3/25/2014 1:50 PM, Matt Wagner wrote: This would argue to me that the easiest thing for Ceilometer might be to query us for IPMI stats, if the credential store is pluggable. Fetch these bare metal statistics doesn't seem too off-course for Ironic to me. The alternative is that Ceilometer and Ironic would both have to be configured for the same pluggable credential store. There is already a blueprint with a proposed patch here for Ironic to do the querying: https://blueprints.launchpad.net/ironic/+spec/send-data-to-ceilometer. I think, for terms of credential storage (and, for that matter, metrics gathering as I noted in that blueprint), it's very useful to have things pluggable. Ironic, in particular, has many different use cases: bare metal private cloud, bare metal public cloud, and triple-o. I could easily see all three being different enough to call for different forms of credential storage. -Jay Faulkner ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] SR-IOV and IOMMU check
Hi, all Currently openstack can support SR-IOV device pass-through (at least there are some patches for this), but the prerequisite to this is both IOMMU and SR-IOV must be enabled correctly, it seems there is not a robust way to check this in openstack, I have implemented a way to do this and hope it can be committed into upstream, this can help find the issue beforehand, instead of letting kvm report the issue no IOMMU found until the VM is started. I didn't find an appropriate place to put into this, do you think this is necessary? Where can it be put into? Welcome your advice and thank you in advance. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] SR-IOV and IOMMU check
Hi, Yang, Yi y Agree with you, IOMMU and SR-IOV need to be checked beforehand. I think it should be checked before booting a instance with the pci flavor, that means when the flavor contains some normal pci cards or SR-IOV cards. Just like when you find there are pci_requests in the instance system_metadata. The details are out of my current knows. Hope can help you. From: Yang, Yi Y [mailto:yi.y.y...@intel.com] Sent: Wednesday, March 26, 2014 10:51 AM To: openstack-dev@lists.openstack.org Subject: [openstack-dev] SR-IOV and IOMMU check Hi, all Currently openstack can support SR-IOV device pass-through (at least there are some patches for this), but the prerequisite to this is both IOMMU and SR-IOV must be enabled correctly, it seems there is not a robust way to check this in openstack, I have implemented a way to do this and hope it can be committed into upstream, this can help find the issue beforehand, instead of letting kvm report the issue no IOMMU found until the VM is started. I didn't find an appropriate place to put into this, do you think this is necessary? Where can it be put into? Welcome your advice and thank you in advance. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [All][Keystone] Deprecation of the v2 API
Excerpts from Dolph Mathews's message of 2014-03-25 19:01:17 -0700: On Tue, Mar 25, 2014 at 5:50 PM, Russell Bryant rbry...@redhat.com wrote: We discussed the deprecation of the v2 keystone API in the cross-project meeting today [1]. This thread is to recap and bring that discussion to some consensus. The issue is that Keystone has marked the v2 API as deprecated in Icehouse: https://blueprints.launchpad.net/keystone/+spec/deprecate-v2-api If you use the API, deployments will get this in their logs: WARNING keystone.openstack.common.versionutils [-] Deprecated: v2 API is deprecated as of Icehouse in favor of v3 API and may be removed in K. The deprecation status is reflected in the API for end users, as well. For example, from the CLI: $ keystone discover Keystone found at http://172.16.12.38:5000/v2.0 - supports version v2.0 (deprecated) here http://172.16.12.38:5000/v2.0/ My proposal is that this deprecation be reverted. Here's why: First, it seems there isn't a common use of deprecated. To me, marking something deprecated means that the deprecated feature: - has been completely replaced by something else - end users / deployers should take action to migrate to the new thing immediately. - The project has provided a documented migration path - the old thing will be removed at a specific date/release Agree on all points. Unfortunately, we have yet to succeed on the documentation front: https://blueprints.launchpad.net/keystone/+spec/document-v2-to-v3-transition The problem with the above is that most OpenStack projects do not support the v3 API yet. From talking to Dolph in the meeting, it sounds like the intention is: - fully support v2, just don't add features - signal to other projects that they should be migrating to v3 Above all else, this was our primary goal: to raise awareness about our path forward, and to identify the non-obvious stakeholders that we needed to work with in order to drop support for v2. With today's discussion as evidence, I think we've succeeded in that regard :) Given that intention, I believe the proper thing to do is to actually leave the API marked as fully supported / stable. Keystone should be working with other OpenStack projects to migrate them to v3. Once that is complete, deprecation can be re-visited. Happy to! Revert deprecation of the v2 API: https://review.openstack.org/#/c/82963/ Although I'd prefer to apply this patch directly to milestone-proposed, so we can continue into Juno with the deprecation in master. As somebody maintaining a few master-chasing CD clouds, I'd like to ask you to please stop the squawking about deprecation until it has a definite replacement and most if not all OpenStack core projects are using it. 1 out of every 2 API calls on these clouds produces one of these errors in Keystone. That is just pointless. :-P ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev