[openstack-dev] [Nova][VMware] Deploy from vCenter template
I saw a commit for Deploying from VMware vCenter template and found it's abandoned. https://review.openstack.org/#/c/34903 Anyone knows the plan to support the deployment from VMware vCenter template? Thanks! Best Regards___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [climate] Weekly meeting accidental moving to Tuesday, 10:00 UTC
Hello! Guys, I have no opportunity to hold our IRC meeting today. I propose to move it tomorrow, the same time. Please let me know if you are OK with that. Thank you! Best regards, Dina Belova Software Engineer Mirantis Inc. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [climate] Weekly meeting accidental moving to Tuesday, 10:00 UTC
Le 16/12/2013 10:41, Dina Belova a écrit : Hello! Guys, I have no opportunity to hold our IRC meeting today. I propose to move it tomorrow, the same time. Please let me know if you are OK with that. Thank you! +2 to this. No regular meetings planned on Tuesdays 1000UTC as per https://wiki.openstack.org/wiki/Meetings, so we can take #openstack-meeting for one hour. -Sylvain ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [climate] Weekly meeting accidental moving to Tuesday, 10:00 UTC
+2 too Nikolay Starodubtsev Software Engineer Mirantis Inc. Skype: dark_harlequine1 2013/12/16 Sylvain Bauza sylvain.ba...@bull.net Le 16/12/2013 10:41, Dina Belova a écrit : Hello! Guys, I have no opportunity to hold our IRC meeting today. I propose to move it tomorrow, the same time. Please let me know if you are OK with that. Thank you! +2 to this. No regular meetings planned on Tuesdays 1000UTC as per https://wiki.openstack.org/wiki/Meetings, so we can take #openstack-meeting for one hour. -Sylvain ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [oslo]: implementing olso.messaging over amqp 1.0
On 12/12/2013 02:14 PM, Flavio Percoco wrote: I've a draft in my head of how the amqp 1.0 driver could be implemented and how to map the current expectations of the messaging layer to the new protocol. I think a separate thread to discuss this mapping is worth it. There are some critical areas that definitely need more discussion I have also been looking at this, and trying to write up a simple design notes. Some of the questions that occurred to me while doing so are: * Use one link for all sends, with 'to' field set, or use a link for each target? * How to handle calls to one of a group of servers? * Use a distinct response address per request, or allow an address to be shared by multiple requests in conjunction with correlation id on responses? * Support both intermediated and direct communication? For all patterns? The aim in my view should be to have the driver support as many alternatives in deployment as possible without overcomplicating things, distorting the mapping or introducing server specific extensions. I have some notes to share if anyone is interested. I can send them to this list or put them upon the wiki or an etherpad or something. --Gordon. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Nova] All I want for Christmas is one more +2 ...
I've had to rebase this a couple of times since to keep ahead of RPC version numbers - sitting there with a +1 from Jenkins at the moment: https://review.openstack.org/#/c/35303/ Phil -Original Message- From: Russell Bryant [mailto:rbry...@redhat.com] Sent: 12 December 2013 14:38 To: openstack-dev@lists.openstack.org Subject: Re: [openstack-dev] [Nova] All I want for Christmas is one more +2 ... On 12/12/2013 09:22 AM, Day, Phil wrote: Hi Cores, The Stop, Rescue, and Delete should give guest a chance to shutdown change https://review.openstack.org/#/c/35303/ was approved a couple of days ago, but failed to merge because the RPC version had moved on. Its rebased and sitting there with one +2 and a bunch of +1s -would be really nice if it could land before it needs another rebase please ? Approved. FWIW, I'm fine with folks approving with a single +2 for cases where a patch is approved but needed a simple rebase. This happens pretty often. We even have a script that generates a list of patches still open that were previously approved: http://russellbryant.net/openstack-stats/nova-openapproved.txt -- Russell Bryant ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [climate] Weekly meeting accidental moving to Tuesday, 10:00 UTC
ok On Mon, Dec 16, 2013 at 2:14 PM, Nikolay Starodubtsev nstarodubt...@mirantis.com wrote: +2 too Nikolay Starodubtsev Software Engineer Mirantis Inc. Skype: dark_harlequine1 2013/12/16 Sylvain Bauza sylvain.ba...@bull.net Le 16/12/2013 10:41, Dina Belova a écrit : Hello! Guys, I have no opportunity to hold our IRC meeting today. I propose to move it tomorrow, the same time. Please let me know if you are OK with that. Thank you! +2 to this. No regular meetings planned on Tuesdays 1000UTC as per https://wiki.openstack.org/wiki/Meetings, so we can take #openstack-meeting for one hour. -Sylvain ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev -- Sincerely yours, Sergey Lukjanov Savanna Technical Lead Mirantis Inc. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] Performance Regression in Neutron/Havana compared to Quantum/Grizzly
Hello Sreedhar, I am focusing only on the OVS agent at the moment. Armando fixed a few issues recently with the DHCP agent; those issues were triggering a perennial resync; with his fixes I reckon DHCP agent response times should be better. I reckon Maru is also working on architectural improvements for the DHCP agent (see thread on DHCP agent reliability). Regards, Salvatore On 13 December 2013 20:26, Nathani, Sreedhar (APS) sreedhar.nath...@hp.comwrote: Hello All, Update with my testing. I have installed one more VM as neutron-server host and configured under the Load Balancer. Currently I have 2 VMs running neutron-server process (one is Controller and other is dedicated neutron-server VM) With this configuration during the batch instance deployment with a batch size of 30 and sleep time of 20min, 180 instances could get an IP during the first boot. During 181-210 instance creation some instances could not get an IP. This is much better than when running with single neutron server where only 120 instances could get an IP during the first boot in Havana. When the instances are getting created, parent neutron-server process spending close to 90% of the cpu time on both the servers, While rest of the neutron-server process (APIs) are spending very low CPU utilization. I think it’s good idea to expand the current multiple neutron-server api process to support rpc messages as well. Even with current setup (multiple neutron-server hosts), we still see rpc timeouts in DHCP, L2 agents and dnsmasq process is getting restarted due to SIGKILL though. Thanks Regards, Sreedhar Nathani *From:* Nathani, Sreedhar (APS) *Sent:* Friday, December 13, 2013 12:08 AM *To:* OpenStack Development Mailing List (not for usage questions) *Subject:* RE: [openstack-dev] Performance Regression in Neutron/Havana compared to Quantum/Grizzly Hello Salvatore, Thanks for your feedback. Does the patch https://review.openstack.org/#/c/57420/ which you are working on bug https://bugs.launchpad.net/neutron/+bug/1253993 will help to correct the OVS agent loop slowdown issue? Does this patch address the DHCP agent updating the host file once in a minute and finally sending SIGKILL to dnsmasq process? I have tested with Marun’s patch https://review.openstack.org/#/c/61168/regarding ‘Send DHCP notifications regardless of agent status’ but this patch Also observed the same behavior. Thanks Regards, Sreedhar Nathani *From:* Salvatore Orlando [mailto:sorla...@nicira.comsorla...@nicira.com] *Sent:* Thursday, December 12, 2013 6:21 PM *To:* OpenStack Development Mailing List (not for usage questions) *Subject:* Re: [openstack-dev] Performance Regression in Neutron/Havana compared to Quantum/Grizzly I believe your analysis is correct and inline with the findings reported in the bug concerning OVS agent loop slowdown. The issue has become even more prominent with the ML2 plugin due to an increased number of notifications sent. Another issue which makes delays on the DHCP agent worse is that instances send a discover message once a minute. Salvatore Il 11/dic/2013 11:50 Nathani, Sreedhar (APS) sreedhar.nath...@hp.com ha scritto: Hello Peter, Here are the tests I have done. Already have 240 instances active across all the 16 compute nodes. To make the tests and data collection easy, I have done the tests on single compute node First Test - * 240 instances already active, 16 instances on the compute node where I am going to do the tests * deploy 10 instances concurrently using nova boot command with num-instances option in single compute node * All the instances could get IP during the instance boot time. - Instances are created at 2013-12-10 13:41:01 - From the compute host, DHCP requests are sent from 13:41:20 but those are not reaching the DHCP server Reply from the DHCP server got at 13:43:08 (A delay of 108 seconds) - DHCP agent updated the host file from 13:41:06 till 13:42:54. Dnsmasq process got SIGHUP message every time the hosts file is updated - In compute node tap devices are created between 13:41:08 and 13:41:18 Security group rules are received between 13:41:45 and 13:42:56 IP table rules were updated between 13:41:50 and 13:43:04 Second Test - * Deleted the newly created 10 instances. * 240 instances already active, 16 instances on the compute node where I am going to do the tests * Deploy 30 instances concurrently using nova boot command with num-instances option in single compute node * None of the instances could get the IP during the instance boot. - Instances are created at 2013-12-10 14:13:50 - From the compute host, DHCP Requests are sent from 14:14:14 but those are not reaching the DHCP Server (don't see any DHCP requests are reaching the
[openstack-dev] [Nova][libvirt]when deleting instance which is in migrating state, instance files can be stay in destination node forever
Hi all. When I try to fix a bug:https://bugs.launchpad.net/nova/+bug/1242961, I get a trouble. To reproduce the bug is very easy. Live migrate a vm in block_migration mode, and then delelte the vm immediately. The reason of this bug is as follow: 1. Because live migrate costs more time, so the vm will be deleted sucessfully before live migrate complete. And then, we will get an exception while live migrating. 2. After live migrate failed, we start to rollback. But, in the rollback method we will get or modify the info of vm from db. Because the vm has been deleted already, so we will get instance_not_found exception and rollback will be faild too. I have two ways to fix the bug: i)Add check in nova-api. When try to delete a vm, we return an error message if the vm_state is LIVE_MIGRATING. This way is very simple, but need to carefully consider. I have found a related discussion: http://lists.openstack.org/pipermail/openstack-dev/2013-October/017454.html, but it has no result in the discussion. ii)Before live migrate we get all the data needed by rollback method, and add a new rollback method. The new method will clean up resources at destination based on the above data(The resouces at source has been already cleaned up by deleting). I have no idea whitch one I should choose. Or, any other ideas?:) Regards, wanghong ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [multipath] Could I use the multipath software provided by the SAN vendors instead of dm-multipath in openstack?
Hi,all The storage array used by cinder in my experiment is produced by Huawei. The vendor releases its own multipath named Ultrapath with the SAN. Could I use the Ultrapath instead of dm-multipath in openstack? Best wishes, Qi Qi Xiaozhen CLOUD OS PDU, IT Product Line, Huawei Enterprise Business Group enterprise.huawei.com ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] Performance Regression in Neutron/Havana compared to Quantum/Grizzly
Hello Salvatore, Thanks for the updates. All the changes which you talked is from the agent side. From my tests, with multiple L2 agents running and sending/requesting messages at the same time from the single neutron rpc server process is not able to handle All the load fast enough and causing the bottleneck. With the Carl's patch (https://review.openstack.org/#/c/60082), we now support multiple neutron API process, My question is why can't we support multiple neutron rpc server process as well? Horizontal scaling with multiple neutron-server hosts would be one option, but having support of multiple neutron rpc servers process in in the same System would be really helpful for the scaling of neutron server especially during concurrent instance deployments. Thanks Regards, Sreedhar Nathani From: Salvatore Orlando [mailto:sorla...@nicira.com] Sent: Monday, December 16, 2013 4:55 PM To: OpenStack Development Mailing List (not for usage questions) Subject: Re: [openstack-dev] Performance Regression in Neutron/Havana compared to Quantum/Grizzly Hello Sreedhar, I am focusing only on the OVS agent at the moment. Armando fixed a few issues recently with the DHCP agent; those issues were triggering a perennial resync; with his fixes I reckon DHCP agent response times should be better. I reckon Maru is also working on architectural improvements for the DHCP agent (see thread on DHCP agent reliability). Regards, Salvatore On 13 December 2013 20:26, Nathani, Sreedhar (APS) sreedhar.nath...@hp.commailto:sreedhar.nath...@hp.com wrote: Hello All, Update with my testing. I have installed one more VM as neutron-server host and configured under the Load Balancer. Currently I have 2 VMs running neutron-server process (one is Controller and other is dedicated neutron-server VM) With this configuration during the batch instance deployment with a batch size of 30 and sleep time of 20min, 180 instances could get an IP during the first boot. During 181-210 instance creation some instances could not get an IP. This is much better than when running with single neutron server where only 120 instances could get an IP during the first boot in Havana. When the instances are getting created, parent neutron-server process spending close to 90% of the cpu time on both the servers, While rest of the neutron-server process (APIs) are spending very low CPU utilization. I think it's good idea to expand the current multiple neutron-server api process to support rpc messages as well. Even with current setup (multiple neutron-server hosts), we still see rpc timeouts in DHCP, L2 agents and dnsmasq process is getting restarted due to SIGKILL though. Thanks Regards, Sreedhar Nathani From: Nathani, Sreedhar (APS) Sent: Friday, December 13, 2013 12:08 AM To: OpenStack Development Mailing List (not for usage questions) Subject: RE: [openstack-dev] Performance Regression in Neutron/Havana compared to Quantum/Grizzly Hello Salvatore, Thanks for your feedback. Does the patch https://review.openstack.org/#/c/57420/ which you are working on bug https://bugs.launchpad.net/neutron/+bug/1253993 will help to correct the OVS agent loop slowdown issue? Does this patch address the DHCP agent updating the host file once in a minute and finally sending SIGKILL to dnsmasq process? I have tested with Marun's patch https://review.openstack.org/#/c/61168/ regarding 'Send DHCP notifications regardless of agent status' but this patch Also observed the same behavior. Thanks Regards, Sreedhar Nathani From: Salvatore Orlando [mailto:sorla...@nicira.com] Sent: Thursday, December 12, 2013 6:21 PM To: OpenStack Development Mailing List (not for usage questions) Subject: Re: [openstack-dev] Performance Regression in Neutron/Havana compared to Quantum/Grizzly I believe your analysis is correct and inline with the findings reported in the bug concerning OVS agent loop slowdown. The issue has become even more prominent with the ML2 plugin due to an increased number of notifications sent. Another issue which makes delays on the DHCP agent worse is that instances send a discover message once a minute. Salvatore Il 11/dic/2013 11:50 Nathani, Sreedhar (APS) sreedhar.nath...@hp.commailto:sreedhar.nath...@hp.com ha scritto: Hello Peter, Here are the tests I have done. Already have 240 instances active across all the 16 compute nodes. To make the tests and data collection easy, I have done the tests on single compute node First Test - * 240 instances already active, 16 instances on the compute node where I am going to do the tests * deploy 10 instances concurrently using nova boot command with num-instances option in single compute node * All the instances could get IP during the instance boot time. - Instances are created at 2013-12-10 13:41:01 - From the compute host, DHCP requests are sent from 13:41:20 but those are not reaching the DHCP server Reply from the DHCP
[openstack-dev] [Mistral] Community meeting agenda - 12/16/2013
Hi! This is a reminder that we will have another community meeting today in IRC (#openstack-meeting) at 16.00 UTC. Here’s the agenda: https://wiki.openstack.org/wiki/Meetings/MistralAgenda As usually, you’re welcome to join! Renat Akhmerov @ Mirantis Inc. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [tuskar] How to install tuskar-ui from packaging point of view
On 12/16/2013 08:52 AM, Stephen Gran wrote: On 16/12/13 03:47, Thomas Goirand wrote: Hi, I've been working over the last 2 months to get Ironic, TripleO and Tuskar ready for an upload in Debian. However, for tuskar-ui, I'm facing the fact that there's a lack of documentation. It was easy to get Tuskar packaged. If I understand well, it only needs 2 daemons: tuskar-api, and tuskar-manager. Is this right? If not, what did I miss? Is tuskar-manager really a daemon? (I have to admit that I didn't find yet the time to try, so I would appreciate some guidance here) I think you are right here. As for tuskar-ui, the install.rst is quite vague about how to install. I got the python-tuskar-ui binary package done, with egg-info and all, that's not the problem. What worries me is this part: Go into horizon and create a symlink to the tuskar-ui code: cd horizon ln -s ../tuskar-ui/tuskar_ui Then, install a virtual environment for your setup: Add this to debian/links or something? It sounds like it needs a dependency on horizon to make sure that the directory exists. Not sure how it translates to Debian packaging but you need to copy/symlink the Tuskar-UI source *inside* the Horizon directory. python tools/install_venv.py This means turn the list of dependencies in the source package into dependencies in the debian package, I would think. Yes, that's correct. 3/ The install.rst has: If everything has gone according to plan, you should be able to run: tools/with_venv.sh ./manage.py runserver and have the application start on port 8080. The Tuskar dashboard will be located at http://localhost:8080/infrastructure does this mean that on top of Horizon running through Apache, tuskar-ui needs to run independently? Why is that? Can't we just have tuskar-ui simply integrated with the rest of Horizon? Yes, Tuskar-UI runs on top of Horizon. You don't have to create a separate Horizon+Tuskar-UI installation, it does not run independently of the existing Horizon installation but you have to modify it. When you create a symlink into the Horizon source, that makes the Infrastructure dashboard provided by Tuskar-UI autodiscovered when the Horizon application boots up. Tuskar-UI creates and additional tab inside the Horizon application, which will be available at http://localhost:8080/infrastructure (or on whatever port you set Horizon up) and at http://localhost:8080/ you can access the Project and Admin dashboards provided by Horizon. It's not stated in tuskar-ui/install.rst but this guide is meant to set up the development environment. It is also worth mentioning that the current solution is only temporary, in the long term Tuskar-UI will be a part of Horizon (see the Horizon and Tuskar-UI merge thread). Imre ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] Performance Regression in Neutron/Havana compared to Quantum/Grizzly
Multiple RPC servers is something we should definitely look at. I don't see a show-stopper reason for which this would not work, although I recall we found out a few caveats one should be aware of when doing multiple RPC servers when reviewing the patch for multiple API server (I wrote them in some other ML thread, I will dig them later). If you are thinking of implementing this support, you might want to sync up with Mark McClain who's working on splitting API and RPC servers. While horizontal scaling is surely desirable, evidence we gathered from analysis like the one you did showed that probably we can make the interactions between the neutron server and the agents a lot more efficient and reliable. I reckon both items are needed and can be implemented independently. Regards, Salvatore On 16 December 2013 12:42, Nathani, Sreedhar (APS) sreedhar.nath...@hp.comwrote: Hello Salvatore, Thanks for the updates. All the changes which you talked is from the agent side. From my tests, with multiple L2 agents running and sending/requesting messages at the same time from the single neutron rpc server process is not able to handle All the load fast enough and causing the bottleneck. With the Carl’s patch (https://review.openstack.org/#/c/60082), we now support multiple neutron API process, My question is why can’t we support multiple neutron rpc server process as well? Horizontal scaling with multiple neutron-server hosts would be one option, but having support of multiple neutron rpc servers process in in the same System would be really helpful for the scaling of neutron server especially during concurrent instance deployments. Thanks Regards, Sreedhar Nathani *From:* Salvatore Orlando [mailto:sorla...@nicira.com] *Sent:* Monday, December 16, 2013 4:55 PM *To:* OpenStack Development Mailing List (not for usage questions) *Subject:* Re: [openstack-dev] Performance Regression in Neutron/Havana compared to Quantum/Grizzly Hello Sreedhar, I am focusing only on the OVS agent at the moment. Armando fixed a few issues recently with the DHCP agent; those issues were triggering a perennial resync; with his fixes I reckon DHCP agent response times should be better. I reckon Maru is also working on architectural improvements for the DHCP agent (see thread on DHCP agent reliability). Regards, Salvatore On 13 December 2013 20:26, Nathani, Sreedhar (APS) sreedhar.nath...@hp.com wrote: Hello All, Update with my testing. I have installed one more VM as neutron-server host and configured under the Load Balancer. Currently I have 2 VMs running neutron-server process (one is Controller and other is dedicated neutron-server VM) With this configuration during the batch instance deployment with a batch size of 30 and sleep time of 20min, 180 instances could get an IP during the first boot. During 181-210 instance creation some instances could not get an IP. This is much better than when running with single neutron server where only 120 instances could get an IP during the first boot in Havana. When the instances are getting created, parent neutron-server process spending close to 90% of the cpu time on both the servers, While rest of the neutron-server process (APIs) are spending very low CPU utilization. I think it’s good idea to expand the current multiple neutron-server api process to support rpc messages as well. Even with current setup (multiple neutron-server hosts), we still see rpc timeouts in DHCP, L2 agents and dnsmasq process is getting restarted due to SIGKILL though. Thanks Regards, Sreedhar Nathani *From:* Nathani, Sreedhar (APS) *Sent:* Friday, December 13, 2013 12:08 AM *To:* OpenStack Development Mailing List (not for usage questions) *Subject:* RE: [openstack-dev] Performance Regression in Neutron/Havana compared to Quantum/Grizzly Hello Salvatore, Thanks for your feedback. Does the patch https://review.openstack.org/#/c/57420/ which you are working on bug https://bugs.launchpad.net/neutron/+bug/1253993 will help to correct the OVS agent loop slowdown issue? Does this patch address the DHCP agent updating the host file once in a minute and finally sending SIGKILL to dnsmasq process? I have tested with Marun’s patch https://review.openstack.org/#/c/61168/regarding ‘Send DHCP notifications regardless of agent status’ but this patch Also observed the same behavior. Thanks Regards, Sreedhar Nathani *From:* Salvatore Orlando [mailto:sorla...@nicira.comsorla...@nicira.com] *Sent:* Thursday, December 12, 2013 6:21 PM *To:* OpenStack Development Mailing List (not for usage questions) *Subject:* Re: [openstack-dev] Performance Regression in Neutron/Havana compared to Quantum/Grizzly I believe your analysis is correct and inline with the findings reported in the bug concerning OVS agent loop slowdown. The issue has become even
Re: [openstack-dev] [TripleO] [Horizon] [Tuskar] [UI] Horizon and Tuskar-UI merge
On 12/13/2013 03:08 PM, Ladislav Smola wrote: Horizoners, As discussed in TripleO and Horizon meetings, we are proposing to move Tuskar UI under the Horizon umbrella. Since we are building our UI solution on top of Horizon, we think this is a good fit. It will allow us to get feedback and reviews from the appropriate group of developers. I don't think, we really disagree here. My main concern would be more: what do we get, if we make up another project under the umbrella of horizon? I mean, what does that mean at all? My proposal would be, to send patches directly to horizon. As discussed in last weeks horizon meeting, tuskar UI would become integrated in Horizon, but disabled by default. This would enable a faster integration in Horizon and would reduce the overhead of creating a separate repositoy, installation instructions, packaging etc. etc. From the horizon side: we would get some new contributors (and hopefully reviewers), which is very much appreciated. Matthias ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Nova][libvirt]when deleting instance which is in migrating state, instance files can be stay in destination node forever
Isn’t just handling the exception instance_not_found enough? By this time source would’ve been cleaned up. Destination VM resources will get cleaned up by the periodic task since the VM is not associated with this host. Am I missing something here? From: 王宏 [mailto:w.wangho...@gmail.com] Sent: 16 December 2013 11:32 To: openstack-dev@lists.openstack.org Subject: [openstack-dev] [Nova][libvirt]when deleting instance which is in migrating state, instance files can be stay in destination node forever Hi all. When I try to fix a bug:https://bugs.launchpad.net/nova/+bug/1242961, I get a trouble. To reproduce the bug is very easy. Live migrate a vm in block_migration mode, and then delelte the vm immediately. The reason of this bug is as follow: 1. Because live migrate costs more time, so the vm will be deleted sucessfully before live migrate complete. And then, we will get an exception while live migrating. 2. After live migrate failed, we start to rollback. But, in the rollback method we will get or modify the info of vm from db. Because the vm has been deleted already, so we will get instance_not_found exception and rollback will be faild too. I have two ways to fix the bug: i)Add check in nova-api. When try to delete a vm, we return an error message if the vm_state is LIVE_MIGRATING. This way is very simple, but need to carefully consider. I have found a related discussion: http://lists.openstack.org/pipermail/openstack-dev/2013-October/017454.html, but it has no result in the discussion. ii)Before live migrate we get all the data needed by rollback method, and add a new rollback method. The new method will clean up resources at destination based on the above data(The resouces at source has been already cleaned up by deleting). I have no idea whitch one I should choose. Or, any other ideas?:) Regards, wanghong ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [TripleO] [Tuskar] [UI] Icehouse Requirements - Summary, Milestones
On 12/13/2013 05:22 PM, James Slagle wrote: On Fri, Dec 13, 2013 at 03:04:09PM +0100, Imre Farkas wrote: One note to deploy: It's not done only by Heat and Nova. If we expect a fully functional OpenStack installation as a result, we are missing a few steps like creating users, initializing and registering the service endpoints with Keystone. In TripleO this is done by the init-keystone and setup-endpoints scripts. Check devtest for more details: http://docs.openstack.org/developer/tripleo-incubator/devtest_undercloud.html Excellent point Imre, as the deployment isn't really useable until those steps are done. The link to the overcloud setup steps is actually: http://docs.openstack.org/developer/tripleo-incubator/devtest_overcloud.html Very similar to what is done for the undercloud. You are right, that's the correct link for the overcloud setup. However, I intentionally picked the one for the undercloud because I wanted to focus on the keystone configuration part and that's the same for the both (the init-keystone, setup-endpoints and keystone role-create workflow). There are some other stuff going on in the overcloud setup (eg. creating a vm for a user) which might distract those who are not familiar with devtest from what really is needed to deploy OpenStack. But it would have been better from me to note that the link is not for the overcloud. I think most of that logic could be reimplemented to be done via direct calls to the API using the client libs vs using a CLI. Not sure about keystone-manage pki_setup though, would need to look into that. Yeah, we can put big part of the needed configuration steps into Tuskar as most of it just uses CLI client of Keystone which can be replaced by using the direct API calls of the same library. The rest might go to Heat or os-refresh-config or else. Imre ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [governance] Becoming a Program, before applying for incubation
On 13/12/13 16:37 +0100, Flavio Percoco wrote: On 13/12/13 15:53 +0100, Thierry Carrez wrote: Hi everyone, TL;DR: Incubation is getting harder, why not ask efforts to apply for a new program first to get the visibility they need to grow. Long version: Last cycle we introduced the concept of Programs to replace the concept of Official projects which was no longer working that well for us. This was recognizing the work of existing teams, organized around a common mission, as an integral part of delivering OpenStack. Contributors to programs become ATCs, so they get to vote in Technical Committee (TC) elections. In return, those teams place themselves under the authority of the TC. This created an interesting corner case. Projects applying for incubation would actually request two concurrent things: be considered a new Program, and give incubated status to a code repository under that program. Over the last months we significantly raised the bar for accepting new projects in incubation, learning from past integration and QA mistakes. The end result is that a number of promising projects applied for incubation but got rejected on maturity, team size, team diversity, or current integration level grounds. At that point I called for some specific label, like Emerging Technology that the TC could grant to promising projects that just need more visibility, more collaboration, more crystallization before they can make good candidates to be made part of our integrated releases. However, at the last TC meeting it became apparent we could leverage Programs to achieve the same result. Promising efforts would first get their mission, scope and existing results blessed and recognized as something we'd really like to see in OpenStack one day. Then when they are ready, they could have one of their deliveries apply for incubation if that makes sense. The consequences would be that the effort would place itself under the authority of the TC. Their contributors would be ATCs and would vote in TC elections, even if their deliveries never make it to incubation. They would get (some) space at Design Summits. So it's not free, we still need to be pretty conservative about accepting them, but it's probably manageable. I'm still weighing the consequences, but I think it's globally nicer than introducing another status. As long as the TC feels free to revoke Programs that do not deliver the expected results (or that no longer make sense in the new world order) I think this approach would be fine. Comments, thoughts ? With the above, I'm basically saying that a Queuing ;) program shouldn't exist until there's an integrated team of folks working on queuing. Incubation doesn't guarantees integration and emerging technology doesn't guarantees incubation. Both stages mean there's interest about that technology and that we're looking forward to see it being part of OpenStack, period. Each stage probably means a bit more than that but, IMHO, that's the 'community' point of view of those stages. What if we have a TC-managed* Program incubation period? The Program won't be managed by the team working on the emerging technology, nor the team working on the incubated project. Until those projects don't graduate, the program won't be official nor will have the 'rights' of other programs. And if the project fits into another program, then it won't be officially part of it until it graduates. Since I, most likely, won't make it to tomorrow's TC meeting, I'd like to extend this argument a bit more and make sure I share my thoughts about it. Hopefully they'll be of help. What I'm arguing here is: 1. Programs that are not part of OpenStack's release cycle shouldn't be considered official nor they should have the rights that integrated projects have. 2. I think requesting Programs to exist at the early stages of the project is not necessary. I don't even think incubated projects should have programs. I do agree the project's mission and goals have to be clear but the program should be officially created *after* the project graduates from incubation. The reasoning here is that anything could happen during incubation. For example, a program created for project A - which is incubated - may change to cover a broader mission that will allow a newborn project B to fall under its umbrella, hence my previous proposal of having a incubation stage for programs as well. My proposal is to either not requesting any program to be created for incubated projects / emerging technologies or to have a program called 'Emerging Technologies' were all these projects could fit in. The only difference is that, IMHO, projects under this program should not have all the rights that integrated projects and other programs have, although the program will definitely fall under the TCs authority. For example, projects under this program shouldn't be able to vote on the TCs elections. Hope this make sense and that is of help during the upcoming
[openstack-dev] [Nova][Docker] Environment variables
Hi All, I have submitted a new blueprint which addresses the a common pattern in the docker world. A usual pattern in the docker world is to use environment variables to configure a container. docker run -e SQL_URL=postgres://user:password@/db my-app The nova docker driver doesn't support to set any environment variables. To work around this issue I used cloud-init which works fine. But this approach has of course the drawback that a) I have to install the cloud init service. and b) my docker container doesn't work outside of openstack. I propose to allow a user to set docker environment variables via nova instance metadata. The metadata key should have a prefix like ENV_ which can be used to determine all environment variables. The prefix should be removed and the remainder key and vaule will be injected. The metadata can unfortunately not be set in horizon but can be used from the nova command line tool and from heat. Example heat: myapp: Type: OS::Nova::Server Properties: flavor: m1.small image: my-app:latest meta-data: - ENV_SQL_URL: postgres://user:password@/db - ENV_SOMETHING_ELSE: Value Let me know what you think about that. Blueprint: https://blueprints.launchpad.net/nova/+spec/docker-env-via-meta-data regards, Daniel ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] Performance Regression in Neutron/Havana compared to Quantum/Grizzly
Hello Salvatore, I agree with you on we need both items to improve the scaling and performance of neutron server. I am not a developer so can't implement the changes myself. If somebody is going to implement I am more than happy to do the tests. Thanks Regards, Sreedhar Nathani From: Salvatore Orlando [mailto:sorla...@nicira.com] Sent: Monday, December 16, 2013 6:18 PM To: OpenStack Development Mailing List (not for usage questions) Subject: Re: [openstack-dev] Performance Regression in Neutron/Havana compared to Quantum/Grizzly Multiple RPC servers is something we should definitely look at. I don't see a show-stopper reason for which this would not work, although I recall we found out a few caveats one should be aware of when doing multiple RPC servers when reviewing the patch for multiple API server (I wrote them in some other ML thread, I will dig them later). If you are thinking of implementing this support, you might want to sync up with Mark McClain who's working on splitting API and RPC servers. While horizontal scaling is surely desirable, evidence we gathered from analysis like the one you did showed that probably we can make the interactions between the neutron server and the agents a lot more efficient and reliable. I reckon both items are needed and can be implemented independently. Regards, Salvatore On 16 December 2013 12:42, Nathani, Sreedhar (APS) sreedhar.nath...@hp.commailto:sreedhar.nath...@hp.com wrote: Hello Salvatore, Thanks for the updates. All the changes which you talked is from the agent side. From my tests, with multiple L2 agents running and sending/requesting messages at the same time from the single neutron rpc server process is not able to handle All the load fast enough and causing the bottleneck. With the Carl's patch (https://review.openstack.org/#/c/60082), we now support multiple neutron API process, My question is why can't we support multiple neutron rpc server process as well? Horizontal scaling with multiple neutron-server hosts would be one option, but having support of multiple neutron rpc servers process in in the same System would be really helpful for the scaling of neutron server especially during concurrent instance deployments. Thanks Regards, Sreedhar Nathani From: Salvatore Orlando [mailto:sorla...@nicira.commailto:sorla...@nicira.com] Sent: Monday, December 16, 2013 4:55 PM To: OpenStack Development Mailing List (not for usage questions) Subject: Re: [openstack-dev] Performance Regression in Neutron/Havana compared to Quantum/Grizzly Hello Sreedhar, I am focusing only on the OVS agent at the moment. Armando fixed a few issues recently with the DHCP agent; those issues were triggering a perennial resync; with his fixes I reckon DHCP agent response times should be better. I reckon Maru is also working on architectural improvements for the DHCP agent (see thread on DHCP agent reliability). Regards, Salvatore On 13 December 2013 20:26, Nathani, Sreedhar (APS) sreedhar.nath...@hp.commailto:sreedhar.nath...@hp.com wrote: Hello All, Update with my testing. I have installed one more VM as neutron-server host and configured under the Load Balancer. Currently I have 2 VMs running neutron-server process (one is Controller and other is dedicated neutron-server VM) With this configuration during the batch instance deployment with a batch size of 30 and sleep time of 20min, 180 instances could get an IP during the first boot. During 181-210 instance creation some instances could not get an IP. This is much better than when running with single neutron server where only 120 instances could get an IP during the first boot in Havana. When the instances are getting created, parent neutron-server process spending close to 90% of the cpu time on both the servers, While rest of the neutron-server process (APIs) are spending very low CPU utilization. I think it's good idea to expand the current multiple neutron-server api process to support rpc messages as well. Even with current setup (multiple neutron-server hosts), we still see rpc timeouts in DHCP, L2 agents and dnsmasq process is getting restarted due to SIGKILL though. Thanks Regards, Sreedhar Nathani From: Nathani, Sreedhar (APS) Sent: Friday, December 13, 2013 12:08 AM To: OpenStack Development Mailing List (not for usage questions) Subject: RE: [openstack-dev] Performance Regression in Neutron/Havana compared to Quantum/Grizzly Hello Salvatore, Thanks for your feedback. Does the patch https://review.openstack.org/#/c/57420/ which you are working on bug https://bugs.launchpad.net/neutron/+bug/1253993 will help to correct the OVS agent loop slowdown issue? Does this patch address the DHCP agent updating the host file once in a minute and finally sending SIGKILL to dnsmasq process? I have tested with Marun's patch https://review.openstack.org/#/c/61168/ regarding 'Send DHCP notifications regardless of agent status' but this patch Also observed
Re: [openstack-dev] [TripleO] [Horizon] [Tuskar] [UI] Horizon and Tuskar-UI merge
On 2013/16/12 14:03, Matthias Runge wrote: On 12/13/2013 03:08 PM, Ladislav Smola wrote: Horizoners, As discussed in TripleO and Horizon meetings, we are proposing to move Tuskar UI under the Horizon umbrella. Since we are building our UI solution on top of Horizon, we think this is a good fit. It will allow us to get feedback and reviews from the appropriate group of developers. I don't think, we really disagree here. My main concern would be more: what do we get, if we make up another project under the umbrella of horizon? I mean, what does that mean at all? My proposal would be, to send patches directly to horizon. As discussed in last weeks horizon meeting, tuskar UI would become integrated in Horizon, but disabled by default. This would enable a faster integration in Horizon and would reduce the overhead of creating a separate repositoy, installation instructions, packaging etc. etc. From the horizon side: we would get some new contributors (and hopefully reviewers), which is very much appreciated. Matthias This is important note. From information architecture and user interaction point of view, I don't think it makes sense to keep all the three tabs visible together (Project, Admin, Infrastructure). There are lot of reasons, but main points: * Infrastructure itself is undercloud concept running in different instance of Horizon. * Users dealing with deployment and infrastructure management are not the users of OpenStack UI / Dashboard. It is different set of users. So it doesn't make sense to have giant application, which provides each and every possible feature. I think we need to keep focused. So by default, I would say that there should exist Project + Admin tab together or Infrastructure. But never all three together. So when Matthias say 'disabled by default', I would mean completely hidden for user and if user wants to use Infrastructure management, he can enable it in different horizon instance, but it will be the only visible tab for him. So it will be sort of separate application, but still running on top of Horizon. -- Jarda ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [governance] Becoming a Program, before applying for incubation
Flavio Percoco wrote: What I'm arguing here is: 1. Programs that are not part of OpenStack's release cycle shouldn't be considered official nor they should have the rights that integrated projects have. 2. I think requesting Programs to exist at the early stages of the project is not necessary. I don't even think incubated projects should have programs. I do agree the project's mission and goals have to be clear but the program should be officially created *after* the project graduates from incubation. The reasoning here is that anything could happen during incubation. For example, a program created for project A - which is incubated - may change to cover a broader mission that will allow a newborn project B to fall under its umbrella, hence my previous proposal of having a incubation stage for programs as well. I think your concerns can be covered if we consider that programs covering incubated or promising projects should also somehow incubate. To avoid confusion I'd use a different term, let's say incoming programs for the sake of the discussion. Incoming programs would automatically graduate when one of their deliveries graduate to integrated status (for projects with such deliveries), or when the TC decides so (think: for horizontal programs like Documentation or Deployment). That doesn't change most of this proposal, which is that we'd encourage teams to ask to become an (incoming) program before they consider filing one of their projects for incubation. FWIW we already distinguish (on https://wiki.openstack.org/wiki/Programs) programs that are born out of an incubated project from other programs, so adding this incoming status would not change much. My proposal is to either not requesting any program to be created for incubated projects / emerging technologies or to have a program called 'Emerging Technologies' were all these projects could fit in. I don't think an Emerging Technologies program would make sense, since that would just be a weird assemblage of separate teams (how would that program elect a PTL ?). I prefer that they act as separate teams (which they are) and use the incoming Program concept described above. The only difference is that, IMHO, projects under this program should not have all the rights that integrated projects and other programs have, although the program will definitely fall under the TCs authority. For example, projects under this program shouldn't be able to vote on the TCs elections. So *that* would be a change from where we stand today, which is that incubated project contributors get ATC status and vote on TC elections. We can go either way, consider incoming programs to be OpenStack programs in the sense of the TC charter, or not. I'm not convinced there is so much value in restricting TC voting access (or ATC status) to OpenStack programs. Incoming programs would all be placed under the authority of the TC so it's only fair that they have a vote. Also giving them ATC status gets them automatically invited to Design Summits, and getting incoming programs in Design Summits sounds like a good thing to do... -- Thierry Carrez (ttx) signature.asc Description: OpenPGP digital signature ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [nova] VM diagnostics - V3 proposal
Hi, At the moment the administrator is able to retrieve diagnostics for a running VM. Currently the implementation is very loosely defined, that is, each driver returns whatever they have to return. This is problematic in a number of respects: 1. The tempest tests were written specifically for one driver and break with all other drivers (the test was removed to prevent this – bug 1240043) 2. An admin is unable to write tools that may work with a hybrid cloud 3. Adding support for get_diagnostics for drivers that do not support is painful I'd like to propose the following for the V3 API (we will not touch V2 in case operators have applications that are written against this – this may be the case for libvirt or xen. The VMware API support was added in I1): 1. We formalize the data that is returned by the API [1] 2. We enable the driver to add extra information that will assist the administrators in troubleshooting problems for VM's I have proposed a BP for this - https://blueprints.launchpad.net/nova/+spec/diagnostics-namespace (I'd like to change the name to v3-api-diagnostics – which is more apt) And as Nelson Mandel would have said: “It always seems impossible until it's done.” Moving forwards we decide to provide administrator the option of using the for V2 (it may be very helpful with debugging issues). But that is another discussion. Thanks Gary [1] https://etherpad.openstack.org/p/vm-diagnostics ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Tempest][Neutron] Network client and API tests refactoring.
Hi Eugene, as you have already noticed there's already some overlap with your work and the current tests development. We should find a productive way to coordinate the efforts. Thanks for starting the refactoring, in my opinion it's needed. cheers, Rossella On Sat, Dec 14, 2013 at 4:53 PM, Jay Pipes jaypi...@gmail.com wrote: On Sat, 2013-12-14 at 19:09 +0400, Eugene Nikanorov wrote: Hi Jay, Sure, that is understood. In fact such refactoring could be a big change so I'd split it to two or more patches. Hope this will not overlap with ongoing neutron API tests development. Hehe, given the sheer number of new tests that get added to Tempest every week, I'd say that any sort of base refactoring like this will need to be heavily coordinated with the Tempest core reviewers and other contributors pushing code! Best, -jay ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Nova][Docker] Environment variables
On 12/16/2013 09:27 AM, Daniel Kuffner wrote: Hi All, I have submitted a new blueprint which addresses the a common pattern in the docker world. A usual pattern in the docker world is to use environment variables to configure a container. docker run -e SQL_URL=postgres://user:password@/db my-app The nova docker driver doesn't support to set any environment variables. To work around this issue I used cloud-init which works fine. But this approach has of course the drawback that a) I have to install the cloud init service. and b) my docker container doesn't work outside of openstack. I propose to allow a user to set docker environment variables via nova instance metadata. The metadata key should have a prefix like ENV_ which can be used to determine all environment variables. The prefix should be removed and the remainder key and vaule will be injected. The metadata can unfortunately not be set in horizon but can be used from the nova command line tool and from heat. Example heat: myapp: Type: OS::Nova::Server Properties: flavor: m1.small image: my-app:latest meta-data: - ENV_SQL_URL: postgres://user:password@/db - ENV_SOMETHING_ELSE: Value Let me know what you think about that. Blueprint: https://blueprints.launchpad.net/nova/+spec/docker-env-via-meta-data Thanks for starting the discussion. More people should do this for their blueprints. :-) One of the things we should be striving for is to provide as consistent of an experience as we can across drivers. Right now, we have the metadata service and config drive, and neither of those are driver specific. In the case of config drive, whether it's used or not is exposed through the API. As you point out, the meta-data service does technically work with the docker driver. I don't think we should support environment variables like this automatically. Instead, I think it would be more appropriate to add an API extension for specifying env vars. That way the behavior is more explicit and communicated through the API. The env vars would be passed through all of the appropriate plumbing and down to drivers that are able to support it. This is all also assuming that containers support is staying in Nova and not a new service. That discussion seems to have stalled. Is anyone still pushing on that? Any updates? -- Russell Bryant ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Nova][Docker] Environment variables
Hi Russel, I have something that is pushing it for to stay in nova (at least the compute drivers). I should have a gerrit branch for people to review soon. Regards chuck On Mon, Dec 16, 2013 at 10:07 AM, Russell Bryant rbry...@redhat.com wrote: On 12/16/2013 09:27 AM, Daniel Kuffner wrote: Hi All, I have submitted a new blueprint which addresses the a common pattern in the docker world. A usual pattern in the docker world is to use environment variables to configure a container. docker run -e SQL_URL=postgres://user:password@/db my-app The nova docker driver doesn't support to set any environment variables. To work around this issue I used cloud-init which works fine. But this approach has of course the drawback that a) I have to install the cloud init service. and b) my docker container doesn't work outside of openstack. I propose to allow a user to set docker environment variables via nova instance metadata. The metadata key should have a prefix like ENV_ which can be used to determine all environment variables. The prefix should be removed and the remainder key and vaule will be injected. The metadata can unfortunately not be set in horizon but can be used from the nova command line tool and from heat. Example heat: myapp: Type: OS::Nova::Server Properties: flavor: m1.small image: my-app:latest meta-data: - ENV_SQL_URL: postgres://user:password@/db - ENV_SOMETHING_ELSE: Value Let me know what you think about that. Blueprint: https://blueprints.launchpad.net/nova/+spec/docker-env-via-meta-data Thanks for starting the discussion. More people should do this for their blueprints. :-) One of the things we should be striving for is to provide as consistent of an experience as we can across drivers. Right now, we have the metadata service and config drive, and neither of those are driver specific. In the case of config drive, whether it's used or not is exposed through the API. As you point out, the meta-data service does technically work with the docker driver. I don't think we should support environment variables like this automatically. Instead, I think it would be more appropriate to add an API extension for specifying env vars. That way the behavior is more explicit and communicated through the API. The env vars would be passed through all of the appropriate plumbing and down to drivers that are able to support it. This is all also assuming that containers support is staying in Nova and not a new service. That discussion seems to have stalled. Is anyone still pushing on that? Any updates? -- Russell Bryant ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [nova] time for a new major network rpc api version?
On 12/15/2013 05:12 PM, Robert Collins wrote: That said, doing anything to the network RPC API seems premature until the Neutron question is resolved. This. I've been pretty much ignoring this API since it has been frozen and almost deprecated for a long time. My plan was to revisit the status of nova-network after the release of icehouse-2. -- Russell Bryant ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Nova][Docker] Environment variables
On 12/16/2013 10:12 AM, Chuck Short wrote: I have something that is pushing it for to stay in nova (at least the compute drivers). I should have a gerrit branch for people to review soon. OK. Do you have any design notes for whatever you're proposing? That would probably be easier to review and discuss. -- Russell Bryant ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Nova][Docker] Environment variables
Hi Russell, You actually propose to extend the whole nova stack to support environment variables. Would any other driver benefit from this API extension? Is that what you imagine? nova --env SQL_URL=postgres://user:password --image Regarding the discussion you mentioned. Are there any public resources to read. I kind of missed it. Most likely it was before I was part of this community :) thanks, Daniel On Mon, Dec 16, 2013 at 4:07 PM, Russell Bryant rbry...@redhat.com wrote: On 12/16/2013 09:27 AM, Daniel Kuffner wrote: Hi All, I have submitted a new blueprint which addresses the a common pattern in the docker world. A usual pattern in the docker world is to use environment variables to configure a container. docker run -e SQL_URL=postgres://user:password@/db my-app The nova docker driver doesn't support to set any environment variables. To work around this issue I used cloud-init which works fine. But this approach has of course the drawback that a) I have to install the cloud init service. and b) my docker container doesn't work outside of openstack. I propose to allow a user to set docker environment variables via nova instance metadata. The metadata key should have a prefix like ENV_ which can be used to determine all environment variables. The prefix should be removed and the remainder key and vaule will be injected. The metadata can unfortunately not be set in horizon but can be used from the nova command line tool and from heat. Example heat: myapp: Type: OS::Nova::Server Properties: flavor: m1.small image: my-app:latest meta-data: - ENV_SQL_URL: postgres://user:password@/db - ENV_SOMETHING_ELSE: Value Let me know what you think about that. Blueprint: https://blueprints.launchpad.net/nova/+spec/docker-env-via-meta-data Thanks for starting the discussion. More people should do this for their blueprints. :-) One of the things we should be striving for is to provide as consistent of an experience as we can across drivers. Right now, we have the metadata service and config drive, and neither of those are driver specific. In the case of config drive, whether it's used or not is exposed through the API. As you point out, the meta-data service does technically work with the docker driver. I don't think we should support environment variables like this automatically. Instead, I think it would be more appropriate to add an API extension for specifying env vars. That way the behavior is more explicit and communicated through the API. The env vars would be passed through all of the appropriate plumbing and down to drivers that are able to support it. This is all also assuming that containers support is staying in Nova and not a new service. That discussion seems to have stalled. Is anyone still pushing on that? Any updates? -- Russell Bryant ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Nova][Docker] Environment variables
Hi Chuck, yes please, I'm eager to see what you have. :) Daniel On Mon, Dec 16, 2013 at 4:12 PM, Chuck Short chuck.sh...@canonical.com wrote: Hi Russel, I have something that is pushing it for to stay in nova (at least the compute drivers). I should have a gerrit branch for people to review soon. Regards chuck On Mon, Dec 16, 2013 at 10:07 AM, Russell Bryant rbry...@redhat.com wrote: On 12/16/2013 09:27 AM, Daniel Kuffner wrote: Hi All, I have submitted a new blueprint which addresses the a common pattern in the docker world. A usual pattern in the docker world is to use environment variables to configure a container. docker run -e SQL_URL=postgres://user:password@/db my-app The nova docker driver doesn't support to set any environment variables. To work around this issue I used cloud-init which works fine. But this approach has of course the drawback that a) I have to install the cloud init service. and b) my docker container doesn't work outside of openstack. I propose to allow a user to set docker environment variables via nova instance metadata. The metadata key should have a prefix like ENV_ which can be used to determine all environment variables. The prefix should be removed and the remainder key and vaule will be injected. The metadata can unfortunately not be set in horizon but can be used from the nova command line tool and from heat. Example heat: myapp: Type: OS::Nova::Server Properties: flavor: m1.small image: my-app:latest meta-data: - ENV_SQL_URL: postgres://user:password@/db - ENV_SOMETHING_ELSE: Value Let me know what you think about that. Blueprint: https://blueprints.launchpad.net/nova/+spec/docker-env-via-meta-data Thanks for starting the discussion. More people should do this for their blueprints. :-) One of the things we should be striving for is to provide as consistent of an experience as we can across drivers. Right now, we have the metadata service and config drive, and neither of those are driver specific. In the case of config drive, whether it's used or not is exposed through the API. As you point out, the meta-data service does technically work with the docker driver. I don't think we should support environment variables like this automatically. Instead, I think it would be more appropriate to add an API extension for specifying env vars. That way the behavior is more explicit and communicated through the API. The env vars would be passed through all of the appropriate plumbing and down to drivers that are able to support it. This is all also assuming that containers support is staying in Nova and not a new service. That discussion seems to have stalled. Is anyone still pushing on that? Any updates? -- Russell Bryant ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [TripleO] [Horizon] [Tuskar] [UI] Horizon and Tuskar-UI merge
On 12/16/2013 03:32 PM, Jaromir Coufal wrote: On 2013/16/12 14:03, Matthias Runge wrote: On 12/13/2013 03:08 PM, Ladislav Smola wrote: Horizoners, As discussed in TripleO and Horizon meetings, we are proposing to move Tuskar UI under the Horizon umbrella. Since we are building our UI solution on top of Horizon, we think this is a good fit. It will allow us to get feedback and reviews from the appropriate group of developers. I don't think, we really disagree here. My main concern would be more: what do we get, if we make up another project under the umbrella of horizon? I mean, what does that mean at all? My proposal would be, to send patches directly to horizon. As discussed in last weeks horizon meeting, tuskar UI would become integrated in Horizon, but disabled by default. This would enable a faster integration in Horizon and would reduce the overhead of creating a separate repositoy, installation instructions, packaging etc. etc. From the horizon side: we would get some new contributors (and hopefully reviewers), which is very much appreciated. Matthias This is important note. From information architecture and user interaction point of view, I don't think it makes sense to keep all the three tabs visible together (Project, Admin, Infrastructure). There are lot of reasons, but main points: * Infrastructure itself is undercloud concept running in different instance of Horizon. * Users dealing with deployment and infrastructure management are not the users of OpenStack UI / Dashboard. It is different set of users. So it doesn't make sense to have giant application, which provides each and every possible feature. I think we need to keep focused. So by default, I would say that there should exist Project + Admin tab together or Infrastructure. But never all three together. So when Matthias say 'disabled by default', I would mean completely hidden for user and if user wants to use Infrastructure management, he can enable it in different horizon instance, but it will be the only visible tab for him. So it will be sort of separate application, but still running on top of Horizon. -- Jarda ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev Thanks for pointing this out, In Horizon you can easily decide which dashboards to show, so the Infrastructure management Horizon instance can have Project and Admin dashboards disabled. I think there has been discussed that some panels of Admin dashboard should be required for infrastructure management. We can solve this by adding those selected Admin panels also into Infrastructure dashboard. Jirka ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [Solum] Third working group meeting on language packs today
Hi, We will hold our third Git Integration working group meeting on IRC in #solum on Monday, December 16, 2013 1700 UTC / 0900 PST. Previous meeting notes are here [4] Agenda for today's meeting: * Administrative: * Topics: * Discuss lang-pack-examples spec for inclusion into M1 [1] * Discuss specify-lang-pack proposal [2] * Feedback on etherpad items created last week [3] * Discuss alternative names for language-pack and set up a poll * General discussion [1] https://blueprints.launchpad.net/solum/+spec/lang-pack-examples [2] https://blueprints.launchpad.net/solum/+spec/specify-lang-pack [3] https://etherpad.openstack.org/p/Solum-Language-pack-json-format [4] http://irclogs.solum.io/2013/solum.2013-12-09-17.01.html ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [nova] VM diagnostics - V3 proposal
On Mon, Dec 16, 2013 at 06:58:24AM -0800, Gary Kotton wrote: Hi, At the moment the administrator is able to retrieve diagnostics for a running VM. Currently the implementation is very loosely defined, that is, each driver returns whatever they have to return. This is problematic in a number of respects: 1. The tempest tests were written specifically for one driver and break with all other drivers (the test was removed to prevent this – bug 1240043) 2. An admin is unable to write tools that may work with a hybrid cloud 3. Adding support for get_diagnostics for drivers that do not support is painful Technically 3 is currently easy, since currently you don't need to care about what the other drivers have done - you can return any old info for your new driver's get_diagnostics API impl ;-) Seriously though, I agree the current API is a big trainwreck. I'd like to propose the following for the V3 API (we will not touch V2 in case operators have applications that are written against this – this may be the case for libvirt or xen. The VMware API support was added in I1): 1. We formalize the data that is returned by the API [1] Before we debate what standard data should be returned we need detail of exactly what info the current 3 virt drivers return. IMHO it would be better if we did this all in the existing wiki page associated with the blueprint, rather than etherpad, so it serves as a permanent historical record for the blueprint design. While we're doing this I think we should also consider whether the 'get_diagnostics' API is fit for purpose more generally. eg currently it is restricted to administrators. Some, if not all, of the data libvirt returns is relevant to the owner of the VM but they can not get at it. For a cloud administrator it might be argued that the current API is too inefficient to be useful in many troubleshooting scenarios since it requires you to invoke it once per instance if you're collecting info on a set of guests, eg all VMs on one host. It could be that cloud admins would be better served by an API which returned info for all VMs ona host at once, if they're monitoring say, I/O stats across VM disks to identify one that is causing I/O trouble ? IOW, I think we could do with better identifying the usage scenarios for this API if we're to improve its design / impl. 2. We enable the driver to add extra information that will assist the administrators in troubleshooting problems for VM's I have proposed a BP for this - https://blueprints.launchpad.net/nova/+spec/diagnostics-namespace (I'd like to change the name to v3-api-diagnostics – which is more apt) The bp rename would be a good idea. [1] https://etherpad.openstack.org/p/vm-diagnostics Regards, Daniel -- |: http://berrange.com -o-http://www.flickr.com/photos/dberrange/ :| |: http://libvirt.org -o- http://virt-manager.org :| |: http://autobuild.org -o- http://search.cpan.org/~danberr/ :| |: http://entangle-photo.org -o- http://live.gnome.org/gtk-vnc :| ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Nova][Docker] Environment variables
That would be great, I have also a couple of change request waiting for approval. Would be good to know if it has any relevance in the future. https://review.openstack.org/#/c/59824/ https://review.openstack.org/#/c/62182/ https://review.openstack.org/#/c/62183/ https://review.openstack.org/#/c/62220/ Daniel On Mon, Dec 16, 2013 at 4:17 PM, Russell Bryant rbry...@redhat.com wrote: On 12/16/2013 10:12 AM, Chuck Short wrote: I have something that is pushing it for to stay in nova (at least the compute drivers). I should have a gerrit branch for people to review soon. OK. Do you have any design notes for whatever you're proposing? That would probably be easier to review and discuss. -- Russell Bryant ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [nova] VM diagnostics - V3 proposal
On 16 December 2013 15:25, Daniel P. Berrange berra...@redhat.com wrote: On Mon, Dec 16, 2013 at 06:58:24AM -0800, Gary Kotton wrote: Hi, At the moment the administrator is able to retrieve diagnostics for a running VM. Currently the implementation is very loosely defined, that is, each driver returns whatever they have to return. This is problematic in a number of respects: 1. The tempest tests were written specifically for one driver and break with all other drivers (the test was removed to prevent this – bug 1240043) 2. An admin is unable to write tools that may work with a hybrid cloud 3. Adding support for get_diagnostics for drivers that do not support is painful Technically 3 is currently easy, since currently you don't need to care about what the other drivers have done - you can return any old info for your new driver's get_diagnostics API impl ;-) Seriously though, I agree the current API is a big trainwreck. +1 I'd like to propose the following for the V3 API (we will not touch V2 in case operators have applications that are written against this – this may be the case for libvirt or xen. The VMware API support was added in I1): 1. We formalize the data that is returned by the API [1] Before we debate what standard data should be returned we need detail of exactly what info the current 3 virt drivers return. IMHO it would be better if we did this all in the existing wiki page associated with the blueprint, rather than etherpad, so it serves as a permanent historical record for the blueprint design. +1 While we're doing this I think we should also consider whether the 'get_diagnostics' API is fit for purpose more generally. eg currently it is restricted to administrators. Some, if not all, of the data libvirt returns is relevant to the owner of the VM but they can not get at it. Ceilometer covers that ground, we should ask them about this API. For a cloud administrator it might be argued that the current API is too inefficient to be useful in many troubleshooting scenarios since it requires you to invoke it once per instance if you're collecting info on a set of guests, eg all VMs on one host. It could be that cloud admins would be better served by an API which returned info for all VMs ona host at once, if they're monitoring say, I/O stats across VM disks to identify one that is causing I/O trouble ? IOW, I think we could do with better identifying the usage scenarios for this API if we're to improve its design / impl. I like the API that helps you dig into info for a specific host that other system highlight as problematic. You can do things that could be expensive to compute, but useful for troubleshooting. But you are right, we should think about it first. 2. We enable the driver to add extra information that will assist the administrators in troubleshooting problems for VM's I think we need to version this information, if possible. I don't like the idea of the driver just changing the public API as it wishes. I have proposed a BP for this - https://blueprints.launchpad.net/nova/+spec/diagnostics-namespace (I'd like to change the name to v3-api-diagnostics – which is more apt) The bp rename would be a good idea. +1 [1] https://etherpad.openstack.org/p/vm-diagnostics John ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Nova][Docker] Environment variables
On 12/16/2013 10:18 AM, Daniel Kuffner wrote: Hi Russell, You actually propose to extend the whole nova stack to support environment variables. Would any other driver benefit from this API extension? Is that what you imagine? nova --env SQL_URL=postgres://user:password --image Yes. Regarding the discussion you mentioned. Are there any public resources to read. I kind of missed it. Most likely it was before I was part of this community :) It started here back in November: http://lists.openstack.org/pipermail/openstack-dev/2013-November/019637.html and then there have been a few messages on that thread this month, too. -- Russell Bryant ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Nova][Docker] Environment variables
On Mon, Dec 16, 2013 at 04:18:52PM +0100, Daniel Kuffner wrote: Hi Russell, You actually propose to extend the whole nova stack to support environment variables. Would any other driver benefit from this API extension? Is that what you imagine? nova --env SQL_URL=postgres://user:password --image Regarding the discussion you mentioned. Are there any public resources to read. I kind of missed it. Most likely it was before I was part of this community :) With glance images we have a way to associate arbitrary metadata attributes with the image. I could see using this mechanism to associate some default set of environment variables. eg use a 'env_' prefix for glance image attributes We've got a couple of cases now where we want to overrides these same things on a per-instance basis. Kernel command line args is one other example. Other hardware overrides like disk/net device types are another possibility Rather than invent new extensions for each, I think we should have a way to pass arbitrary attributes alon with the boot API call, that a driver would handle in much the same way as they do for glance image properties. Basically think of it as a way to custom any image property per instance created. Daniel -- |: http://berrange.com -o-http://www.flickr.com/photos/dberrange/ :| |: http://libvirt.org -o- http://virt-manager.org :| |: http://autobuild.org -o- http://search.cpan.org/~danberr/ :| |: http://entangle-photo.org -o- http://live.gnome.org/gtk-vnc :| ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Nova][Docker] Environment variables
On 12/16/2013 10:39 AM, Daniel P. Berrange wrote: On Mon, Dec 16, 2013 at 04:18:52PM +0100, Daniel Kuffner wrote: Hi Russell, You actually propose to extend the whole nova stack to support environment variables. Would any other driver benefit from this API extension? Is that what you imagine? nova --env SQL_URL=postgres://user:password --image Regarding the discussion you mentioned. Are there any public resources to read. I kind of missed it. Most likely it was before I was part of this community :) With glance images we have a way to associate arbitrary metadata attributes with the image. I could see using this mechanism to associate some default set of environment variables. eg use a 'env_' prefix for glance image attributes We've got a couple of cases now where we want to overrides these same things on a per-instance basis. Kernel command line args is one other example. Other hardware overrides like disk/net device types are another possibility Rather than invent new extensions for each, I think we should have a way to pass arbitrary attributes alon with the boot API call, that a driver would handle in much the same way as they do for glance image properties. Basically think of it as a way to custom any image property per instance created. That's a pretty nice idea. I like it. -- Russell Bryant ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [TripleO][Tuskar] Icehouse Requirements
On 13/12/13 19:06 +1300, Robert Collins wrote: On 13 December 2013 06:24, Will Foster wfos...@redhat.com wrote: I just wanted to add a few thoughts: Thank you! For some comparative information here from the field I work extensively on deployments of large OpenStack implementations, most recently with a ~220node/9rack deployment (scaling up to 42racks / 1024 nodes soon). My primary role is of a Devops/Sysadmin nature, and not a specific development area so rapid provisioning/tooling/automation is an area I almost exclusively work within (mostly using API-driven using Foreman/Puppet). The infrastructure our small team designs/builds supports our development and business. I am the target user base you'd probably want to cater to. Absolutely! I can tell you the philosophy and mechanics of Tuskar/OOO are great, something I'd love to start using extensively but there are some needed aspects in the areas of control that I feel should be added (though arguably less for me and more for my ilk who are looking to expand their OpenStack footprint). * ability to 'preview' changes going to the scheduler What does this give you? How detailed a preview do you need? What information is critical there? Have you seen the proposed designs for a heat template preview feature - would that be sufficient? Thanks for the reply. Preview-wise it'd be useful to see node allocation prior to deployment - nothing too in-depth. I have not seen the heat template preview features, are you referring to the YAML templating[1] or something else[2]? I'd like to learn more. [1] - http://docs.openstack.org/developer/heat/template_guide/hot_guide.html [2] - https://github.com/openstack/heat-templates * ability to override/change some aspects within node assignment What would this be used to do? How often do those situations turn up? Whats the impact if you can't do that? One scenario might be that autodiscovery does not pick up an available node in your pool of resources, or detects incorrectly - you could manually change things as you like it. Another (more common) scenario is that you don't have an isolated, flat network with which to deploy and nodes are picked that you do not want included in the provisioning - you could remove those from the set of resources prior to launching overcloud creation. The impact would be that the tooling would seem inflexible to those lacking a thoughtfully prepared network/infrastructure, or more commonly in cases where the existing network design is too inflexible the usefulness and quick/seamless provisioning benefits would fall short. * ability to view at least minimal logging from within Tuskar UI Logging of what - the deployment engine? The heat event-log? Nova undercloud logs? Logs from the deployed instances? If it's not there in V1, but you can get, or already have credentials for the [instances that hold the logs that you wanted] would that be a big adoption blocker, or just a nuisance? Logging of the deployment engine status during the bootstrapping process initially, and some rudimentary node success/failure indication. It should be simplistic enough to not rival existing monitoring/log systems but at least provide deployment logs as the overcloud is being built and a general node/health 'check-in' that it's complete. Afterwards as you mentioned the logs are available on the deployed systems. Think of it as providing some basic written navigational signs for people crossing a small bridge before they get to the highway, there's continuity from start - finish and a clear sense of what's occurring. From my perspective, absence of this type of verbosity may impede adoption of new users (who are used to this type of information with deployment tooling). Here's the main reason - most new adopters of OpenStack/IaaS are going to be running legacy/mixed hardware and while they might have an initiative to explore and invest and even a decent budget most of them are not going to have completely identical hardware, isolated/flat networks and things set aside in such a way that blind auto-discovery/deployment will just work all the time. Thats great information (and something I reasonably well expected, to a degree). We have a hard dependency on no wildcard DHCP servers in the environment (or we can't deploy). Autodiscovery is something we don't have yet, but certainly debugging deployment failures is a very important use case and one we need to improve both at the plumbing layer and in the stories around it in the UI. There will be a need to sometimes adjust, and those coming from a more vertically-scaling infrastructure (most large orgs.) will not have 100% matching standards in place of vendor, machine spec and network design which may make Tuscar/OOO seem inflexible and 'one-way'. This may just be a carry-over or fear of the old ways of deployment but nonetheless it is present. I'm not sure what you mean by matching standards here :). Ironic is
Re: [openstack-dev] [nova] VM diagnostics - V3 proposal
On Mon, Dec 16, 2013 at 03:37:39PM +, John Garbutt wrote: On 16 December 2013 15:25, Daniel P. Berrange berra...@redhat.com wrote: On Mon, Dec 16, 2013 at 06:58:24AM -0800, Gary Kotton wrote: I'd like to propose the following for the V3 API (we will not touch V2 in case operators have applications that are written against this – this may be the case for libvirt or xen. The VMware API support was added in I1): 1. We formalize the data that is returned by the API [1] Before we debate what standard data should be returned we need detail of exactly what info the current 3 virt drivers return. IMHO it would be better if we did this all in the existing wiki page associated with the blueprint, rather than etherpad, so it serves as a permanent historical record for the blueprint design. +1 While we're doing this I think we should also consider whether the 'get_diagnostics' API is fit for purpose more generally. eg currently it is restricted to administrators. Some, if not all, of the data libvirt returns is relevant to the owner of the VM but they can not get at it. Ceilometer covers that ground, we should ask them about this API. If we consider what is potentially in scope for ceilometer and subtract that from what the libvirt get_diagnostics impl currently returns, you pretty much end up with the empty set. This might cause us to question if 'get_diagnostics' should exist at all from the POV of the libvirt driver's impl. Perhaps vmware/xen return data that is out of scope for ceilometer ? For a cloud administrator it might be argued that the current API is too inefficient to be useful in many troubleshooting scenarios since it requires you to invoke it once per instance if you're collecting info on a set of guests, eg all VMs on one host. It could be that cloud admins would be better served by an API which returned info for all VMs ona host at once, if they're monitoring say, I/O stats across VM disks to identify one that is causing I/O trouble ? IOW, I think we could do with better identifying the usage scenarios for this API if we're to improve its design / impl. I like the API that helps you dig into info for a specific host that other system highlight as problematic. You can do things that could be expensive to compute, but useful for troubleshooting. If things get expensive to compute, then it may well be preferrable to have separate APIs for distinct pieces of interesting diagnostic data. eg If they only care about one particular thing, they don't want to trigger expensive computations of things they don't care about seeing. Regards, Daniel -- |: http://berrange.com -o-http://www.flickr.com/photos/dberrange/ :| |: http://libvirt.org -o- http://virt-manager.org :| |: http://autobuild.org -o- http://search.cpan.org/~danberr/ :| |: http://entangle-photo.org -o- http://live.gnome.org/gtk-vnc :| ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [nova] VM diagnostics - V3 proposal
On 12/16/13 5:25 PM, Daniel P. Berrange berra...@redhat.com wrote: On Mon, Dec 16, 2013 at 06:58:24AM -0800, Gary Kotton wrote: Hi, At the moment the administrator is able to retrieve diagnostics for a running VM. Currently the implementation is very loosely defined, that is, each driver returns whatever they have to return. This is problematic in a number of respects: 1. The tempest tests were written specifically for one driver and break with all other drivers (the test was removed to prevent this bug 1240043) 2. An admin is unable to write tools that may work with a hybrid cloud 3. Adding support for get_diagnostics for drivers that do not support is painful Technically 3 is currently easy, since currently you don't need to care about what the other drivers have done - you can return any old info for your new driver's get_diagnostics API impl ;-) To be honest it was not easy at all. Seriously though, I agree the current API is a big trainwreck. I'd like to propose the following for the V3 API (we will not touch V2 in case operators have applications that are written against this this may be the case for libvirt or xen. The VMware API support was added in I1): 1. We formalize the data that is returned by the API [1] Before we debate what standard data should be returned we need detail of exactly what info the current 3 virt drivers return. IMHO it would be better if we did this all in the existing wiki page associated with the blueprint, rather than etherpad, so it serves as a permanent historical record for the blueprint design. I will add this to the wiki. Not sure what this will achieve - other than the fact that it will crystalize the fact that we need to have common data returned. While we're doing this I think we should also consider whether the 'get_diagnostics' API is fit for purpose more generally. eg currently it is restricted to administrators. Some, if not all, of the data libvirt returns is relevant to the owner of the VM but they can not get at it. This is configurable. The default is for an admin user. This is in the policy.json file - https://github.com/openstack/nova/blob/master/etc/nova/policy.json#L202 For a cloud administrator it might be argued that the current API is too inefficient to be useful in many troubleshooting scenarios since it requires you to invoke it once per instance if you're collecting info on a set of guests, eg all VMs on one host. It could be that cloud admins would be better served by an API which returned info for all VMs ona host at once, if they're monitoring say, I/O stats across VM disks to identify one that is causing I/O trouble ? IOW, I think we could do with better identifying the usage scenarios for this API if we're to improve its design / impl. Host diagnostics would be a nice feature to have. I do not think that it is part of the scope of what we want to achieve here but I will certainly be happy to work on this afterwards. 2. We enable the driver to add extra information that will assist the administrators in troubleshooting problems for VM's I have proposed a BP for this - https://urldefense.proofpoint.com/v1/url?u=https://blueprints.launchpad.n et/nova/%2Bspec/diagnostics-namespacek=oIvRg1%2BdGAgOoM1BIlLLqw%3D%3D%0A r=eH0pxTUZo8NPZyF6hgoMQu%2BfDtysg45MkPhCZFxPEq8%3D%0Am=xY7Bdz7UGQQFxbe2 g6zVO%2FBUpkZfN%2BImM564xugLjsk%3D%0As=9d0ecd519b919b6c87bbdd2e1e1bf9a51 6f469143d15797d272cfd8c7e2d0686 (I'd like to change the name to v3-api-diagnostics which is more apt) The bp rename would be a good idea. [1] https://urldefense.proofpoint.com/v1/url?u=https://etherpad.openstack.org /p/vm-diagnosticsk=oIvRg1%2BdGAgOoM1BIlLLqw%3D%3D%0Ar=eH0pxTUZo8NPZyF6h goMQu%2BfDtysg45MkPhCZFxPEq8%3D%0Am=xY7Bdz7UGQQFxbe2g6zVO%2FBUpkZfN%2BIm M564xugLjsk%3D%0As=d1386b91ca07f5504844e7f4312dc5b53b709660fe71ca96c76c3 8d447bec2e5 Regards, Daniel -- |: https://urldefense.proofpoint.com/v1/url?u=http://berrange.com/k=oIvRg1%2 BdGAgOoM1BIlLLqw%3D%3D%0Ar=eH0pxTUZo8NPZyF6hgoMQu%2BfDtysg45MkPhCZFxPEq8% 3D%0Am=xY7Bdz7UGQQFxbe2g6zVO%2FBUpkZfN%2BImM564xugLjsk%3D%0As=c421c25857 f1ca0294b5cc318e87a758a2b49ecc35b3ca9f75b57be574ce0299 -o- https://urldefense.proofpoint.com/v1/url?u=http://www.flickr.com/photos/db errange/k=oIvRg1%2BdGAgOoM1BIlLLqw%3D%3D%0Ar=eH0pxTUZo8NPZyF6hgoMQu%2BfD tysg45MkPhCZFxPEq8%3D%0Am=xY7Bdz7UGQQFxbe2g6zVO%2FBUpkZfN%2BImM564xugLjsk %3D%0As=281520d30342d840da18dac821fdc235faf903c0bb7e8fcb51620217bf7b236a :| |: https://urldefense.proofpoint.com/v1/url?u=http://libvirt.org/k=oIvRg1%2B dGAgOoM1BIlLLqw%3D%3D%0Ar=eH0pxTUZo8NPZyF6hgoMQu%2BfDtysg45MkPhCZFxPEq8%3 D%0Am=xY7Bdz7UGQQFxbe2g6zVO%2FBUpkZfN%2BImM564xugLjsk%3D%0As=9424295e978 7fe72415305745f36556f0b8167ba0da8ac9632a4f8a183a926aa -o- https://urldefense.proofpoint.com/v1/url?u=http://virt-manager.org/k=oIvR g1%2BdGAgOoM1BIlLLqw%3D%3D%0Ar=eH0pxTUZo8NPZyF6hgoMQu%2BfDtysg45MkPhCZFxP
Re: [openstack-dev] [Nova][libvirt]when deleting instance which is in migrating state, instance files can be stay in destination node forever
I would block it in the API or have the API cancelling the migration first. I don't see a reason why to start an operation that is meant to fail, which also has a complex chain of event, following it failure. Regardless of the above, I think that the suggested exception handling is needed in any case. Vladik - Original Message - From: Loganathan Parthipan parthi...@hp.com To: OpenStack Development Mailing List (not for usage questions) openstack-dev@lists.openstack.org Sent: Monday, 16 December, 2013 8:25:09 AM Subject: Re: [openstack-dev] [Nova][libvirt]when deleting instance which is in migrating state, instance files can be stay in destination node forever Isn’t just handling the exception instance_not_found enough? By this time source would’ve been cleaned up. Destination VM resources will get cleaned up by the periodic task since the VM is not associated with this host. Am I missing something here? From: 王宏 [mailto:w.wangho...@gmail.com] Sent: 16 December 2013 11:32 To: openstack-dev@lists.openstack.org Subject: [openstack-dev] [Nova][libvirt]when deleting instance which is in migrating state, instance files can be stay in destination node forever Hi all. When I try to fix a bug: https://bugs.launchpad.net/nova/+bug/1242961 , I get a trouble. To reproduce the bug is very easy. Live migrate a vm in block_migration mode, and then delelte the vm immediately. The reason of this bug is as follow: 1. Because live migrate costs more time, so the vm will be deleted sucessfully before live migrate complete. And then, we will get an exception while live migrating. 2. After live migrate failed, we start to rollback. But, in the rollback method we will get or modify the info of vm from db. Because the vm has been deleted already, so we will get instance_not_found exception and rollback will be faild too. I have two ways to fix the bug: i)Add check in nova-api. When try to delete a vm, we return an error message if the vm_state is LIVE_MIGRATING. This way is very simple, but need to carefully consider. I have found a related discussion: http://lists.openstack.org/pipermail/openstack-dev/2013-October/017454.html , but it has no result in the discussion. ii)Before live migrate we get all the data needed by rollback method, and add a new rollback method. The new method will clean up resources at destination based on the above data(The resouces at source has been already cleaned up by deleting). I have no idea whitch one I should choose. Or, any other ideas?:) Regards, wanghong ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Ceilometer] Nomination of Sandy Walsh to core team
+1. Sandy has been both helpful and insightful, plus he happens to have a good handle on the many moving parts that make up this project. :-) -Phil -Original Message- From: Lu, Lianhao [mailto:lianhao...@intel.com] Sent: Sunday, December 15, 2013 5:29 PM To: OpenStack Development Mailing List (not for usage questions) Subject: Re: [openstack-dev] [Ceilometer] Nomination of Sandy Walsh to core team +1 in support. -Lianhao Nicolas Barcet wrote on 2013-12-13: +1 in support of Sandy. He is a proven contributor and reviewer and he brings a great business vision and experience to the team. Cheers, Nick On Wed, Dec 11, 2013 at 8:18 PM, Gordon Chung chu...@ca.ibm.com wrote: To that end, I would like to nominate Sandy Walsh from Rackspace to ceilometer-core. Sandy is one of the original authors of StackTach, and spearheaded the original stacktach-ceilometer integration. He has been instrumental in many of my codes reviews, and has contributed much of the existing event storage and querying code. +1 in support of Sandy. the Event work he's led in Ceilometer has been an important feature and i think he has some valuable ideas. cheers, gordon chungopenstack, ibm software standards ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Nova][libvirt]when deleting instance which is in migrating state, instance files can be stay in destination node forever
could we use Taskflow https://wiki.openstack.org/wiki/TaskFlow to manage task state and resource for this kind of tasks in Nova? Cinder has been an pilot to use Taskflow for volume backup tasks. anyone interested in this suggestion or has done some research to improve the live migration workflow? 2013/12/17 Vladik Romanovsky vladik.romanov...@enovance.com I would block it in the API or have the API cancelling the migration first. I don't see a reason why to start an operation that is meant to fail, which also has a complex chain of event, following it failure. Regardless of the above, I think that the suggested exception handling is needed in any case. Vladik - Original Message - From: Loganathan Parthipan parthi...@hp.com To: OpenStack Development Mailing List (not for usage questions) openstack-dev@lists.openstack.org Sent: Monday, 16 December, 2013 8:25:09 AM Subject: Re: [openstack-dev] [Nova][libvirt]when deleting instance which is in migrating state, instance files can be stay in destination node forever Isn’t just handling the exception instance_not_found enough? By this time source would’ve been cleaned up. Destination VM resources will get cleaned up by the periodic task since the VM is not associated with this host. Am I missing something here? From: 王宏 [mailto:w.wangho...@gmail.com] Sent: 16 December 2013 11:32 To: openstack-dev@lists.openstack.org Subject: [openstack-dev] [Nova][libvirt]when deleting instance which is in migrating state, instance files can be stay in destination node forever Hi all. When I try to fix a bug: https://bugs.launchpad.net/nova/+bug/1242961 , I get a trouble. To reproduce the bug is very easy. Live migrate a vm in block_migration mode, and then delelte the vm immediately. The reason of this bug is as follow: 1. Because live migrate costs more time, so the vm will be deleted sucessfully before live migrate complete. And then, we will get an exception while live migrating. 2. After live migrate failed, we start to rollback. But, in the rollback method we will get or modify the info of vm from db. Because the vm has been deleted already, so we will get instance_not_found exception and rollback will be faild too. I have two ways to fix the bug: i)Add check in nova-api. When try to delete a vm, we return an error message if the vm_state is LIVE_MIGRATING. This way is very simple, but need to carefully consider. I have found a related discussion: http://lists.openstack.org/pipermail/openstack-dev/2013-October/017454.html, but it has no result in the discussion. ii)Before live migrate we get all the data needed by rollback method, and add a new rollback method. The new method will clean up resources at destination based on the above data(The resouces at source has been already cleaned up by deleting). I have no idea whitch one I should choose. Or, any other ideas?:) Regards, wanghong ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev -- Tang Yaguang Canonical Ltd. | www.ubuntu.com | www.canonical.com Mobile: +86 152 1094 6968 gpg key: 0x187F664F ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] Unified Guest Agent proposal
On Fri, Dec 13, 2013 at 11:32:01AM -0800, Fox, Kevin M wrote: I hadn't thought about that use case, but that does sound like it would be a problem. That, at least, is not much of a problem, because you can block access to the metadata via a blackhole route or similar after you complete your initial configuration: ip route add blackhole 169.254.169.254 This prevents access to the metadata unless someone already has root access on the instance. -- Lars Kellogg-Stedman l...@redhat.com | larsks @ irc Cloud Engineering / OpenStack | @ twitter pgp4mFXCAneZr.pgp Description: PGP signature ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [infra] Meeting Tuesday December 17th at 19:00 UTC
The OpenStack Infrastructure (Infra) team is hosting our weekly meeting tomorrow, Tuesday December 17th, at 19:00 UTC in #openstack-meeting Meeting agenda available here: https://wiki.openstack.org/wiki/Meetings/InfraTeamMeeting (anyone is welcome to to add agenda items) Everyone interested in infrastructure and process surrounding automated testing and deployment is encouraged to attend. Note: As discussed at our last meeting, with our next two meeting dates landing on Christmas Eve and New Years Eve, this is likely to be the last formal team meeting team of the year. -- Elizabeth Krumbach Joseph || Lyz || pleia2 http://www.princessleia.com ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [nova][db] Thoughts on making instances.uuid non-nullable?
I've got a blueprint [1] scheduled for icehouse-3 to add DB2 support to Nova. That's blocked by a patch working it's way through sqlalchemy-migrate to add DB2 support [2] there. I've held off pushing any nova patches up until the sqlalchemy-migrate DB2 support is merged (which is also blocked by 3rd party CI, which is a WIP of it's own). Thinking ahead though for nova, one of the main issues with DB2 in the migration scripts is DB2 10.5 doesn't support unique constraints over nullable columns. The sqlalchemy-migrate code will instead create a unique index, since that's DB2's alternative. However, since a unique index is not a unique constraint, the FK creation fails if the UC doesn't exist. There are a lot of foreign keys in nova based on the instances.uuid column [3]. I need to figure out how I'm going to solve the UC problem for DB2 in that case. Here are the options as I see them, looking for input on the best way to go. 1. Add a migration to change instances.uuid to non-nullable. Besides the obvious con of having yet another migration script, this seems the most straight-forward. The instance object class already defines the uuid field as non-nullable, so it's constrained at the objects layer, just not in the DB model. Plus I don't think we'd ever have a case where instance.uuid is null, right? Seems like a lot of things would break down if that happened. With this option I can build on top of it for the DB2 migration support to add the same FKs as the other engines. 2. When I push up the migration script changes for DB2, I make the instances.uuid (and any other similar cases) work in the DB2 case only, i.e. if the engine is 'ibm_db_sa', then instances.uuid is non-nullable. This could be done in the 160_havana migration script since moving to DB2 with nova is going to require a fresh migration anyway (there are some other older scripts that I'll have to change to work with migrating to DB2). I don't particularly care for this option since it makes the model inconsistent between backends, but the upside is it doesn't require a new migration for any other backend, only DB2 - and you'd have to run the migrations for DB2 support anyway. I'm trying to flesh this out early since I could start working on option 1 at any time if it's the agreed upon solution, but looking for input first because I don't want to make assumptions about what everyone thinks here. [1] https://blueprints.launchpad.net/nova/+spec/db2-database [2] https://review.openstack.org/#/c/55572/ [3] https://github.com/openstack/nova/blob/master/nova/db/sqlalchemy/migrate_repo/versions/160_havana.py#L1335 -- Thanks, Matt Riedemann ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] Re-initializing or dynamically configuring cinder driver
Ah, u might be able to do what u said. Try it out and see how far u can get :) I would be interested to know how u plan on waiting for all existing operations to finish. Maybe it's not so hard, not really sure... Sent from my really tiny device... On Dec 15, 2013, at 9:43 PM, iKhan ik.ibadk...@gmail.commailto:ik.ibadk...@gmail.com wrote: Ok, I though we can make make cinder-volume aware of SIGTERM call and make sure it terminates with cleaning all the existing operations. If its not possible then probably SIGHUB is the only solution. :( On Mon, Dec 16, 2013 at 10:25 AM, Joshua Harlow harlo...@yahoo-inc.commailto:harlo...@yahoo-inc.com wrote: It depends on the corruption that u are willing to tolerate. Sigterm means the process just terminates, what if said process was 3/4 through some operation (create_volume for example)?? Personally I am willing to tolerate zero corruption, reliability and consistency are foundational things for me. Others may be more tolerant though, seems worth further discussion IMHO. Sent from my really tiny device... On Dec 15, 2013, at 8:39 PM, iKhan ik.ibadk...@gmail.commailto:ik.ibadk...@gmail.com wrote: How about sending SIGTERM to child processes and then starting them? I know this is the hard way of achieving the objective and SIGHUP approach will handle it more gracefully. As you mentioned it is a major change, tentatively can we use SIGTERM to achieve the objective? On Mon, Dec 16, 2013 at 9:50 AM, Joshua Harlow harlo...@yahoo-inc.commailto:harlo...@yahoo-inc.com wrote: In your proposal does it means that the child process will be restarted (that means kill -9 or sigint??). If so, without taskflow to help (or other solution) that means operations in progress will be corrupted/lost. That seems bad... A SIGHUP approach could be handled more gracefully (but it does require some changes in the underlying codebase to do this refresh). Sent from my really tiny device... On Dec 15, 2013, at 3:11 AM, iKhan ik.ibadk...@gmail.commailto:ik.ibadk...@gmail.com wrote: I don't know if this is being planned in Icehouse, if not probably proposing an approach will help. We have seen cinder-volume service initialization part. Similarly if we can get our hands on child process that are running under cinder-volume service, if we terminate those process and restart them along with newly added backends. It might help us achieve the target. On Sun, Dec 15, 2013 at 12:49 PM, Joshua Harlow harlo...@yahoo-inc.commailto:harlo...@yahoo-inc.com wrote: I don't currently know of a one size fits all solution here. There was talk at the summit of having the cinder app respond to a SIGHUP signal and attempting to reload config on this signal. Dynamic reloading is tricky business (basically u need to unravel anything holding references to the old config values/affected by the old config values). I would start with a simple trial of this if u want to so it, part if the issue will likely be oslo.config (can that library understand dynamic reloading?) and then cinder drivers themselves (perhaps u need to create a registry of drivers that can dynamically reload on config reloads?). Start out with something simple, isolate the reloading as much as u can to a single area (something like the mentioned registry of objects that can be reloaded when a SIGHUP arrives) and see how it goes. It does seem like a nice feature if u can get it right :-) Sent from my really tiny device... On Dec 13, 2013, at 8:57 PM, iKhan ik.ibadk...@gmail.commailto:ik.ibadk...@gmail.com wrote: Hi All, At present cinder driver can be only configured with adding entries in conf file. Once these driver related entries are modified or added in conf file, we need to restart cinder-volume service to validate the conf entries and create a child process that runs in background. I am thinking of a way to re-initialize or dynamically configure cinder driver. So that I can accept the configuration from user on fly and perform operations. I think solution lies somewhere around oslo.config.cfg, but I am still unclear about how re-initializing can be achieved. Let know if anyone here is aware of any approach to re-initialize or dynamically configure a driver. -- Thanks, IK ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.orgmailto:OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.orgmailto:OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev -- Thanks, Ibad Khan 9686594607tel:9686594607 ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.orgmailto:OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [nova][db] Thoughts on making instances.uuid non-nullable?
+1 on a migration to make uuid a non-nullable column. I advocated a few patches back in Havana that make assumptions based on the UUID being present and unique per instance. If it gets nulled the VMware drivers will have have breakage and I have no idea how to avoid that reasonably without the UUID. On Mon, Dec 16, 2013 at 11:59 AM, Russell Bryant rbry...@redhat.com wrote: On 12/16/2013 11:45 AM, Matt Riedemann wrote: 1. Add a migration to change instances.uuid to non-nullable. Besides the obvious con of having yet another migration script, this seems the most straight-forward. The instance object class already defines the uuid field as non-nullable, so it's constrained at the objects layer, just not in the DB model. Plus I don't think we'd ever have a case where instance.uuid is null, right? Seems like a lot of things would break down if that happened. With this option I can build on top of it for the DB2 migration support to add the same FKs as the other engines. Yeah, having instance.uuid nullable doesn't seem valuable to me, so this seems OK. -- Russell Bryant ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev -- # Shawn.Hartsock - twitter: @hartsock - plus.google.com/+ShawnHartsock ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [nova][db] Thoughts on making instances.uuid non-nullable?
On 12/16/2013 11:45 AM, Matt Riedemann wrote: 1. Add a migration to change instances.uuid to non-nullable. Besides the obvious con of having yet another migration script, this seems the most straight-forward. The instance object class already defines the uuid field as non-nullable, so it's constrained at the objects layer, just not in the DB model. Plus I don't think we'd ever have a case where instance.uuid is null, right? Seems like a lot of things would break down if that happened. With this option I can build on top of it for the DB2 migration support to add the same FKs as the other engines. Yeah, having instance.uuid nullable doesn't seem valuable to me, so this seems OK. -- Russell Bryant ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [Mistral] Community meeting minutes - 12/16/2013
Hello, Thanks for joining us today in IRC, here are the links to meeting minutes and logs: Minutes: http://eavesdrop.openstack.org/meetings/mistral/2013/mistral.2013-12-16-16.00.html Logs: http://eavesdrop.openstack.org/meetings/mistral/2013/mistral.2013-12-16-16.00.log.html Join us next time. Renat Akhmerov @ Mirantis Inc. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [oslo]: implementing olso.messaging over amqp 1.0
On 12/16/2013 10:37 AM, Gordon Sim wrote: On 12/12/2013 02:14 PM, Flavio Percoco wrote: I've a draft in my head of how the amqp 1.0 driver could be implemented and how to map the current expectations of the messaging layer to the new protocol. I think a separate thread to discuss this mapping is worth it. There are some critical areas that definitely need more discussion I have also been looking at this, and trying to write up a simple design notes. Some of the questions that occurred to me while doing so are: * Use one link for all sends, with 'to' field set, or use a link for each target? * How to handle calls to one of a group of servers? * Use a distinct response address per request, or allow an address to be shared by multiple requests in conjunction with correlation id on responses? * Support both intermediated and direct communication? For all patterns? The aim in my view should be to have the driver support as many alternatives in deployment as possible without overcomplicating things, distorting the mapping or introducing server specific extensions. I have some notes to share if anyone is interested. I can send them to this list or put them upon the wiki or an etherpad or something. I've pasted these into an etherpad[1] for anyone interested. Please feel free to edit/augment etc, or even to query anything on this list. It's really just an initial draft to get the ball rolling. --Gordon [1] https://etherpad.openstack.org/p/olso.messaging_amqp_1.0 ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] Unified Guest Agent proposal
The idea being discussed is using 169.254.169.254 for long term messaging between a vm and some other process. For example, Trove - TroveVM. I guess this thread is getting too long. The details are getting lost. Thanks, Kevin From: Lars Kellogg-Stedman [l...@redhat.com] Sent: Monday, December 16, 2013 8:18 AM To: OpenStack Development Mailing List (not for usage questions) Subject: Re: [openstack-dev] Unified Guest Agent proposal On Fri, Dec 13, 2013 at 11:32:01AM -0800, Fox, Kevin M wrote: I hadn't thought about that use case, but that does sound like it would be a problem. That, at least, is not much of a problem, because you can block access to the metadata via a blackhole route or similar after you complete your initial configuration: ip route add blackhole 169.254.169.254 This prevents access to the metadata unless someone already has root access on the instance. -- Lars Kellogg-Stedman l...@redhat.com | larsks @ irc Cloud Engineering / OpenStack | @ twitter ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [marconi] Meeting agenda for tomorrow at 1500 UTC
The Marconi project team holds a weekly meeting in #openstack-meeting-alt on Tuesdays, 1500 UTChttp://www.timeanddate.com/worldclock/fixedtime.html?hour=15min=0sec=0. The next meeting is Tomorrow, Dec. 10. Everyone is welcome, but please take a minute to review the wiki before attending for the first time: http://wiki.openstack.org/marconi This week we will discuss progress made toward graduation, and should have some time to triage new bugs/bps. |-..._ '-._'`| \ ``` ``---... _ | | / /#\ }--..__..-{ ### } _ _ { 6 6 {^} {{\ -=- /}} {{{;.___.;}}} {{{) (}}}' `'': :'''` after/jgs `@` Proposed Agenda: * Review actions from last time * Review Graduation BPs/Bugs * Updates on bugs * Updates on blueprints * SQLAlchemy storage driver strategy * Open discussion (time permitting) If you have additions to the agenda, please add them to the wiki and note your IRC name so we can call on you during the meeting: http://wiki.openstack.org/Meetings/Marconi Cheers, --- @kgriffs Kurt Giffiths ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Nova][VMware] Deploy from vCenter template
IIRC someone who shows up at https://wiki.openstack.org/wiki/Meetings/VMwareAPI#Meetings is planning on working on that again for Icehouse-3 but there's some new debate on the best way to implement the desired effect. The goal of that change would be to avoid streaming the disk image out of vCenter for the purpose of then streaming the same image back into the same vCenter. That's really inefficient. So there's a Nova level change that could happen (that's the patch you saw) and there's a Glance level change that could happen, and there's a combination of both approaches that could happen. If you want to discuss it informally with the group that's looking into the problem I could probably make sure you end up talking to the right people on #openstack-vmware or if you pop into the weekly team meeting on IRC you could mention it during open discussion time. On Mon, Dec 16, 2013 at 3:27 AM, Qing Xin Meng mengq...@cn.ibm.com wrote: I saw a commit for Deploying from VMware vCenter template and found it's abandoned. *https://review.openstack.org/#/c/34903*https://review.openstack.org/#/c/34903 Anyone knows the plan to support the deployment from VMware vCenter template? Thanks! Best Regards ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev -- # Shawn.Hartsock - twitter: @hartsock - plus.google.com/+ShawnHartsock ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] Unified Guest Agent proposal
On 12/16/2013 10:29 AM, Fox, Kevin M wrote: Yeah, this is similar to what I am proposing. I think we just about have just about everything we need already. Thread started out discussing a slightly different use case then below. The use case is processing events like: User performs backup database B in Trove UI, Trove sends event backup-database with params B to the vm, vm response sometime later with done backup database B, Trove UI updates. The idea is we need a unified agent to receive the messages, perform the action and respond back to the event,. The main issues are, as I see it: * The VM might be on a private neutron network only. This is desirable for increased security. * We want the agent to be minimal so as not to have to maintain much in the VM's. Its hard to keep all those ducks in a row. * There is a desire not to have the agent allow arbitrary commands to execute in the VM for security reasons. If security is a concern of the unified agent, the best way to reduce the attack surface is to limit the number of interactions the agent can actually do. Special purpose code for each operation could easily be implemented. I know salt was mentioned as a possibility to solving this problem, but brings a whole host of new problems to content with. Having a unified agent doesn't mean we can't put special-purpose code for each service (eg trove) for each operation (eg backup) in said unified agent. We could even do this using cloud-init using the part-handler logic. We really need someone from the community to step up and drive this effort, as opposed to beating this thread into too much complexity, as mentioned previously by Clint. Regards -steve Thanks, Kevin From: Robert Collins [robe...@robertcollins.net] Sent: Sunday, December 15, 2013 6:44 PM To: OpenStack Development Mailing List (not for usage questions) Subject: Re: [openstack-dev] Unified Guest Agent proposal On 15 December 2013 21:17, Clint Byrum cl...@fewbar.com wrote: Excerpts from Steven Dake's message of 2013-12-14 09:00:53 -0800: On 12/13/2013 01:13 PM, Clint Byrum wrote: Excerpts from Dmitry Mescheryakov's message of 2013-12-13 12:01:01 -0800: Still, what about one more server process users will have to run? I see unified agent as library which can be easily adopted by both exiting and new OpenStack projects. The need to configure and maintain Salt server process is big burden for end users. That idea will definitely scare off adoption of the agent. And at the same time what are the gains of having that server process? I don't really see to many of them. I tend to agree, I don't see a big advantage to using something like salt, when the current extremely simplistic cfn-init + friends do the job. What specific problem does salt solve? I guess I missed that context in this long thread. Yes you missed the crux of the thread. There is a need to have agents that are _not_ general purpose like cfn-init and friends. They specifically need to be narrow in focus and not give the higher level service operator backdoor access to everything via SSH-like control. So, just spitballing, but: We have a metadata service. We want low-latency updates there (e.g. occ listening on long-poll). Ignore implementation for now. I assert that agent restrictness is really up to the agent. For instance, an agent that accepts one command 'do something' with args 'something', is clearly not restricted. So - mainly to tease requirements out: How would salt be different to: - heat-metadata with push notification of updates - an ORC script that looks for a list of requests in post-configure.d and executes them. trove-agent: - 'backup': db-id: '52' - 'backup': db-id: '43' - 'create': db-id: '93' initial-schema: [.] etc. ? -- Robert Collins rbtcoll...@hp.com Distinguished Technologist HP Converged Cloud ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [nova][db] Thoughts on making instances.uuid non-nullable?
On 12/16/2013 11:59 AM, Russell Bryant wrote: On 12/16/2013 11:45 AM, Matt Riedemann wrote: 1. Add a migration to change instances.uuid to non-nullable. Besides the obvious con of having yet another migration script, this seems the most straight-forward. The instance object class already defines the uuid field as non-nullable, so it's constrained at the objects layer, just not in the DB model. Plus I don't think we'd ever have a case where instance.uuid is null, right? Seems like a lot of things would break down if that happened. With this option I can build on top of it for the DB2 migration support to add the same FKs as the other engines. Yeah, having instance.uuid nullable doesn't seem valuable to me, so this seems OK. +1 -jay ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Neutron] Does any plugin require hairpinning to be enabled?
Hi, I have registered two blueprints, one in Nova and one in Neutron to make it a VIF attribute that the libvirt driver in Nova will honor. https://blueprints.launchpad.net/neutron/+spec/vif-attribute-for-hairpinning https://blueprints.launchpad.net/nova/+spec/nova-hairpin-vif-attribute -- Sean M. Collins ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [TripleO] [Horizon] [Tuskar] [UI] Horizon and Tuskar-UI merge
On 12/16/2013 04:22 PM, Jiri Tomasek wrote: Thanks for pointing this out, In Horizon you can easily decide which dashboards to show, so the Infrastructure management Horizon instance can have Project and Admin dashboards disabled. I think there has been discussed that some panels of Admin dashboard should be required for infrastructure management. We can solve this by adding those selected Admin panels also into Infrastructure dashboard. Jirka Oh, I would expect a new role for an infrastructure admin; that role shouldn't necessarily see running instances or tenants etc. at all. Matthias ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Heat] [Murano] [Solum] [Glance]Metadata repository initiative discussion for Glance
Hi, Doodle shows that the most suitable time is 10AM PST on Tuesday. Lets keep this time for Metadata Repository\Catalog meeting in #openstack-glance IRC channel. See you tomorrow! Thanks Georgy On Fri, Dec 13, 2013 at 12:09 PM, Georgy Okrokvertskhov gokrokvertsk...@mirantis.com wrote: Hi, It looks like I forgot to add Glance. Fixing this now. I am sorry for duplicating the thread. Thanks Georgy On Fri, Dec 13, 2013 at 12:02 PM, Georgy Okrokvertskhov gokrokvertsk...@mirantis.com wrote: Yes. It is a Pacific Standard Time. Thanks Georgy On Fri, Dec 13, 2013 at 12:01 PM, Keith Bray keith.b...@rackspace.comwrote: PT as in Pacific Standard Time? -Keith On Dec 13, 2013 1:56 PM, Georgy Okrokvertskhov gokrokvertsk...@mirantis.com wrote: Hi, It is PT. I will add this info to the doodle pool. Thanks Georgy On Fri, Dec 13, 2013 at 11:50 AM, Keith Bray keith.b...@rackspace.comwrote: What timezone is the poll in? It doesn't say on the Doodle page. Thanks, -Keith From: Georgy Okrokvertskhov gokrokvertsk...@mirantis.com Reply-To: OpenStack Development Mailing List (not for usage questions) openstack-dev@lists.openstack.org Date: Friday, December 13, 2013 12:21 PM To: OpenStack Development Mailing List openstack-dev@lists.openstack.org Subject: [openstack-dev] [Heat] [Murano] [Solum] Metadata repository initiative discussion for Glance Hi, Recently a Heater proposal was announced in openstack-dev mailing list. This discussion lead to a decision to add unified metadata service \ catalog capabilities into Glance. On the Glance weekly meeting this initiative was discussed and Glance team agreed to take a look onto BPs and API documents for metadata repository\catalog, in order to understand what can be done during Icehouse release and how to organize this work in general. There will be a separate meeting devoted to this initiative on Tuesday 12/17 in #openstack-glance channel. Exact time is not defined yet and I need time preferences from all parties. Here is a link to a doodle poll http://doodle.com/9f2vxrftizda9pun . Please select time slot which will be suitable for you. The agenda for this meeting is the following: 1. Define project goals in general 2. Discuss API for this service and find out what can be implemented during IceHouse release. 3. Define organizational stuff like how this initiative should be developed (branch of Glance or separate project within Glance program) Here is an etherpad https://etherpad.openstack.org/p/MetadataRepository-API for initial API version for this service. All project which are interested in metadata repository are welcome to discuss API and service itself. Currently there are several possible use cases for this service: 1. Heat template catalog 2. HOT Software orchestration scripts\recipes storage 3. Murano Application Catalog object storage 4. Solum assets storage Thanks Georgy -- Georgy Okrokvertskhov Technical Program Manager, Cloud and Infrastructure Services, Mirantis http://www.mirantis.com Tel. +1 650 963 9828 Mob. +1 650 996 3284 -- Georgy Okrokvertskhov Technical Program Manager, Cloud and Infrastructure Services, Mirantis http://www.mirantis.com Tel. +1 650 963 9828 Mob. +1 650 996 3284 -- Georgy Okrokvertskhov Technical Program Manager, Cloud and Infrastructure Services, Mirantis http://www.mirantis.com Tel. +1 650 963 9828 Mob. +1 650 996 3284 -- Georgy Okrokvertskhov Technical Program Manager, Cloud and Infrastructure Services, Mirantis http://www.mirantis.com Tel. +1 650 963 9828 Mob. +1 650 996 3284 ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [TripleO] UI Wireframes for Resource Management - ready for implementation
On 12/13/2013 01:53 PM, Tzu-Mainn Chen wrote: On 2013/13/12 11:20, Tzu-Mainn Chen wrote: These look good! Quick question - can you explain the purpose of Node Tags? Are they an additional way to filter nodes through nova-scheduler (is that even possible?), or are they there solely for display in the UI? Mainn We start easy, so that's solely for UI needs of filtering and monitoring (grouping of nodes). It is already in Ironic, so there is no reason why not to take advantage of it. -- Jarda Okay, great. Just for further clarification, are you expecting this UI filtering to be present in release 0? I don't think Ironic natively supports filtering by node tag, so that would be further work that would have to be done. Mainn I might be getting ahead of things, but will the tags be free-form entered by the user, pre-entered in a separate settings and selectable at node register/update time, or locked into a select few that we specify? ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] Project-Scoped Service Catalog Entries
I've run into a use case that doesn't currently seem to have a great solution: Let's say my users want to use a top-of-stack OpenStack project such as Heat, Trove, etc. that I don't currently support in my deployment. There's absolutely no reason these services can't live happily in a VM talking to Nova, etc. via the normal APIs. However, in order to have a good experience (Horizon integration, seamless CLI integration) the service needs to be in the Service Catalog. One user could have their service added to the catalog by an admin, but then everyone in the cloud would be using their VM. And if you have multiple users all doing the same thing in their own projects, you've got collisions! So, I submit to you all that there is value in having a way to scope Service Catalog entries to specific projects, and to allow users with appropriate permissions on their project to add/remove those project-level service catalog entries. This could be accomplished in a number of ways: * Adding a new field to the model to store a Project ID. * Adding it in a standardized manner to service metadata as with https://blueprints.launchpad.net/keystone/+spec/service-metadata * Adding it as an additional requirement as proposed by https://blueprints.launchpad.net/keystone/+spec/auth-mechanisms-for-services * Use the existing Region field to track project scope as a hack. * Something else... I see this as analogous to Nova's concept of per-project flavors, or Glance's private/public/shared image capabilities. Allowing explicit sharing would even be an interesting option for service endpoints. It all depends how far we would want to go with it. Feel free to offer feedback or other suggestions. Thanks! - Gabriel ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] Project-Scoped Service Catalog Entries
+1 There is also the use case where a new service is being introduced for everyone eventually but you wish to start with a few friends. In the event of problems, the effort to tidy up is much less. Documentation can be updated with the production environment. Tim -Original Message- From: Gabriel Hurley [mailto:gabriel.hur...@nebula.com] Sent: 16 December 2013 20:58 To: OpenStack Development Mailing List (openstack-dev@lists.openstack.org) Subject: [openstack-dev] Project-Scoped Service Catalog Entries I've run into a use case that doesn't currently seem to have a great solution: Let's say my users want to use a top-of-stack OpenStack project such as Heat, Trove, etc. that I don't currently support in my deployment. There's absolutely no reason these services can't live happily in a VM talking to Nova, etc. via the normal APIs. However, in order to have a good experience (Horizon integration, seamless CLI integration) the service needs to be in the Service Catalog. One user could have their service added to the catalog by an admin, but then everyone in the cloud would be using their VM. And if you have multiple users all doing the same thing in their own projects, you've got collisions! So, I submit to you all that there is value in having a way to scope Service Catalog entries to specific projects, and to allow users with appropriate permissions on their project to add/remove those project-level service catalog entries. This could be accomplished in a number of ways: * Adding a new field to the model to store a Project ID. * Adding it in a standardized manner to service metadata as with https://blueprints.launchpad.net/keystone/+spec/service-metadata * Adding it as an additional requirement as proposed by https://blueprints.launchpad.net/keystone/+spec/auth-mechanisms-for- services * Use the existing Region field to track project scope as a hack. * Something else... I see this as analogous to Nova's concept of per-project flavors, or Glance's private/public/shared image capabilities. Allowing explicit sharing would even be an interesting option for service endpoints. It all depends how far we would want to go with it. Feel free to offer feedback or other suggestions. Thanks! - Gabriel ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [Ceilometer] Complex query BP implementation
Hi guys, The first working version of the Complex filter expressions in API queries blueprint [1] was pushed for review[2]. We implemented a new query REST resource in order to provide rich query functionality for samples, alarms and alarm history. The future plans (in separated blueprints) with this new functionality is extending it to support Statistics and stored queries. The new feature is documented on Launchpad wiki[3], with an example for how to use the new query on the API. What is your opinion about this solution? I would appreciate some review comments and/or feedback on the implementation. :) [1] https://blueprints.launchpad.net/ceilometer/+spec/complex-filter-expressions-in-api-queries [2] https://review.openstack.org/#/q/status:open+project:openstack/ceilometer+branch:master+topic:bp/complex-filter-expressions-in-api-queries,n,z [3] https://wiki.openstack.org/wiki/Ceilometer/ComplexFilterExpressionsInAPIQueries Thanks and Best Regards, Ildiko ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [governance] Becoming a Program, before applying for incubation
On Mon, Dec 16, 2013 at 9:49 AM, Thierry Carrez thie...@openstack.orgwrote: Flavio Percoco wrote: What I'm arguing here is: 1. Programs that are not part of OpenStack's release cycle shouldn't be considered official nor they should have the rights that integrated projects have. 2. I think requesting Programs to exist at the early stages of the project is not necessary. I don't even think incubated projects should have programs. I do agree the project's mission and goals have to be clear but the program should be officially created *after* the project graduates from incubation. The reasoning here is that anything could happen during incubation. For example, a program created for project A - which is incubated - may change to cover a broader mission that will allow a newborn project B to fall under its umbrella, hence my previous proposal of having a incubation stage for programs as well. I think your concerns can be covered if we consider that programs covering incubated or promising projects should also somehow incubate. To avoid confusion I'd use a different term, let's say incoming programs for the sake of the discussion. Incoming programs would automatically graduate when one of their deliveries graduate to integrated status (for projects with such deliveries), or when the TC decides so (think: for horizontal programs like Documentation or Deployment). That doesn't change most of this proposal, which is that we'd encourage teams to ask to become an (incoming) program before they consider filing one of their projects for incubation. It seems like the implications of the incoming designation is the same as the emerging designation you suggested previously. :-) I like the idea of some sort of acknowledgement that there is a group working on a solution to a problem and that the solution hasn't reached sufficient maturity to be an incubated project. I prefer the name emerging over incoming but not strongly. The status of a fledgeling program in this state should be re-evaluated periodically, as we do with incubated projects, so I don't see a problem with creating such working groups (maybe that's a better name?) when there is sufficient interest and participation early on. I do like the idea of asking them to produce *something* -- a design doc, requirements list, some sort of detailed plan for doing whatever the program's mission would be -- before being granted this new official designation, to show that the people involved are prepared to spend time and effort, more than just saying yes, I'm interested, too. FWIW we already distinguish (on https://wiki.openstack.org/wiki/Programs) programs that are born out of an incubated project from other programs, so adding this incoming status would not change much. My proposal is to either not requesting any program to be created for incubated projects / emerging technologies or to have a program called 'Emerging Technologies' were all these projects could fit in. I don't think an Emerging Technologies program would make sense, since that would just be a weird assemblage of separate teams (how would that program elect a PTL ?). I prefer that they act as separate teams (which they are) and use the incoming Program concept described above. +1 The only difference is that, IMHO, projects under this program should not have all the rights that integrated projects and other programs have, although the program will definitely fall under the TCs authority. For example, projects under this program shouldn't be able to vote on the TCs elections. So *that* would be a change from where we stand today, which is that incubated project contributors get ATC status and vote on TC elections. We can go either way, consider incoming programs to be OpenStack programs in the sense of the TC charter, or not. I'm not convinced there is so much value in restricting TC voting access (or ATC status) to OpenStack programs. Incoming programs would all be placed under the authority of the TC so it's only fair that they have a vote. Also giving them ATC status gets them automatically invited to Design Summits, and getting incoming programs in Design Summits sounds like a good thing to do... Right, bringing them to the summits is a big goal, isn't it? Doug -- Thierry Carrez (ttx) ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [Openstack][Heat] Creating an Openstack project using Heat
Hi, I have installed Openstack with heat using packstack. One thing I noticed is that the Orchestration Heat option is only available inside a project view. Is this by design ? My use case it to create a project with images, networks,routers and firewall rules in a single workflow. I looked at the documentation and at this point there is no resource available to create a project or upload an image. Regards, Sayaji ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] Announcing Fuel
Thanks for support, as a starting point we have chosen Ironic - we really want to see it as a replacement of Fuel's existing provisioning layer, with an intention of participation and delivery of many features which we already know from the real-world installations with Fuel. We are trying to be more test-driven, so starting our contributions as proof of concept for functional testing framework for Ironic: https://review.openstack.org/#/c/62410/2, see the description in README file: https://review.openstack.org/#/c/62410/2/irci/README.md Distributed envs testing, especially PXE booting is the area of our interest. Next, we plan to collaboratively work on torrent-based provisioning driver. Fuel UI Tuskar UI are also very interesting topic. Mark - thanks for provided links. Let us get a few cycles to research what you have so far and get back to you. On Fri, Dec 13, 2013 at 6:11 PM, Liz Blanchard lsure...@redhat.com wrote: On Dec 13, 2013, at 8:04 AM, Jaromir Coufal jcou...@redhat.com wrote: On 2013/12/12 15:31, Mike Scherbakov wrote: Folks, Most of you by now have heard of Fuel, which we’ve been working on as a related OpenStack project for a period of time -https://github.com/stackforge/fuel-mainsee https://launchpad.net/fueland https://wiki.openstack.org/wiki/Fuel. The aim of the project is to provide a distribution agnostic and plug-in agnostic engine for preparing, configuring and ultimately deploying various “flavors” of OpenStack in production. We’ve also used Fuel in most of our customer engagements to stand up an OpenStack cloud. ... We’d love to open discussion on this and hear everybody’s thoughts on this direction. Hey Mike, it sounds all great. I'll be very happy to discuss all the UX efforts going on in TripleO/Tuskar UI together with intentions and future steps of Fuel. +1. The Fuel wizard has some great UX ideas to bring to our thoughts around deployment in the Tuskar UI! Great to hear these will be brought together, Liz Cheers -- Jarda --- Jaromir Coufal (jcoufal) --- OpenStack User Experience --- IRC: #openstack-ux (at FreeNode) --- Forum: http://ask-openstackux.rhcloud.com --- Wiki: https://wiki.openstack.org/wiki/UX ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev -- Mike Scherbakov #mihgen ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Nova][Docker] Environment variables
eg use a 'env_' prefix for glance image attributes We've got a couple of cases now where we want to overrides these same things on a per-instance basis. Kernel command line args is one other example. Other hardware overrides like disk/net device types are another possibility Rather than invent new extensions for each, I think we should have a way to pass arbitrary attributes alon with the boot API call, that a driver would handle in much the same way as they do for glance image properties. Basically think of it as a way to custom any image property per instance created. Personally, I think having a bunch of special case magic namespaces (even if documented) is less desirable than a proper API to do something like this. Especially a namespace that someone else could potentially use legitimately that would conflict. To me, this feels a lot like what I'm worried this effort will turn into, which is making containers support in Nova look like a bolt-on thing with a bunch of specialness required to make it behave. Anyone remember this bolt-on gem? nova boot --block-device-mapping vda=965453c9-02b5-4d5b-8ec0-3164a89bf6f4:::0 --flavor=m1.tiny --image=6415797a-7c03-45fe-b490-f9af99d2bae0 BFV I found that one amidst hundreds of forum threads of people confused about what incantation of magic they were supposed to do to make it actually boot from volume. Just MHO. --Dan ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] Announcing Fuel
On 17 December 2013 09:59, Mike Scherbakov mscherba...@mirantis.com wrote: Thanks for support, as a starting point we have chosen Ironic - we really want to see it as a replacement of Fuel's existing provisioning layer, with an intention of participation and delivery of many features which we already know from the real-world installations with Fuel. We are trying to be more test-driven, so starting our contributions as proof of concept for functional testing framework for Ironic: https://review.openstack.org/#/c/62410/2, see the description in README file: https://review.openstack.org/#/c/62410/2/irci/README.md Distributed envs testing, especially PXE booting is the area of our interest. Next, we plan to collaboratively work on torrent-based provisioning driver. Cool. Rather than torrents you may prefer a multicast driver, it should be about 50% network utilisation, linear IO for the target disk - much better. Rob -- Robert Collins rbtcoll...@hp.com Distinguished Technologist HP Converged Cloud ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Openstack][Heat] Creating an Openstack project using Heat
On 16/12/13 15:57, Sayaji Patil wrote: Hi, I have installed Openstack with heat using packstack. One thing I noticed is that the Orchestration Heat option is only available inside a project view. Is this by design ? My use case it to create a project with images, networks,routers and firewall rules in a single workflow. I looked at the documentation and at this point there is no resource available to create a project or upload an image. Regards, Sayaji This is the development mailing list; the appropriate forum for these kinds of questions would be the general OpenStack mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack cheers, Zane. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [cinder][qa] release notes for cinder v1 to v2?
Sorry for lost subject in last message. Is there a document that describes the api changes from v1 to v2, similar to the one documenting nova v2 to v3? -David ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] Project-Scoped Service Catalog Entries
On 12/16/2013 04:17 PM, Georgy Okrokvertskhov wrote: By they way, there is an initiative to create generic metadata repository based on Glance project. As services endpoints are just URLs they also can be stored in this Glance metadata repository and have all features related to visibility and access control provides by this repository. Well, hold on there :) The proposed enhancements to Glance for a generic application metadata repository are a bit different from a service catalog, for one main reason: the service catalog is returned by Keystone in the Keystone token, and the service catalog will be [1] scoped to the owner of the token that was authenticated by Keystone. I think because of the current relationship between the service catalog actually being within the construct of a returned Keystone token, Keystone, at least for the foreseeable future, is the appropriate place to do this kind of project-scoped service catalog. I'm actually pretty familiar with the code in Keystone, and I was planning on implementing the proper region resource in the 3.2 API [2]. I would not mind doing the coding on Gabe's proposed project-scoped catalog once [1] has been implemented (API spec: [3]). Best, -jay [1] https://blueprints.launchpad.net/keystone/+spec/service-scoped-tokens [2] https://review.openstack.org/#/c/54215/ [3] https://review.openstack.org/#/c/61869/ On Mon, Dec 16, 2013 at 12:21 PM, Tim Bell tim.b...@cern.ch mailto:tim.b...@cern.ch wrote: +1 There is also the use case where a new service is being introduced for everyone eventually but you wish to start with a few friends. In the event of problems, the effort to tidy up is much less. Documentation can be updated with the production environment. Tim -Original Message- From: Gabriel Hurley [mailto:gabriel.hur...@nebula.com mailto:gabriel.hur...@nebula.com] Sent: 16 December 2013 20:58 To: OpenStack Development Mailing List (openstack-dev@lists.openstack.org mailto:openstack-dev@lists.openstack.org) Subject: [openstack-dev] Project-Scoped Service Catalog Entries I've run into a use case that doesn't currently seem to have a great solution: Let's say my users want to use a top-of-stack OpenStack project such as Heat, Trove, etc. that I don't currently support in my deployment. There's absolutely no reason these services can't live happily in a VM talking to Nova, etc. via the normal APIs. However, in order to have a good experience (Horizon integration, seamless CLI integration) the service needs to be in the Service Catalog. One user could have their service added to the catalog by an admin, but then everyone in the cloud would be using their VM. And if you have multiple users all doing the same thing in their own projects, you've got collisions! So, I submit to you all that there is value in having a way to scope Service Catalog entries to specific projects, and to allow users with appropriate permissions on their project to add/remove those project-level service catalog entries. This could be accomplished in a number of ways: * Adding a new field to the model to store a Project ID. * Adding it in a standardized manner to service metadata as with https://blueprints.launchpad.net/keystone/+spec/service-metadata * Adding it as an additional requirement as proposed by https://blueprints.launchpad.net/keystone/+spec/auth-mechanisms-for- services * Use the existing Region field to track project scope as a hack. * Something else... I see this as analogous to Nova's concept of per-project flavors, or Glance's private/public/shared image capabilities. Allowing explicit sharing would even be an interesting option for service endpoints. It all depends how far we would want to go with it. Feel free to offer feedback or other suggestions. Thanks! - Gabriel ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org mailto:OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org mailto:OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev -- Georgy Okrokvertskhov Technical Program Manager, Cloud and Infrastructure Services, Mirantis http://www.mirantis.com http://www.mirantis.com/ Tel. +1 650 963 9828 Mob. +1 650 996 3284 ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [savanna][qa] How will tempest tests run?
So it's great to see a submission of savanna tests for tempest. We would like to see these tests run before reviewing them. Is the intent that savanna will be enabled by default in devstack? If not, then I guess there will need to be separate savanna jobs. I see that right now there are savanna-enabled devstack jobs on the tempest experimental queue. In the long run, what is the intent for how these jobs should run? This is similar to the issue with ironic. It doesn't seem very scalable to set up separate complete tempest jobs for every project that is not turned on by default in devstack. Thoughts? -David ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [Neutron] Enable to set DHCP port attributes
Hi Neutron developers, I submitted the following blue print. https://blueprints.launchpad.net/neutron/+spec/enable-to-set-dhcp-port-attributes It is a proposal to be enable to control dhcp port attributes (especially ip address) by a user. This is based on a real requirement from our customer. I don't know there is a consensus that dhcp port attributes should not be enable to set by a user. Comments are welcome. Thanks. -- Itsuro ODA o...@valinux.co.jp ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [qa][ceilometer] Who can be a asked to help review tempest tests?
On 12/16/2013 05:50 PM, Doug Hellmann wrote: It might be good, to start out, if we all look at them. That way we can all learn a bit about tempest, too. If you add ceilometer-core as a reviewer gerrit will expand the group name. Doug Thanks, Doug. That makes sense. The only reason I asked this was because I didn't know it was possible to ask all of you like that! -David On Mon, Dec 16, 2013 at 5:33 PM, David Kranz dkr...@redhat.com mailto:dkr...@redhat.com wrote: Ceilometer team, we are reviewing tempest tests and hope to see more. The tempest review team is hoping to identify some ceilometer devs who could help answer questions or provide a review if needed for ceilometer patches. Since ceilometer is new we are not all familiar with many of the details. Can any one on the ceilometer team volunteer? -David ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org mailto:OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Neutron] DHCP Agent Reliability
On Dec 13, 2013, at 8:06 PM, Isaku Yamahata isaku.yamah...@gmail.com wrote: On Fri, Dec 06, 2013 at 04:30:17PM +0900, Maru Newby ma...@redhat.com wrote: On Dec 5, 2013, at 5:21 PM, Isaku Yamahata isaku.yamah...@gmail.com wrote: On Wed, Dec 04, 2013 at 12:37:19PM +0900, Maru Newby ma...@redhat.com wrote: In the current architecture, the Neutron service handles RPC and WSGI with a single process and is prone to being overloaded such that agent heartbeats can be delayed beyond the limit for the agent being declared 'down'. Even if we increased the agent timeout as Yongsheg suggests, there is no guarantee that we can accurately detect whether an agent is 'live' with the current architecture. Given that amqp can ensure eventual delivery - it is a queue - is sending a notification blind such a bad idea? In the best case the agent isn't really down and can process the notification. In the worst case, the agent really is down but will be brought up eventually by a deployment's monitoring solution and process the notification when it returns. What am I missing? Do you mean overload of neutron server? Not neutron agent. So event agent sends periodic 'live' report, the reports are piled up unprocessed by server. When server sends notification, it considers agent dead wrongly. Not because agent didn't send live reports due to overload of agent. Is this understanding correct? Your interpretation is likely correct. The demands on the service are going to be much higher by virtue of having to field RPC requests from all the agents to interact with the database on their behalf. Is this strongly indicating thread-starvation. i.e. too much unfair thread scheduling. Given that eventlet is cooperative threading, should sleep(0) to hogging thread? I'm afraid that's a question for a profiler: https://github.com/colinhowe/eventlet_profiler m. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Nova][libvirt]when deleting instance which is in migrating state, instance files can be stay in destination node forever
Hi! Actually, I have already filed cancel of LiveMigration with using taskflow [0]. But my approach to blueprint was not so good at that time. So, I rewrote this blueprint from also point of russelb. I want to repush and wanted to approved. If you have any suggestions, ideas etc... I appreciate it :) Sincerely, Haruka Tanizawa [0] https://blueprints.launchpad.net/nova/+spec/feature-of-cancel 2013/12/17 Yaguang Tang yaguang.t...@canonical.com could we use Taskflow https://wiki.openstack.org/wiki/TaskFlow to manage task state and resource for this kind of tasks in Nova? Cinder has been an pilot to use Taskflow for volume backup tasks. anyone interested in this suggestion or has done some research to improve the live migration workflow? 2013/12/17 Vladik Romanovsky vladik.romanov...@enovance.com I would block it in the API or have the API cancelling the migration first. I don't see a reason why to start an operation that is meant to fail, which also has a complex chain of event, following it failure. Regardless of the above, I think that the suggested exception handling is needed in any case. Vladik - Original Message - From: Loganathan Parthipan parthi...@hp.com To: OpenStack Development Mailing List (not for usage questions) openstack-dev@lists.openstack.org Sent: Monday, 16 December, 2013 8:25:09 AM Subject: Re: [openstack-dev] [Nova][libvirt]when deleting instance which is in migrating state, instance files can be stay in destination node forever Isn’t just handling the exception instance_not_found enough? By this time source would’ve been cleaned up. Destination VM resources will get cleaned up by the periodic task since the VM is not associated with this host. Am I missing something here? From: 王宏 [mailto:w.wangho...@gmail.com] Sent: 16 December 2013 11:32 To: openstack-dev@lists.openstack.org Subject: [openstack-dev] [Nova][libvirt]when deleting instance which is in migrating state, instance files can be stay in destination node forever Hi all. When I try to fix a bug: https://bugs.launchpad.net/nova/+bug/1242961 , I get a trouble. To reproduce the bug is very easy. Live migrate a vm in block_migration mode, and then delelte the vm immediately. The reason of this bug is as follow: 1. Because live migrate costs more time, so the vm will be deleted sucessfully before live migrate complete. And then, we will get an exception while live migrating. 2. After live migrate failed, we start to rollback. But, in the rollback method we will get or modify the info of vm from db. Because the vm has been deleted already, so we will get instance_not_found exception and rollback will be faild too. I have two ways to fix the bug: i)Add check in nova-api. When try to delete a vm, we return an error message if the vm_state is LIVE_MIGRATING. This way is very simple, but need to carefully consider. I have found a related discussion: http://lists.openstack.org/pipermail/openstack-dev/2013-October/017454.html, but it has no result in the discussion. ii)Before live migrate we get all the data needed by rollback method, and add a new rollback method. The new method will clean up resources at destination based on the above data(The resouces at source has been already cleaned up by deleting). I have no idea whitch one I should choose. Or, any other ideas?:) Regards, wanghong ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev -- Tang Yaguang Canonical Ltd. | www.ubuntu.com | www.canonical.com Mobile: +86 152 1094 6968 gpg key: 0x187F664F ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] Project-Scoped Service Catalog Entries
On 12/16/2013 02:57 PM, Gabriel Hurley wrote: I've run into a use case that doesn't currently seem to have a great solution: Let's say my users want to use a top-of-stack OpenStack project such as Heat, Trove, etc. that I don't currently support in my deployment. There's absolutely no reason these services can't live happily in a VM talking to Nova, etc. via the normal APIs. However, in order to have a good experience (Horizon integration, seamless CLI integration) the service needs to be in the Service Catalog. One user could have their service added to the catalog by an admin, but then everyone in the cloud would be using their VM. And if you have multiple users all doing the same thing in their own projects, you've got collisions! So, I submit to you all that there is value in having a way to scope Service Catalog entries to specific projects, and to allow users with appropriate permissions on their project to add/remove those project-level service catalog entries. This could be accomplished in a number of ways: * Adding a new field to the model to store a Project ID. * Adding it in a standardized manner to service metadata as with https://blueprints.launchpad.net/keystone/+spec/service-metadata * Adding it as an additional requirement as proposed by https://blueprints.launchpad.net/keystone/+spec/auth-mechanisms-for-services * Use the existing Region field to track project scope as a hack. * Something else... I see this as analogous to Nova's concept of per-project flavors, or Glance's private/public/shared image capabilities. Allowing explicit sharing would even be an interesting option for service endpoints. It all depends how far we would want to go with it. Feel free to offer feedback or other suggestions. Thanks! - Gabriel ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev See the endpoint filtering blueprint from the Havana release as a starting point. I think the difference between that and what you have here is that these endpoints should only show up in a subset of the service catalogs returned? https://github.com/openstack/keystone/commit/5dc50bbf0fb94506a06ae325d46bcf3ac1c4ad0a ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] Multidomain User Ids
On 12/04/2013 12:35 PM, Henry Nash wrote: On 4 Dec 2013, at 13:28, Dolph Mathews dolph.math...@gmail.com mailto:dolph.math...@gmail.com wrote: On Sun, Nov 24, 2013 at 9:39 PM, Adam Young ayo...@redhat.com mailto:ayo...@redhat.com wrote: The #1 pain point I hear from people in the field is that they need to consume read only LDAP but have service users in something Keystone specific. We are close to having this, but we have not closed the loop. This was something that was Henry's to drive home to completion. Do we have a plan? Federation depends on this, I think, but this problem stands alone. I'm still thinking through the idea of having keystone natively federate to itself out of the box, where keystone presents itself as an IdP (primarily for service users). It sounds like a simpler architectural solution than having to shuffle around code paths for both federated identities and local identities. Two Solutions: 1 always require domain ID along with the user id for role assignments. From an API perspective, how? (while still allowing for cross-domain role assignments) 2 provide some way to parse from the user ID what domain it is. I think you meant this one the other way around: Determine the domain given the user ID. I was thinking that we could do something along the lines of 2 where we provide domain specific user_id prefix for example, if there is just one ldpa service, and they wanted to prefix anyting out of ldap with ldap@, then an id would be prefix field from LDAP. And would be configured on a per domain basis. THis would be optional. The weakness is that itbe Log N to determine which Domain a user_id came from. A better approach would be to use a divider, like '@' and then prefix would be the key for a hashtable lookup. Since it is optional, domains could still be stored in SQL and user_ids could be uuids. One problem is if someone comes by later an must use email address as the userid, the @ would mess them up. So The default divider should be something URL safe but no likely to be part of a userid. I realize that it might be impossible to match this criterion. I know this sounds a bit like back to the future', but how about we make a user_id passed via the API a structured binary field, containing a concatenation of domain_id and (the actual) user_id, but rather than have a separator, encode the start positions in the first few digits, e.g. something like: This might be the most insane idea I have heard all day. I love it. Digit #Meaning 0-1Start position of domain_id, (e.g. this will usually be 4) 2-3Start position of user_id 4-Ndomain_id M-enduser_id I suspect it is more of a brainstorming attempt than as an actual proposal. It can't be binary for many reasons, and strings parsing gets wonky, especially if you assume utf-8 is in there (how many bytes per character?) The interesting idea is appending the domain id instead of prepending it. It may be an irrelevant change, but worth mulling. An interesting approach would be to do domain prepended user ids using / so that user/domain is the ID, and then the URL would be automagically segmented: If they leave off the domain, then the userid by itself would still be valid. We would run a migration that would convert all existing mappings. Further, we would ensure (with padding if necessary) that this new user_id is ALWAYS larger than 64chars - hence we could easily detect which type of ID we had. For usernames, sure... but I don't know why anyone would care to use email addresses as ID's. Actually, there might be other reasons to forbid @ signs from IDs, as they look like phishing attempts in URLs. Phishing attempts?? They need to be encoded anyway... ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org mailto:OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev -- -Dolph ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org mailto:OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] Project-Scoped Service Catalog Entries
On 12/16/2013 08:39 PM, Adam Young wrote: On 12/16/2013 02:57 PM, Gabriel Hurley wrote: I've run into a use case that doesn't currently seem to have a great solution: Let's say my users want to use a top-of-stack OpenStack project such as Heat, Trove, etc. that I don't currently support in my deployment. There's absolutely no reason these services can't live happily in a VM talking to Nova, etc. via the normal APIs. However, in order to have a good experience (Horizon integration, seamless CLI integration) the service needs to be in the Service Catalog. One user could have their service added to the catalog by an admin, but then everyone in the cloud would be using their VM. And if you have multiple users all doing the same thing in their own projects, you've got collisions! So, I submit to you all that there is value in having a way to scope Service Catalog entries to specific projects, and to allow users with appropriate permissions on their project to add/remove those project-level service catalog entries. This could be accomplished in a number of ways: * Adding a new field to the model to store a Project ID. * Adding it in a standardized manner to service metadata as with https://blueprints.launchpad.net/keystone/+spec/service-metadata * Adding it as an additional requirement as proposed by https://blueprints.launchpad.net/keystone/+spec/auth-mechanisms-for-services * Use the existing Region field to track project scope as a hack. * Something else... I see this as analogous to Nova's concept of per-project flavors, or Glance's private/public/shared image capabilities. Allowing explicit sharing would even be an interesting option for service endpoints. It all depends how far we would want to go with it. Feel free to offer feedback or other suggestions. Thanks! - Gabriel ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev See the endpoint filtering blueprint from the Havana release as a starting point. I think the difference between that and what you have here is that these endpoints should only show up in a subset of the service catalogs returned? https://github.com/openstack/keystone/commit/5dc50bbf0fb94506a06ae325d46bcf3ac1c4ad0a Also note the above is an admin API extension... -jay ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] Project-Scoped Service Catalog Entries
On 12/16/2013 08:39 PM, Adam Young wrote: snip See the endpoint filtering blueprint from the Havana release as a starting point. I think the difference between that and what you have here is that these endpoints should only show up in a subset of the service catalogs returned? https://github.com/openstack/keystone/commit/5dc50bbf0fb94506a06ae325d46bcf3ac1c4ad0a Unfortunately, this functionality was added as an API extension that isn't enabled by default in the WSGI pipeline and, by the nature of it being an extension, isn't guaranteed to be the same across deployments :( So, in order to use this functionality, Horizon will need to query for the existence of an OS-EP-FILTER extenstion to the Keystone API, and enable functionality based on this. Yet another reason I hate the idea of API extensions. -jay ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [cinder] weekly meeting
Hi All, Prompted by a recent suggestion from Tom Fifield, I thought I'd gauge some interest in either changing the weekly Cinder meeting time, or proposing a second meeting to accomodate folks in other time-zones. A large number of folks are already in time-zones that are not friendly to our current meeting time. I'm wondering if there is enough of an interest to move the meeting time from 16:00 UTC on Wednesdays, to 04:00 or 05:00 UTC? Depending on the interest I'd be willing to look at either moving the meeting for a trial period or holding a second meeting to make sure folks in other TZ's had a chance to be heard. Let me know your thoughts, if there are folks out there that feel unable to attend due to TZ conflicts and we can see what we might be able to do. Thanks, John ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [cinder] weekly meeting
04:00 or 05:00 UTC works for me. On Tue, Dec 17, 2013 at 11:04 AM, John Griffith john.griff...@solidfire.com wrote: Hi All, Prompted by a recent suggestion from Tom Fifield, I thought I'd gauge some interest in either changing the weekly Cinder meeting time, or proposing a second meeting to accomodate folks in other time-zones. A large number of folks are already in time-zones that are not friendly to our current meeting time. I'm wondering if there is enough of an interest to move the meeting time from 16:00 UTC on Wednesdays, to 04:00 or 05:00 UTC? Depending on the interest I'd be willing to look at either moving the meeting for a trial period or holding a second meeting to make sure folks in other TZ's had a chance to be heard. Let me know your thoughts, if there are folks out there that feel unable to attend due to TZ conflicts and we can see what we might be able to do. Thanks, John ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev -- Regards Huang Zhiteng ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [cinder] weekly meeting
Hi John, I think the current meeting schedule, UTC 16:00, basically works for China TZ (12AM), although it is not perfect. If we need to reschedule, I think UTC 05:00 is better than UTC 04:00, since UTC 04:00 (China 12PM) is our lunch time. On Tue, Dec 17, 2013 at 11:04 AM, John Griffith john.griff...@solidfire.com wrote: Hi All, Prompted by a recent suggestion from Tom Fifield, I thought I'd gauge some interest in either changing the weekly Cinder meeting time, or proposing a second meeting to accomodate folks in other time-zones. A large number of folks are already in time-zones that are not friendly to our current meeting time. I'm wondering if there is enough of an interest to move the meeting time from 16:00 UTC on Wednesdays, to 04:00 or 05:00 UTC? Depending on the interest I'd be willing to look at either moving the meeting for a trial period or holding a second meeting to make sure folks in other TZ's had a chance to be heard. Let me know your thoughts, if there are folks out there that feel unable to attend due to TZ conflicts and we can see what we might be able to do. Thanks, John ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [cinder] weekly meeting
On Mon, Dec 16, 2013 at 8:57 PM, 赵钦 chaoc...@gmail.com wrote: Hi John, I think the current meeting schedule, UTC 16:00, basically works for China TZ (12AM), although it is not perfect. If we need to reschedule, I think UTC 05:00 is better than UTC 04:00, since UTC 04:00 (China 12PM) is our lunch time. On Tue, Dec 17, 2013 at 11:04 AM, John Griffith john.griff...@solidfire.com wrote: Hi All, Prompted by a recent suggestion from Tom Fifield, I thought I'd gauge some interest in either changing the weekly Cinder meeting time, or proposing a second meeting to accomodate folks in other time-zones. A large number of folks are already in time-zones that are not friendly to our current meeting time. I'm wondering if there is enough of an interest to move the meeting time from 16:00 UTC on Wednesdays, to 04:00 or 05:00 UTC? Depending on the interest I'd be willing to look at either moving the meeting for a trial period or holding a second meeting to make sure folks in other TZ's had a chance to be heard. Let me know your thoughts, if there are folks out there that feel unable to attend due to TZ conflicts and we can see what we might be able to do. Thanks, John ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev Hi Chaochin, Thanks for the feedback, I think the alternate time would have to be moved up an hour or two anyway (between the lunch hour in your TZ and the fact that it just moves the problem of being at midnight to the folks in US Eastern TZ). Also, I think if there is interest that a better solution might be to implement something like the Ceilometer team does and alternate the time each week. John ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] Scheduler sub-group agenda 12/17
1) Memcached based scheduler updates 2) Scheduler code forklift 3) Instance groups -- Don Dugger Censeo Toto nos in Kansa esse decisse. - D. Gale Ph: 303/443-3786 ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [cinder] weekly meeting
Hello John, 04:00 or 05:00 UTC works for me too. On Tue, Dec 17, 2013 at 12:05 PM, John Griffith john.griff...@solidfire.com wrote: On Mon, Dec 16, 2013 at 8:57 PM, 赵钦 chaoc...@gmail.com wrote: Hi John, I think the current meeting schedule, UTC 16:00, basically works for China TZ (12AM), although it is not perfect. If we need to reschedule, I think UTC 05:00 is better than UTC 04:00, since UTC 04:00 (China 12PM) is our lunch time. On Tue, Dec 17, 2013 at 11:04 AM, John Griffith john.griff...@solidfire.com wrote: Hi All, Prompted by a recent suggestion from Tom Fifield, I thought I'd gauge some interest in either changing the weekly Cinder meeting time, or proposing a second meeting to accomodate folks in other time-zones. A large number of folks are already in time-zones that are not friendly to our current meeting time. I'm wondering if there is enough of an interest to move the meeting time from 16:00 UTC on Wednesdays, to 04:00 or 05:00 UTC? Depending on the interest I'd be willing to look at either moving the meeting for a trial period or holding a second meeting to make sure folks in other TZ's had a chance to be heard. Let me know your thoughts, if there are folks out there that feel unable to attend due to TZ conflicts and we can see what we might be able to do. Thanks, John ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev Hi Chaochin, Thanks for the feedback, I think the alternate time would have to be moved up an hour or two anyway (between the lunch hour in your TZ and the fact that it just moves the problem of being at midnight to the folks in US Eastern TZ). Also, I think if there is interest that a better solution might be to implement something like the Ceilometer team does and alternate the time each week. Agreed, like Glance team also. zhiyan John ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Openstack][Heat] Creating an Openstack project using Heat
Thanks Steve. Regards, Sayaji On Mon, Dec 16, 2013 at 2:57 PM, Steve Baker sba...@redhat.com wrote: On 12/17/2013 09:57 AM, Sayaji Patil wrote: Hi, I have installed Openstack with heat using packstack. One thing I noticed is that the Orchestration Heat option is only available inside a project view. Is this by design ? Yes, heat stacks are scoped to a project/tenant. My use case it to create a project with images, networks,routers and firewall rules in a single workflow. I looked at the documentation and at this point there is no resource available to create a project or upload an image. It wouldn't be hard to write a resource which creates a tenant/project, however there will be more changes required before the other resources in your stack can be created in the context of your new project. For now you need to create your project and user outside of heat. As for image upload, a glance resource could be written which registers an image from a URL. Feel free to file a blueprint for that describing your use case. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [neutron] Re: [Blueprint vlan-aware-vms] VLAN aware VMs
Added openstack-dev On Mon, Dec 16, 2013 at 11:34:05PM +0100, Erik Moe emoe...@gmail.com wrote: Hi, I have added a new document to the launchpad. Document should now be more in line with what we discussed at the Icehouse summit. https://blueprints.launchpad.net/neutron/+spec/vlan-aware-vms Doc: https://docs.google.com/document/d/1lDJ31-XqkjjWC-IBq-_wV1KVhi7DzPkKYlCxTIPs_9U/edit?usp=sharing You are very welcome to give feedback if this is the solution you had in mind. The document is view-only. So I commented below. - 2 Modeling proposal What's the purpose of trunk network? Can you please add a use case that trunk network can't be optimized away? - 4 IP address management nitpick Can you please clarify what the L2 gateway ports in section 2 modeling proposal, figure 1? - Table 3 Will this be same to l2-gateway one? https://blueprints.launchpad.net/neutron/+spec/l2-gateway - Figure 5 What's the purpose of br-int local VID? VID can be directly converted from br-eth1 VID to VM VID untagged. -- Isaku Yamahata isaku.yamah...@gmail.com ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [cinder] weekly meeting
Hi John, Yes, alternating the time for each week should be fine. I just change my gmail name to English... I think you can see my name now... On Tue, Dec 17, 2013 at 12:05 PM, John Griffith john.griff...@solidfire.com wrote: On Mon, Dec 16, 2013 at 8:57 PM, 赵钦 chaoc...@gmail.com wrote: Hi John, I think the current meeting schedule, UTC 16:00, basically works for China TZ (12AM), although it is not perfect. If we need to reschedule, I think UTC 05:00 is better than UTC 04:00, since UTC 04:00 (China 12PM) is our lunch time. On Tue, Dec 17, 2013 at 11:04 AM, John Griffith john.griff...@solidfire.com wrote: Hi All, Prompted by a recent suggestion from Tom Fifield, I thought I'd gauge some interest in either changing the weekly Cinder meeting time, or proposing a second meeting to accomodate folks in other time-zones. A large number of folks are already in time-zones that are not friendly to our current meeting time. I'm wondering if there is enough of an interest to move the meeting time from 16:00 UTC on Wednesdays, to 04:00 or 05:00 UTC? Depending on the interest I'd be willing to look at either moving the meeting for a trial period or holding a second meeting to make sure folks in other TZ's had a chance to be heard. Let me know your thoughts, if there are folks out there that feel unable to attend due to TZ conflicts and we can see what we might be able to do. Thanks, John ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev Hi Chaochin, Thanks for the feedback, I think the alternate time would have to be moved up an hour or two anyway (between the lunch hour in your TZ and the fact that it just moves the problem of being at midnight to the folks in US Eastern TZ). Also, I think if there is interest that a better solution might be to implement something like the Ceilometer team does and alternate the time each week. John ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev -- Qin Zhao ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [oslo][glance] Oslo.cfg resets not really resetting the CONF
Hi Mark, Ben The reset() method in turn calls the *clear()* method which does an *unregister_opt()*. However the unregister_opt only unregisters the *config_opts*. The entire set of options inside *_opts* remain as is. We've filed a bug https://bugs.launchpad.net/oslo/+bug/1261376 on the oslo end. On Tue, Dec 17, 2013 at 5:27 AM, Mark McLoughlin mar...@redhat.com wrote: Hi On Fri, 2013-12-13 at 14:14 +0530, Amala Basha Alungal wrote: Hi, I stumbled into a situation today where in I had to write few tests that modifies the oslo.config.cfg and in turn resets the values back in a tear down. Acc to the docs, oslo.cfg reset() *Clears the object state and unsets overrides and defaults. *but, it doesn't seem to be happening, as the subsequent tests that are run retains these modified values and tests behave abnormally. The patch has been submitted for review herehttps://review.openstack.org/#/c/60188/1. Am I missing something obvious? From https://bugs.launchpad.net/oslo/+bug/1261376 : reset() will clear any values read from the command line or config files and it will also remove any values set with set_default() or set_override() However, it will not undo register_opt() - there is unregister_opt() for that purpose Maybe if you pushed a version of https://review.openstack.org/60188 which uses reset() and explain how it's not working as you expected? Thanks, Mark. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev -- Thanks And Regards Amala Basha +91-7760972008 ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [cinder][qa] release notes for cinder v1 to v2?
It seems that there's no document about the change from v1 to v2. Maybe the change is very small. Only found some info in OpenStack release notes. Cinder API v2 - List volumes/snapshots summary actually is a summary view. In v1 it was the same as detail view. - List volumes/snapshots detail and summary has display_name key changed to name. - List volumes/snapshots detail and summary has display_description key changed to description. https://wiki.openstack.org/wiki/ReleaseNotes/Grizzly#OpenStack_Block_Storage_.28Cinder.29 https://wiki.openstack.org/wiki/ReleaseNotes/Havana#OpenStack_Block_Storage_.28Cinder.29 2013/12/17 David Kranz dkr...@redhat.com Sorry for lost subject in last message. Is there a document that describes the api changes from v1 to v2, similar to the one documenting nova v2 to v3? -David ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev -- Regards, Zhi Kun Liu ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [oslo][glance] Oslo.cfg resets not really resetting the CONF
On Tue, 2013-12-17 at 11:17 +0530, Amala Basha Alungal wrote: Hi Mark, Ben The reset() method in turn calls the clear() method which does an unregister_opt(). However the unregister_opt only unregisters the config_opts. The entire set of options inside _opts remain as is. We've filed a bug on the oslo end. Yes, that's working as designed. Those two options are registered by __call__() so reset() unregisters only them. The idea is that you can register lots and then do __call__() and reset() without affecting the registered options. Mark. On Tue, Dec 17, 2013 at 5:27 AM, Mark McLoughlin mar...@redhat.com wrote: Hi On Fri, 2013-12-13 at 14:14 +0530, Amala Basha Alungal wrote: Hi, I stumbled into a situation today where in I had to write few tests that modifies the oslo.config.cfg and in turn resets the values back in a tear down. Acc to the docs, oslo.cfg reset() *Clears the object state and unsets overrides and defaults. *but, it doesn't seem to be happening, as the subsequent tests that are run retains these modified values and tests behave abnormally. The patch has been submitted for review herehttps://review.openstack.org/#/c/60188/1. Am I missing something obvious? From https://bugs.launchpad.net/oslo/+bug/1261376 : reset() will clear any values read from the command line or config files and it will also remove any values set with set_default() or set_override() However, it will not undo register_opt() - there is unregister_opt() for that purpose Maybe if you pushed a version of https://review.openstack.org/60188 which uses reset() and explain how it's not working as you expected? Thanks, Mark. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev -- Thanks And Regards Amala Basha +91-7760972008 ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [cinder] weekly meeting
I agree with Qin here that alternating might be a good option. I'm not opposed to being present to both meetings though. -Mike Perez On Mon, Dec 16, 2013 at 9:31 PM, Qin Zhao chaoc...@gmail.com wrote: Hi John, Yes, alternating the time for each week should be fine. I just change my gmail name to English... I think you can see my name now... On Tue, Dec 17, 2013 at 12:05 PM, John Griffith john.griff...@solidfire.com wrote: On Mon, Dec 16, 2013 at 8:57 PM, 赵钦 chaoc...@gmail.com wrote: Hi John, I think the current meeting schedule, UTC 16:00, basically works for China TZ (12AM), although it is not perfect. If we need to reschedule, I think UTC 05:00 is better than UTC 04:00, since UTC 04:00 (China 12PM) is our lunch time. On Tue, Dec 17, 2013 at 11:04 AM, John Griffith john.griff...@solidfire.com wrote: Hi All, Prompted by a recent suggestion from Tom Fifield, I thought I'd gauge some interest in either changing the weekly Cinder meeting time, or proposing a second meeting to accomodate folks in other time-zones. A large number of folks are already in time-zones that are not friendly to our current meeting time. I'm wondering if there is enough of an interest to move the meeting time from 16:00 UTC on Wednesdays, to 04:00 or 05:00 UTC? Depending on the interest I'd be willing to look at either moving the meeting for a trial period or holding a second meeting to make sure folks in other TZ's had a chance to be heard. Let me know your thoughts, if there are folks out there that feel unable to attend due to TZ conflicts and we can see what we might be able to do. Thanks, John ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev Hi Chaochin, Thanks for the feedback, I think the alternate time would have to be moved up an hour or two anyway (between the lunch hour in your TZ and the fact that it just moves the problem of being at midnight to the folks in US Eastern TZ). Also, I think if there is interest that a better solution might be to implement something like the Ceilometer team does and alternate the time each week. John ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev -- Qin Zhao ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [cinder][qa] release notes for cinder v1 to v2?
So the grizzly API releases notes are dead on. (I put them together and have worked with both API's ;) ) Unfortunately, a lot of the features in v2 got back ported to v1. The main differences are in the upgrade notes. -Mike Perez On Mon, Dec 16, 2013 at 9:57 PM, Zhi Kun Liu liuzhi...@gmail.com wrote: It seems that there's no document about the change from v1 to v2. Maybe the change is very small. Only found some info in OpenStack release notes. Cinder API v2 - List volumes/snapshots summary actually is a summary view. In v1 it was the same as detail view. - List volumes/snapshots detail and summary has display_name key changed to name. - List volumes/snapshots detail and summary has display_description key changed to description. https://wiki.openstack.org/wiki/ReleaseNotes/Grizzly#OpenStack_Block_Storage_.28Cinder.29 https://wiki.openstack.org/wiki/ReleaseNotes/Havana#OpenStack_Block_Storage_.28Cinder.29 2013/12/17 David Kranz dkr...@redhat.com Sorry for lost subject in last message. Is there a document that describes the api changes from v1 to v2, similar to the one documenting nova v2 to v3? -David ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev -- Regards, Zhi Kun Liu ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev