[openstack-dev] Working with Vagrant and packstack
Hi all, I have been trying to set up an openstack environment using vagrant and packstack. I provisioned a Fedora-19 VM through vagrant and used a shell script to take care of installation and other things. The first thing that shell script does is yum install -y openstack-packstack and then packstack --allinone. Now, the issue is that the second command requires me to enter the root's password explicitly. I mean it doesn't matter if I am running this as root or using sudo, I have to enter the password explicitly everytime. I tried to pass the password to the VM through pipes and other methods, but nothing works. Did anyone face the same problem? Is there any way around this? Or does it mean that I can't use puppet/packstack with vagrant? Thanks, ~Peeyush Gupta___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [Neutron][LBaaS] Vote required for certificate as first level citizen - SSL Termination
To summarize: Certificate will be a first level citizen which can be reused and For certificate management nothing sophisticated is required. Can you please Vote (+1, -1)? We can move on if there is consensus around this. -Original Message- From: Stephen Gran [mailto:stephen.g...@guardian.co.uk] Sent: Wednesday, November 20, 2013 3:01 PM To: OpenStack Development Mailing List (not for usage questions) Subject: Re: [openstack-dev] [Neutron][LBaaS] SSL Termination write-up Hi, On Wed, 2013-11-20 at 08:24 +, Samuel Bercovici wrote: Hi, Evgeny has outlined the wiki for the proposed change at: https://wiki.openstack.org/wiki/Neutron/LBaaS/SSL which is in line with what was discussed during the summit. The https://docs.google.com/document/d/1tFOrIa10lKr0xQyLVGsVfXr29NQBq2n YTvMkMJ_inbo/edit discuss in addition Certificate Chains. What would be the benefit of having a certificate that must be connected to VIP vs. embedding it in the VIP? You could reuse the same certificate for multiple loadbalancer VIPs. This is a fairly common pattern - we have a dev wildcard cert that is self- signed, and is used for lots of VIPs. When we get a system that can store certificates (ex: Barbican), we will add support to it in the LBaaS model. It probably doesn't need anything that complicated, does it? Cheers, -- Stephen Gran Senior Systems Integrator - The Guardian Please consider the environment before printing this email. -- Visit theguardian.com On your mobile, download the Guardian iPhone app theguardian.com/iphone and our iPad edition theguardian.com/iPad Save up to 33% by subscribing to the Guardian and Observer - choose the papers you want and get full digital access. Visit subscribe.theguardian.com This e-mail and all attachments are confidential and may also be privileged. If you are not the named recipient, please notify the sender and delete the e- mail and all attachments immediately. Do not disclose the contents to another person. You may not use the information for any purpose, or store, or copy, it in any way. Guardian News Media Limited is not liable for any computer viruses or other material transmitted with or as part of this e-mail. You should employ virus checking software. Guardian News Media Limited A member of Guardian Media Group plc Registered Office PO Box 68164 Kings Place 90 York Way London N1P 2AP Registered in England Number 908396 -- ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [heat][hadoop][template] Does anyone has a hadoop template
Hello Jay, Just in case you've missed it, there is a project Savanna dedicated to deploying Hadoop clusters on OpenStack: https://github.com/openstack/savanna http://savanna.readthedocs.org/en/0.3/ Dmitry 2013/11/29 Jay Lau jay.lau@gmail.com Hi, I'm now trying to deploy a hadoop cluster with heat, just wondering if someone who has a heat template which can help me do the work. Thanks, Jay ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] FreeBSD hypervisor (bhyve) driver
On Fri, Nov 29, 2013 at 8:24 AM, Roman Bogorodskiy rbogorods...@mirantis.com wrote: Hello, Yes, libvirt's qemu driver works almost fine currently, except the fact that it needs a 'real' bridge driver, so all the networking configuration like filtering rules, NAT, etc could be done automatically, like for Linux now, instead of making user to perform all the configuration manually. Networking is actually part of our work for FreeBSD Nova support: we have a freebsd_net.py driver (equivalent to the linux_net.py), which manages bridging in the BSD way and we're in the process of bringing up FlatDHCPManager configuration for nova-network running on the FreeBSD host. Rafal I've been planning to get to bhyve driver as well, but probably after finishing with the bridge driver (but unfortunately, I don't have a full picture what would be the best way to implement that). On Mon, Nov 25, 2013 at 3:50 PM, Daniel P. Berrange berra...@redhat.com wrote: On Fri, Nov 22, 2013 at 10:46:19AM -0500, Russell Bryant wrote: On 11/22/2013 10:43 AM, Rafał Jaworowski wrote: Russell, First, thank you for the whiteboard input regarding the blueprint for FreeBSD hypervisor nova driver: https://blueprints.launchpad.net/nova/+spec/freebsd-compute-node We were considering libvirt support for bhyve hypervisor as well, only wouldn't want to do this as the first approach for FreeBSD+OpenStack integration. We'd rather bring bhyve bindings for libvirt later as another integration option. For FreeBSD host support a native hypervisor driver is important and desired long-term and we would like to have it anyways. Among things to consider are the following: - libvirt package is additional (non-OpenStack), external dependency (maintained in the 'ports' collection, not included in base system), while native API (via libvmmapi.so library) is integral part of the base system. - libvirt license is LGPL, which might be an important aspect for some users. That's perfectly fine if you want to go that route as a first step. However, that doesn't mean it's appropriate for merging into Nova. Unless there are strong technical justifications for why this approach should be taken, I would probably turn down this driver until you were able to go the libvirt route. The idea of a FreeBSD bhyve driver for libvirt has been mentioned a few times. We've already got a FreeBSD port of libvirt being actively maintained to support QEMU (and possibly Xen, not 100% sure on that one), and we'd be more than happy to see further contributions such as a bhyve driver. I am of course biased, as libvirt project maintainer, but I do agree that supporting bhyve via libvirt would make sense, since it opens up opportunities beyond just OpenStack. There are a bunch of applications built on libvirt that could be used to manage bhyve, and a fair few applications which have plugins using libvirt Taking on maint work for a new OpenStack driver is a non-trivial amount of work in itself. If the burden for OpenStack maintainers can be reduced by, pushing work out to / relying on support from, libvirt, that makes sense from OpenStack/Nova's POV. Regards, Daniel -- |: http://berrange.com -o-http://www.flickr.com/photos/dberrange/ :| |: http://libvirt.org -o- http://virt-manager.org :| |: http://autobuild.org -o- http://search.cpan.org/~danberr/ :| |: http://entangle-photo.org -o- http://live.gnome.org/gtk-vnc :| ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Nova][Schduler] Volunteers wanted for a modest proposal for an external scheduler in our lifetime
The first stage is technical - move Nova scheduling code from A to be. What do we achieve - not much - we actually complicate things - there is always churn in Nova and we will have duplicate code bases. In addition to this the only service that can actually make use of they is Nova The second stage is defining an API that other modules can use (we have yet to decide if this will be RPC based or have a interface like Glance, Cinder etc.) We have yet to even talk about the API's. The third stage is adding shiny new features and trying to not have a community tar and feather us. Yup; I look forward to our tar and feathering overlords. :) Prior to copying code we really need to discuss the API's. I don't think we do: it's clear that we need to come up with them - it's necessary, and noone has expressed any doubt about the ability to do that. RPC API evolution is fairly well understood - we add a new method, and have it do the necessary, then we go to the users and get them using it, then we delete the old one. I agree with Robert. I think that nova RPC is fairly enough for the new scheduler right now. Most of the scheduler works focus on nova anyway, so starting from there is reasonable and rather easy for the transition. We can think of enhancing API later (even creating REST API perhaps). This can even be done in parallel if your concern is time and resources. But the point is we need a API to interface with the service. For a start we can just address the Nova use case. We need to at least address: 1. Scheduling interface 2. Statistics updates 3. API's for configuring the scheduling policies If by 2. Statistics update you mean the database issue for scheduler then yes, it is a big issue, especially during the transition period when nova still holds the host state data. Should scheduler get access to nova's DB for the time being, and later fork out the DB to scheduler? According to Boris, Merantis has already studied the separation of host state from nova's DB. I think we can benefit from their experience. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] ml2 and vxlan configurations, neutron-server fails to start
Hi Trinath Please find the server.log and neutron.conf * ** server.log - 2013-11-29 11:21:45.276 13505 INFO neutron.common.config [-] Logging enabled! 2013-11-29 11:21:45.277 13505 WARNING neutron.common.legacy [-] Old class module path in use. Please change 'quantum.openstack.common.rpc.impl_qpid' to 'neutron.openstack.common.rpc.impl_qpid'. 2013-11-29 11:21:45.277 13505 ERROR neutron.common.legacy [-] Skipping unknown group key: firewall_driver 2013-11-29 11:21:45.277 13505 DEBUG neutron.service [-] log_opt_values /usr/lib/python2.7/site-packages/oslo/config/cfg.py:1890 2013-11-29 11:21:45.277 13505 DEBUG neutron.service [-] Configuration options gathered from: log_opt_values /usr/lib/python2.7/site-packages/oslo/config/cfg.py:1891 2013-11-29 11:21:45.278 13505 DEBUG neutron.service [-] command line args: ['--config-file', '/usr/share/neutron/neutron-dist.conf', '--config-file', '/etc/neutron/neutron.conf', '--config-file', '/etc/neutron/plugin.ini', '--log-file', '/var/log/neutron/server.log'] log_opt_values /usr/lib/python2.7/site-packages/oslo/config/cfg.py:1892 2013-11-29 11:21:45.278 13505 DEBUG neutron.service [-] config files: ['/usr/share/neutron/neutron-dist.conf', '/etc/neutron/neutron.conf', '/etc/neutron/plugin.ini'] log_opt_values /usr/lib/python2.7/site-packages/oslo/config/cfg.py:1893 2013-11-29 11:21:45.278 13505 DEBUG neutron.service [-] log_opt_values /usr/lib/python2.7/site-packages/oslo/config/cfg.py:1894 2013-11-29 11:21:45.278 13505 DEBUG neutron.service [-] allow_bulk = True log_opt_values /usr/lib/python2.7/site-packages/oslo/config/cfg.py:1903 2013-11-29 11:21:45.278 13505 DEBUG neutron.service [-] allow_overlapping_ips = True log_opt_values /usr/lib/python2.7/site-packages/oslo/config/cfg.py:1903 2013-11-29 11:21:45.279 13505 DEBUG neutron.service [-] allow_pagination = False log_opt_values /usr/lib/python2.7/site-packages/oslo/config/cfg.py:1903 2013-11-29 11:21:45.279 13505 DEBUG neutron.service [-] allow_sorting = False log_opt_values /usr/lib/python2.7/site-packages/oslo/config/cfg.py:1903 2013-11-29 11:21:45.279 13505 DEBUG neutron.service [-] allowed_rpc_exception_modules = ['neutron.openstack.common.exception', 'nova.exception', 'cinder.exception', 'exceptions'] log_opt_values /usr/lib/python2.7/site-packages/oslo/config/cfg.py:1903 2013-11-29 11:21:45.279 13505 DEBUG neutron.service [-] api_extensions_path= log_opt_values /usr/lib/python2.7/site-packages/oslo/config/cfg.py:1903 2013-11-29 11:21:45.280 13505 DEBUG neutron.service [-] api_paste_config = /etc/neutron/api-paste.ini log_opt_values /usr/lib/python2.7/site-packages/oslo/config/cfg.py:1903 2013-11-29 11:21:45.280 13505 DEBUG neutron.service [-] auth_strategy = keystone log_opt_values /usr/lib/python2.7/site-packages/oslo/config/cfg.py:1903 2013-11-29 11:21:45.280 13505 DEBUG neutron.service [-] backdoor_port = None log_opt_values /usr/lib/python2.7/site-packages/oslo/config/cfg.py:1903 2013-11-29 11:21:45.280 13505 DEBUG neutron.service [-] backlog = 4096 log_opt_values /usr/lib/python2.7/site-packages/oslo/config/cfg.py:1903 2013-11-29 11:21:45.280 13505 DEBUG neutron.service [-] base_mac = fa:16:3e:00:00:00 log_opt_values /usr/lib/python2.7/site-packages/oslo/config/cfg.py:1903 2013-11-29 11:21:45.281 13505 DEBUG neutron.service [-] bind_host = 0.0.0.0 log_opt_values /usr/lib/python2.7/site-packages/oslo/config/cfg.py:1903 2013-11-29 11:21:45.281 13505 DEBUG neutron.service [-] bind_port = 9696 log_opt_values /usr/lib/python2.7/site-packages/oslo/config/cfg.py:1903 2013-11-29 11:21:45.281 13505 DEBUG neutron.service [-] config_dir = None log_opt_values /usr/lib/python2.7/site-packages/oslo/config/cfg.py:1903 2013-11-29 11:21:45.281 13505 DEBUG neutron.service [-] config_file = ['/usr/share/neutron/neutron-dist.conf', '/etc/neutron/neutron.conf', '/etc/neutron/plugin.ini'] log_opt_values /usr/lib/python2.7/site-packages/oslo/config/cfg.py:1903 2013-11-29 11:21:45.282 13505 DEBUG neutron.service [-] control_exchange = rabbit log_opt_values /usr/lib/python2.7/site-packages/oslo/config/cfg.py:1903 2013-11-29 11:21:45.282 13505 DEBUG neutron.service [-] core_plugin = neutron.plugins.ml2.plugin.Ml2Plugin log_opt_values /usr/lib/python2.7/site-packages/oslo/config/cfg.py:1903 2013-11-29 11:21:45.282 13505 DEBUG neutron.service [-] debug = True log_opt_values /usr/lib/python2.7/site-packages/oslo/config/cfg.py:1903 2013-11-29 11:21:45.282 13505 DEBUG neutron.service [-] default_log_levels = ['amqplib=WARN',
Re: [openstack-dev] [UX] Automatically post new threads from AskBot to the list
On 2013/20/11 01:23, Stefano Maffulli wrote: On 11/19/2013 08:19 AM, Julie Pichon wrote: I've been thinking about the AskBot UX website [0] and its lack of visibility, particularly for new community members. Indeed, it's one of the drawbacks of splitting groups: information tends not to flow very well. Yeah, We will try weekly summary from the beginning. I've heard that the UX team will be the first team to dogfood Storyboard: do you have any idea of any ETA/deadline for when this will happen? That's correct. The last information was that it might be around beginning of next year. But there were lot of variables (mostly resources). But yes, I will be very happy to help to test and provide feedback on that. -- Jarda ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] Reg : Security groups implementation using openflows in quantum ovs plugin
On Fri, Nov 29, 2013 at 2:25 PM, Jian Wen jian@canonical.com wrote: I don't think we can implement a stateful firewall[1] now. I don't think we need a stateful firewall, a stateless one should work well. If the stateful conntrack is completed in the future, we can also take benefit from it. Once connection tracking capability[2] is added to the Linux OVS, we could start to implement the ovs-firewall-driver blueprint. [1] http://en.wikipedia.org/wiki/Stateful_firewall [2] http://wiki.xenproject.org/wiki/Xen_Development_Projects#Add_connection_tracking_capability_to_the_Linux_OVS On Tue, Nov 26, 2013 at 2:23 AM, Mike Wilson geekinu...@gmail.com wrote: Adding Jun to this thread since gmail is failing him. On Tue, Nov 19, 2013 at 10:44 AM, Amir Sadoughi amir.sadou...@rackspace.com wrote: Yes, my work has been on ML2 with neutron-openvswitch-agent. I’m interested to see what Jun Park has. I might have something ready before he is available again, but would like to collaborate regardless. Amir On Nov 19, 2013, at 3:31 AM, Kanthi P pavuluri.kan...@gmail.com wrote: Hi All, Thanks for the response! Amir,Mike: Is your implementation being done according to ML2 plugin Regards, Kanthi On Tue, Nov 19, 2013 at 1:43 AM, Mike Wilson geekinu...@gmail.com wrote: Hi Kanthi, Just to reiterate what Kyle said, we do have an internal implementation using flows that looks very similar to security groups. Jun Park was the guy that wrote this and is looking to get it upstreamed. I think he'll be back in the office late next week. I'll point him to this thread when he's back. -Mike On Mon, Nov 18, 2013 at 3:39 PM, Kyle Mestery (kmestery) kmest...@cisco.com wrote: On Nov 18, 2013, at 4:26 PM, Kanthi P pavuluri.kan...@gmail.com wrote: Hi All, We are planning to implement quantum security groups using openflows for ovs plugin instead of iptables which is the case now. Doing so we can avoid the extra linux bridge which is connected between the vnet device and the ovs bridge, which is given as a work around since ovs bridge is not compatible with iptables. We are planning to create a blueprint and work on it. Could you please share your views on this Hi Kanthi: Overall, this idea is interesting and removing those extra bridges would certainly be nice. Some people at Bluehost gave a talk at the Summit [1] in which they explained they have done something similar, you may want to reach out to them since they have code for this internally already. The OVS plugin is in feature freeze during Icehouse, and will be deprecated in favor of ML2 [2] at the end of Icehouse. I would advise you to retarget your work at ML2 when running with the OVS agent instead. The Neutron team will not accept new features into the OVS plugin anymore. Thanks, Kyle [1] http://www.openstack.org/summit/openstack-summit-hong-kong-2013/session-videos/presentation/towards-truly-open-and-commoditized-software-defined-networks-in-openstack [2] https://wiki.openstack.org/wiki/Neutron/ML2 Thanks, Kanthi ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev -- Cheers, Jian ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] ml2 and vxlan configurations, neutron-server fails to start
Hi Thankyou for the info, I downgraded sqlalchemy accordingly, but there were lot of other dependencies I had to take care (as below). The error which still continues to persist in my environment is : RuntimeError: Unable to load quantum from configuration file /etc/neutron/api-paste.ini. are there any other dependencies to be taken care to resolve this error? pip uninstall sqlalchemy (uninstalled -0.8.3) pip install sqlalchemy==0.7.9 pip install jsonrpclib pip uninstall eventlet (uninstalled -0.12.0) pip install eventlet (installed -0.14.0) pip install pyudev for the error Requirement.parse('amqp=1.0.10,1.1.0')) pip uninstall amqp pip install amqp -- but its installing 1.3.3 so, download the source code of 1.0.10 (https://pypi.python.org/pypi/amqp/1.0.10) python setup.py build python setup.py install Regards Gopi Krishna Yongsheng Gong gongysh at unitedstack.com VersionConflict: (SQLAlchemy 0.8.3 (/usr/lib64/python2.7/site-packages), Requirement.parse('SQLAlchemy=0.7.8,=0.7.99')) it seems your SQLAlchemy is newer than required. so pip uninstall sqlalchemy and then install older one: sudo pip install sqlalchemy==0.7.9 ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Nova][Schduler] Volunteers wanted for a modest proposal for an external scheduler in our lifetime
On 11/28/13 11:34 PM, Robert Collins robe...@robertcollins.net wrote: On 29 November 2013 09:44, Gary Kotton gkot...@vmware.com wrote: The first stage is technical - move Nova scheduling code from A to be. What do we achieve - not much - we actually complicate things - there is always churn in Nova and we will have duplicate code bases. In addition to this the only service that can actually make use of they is Nova The second stage is defining an API that other modules can use (we have yet to decide if this will be RPC based or have a interface like Glance, Cinder etc.) We have yet to even talk about the API's. The third stage is adding shiny new features and trying to not have a community tar and feather us. Yup; I look forward to our tar and feathering overlords. :) Prior to copying code we really need to discuss the API's. I don't think we do: it's clear that we need to come up with them - it's necessary, and noone has expressed any doubt about the ability to do that. RPC API evolution is fairly well understood - we add a new method, and have it do the necessary, then we go to the users and get them using it, then we delete the old one. This can even be done in parallel if your concern is time and resources. But the point is we need a API to interface with the service. For a start we can just address the Nova use case. We need to at least address: 1. Scheduling interface 2. Statistics updates 3. API's for configuring the scheduling policies Later these will all need to bode well with all of the existing modules that we want to support - Nova, Cinder and Neutron (if I missed on then feel free to kick me whilst I am down) Ironic perhaps. I do not think that we should focus on failure modes, we should plan it and break it up so that it will be usable and functional and most importantly useful in the near future. How about next week we sit together and draw up a wiki of the flows, data structures and interfaces. Lets go from there. While I disagree about us needing to do it right now, I'm very happy to spend some time on it - I don't want to stop folk doing work that needs to be done! I do not think that discussion will prevent any of the work getting done or not done. It may actually save us a ton of time. I really think that defining the API and interfaces can save us a lot in the short and long run. The V1 API should really be very simple and we should not get bogged down but if we can define an interface that could work with Nova and be extensible to work with the rest then we will be in a very good state. I am thinking of have a notion of a 'scheduling domain' and that will be used with the scheduling request. This could be a host aggregate, a AZ, the feature that Phil is working on - private hosts. If we can define an interface around this and have the Nova - scheduling interface down then we are on the wayŠ. Hosts scan be added to the domain and the scheduling will be able to get the stats etc. For the first phase this will be completely RPC bases so not to get bogged down. Can we talk about this next week? -Rob -- Robert Collins rbtcoll...@hp.com Distinguished Technologist HP Converged Cloud ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org https://urldefense.proofpoint.com/v1/url?u=http://lists.openstack.org/cgi- bin/mailman/listinfo/openstack-devk=oIvRg1%2BdGAgOoM1BIlLLqw%3D%3D%0Ar=e H0pxTUZo8NPZyF6hgoMQu%2BfDtysg45MkPhCZFxPEq8%3D%0Am=m8Unw2sBiRyQsd6ADjYCH CpcxAKBG1Gb3R58ehl3wIw%3D%0As=6c2d9b2f3b884693094cffc0392402f24b50ddac3d9 d956472d26c726c84a2a6 ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] jsonschema version constraint
Hi, Do we need to update the jsonschema constraint in the requirements.txt? Many tests are failing with jsonschema 1.3.0 with the same error. == FAIL: nova.tests.test_api_validation.AdditionalPropertiesDisableTestCase.test_validate_additionalProperties_disable -- _StringException: Empty attachments: pythonlogging:'' stderr stdout Traceback (most recent call last): File /tmp/buildd/nova-2014.1.dev990.g2c795f0/nova/tests/test_api_validation.py, line 133, in setUp @validation.schema(request_body_schema=schema) File /tmp/buildd/nova-2014.1.dev990.g2c795f0/nova/api/validation/__init__.py, line 36, in schema schema_validator = _SchemaValidator(request_body_schema) File /tmp/buildd/nova-2014.1.dev990.g2c795f0/nova/api/validation/validators.py, line 51, in __init__ validator_cls = jsonschema.validators.extend(self.validator_org, AttributeError: 'dict' object has no attribute 'extend' ~parthi ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Solum] Definition feedback
On 11/27/2013 10:15 PM, Adrian Otto wrote: On Nov 27, 2013, at 11:27 AM, Jay Pipes jaypi...@gmail.com wrote: On 11/27/2013 02:03 PM, Adrian Otto wrote: Jay, On Nov 27, 2013, at 10:36 AM, Jay Pipes jaypi...@gmail.com wrote: On 11/27/2013 06:23 AM, Tom Deckers (tdeckers) wrote: I understand the an Assembly can be a larger group of components. However, those together exist to provide a capability which we want to capture in some catalog so the capability becomes discoverable. I'm not sure how the 'listing' mechanism works out in practice. If this can be used in an enterprise ecosystem to discover services then that's fine. We should capture a work item flesh out discoverability of both Applications and Assemblies. I make that distinction because both scenarios should be provided. As a service consumer, I should be able to look at the 'Applications' listed in the Openstack environment and provision them. In that case, we should also support flavors of the service. Depending on the consumer-provider relationship, we might want to provide different configuratons of the same Application. (e.g. gold-silver-bronze tiering). I believe this is covered by the 'listing' you mentioned. Once deployed, there should also be a mechanism to discover the deployed assemblies. One example of such deployed Assembly is a persistence service that can in its turn be used as a Service in another Assembly. The specific details of the capability provided by the Assembly needs to be discoverable in order to allow successful auto-wiring (I've seen a comment about this elsewhere in the project - I believe in last meeting). Another thought around the naming of Assembly... there's no reason why the API cannot just ditch the entire notion of an assembly, and just use Component in a self-referential way. In other words, an Application (or whatever is agree on for that resource name) contains one or more Components. Components may further be composed of one or more (sub)Components, which themselves may be composed of further (sub)Components. That way you keep the notion of a Component as generic and encompassing as possible and allow for an unlimited generic hierarchy of Component resources to comprise an Application. As currently proposed, an Assembly (a top level grouping of Components) requires only one Component, but may contain many. The question is whether we should even have an Assembly. I admit that Assembly is a new term, and therefore requires definition, explanation, and examples. However, I think eliminating it and just using Components is getting a bit too abstract, and requires a bit too much explanation. I consider this subject analogous to the fundamentals concepts of Chemistry. Imagine trying to describe a molecule by only using the concept of an atom. Each atom can be different, and have more or less electrons etc. But if we did not have the concept of a molecule (a top level grouping of atoms), and tried to explain them as atoms contained within other atoms, Chemistry would get harder to teach. We want this API to be understandable to Application Developers. I am afraid of simplifying matters too much, and making things a bit too abstract. Understood, but I actually think that the Component inside Component approach would work quite well with a simple component type attribute of the Component resource. In your particle physics example, it would be the equivalent of saying that an Atom is composed of subatomic particles, with those subatomic particles having different types (hadrons, baryons, mesons, etc) and those subatomic particles being composed of zero or more subatomic particles of various types (neutrons, protons, fermions, bosons, etc). In fact, particle physics has the concept of elementary particles -- those particles whose composition is unknown -- and composite particles -- those particles that are composed of other particles. The congruence between the taxonomy of particles and what I'm proposing is actually remarkable :) Elementary particle is like a Component with no sub Components Composite particle is like a Component with sub Components. Each particle has a type, and each Component would also have a type. Yes, this is precisely my point. I'm aiming for elementary Chemistry, and you're aiming for Particle Physics. LOL. Touché. Other possibility: Call an Assembly exactly what it is: ComponentGroup I'm open to revisiting more possible names for this besides Assembly, but I do strongly believe that the top level grouping should be it's own thing, and should not just be a self referential arrangement of the same type of resources. I'd like it to convey the idea that an Assembly is the running instance of the complete application, and all of its various parts. I'm not convinced that componentGroup conveys that. Fair enough :) -jay ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org
Re: [openstack-dev] [Nova][Schduler] Volunteers wanted for a modest proposal for an external scheduler in our lifetime
Robert Collins wrote: https://etherpad.openstack.org/p/icehouse-external-scheduler Just looked into it with release management / TC hat on and I have a (possibly minor) concern on the deprecation path/timing. Assuming everything goes well, the separate scheduler will be fast-tracked through incubation in I, graduate at the end of the I cycle to be made a fully-integrated project in the J release. Your deprecation path description mentions that the internal scheduler will be deprecated in I, although there is no released (or security-supported) alternative to switch to at that point. It's not until the J release that such an alternative will be made available. So IMHO for the release/security-oriented users, the switch point is when they start upgrading to J, and not the final step of their upgrade to I (as suggested by the deploy the external scheduler and switch over before you consider your migration to I complete wording in the Etherpad). As the first step towards *switching to J* you would install the new scheduler before upgrading Nova itself. That works whether you're a CD user (and start deploying pre-J stuff just after the I release), or a release user (and wait until J final release to switch to it). Maybe we are talking about the same thing (the migration to the separate scheduler must happen after the I release and, at the latest, when you switch to the J release) -- but I wanted to make sure we were on the same page. I also assume that all the other scheduler-consuming projects would develop the capability to talk to the external scheduler during the J cycle, so that their own schedulers would be deprecated in J release and removed at the start of H. That would be, to me, the condition to considering the external scheduler as integrated with (even if not mandatory for) the rest of the common release components. Does that work for you ? -- Thierry Carrez (ttx) ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Keystone][Oslo] Future of Key Distribution Server, Trusted Messaging
Hey Anyone got an update on this? The keystone blueprint for KDS was marked approved on Tuesday: https://blueprints.launchpad.net/keystone/+spec/key-distribution-server and a new keystone review was added on Sunday, but it must be a draft since I can't access it: https://review.openstack.org/58124 Thanks, Mark. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Ceilometer] storage driver testing
Hello Sandy, I'm very interested in performance results for Ceilometer. Now we have successfully installed Ceilometer in the HA-lab with 200 computes and 3 controllers. Now it works pretty good with MySQL. Our next steps are: 1. Configure alarms 2. Try to use Rally for OpenStack performance with MySQL and MongoDB ( https://wiki.openstack.org/wiki/Rally) We are open to any suggestions. Thanks, Nadya On Wed, Nov 27, 2013 at 9:42 PM, Sandy Walsh sandy.wa...@rackspace.comwrote: Hey! We've ballparked that we need to store a million events per day. To that end, we're flip-flopping between sql and no-sql solutions, hybrid solutions that include elastic search and other schemes. Seems every road we go down has some limitations. So, we've started working on test suite for load testing the ceilometer storage drivers. The intent is to have a common place to record our findings and compare with the efforts of others. There's an etherpad where we're tracking our results [1] and a test suite that we're building out [2]. The test suite works against a fork of ceilometer where we can keep our experimental storage driver tweaks [3]. The test suite hits the storage drivers directly, bypassing the api, but still uses the ceilometer models. We've added support for dumping the results to statsd/graphite for charting of performance results in real-time. If you're interested in large scale deployments of ceilometer, we would welcome any assistance. Thanks! -Sandy [1] https://etherpad.openstack.org/p/ceilometer-data-store-scale-testing [2] https://github.com/rackerlabs/ceilometer-load-tests [3] https://github.com/rackerlabs/instrumented-ceilometer ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] request-id in API response
On 11/28/2013 07:45 AM, Akihiro Motoki wrote: Hi, I am working on adding request-id to API response in Neutron. After I checked what header is used in other projects header name varies project by project. It seems there is no consensus what header is recommended and it is better to have some consensus. nova: x-compute-request-id cinder: x-compute-request-id glance: x-openstack-request-id neutron: x-network-request-id (under review) request-id is assigned and used inside of each project now, so x-service-request-id looks good. On the other hand, if we have a plan to enhance request-id across projects, x-openstack-request-id looks better. My vote is for: x-openstack-request-id With an implementation of create a request UUID if none exists yet in some standardized WSGI middleware... Best, -jay ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Ceilometer] storage driver testing
On Fri, Nov 29 2013, Nadya Privalova wrote: I'm very interested in performance results for Ceilometer. Now we have successfully installed Ceilometer in the HA-lab with 200 computes and 3 controllers. Now it works pretty good with MySQL. Our next steps are: What I'd like to know in both your and Sandy's tests, is the number of collector you are running in parallel. -- Julien Danjou /* Free Software hacker * independent consultant http://julien.danjou.info */ signature.asc Description: PGP signature ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] request-id in API response
On 11/29/2013 10:33 AM, Jay Pipes wrote: On 11/28/2013 07:45 AM, Akihiro Motoki wrote: Hi, I am working on adding request-id to API response in Neutron. After I checked what header is used in other projects header name varies project by project. It seems there is no consensus what header is recommended and it is better to have some consensus. nova: x-compute-request-id cinder: x-compute-request-id glance: x-openstack-request-id neutron: x-network-request-id (under review) request-id is assigned and used inside of each project now, so x-service-request-id looks good. On the other hand, if we have a plan to enhance request-id across projects, x-openstack-request-id looks better. My vote is for: x-openstack-request-id With an implementation of create a request UUID if none exists yet in some standardized WSGI middleware... Agreed. I don't think I see any value in having these have different service names, having just x-openstack-request-id across all the services seems a far better idea, and come back through and fix nova and cinder to be that as well. -Sean -- Sean Dague http://dague.net signature.asc Description: OpenPGP digital signature ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Solum] Working group on language packs
On 11/25/2013 12:39 PM, Georgy Okrokvertskhov wrote: Hi, Just for clarification for Windows images, I think Windows image creation is closer to Docker approach. In order to create a special Windows image we use KVM\QEMU VM with initial base image, then install all necessary components, configure them and then run special tool sysprep to remove all machine specific information like passwords and SIDS and then create a snapshot. I got an impression that Docker does the same, installs application on running VM and then creates a snapshot. It looks like that this can be done with using Heat + HOT software orchestration\Deployment tools without any additional services. This solution scales very well as all configuration steps are executed inside a VM. Right. This is essentially what diskimage-builder does now. You don't even need heat for it- it does all of that locally and makes a nice qcow for you - but it starts with a base cloud image, executes commands in it, and saves the results. On Sat, Nov 23, 2013 at 4:30 PM, Clayton Coleman ccole...@redhat.com mailto:ccole...@redhat.com wrote: On Nov 23, 2013, at 6:48 PM, Robert Collins robe...@robertcollins.net mailto:robe...@robertcollins.net wrote: Ok, so no - diskimage-builder builds regular OpenStack full disk disk images. Translating that to a filesystem is easy; doing a diff against another filesystem version is also doable, and if the container service for Nova understands such partial container contents you could certainly glue it all in together, but we don't have any specific glue for that today. I think docker is great, and if the goal of solum is to deploy via docker, I'd suggest using docker - no need to make diskimage-builder into a docker clone. OTOH if you're deploying via heat, I think Diskimage-builder is targeted directly at your needs : we wrote it for deploying OpenStack after all. I think we're targeting all possible deployment paths, rather than just one. Docker simply represents one emerging direction for deployments due to its speed and efficiency (which vms can't match). The base concept (images and image like constructs that can be started by nova) provides a clean abstraction - how those images are created is specific to the ecosystem or organization. An organization that is heavily invested in a particular image creation technology already can still take advantage of Solum, because all that is necessary for Solum to know about is a thin shim around transforming that base image into a deployable image. The developer and administrative support roles can split responsibilities - one that maintains a baseline, and one that consumes that baseline. -Rob On 24 November 2013 12:24, Adrian Otto adrian.o...@rackspace.com mailto:adrian.o...@rackspace.com wrote: On Nov 23, 2013, at 2:39 PM, Robert Collins robe...@robertcollins.net mailto:robe...@robertcollins.net wrote: On 24 November 2013 05:42, Clayton Coleman ccole...@redhat.com mailto:ccole...@redhat.com wrote: Containers will work fine in diskimage-builder. One only needs to hack in the ability to save in the container image format rather than qcow2. That's good to know. Will diskimage-builder be able to break those down into multiple layers? What do you mean? Docker images can be layered. You can have a base image on the bottom, and then an arbitrary number of deltas on top of that. It essentially works like incremental backups do. You can think of it as each layer has a parent image, and if they all collapse together, you get the current state. Keeping track of past layers gives you the potential for rolling back to a particular restore point, or only distributing incremental changes when you know that the previous layer is already on the host. -Rob -- Robert Collins rbtcoll...@hp.com mailto:rbtcoll...@hp.com Distinguished Technologist HP Converged Cloud ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org mailto:OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org mailto:OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev -- Robert Collins rbtcoll...@hp.com mailto:rbtcoll...@hp.com Distinguished Technologist HP Converged Cloud
Re: [openstack-dev] [nova][scheduler][metrics] Additional metrics
Hi Abbass, I am in the process of coding some of this now - take a look at https://blueprints.launchpad.net/nova/+spec/extensible-resource-tracking - now has a specification document attached https://etherpad.openstack.org/p/IcehouseNovaExtensibleSchedulerMetrics - the design summit session on this topic see what you think and feel free to comment - I think it covers exactly what you describe. Paul. Paul Murray HP Cloud Services +44 117 312 9309 Hewlett-Packard Limited registered Office: Cain Road, Bracknell, Berks RG12 1HN Registered No: 690597 England. The contents of this message and any attachments to it are confidential and may be legally privileged. If you have received this message in error, you should delete it from your system immediately and advise the sender. To any recipient of this message within HP, unless otherwise stated you should consider this message and attachments as HP CONFIDENTIAL. -Original Message- From: Lu, Lianhao [mailto:lianhao...@intel.com] Sent: 22 November 2013 02:03 To: OpenStack Development Mailing List (not for usage questions) Subject: Re: [openstack-dev] [nova][scheduler][metrics] Additional metrics Abbass MAROUNI wrote on 2013-11-21: Hello, I'm in the process of writing a new scheduling algorithm for openstack nova. I have a set of compute nodes that I'm going to filter and weigh according to some metrics collected from these compute nodes. I saw nova.compute.resource_tracker and metrics (ram, disk and cpu) that it collects from compute nodes and updates the rows corresponding to compute nodes in the database. I'm planning to write some modules that will collect the new metrics but I'm wondering if I need to modify the database schema by adding more columns in the 'compute_nodes' table for my new metrics. Will this require some modification to the compute model ? Then how can I use these metrics during the scheduling process, do I fetch each compute node row from the database ? Is there any easier way around this problem ? Best Regards, There are currently some effort on this: https://blueprints.launchpad.net/nova/+spec/utilization-aware-scheduling https://blueprints.launchpad.net/nova/+spec/extensible-resource-tracking - Lianhao ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Ceilometer] storage driver testing
On 11/29/2013 11:41 AM, Julien Danjou wrote: On Fri, Nov 29 2013, Nadya Privalova wrote: I'm very interested in performance results for Ceilometer. Now we have successfully installed Ceilometer in the HA-lab with 200 computes and 3 controllers. Now it works pretty good with MySQL. Our next steps are: What I'd like to know in both your and Sandy's tests, is the number of collector you are running in parallel. For our purposes we aren't interested in the collector. We're purely testing the performance of the storage drivers and the underlying databases. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Ceilometer] storage driver testing
On 11/29/2013 11:32 AM, Nadya Privalova wrote: Hello Sandy, I'm very interested in performance results for Ceilometer. Now we have successfully installed Ceilometer in the HA-lab with 200 computes and 3 controllers. Now it works pretty good with MySQL. Our next steps are: 1. Configure alarms 2. Try to use Rally for OpenStack performance with MySQL and MongoDB (https://wiki.openstack.org/wiki/Rally) We are open to any suggestions. Awesome, as a group we really need to start a similar effort as the storage driver tests for ceilometer in general. I assume you're just pulling Samples via the agent? We're really just focused on event storage and retrieval. There seems to be three levels of load testing required: 1. testing through the collectors (either sample or event collection) 2. testing load on the CM api 3. testing the storage drivers. Sounds like you're addressing #1, we're addressing #3 and Tempest integration tests will be handling #2. I should also add that we've instrumented the db and ceilometer hosts using Diamond to statsd/graphite for tracking load on the hosts while the tests are underway. This will help with determining how many collectors we need, where the bottlenecks are coming from, etc. It might be nice to standardize on that so we can compare results? -S Thanks, Nadya On Wed, Nov 27, 2013 at 9:42 PM, Sandy Walsh sandy.wa...@rackspace.com mailto:sandy.wa...@rackspace.com wrote: Hey! We've ballparked that we need to store a million events per day. To that end, we're flip-flopping between sql and no-sql solutions, hybrid solutions that include elastic search and other schemes. Seems every road we go down has some limitations. So, we've started working on test suite for load testing the ceilometer storage drivers. The intent is to have a common place to record our findings and compare with the efforts of others. There's an etherpad where we're tracking our results [1] and a test suite that we're building out [2]. The test suite works against a fork of ceilometer where we can keep our experimental storage driver tweaks [3]. The test suite hits the storage drivers directly, bypassing the api, but still uses the ceilometer models. We've added support for dumping the results to statsd/graphite for charting of performance results in real-time. If you're interested in large scale deployments of ceilometer, we would welcome any assistance. Thanks! -Sandy [1] https://etherpad.openstack.org/p/ceilometer-data-store-scale-testing [2] https://github.com/rackerlabs/ceilometer-load-tests [3] https://github.com/rackerlabs/instrumented-ceilometer ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org mailto:OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Ceilometer] storage driver testing
Sandy, Seems like we should think about how we can combine our approaches. Rally makes load using python clients (e.g. ceilometer python client) using different amount of users/tenatns/active_users/... So it address #2 point. About profiling part. Actually we attempted to make profiling system based on tomograph + zipkin. But after we finished work around it we got complex and unstable solution. So we take a look at ceilometer and seems like it is the perfect solution for storing profiling data. So we are almost done with this part. Only thing that we need now is virtualization system, that could be ported from zipkin. So, it will be nice if you will be able to join our efforts. And help with testing ceilometer build OpenStack profiling system. Best regards, Boris Pavlovic On Fri, Nov 29, 2013 at 9:05 PM, Sandy Walsh sandy.wa...@rackspace.comwrote: On 11/29/2013 11:32 AM, Nadya Privalova wrote: Hello Sandy, I'm very interested in performance results for Ceilometer. Now we have successfully installed Ceilometer in the HA-lab with 200 computes and 3 controllers. Now it works pretty good with MySQL. Our next steps are: 1. Configure alarms 2. Try to use Rally for OpenStack performance with MySQL and MongoDB (https://wiki.openstack.org/wiki/Rally) We are open to any suggestions. Awesome, as a group we really need to start a similar effort as the storage driver tests for ceilometer in general. I assume you're just pulling Samples via the agent? We're really just focused on event storage and retrieval. There seems to be three levels of load testing required: 1. testing through the collectors (either sample or event collection) 2. testing load on the CM api 3. testing the storage drivers. Sounds like you're addressing #1, we're addressing #3 and Tempest integration tests will be handling #2. I should also add that we've instrumented the db and ceilometer hosts using Diamond to statsd/graphite for tracking load on the hosts while the tests are underway. This will help with determining how many collectors we need, where the bottlenecks are coming from, etc. It might be nice to standardize on that so we can compare results? -S Thanks, Nadya On Wed, Nov 27, 2013 at 9:42 PM, Sandy Walsh sandy.wa...@rackspace.com mailto:sandy.wa...@rackspace.com wrote: Hey! We've ballparked that we need to store a million events per day. To that end, we're flip-flopping between sql and no-sql solutions, hybrid solutions that include elastic search and other schemes. Seems every road we go down has some limitations. So, we've started working on test suite for load testing the ceilometer storage drivers. The intent is to have a common place to record our findings and compare with the efforts of others. There's an etherpad where we're tracking our results [1] and a test suite that we're building out [2]. The test suite works against a fork of ceilometer where we can keep our experimental storage driver tweaks [3]. The test suite hits the storage drivers directly, bypassing the api, but still uses the ceilometer models. We've added support for dumping the results to statsd/graphite for charting of performance results in real-time. If you're interested in large scale deployments of ceilometer, we would welcome any assistance. Thanks! -Sandy [1] https://etherpad.openstack.org/p/ceilometer-data-store-scale-testing [2] https://github.com/rackerlabs/ceilometer-load-tests [3] https://github.com/rackerlabs/instrumented-ceilometer ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org mailto:OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Ceilometer] storage driver testing
On Fri, Nov 29 2013, Sandy Walsh wrote: For our purposes we aren't interested in the collector. We're purely testing the performance of the storage drivers and the underlying databases. Then the connection would map to: with how many SQL connection are you injecting things in the DB in parallel? -- Julien Danjou -- Free Software hacker - independent consultant -- http://julien.danjou.info signature.asc Description: PGP signature ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [oslo] maintenance policy for code graduating from the incubator
We have a review up (https://review.openstack.org/#/c/58297/) to add some features to the notification system in the oslo incubator. THe notification system is being moved into oslo.messaging, and so we have the question of whether to accept the patch to the incubated version, move it to oslo.messaging, or carry it in both. As I say in the review, from a practical standpoint I think we can't really support continued development in both places. Given the number of times the topic of just make everything a library has come up, I would prefer that we focus our energy on completing the transition for a given module or library once it the process starts. We also need to avoid feature drift, and provide a clear incentive for projects to update to the new library. Based on that, I would like to say that we do not add new features to incubated code after it starts moving into a library, and only provide stable-like bug fix support until integrated projects are moved over to the graduated library (although even that is up for discussion). After all integrated projects that use the code are using the library instead of the incubator, we can delete the module(s) from the incubator. Before we make this policy official, I want to solicit feedback from the rest of the community and the Oslo core team. Thanks, Doug ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [oslo] maintenance policy for code graduating from the incubator
Based on that, I would like to say that we do not add new features to incubated code after it starts moving into a library, and only provide stable-like bug fix support until integrated projects are moved over to the graduated library (although even that is up for discussion). After all integrated projects that use the code are using the library instead of the incubator, we can delete the module(s) from the incubator. +1 Although never formalized, this is how I had expected we would handle the graduation process. It is also how we have been responding to patches and blueprints offerings improvements and feature requests for oslo.messaging. -- Regards, Eric Windisch ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [oslo] maintenance policy for code graduating from the incubator
Doug, Based on that, I would like to say that we do not add new features to incubated code after it starts moving into a library, and only provide stable-like bug fix support until integrated projects are moved over to the graduated library (although even that is up for discussion). After all integrated projects that use the code are using the library instead of the incubator, we can delete the module(s) from the incubator. Sounds good and right. Best regards, Boris Pavlovic On Fri, Nov 29, 2013 at 10:47 PM, Eric Windisch e...@cloudscaling.comwrote: Based on that, I would like to say that we do not add new features to incubated code after it starts moving into a library, and only provide stable-like bug fix support until integrated projects are moved over to the graduated library (although even that is up for discussion). After all integrated projects that use the code are using the library instead of the incubator, we can delete the module(s) from the incubator. +1 Although never formalized, this is how I had expected we would handle the graduation process. It is also how we have been responding to patches and blueprints offerings improvements and feature requests for oslo.messaging. -- Regards, Eric Windisch ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [ceilometer][qa] Punting ceilometer from whitelist
In preparing to fail builds with log errors I have been trying to make things easier for projects by maintaining a whitelist. But these bugs in ceilometer are coming in so fast that I can't keep up. So I am just putting .* in the white list for any cases I find before gate failing is turned on, hopefully early this week. -David ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [oslo] maintenance policy for code graduating from the incubator
So, as I mention in the branch, what about deployments that haven't transitioned to the library but would like to cherry pick this feature? after it starts moving into a library can leave a very big gap when the functionality isn't available to users. -S From: Eric Windisch [e...@cloudscaling.com] Sent: Friday, November 29, 2013 2:47 PM To: OpenStack Development Mailing List (not for usage questions) Subject: Re: [openstack-dev] [oslo] maintenance policy for code graduating from the incubator Based on that, I would like to say that we do not add new features to incubated code after it starts moving into a library, and only provide stable-like bug fix support until integrated projects are moved over to the graduated library (although even that is up for discussion). After all integrated projects that use the code are using the library instead of the incubator, we can delete the module(s) from the incubator. +1 Although never formalized, this is how I had expected we would handle the graduation process. It is also how we have been responding to patches and blueprints offerings improvements and feature requests for oslo.messaging. -- Regards, Eric Windisch ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [qa] Pinging people for reviews in IRC
Folks, I understand that the review latency can be too long. We just added two core reviewers and I am sure we can do better still. But please, if you feel you must ping some one by name for a review, do so in #openstack-qa rather than pinging on a private channel. That way other people might see it as well. -David ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [oslo] maintenance policy for code graduating from the incubator
On Fri, Nov 29, 2013 at 2:14 PM, Sandy Walsh sandy.wa...@rackspace.comwrote: So, as I mention in the branch, what about deployments that haven't transitioned to the library but would like to cherry pick this feature? after it starts moving into a library can leave a very big gap when the functionality isn't available to users. Are those deployments tracking trunk or a stable branch? Because IIUC, we don't add features like this to stable branches for the main components, either, and if they are tracking trunk then they will get the new feature when it ships in a project that uses it. Are you suggesting something in between? Doug -S From: Eric Windisch [e...@cloudscaling.com] Sent: Friday, November 29, 2013 2:47 PM To: OpenStack Development Mailing List (not for usage questions) Subject: Re: [openstack-dev] [oslo] maintenance policy for code graduating from the incubator Based on that, I would like to say that we do not add new features to incubated code after it starts moving into a library, and only provide stable-like bug fix support until integrated projects are moved over to the graduated library (although even that is up for discussion). After all integrated projects that use the code are using the library instead of the incubator, we can delete the module(s) from the incubator. +1 Although never formalized, this is how I had expected we would handle the graduation process. It is also how we have been responding to patches and blueprints offerings improvements and feature requests for oslo.messaging. -- Regards, Eric Windisch ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [keystone] Service scoped role definition
Hi Arvind I have added my two-penneth to the latest version. I look forward to your comments regards David On 26/11/2013 23:07, Tiwari, Arvind wrote: Hi David, Thanks for your time and valuable comments. I have replied to your comments and try to explain why I am advocating to this BP. Let me know your thoughts, please feel free to update below etherpad https://etherpad.openstack.org/p/service-scoped-role-definition Thanks again, Arvind -Original Message- From: David Chadwick [mailto:d.w.chadw...@kent.ac.uk] Sent: Monday, November 25, 2013 12:12 PM To: Tiwari, Arvind; OpenStack Development Mailing List Cc: Henry Nash; ayo...@redhat.com; dolph.math...@gmail.com; Yee, Guang Subject: Re: [openstack-dev] [keystone] Service scoped role definition Hi Arvind I have just added some comments to your blueprint page regards David On 19/11/2013 00:01, Tiwari, Arvind wrote: Hi, Based on our discussion in design summit , I have redone the service_id binding with roles BP https://blueprints.launchpad.net/keystone/+spec/serviceid-binding-with-role-definition. I have added a new BP (link below) along with detailed use case to support this BP. https://blueprints.launchpad.net/keystone/+spec/service-scoped-role-definition Below etherpad link has some proposals for Role REST representation and pros and cons analysis https://etherpad.openstack.org/p/service-scoped-role-definition Please take look and let me know your thoughts. It would be awesome if we can discuss it in tomorrow's meeting. Thanks, Arvind ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] problems with rabbitmq on HA controller failure...anyone seen this?
Hi, We're currently running Grizzly (going to Havana soon) and we're running into an issue where if the active controller is ungracefully killed then nova-compute on the compute node doesn't properly connect to the new rabbitmq server on the newly-active controller node. I saw a bugfix in Folsom (https://bugs.launchpad.net/nova/+bug/718869) to retry the connection to rabbitmq if it's lost, but it doesn't seem to be properly handling this case. Interestingly, killing and restarting nova-compute on the compute node seems to work, which implies that the retry code is doing something less effective than the initial startup. Has anyone doing HA controller setups run into something similar? Chris ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Neutron] L3 router service integration with Service Type Framework
FYI, I pushed code review for the blueprint. The patch is missing unitest and tempest. It's only submitted for discussion. The patch implements two-step commit process similar to ML2, but it's not intended to solve all race conditions. Another thing I think worth discussing is how to work with agent scheduler. Thanks, Gary On Thu, Oct 24, 2013 at 11:56 AM, Gary Duan gd...@varmour.com wrote: Hi, I've registered a BP for L3 router service integration with service framework. https://blueprints.launchpad.net/neutron/+spec/l3-router-service-type In general, the implementation will align with how LBaaS is integrated with the framework. One consideration we heard from several team members is to be able to support vendor specific features and extensions in the service plugin. Any comment is welcome. Thanks, Gary ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] problems with rabbitmq on HA controller failure...anyone seen this?
On Nov 29, 02:22:17 PM (Friday), Chris Friesen wrote: We're currently running Grizzly (going to Havana soon) and we're running into an issue where if the active controller is ungracefully killed then nova-compute on the compute node doesn't properly connect to the new rabbitmq server on the newly-active controller node. I saw a bugfix in Folsom (https://bugs.launchpad.net/nova/+bug/718869) to retry the connection to rabbitmq if it's lost, but it doesn't seem to be properly handling this case. Interestingly, killing and restarting nova-compute on the compute node seems to work, which implies that the retry code is doing something less effective than the initial startup. Has anyone doing HA controller setups run into something similar? So the rabbitmq server and the controller are on the same node? My guess is that it's related to this bug 856764 (RabbitMQ connections lack heartbeat or TCP keepalives). The gist of it is that since there are no heartbeats between the MQ and nova-compute, if the MQ goes down ungracefully then nova-compute has no way of knowing. If the MQ goes down gracefully then the MQ clients are notified and so the problem doesn't arise. We got bitten by the same bug a while ago when our controller node got hard reset without any warning!. It came down to this bug (which, unfortunately, doesn't have a fix yet). We worked around this bug by implementing our own crude fix - we wrote a simple app to periodically check if the MQ was alive (write a short message into the MQ, then read it out again). When this fails n-times in a row we restart nova-compute. Very ugly, but it worked! -- Koo ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev