Is this the url for the moderator request: 
https://etherpad.openstack.org/p/BCN-ops-meetup 
<https://etherpad.openstack.org/p/BCN-ops-meetup> << at the bottom, is that 
where we are supposed to fill out?


> On Oct 13, 2016, at 12:19 AM, openstack-operators-requ...@lists.openstack.org 
> wrote:
> 
> Send OpenStack-operators mailing list submissions to
>       openstack-operators@lists.openstack.org
> 
> To subscribe or unsubscribe via the World Wide Web, visit
>       http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
> 
> or, via email, send a message with subject or body 'help' to
>       openstack-operators-requ...@lists.openstack.org
> 
> You can reach the person managing the list at
>       openstack-operators-ow...@lists.openstack.org
> 
> When replying, please edit your Subject line so it is more specific
> than "Re: Contents of OpenStack-operators digest..."
> 
> 
> Today's Topics:
> 
>   1. Re: [openstack-operators][ceph][nova] How do you handle Nova
>      on Ceph? (Adam Kijak)
>   2. Ubuntu package for Octavia (Lutz Birkhahn)
>   3. Re: [openstack-operators][ceph][nova] How do you handle Nova
>      on Ceph? (Adam Kijak)
>   4. [nova] Does anyone use the os-diagnostics API? (Matt Riedemann)
>   5. OPNFV delivered its new Colorado release (Ulrich Kleber)
>   6. Re: [nova] Does anyone use the os-diagnostics API? (Tim Bell)
>   7. Re: [nova] Does anyone use the os-diagnostics    API? (Joe Topjian)
>   8. Re: OPNFV delivered its new Colorado release (Jay Pipes)
>   9. glance, nova backed by NFS (Curtis)
>  10. Re: glance, nova backed by NFS (Kris G. Lindgren)
>  11. Re: glance, nova backed by NFS (Tobias Sch?n)
>  12. Re: glance, nova backed by NFS (Kris G. Lindgren)
>  13. Re: glance, nova backed by NFS (Curtis)
>  14. Re: glance, nova backed by NFS (James Penick)
>  15. Re: glance, nova backed by NFS (Curtis)
>  16. Re: Ubuntu package for Octavia (Xav Paice)
>  17. Re: [openstack-operators][ceph][nova] How do you handle Nova
>      on Ceph? (Warren Wang)
>  18. Re: [openstack-operators][ceph][nova] How do     you handle Nova
>      on Ceph? (Clint Byrum)
>  19. Disable console for an instance (Blair Bethwaite)
>  20. host maintenance (Juvonen, Tomi (Nokia - FI/Espoo))
>  21. Ops@Barcelona - Call for Moderators (Tom Fifield)
> 
> 
> ----------------------------------------------------------------------
> 
> Message: 1
> Date: Wed, 12 Oct 2016 12:23:41 +0000
> From: Adam Kijak <adam.ki...@corp.ovh.com>
> To: Xav Paice <xavpa...@gmail.com>,
>       "openstack-operators@lists.openstack.org"
>       <openstack-operators@lists.openstack.org>
> Subject: Re: [Openstack-operators] [openstack-operators][ceph][nova]
>       How do you handle Nova on Ceph?
> Message-ID: <839b3aba73394bf9aae56c801687e...@corp.ovh.com>
> Content-Type: text/plain; charset="iso-8859-1"
> 
>> ________________________________________
>> From: Xav Paice <xavpa...@gmail.com>
>> Sent: Monday, October 10, 2016 8:41 PM
>> To: openstack-operators@lists.openstack.org
>> Subject: Re: [Openstack-operators] [openstack-operators][ceph][nova] How do 
>> you handle Nova on Ceph?
>> 
>> On Mon, 2016-10-10 at 13:29 +0000, Adam Kijak wrote:
>>> Hello,
>>> 
>>> We use a Ceph cluster for Nova (Glance and Cinder as well) and over
>>> time,
>>> more and more data is stored there. We can't keep the cluster so big
>>> because of
>>> Ceph's limitations. Sooner or later it needs to be closed for adding
>>> new
>>> instances, images and volumes. Not to mention it's a big failure
>>> domain.
>> 
>> I'm really keen to hear more about those limitations.
> 
> Basically it's all related to the failure domain ("blast radius") and risk 
> management.
> Bigger Ceph cluster means more users.
> Growing the Ceph cluster temporary slows it down, so many users will be 
> affected.
> There are bugs in Ceph which can cause data corruption. It's rare, but when 
> it happens 
> it can affect many (maybe all) users of the Ceph cluster.
> 
>>> 
>>> How do you handle this issue?
>>> What is your strategy to divide Ceph clusters between compute nodes?
>>> How do you solve VM snapshot placement and migration issues then
>>> (snapshots will be left on older Ceph)?
>> 
>> Having played with Ceph and compute on the same hosts, I'm a big fan of
>> separating them and having dedicated Ceph hosts, and dedicated compute
>> hosts.  That allows me a lot more flexibility with hardware
>> configuration and maintenance, easier troubleshooting for resource
>> contention, and also allows scaling at different rates.
> 
> Exactly, I consider it the best practice as well.
> 
> 
> 
> 
> ------------------------------
> 
> Message: 2
> Date: Wed, 12 Oct 2016 12:25:38 +0000
> From: Lutz Birkhahn <lutz.birkh...@noris.de>
> To: "openstack-operators@lists.openstack.org"
>       <openstack-operators@lists.openstack.org>
> Subject: [Openstack-operators] Ubuntu package for Octavia
> Message-ID: <b1f091b0-61c6-4ab2-aeec-f2f458101...@noris.de>
> Content-Type: text/plain; charset="utf-8"
> 
> Has anyone seen Ubuntu packages for Octavia yet?
> 
> We?re running Ubuntu 16.04 with Newton, but for whatever reason I can not 
> find any Octavia package?
> 
> So far I?ve only found in 
> https://wiki.openstack.org/wiki/Neutron/LBaaS/HowToRun the following:
> 
>     Ubuntu Packages Setup: Install octavia with your favorite distribution: 
> ?pip install octavia?
> 
> That was not exactly what we would like to do in our production cloud?
> 
> Thanks,
> 
> /lutz
> -------------- next part --------------
> A non-text attachment was scrubbed...
> Name: smime.p7s
> Type: application/x-pkcs7-signature
> Size: 6404 bytes
> Desc: not available
> URL: 
> <http://lists.openstack.org/pipermail/openstack-operators/attachments/20161012/7a1bed2f/attachment-0001.bin>
> 
> ------------------------------
> 
> Message: 3
> Date: Wed, 12 Oct 2016 12:35:48 +0000
> From: Adam Kijak <adam.ki...@corp.ovh.com>
> To: Abel Lopez <alopg...@gmail.com>
> Cc: openstack-operators <openstack-operators@lists.openstack.org>
> Subject: Re: [Openstack-operators] [openstack-operators][ceph][nova]
>       How do you handle Nova on Ceph?
> Message-ID: <29c531e1c0614dc1bd1cf587d69aa...@corp.ovh.com>
> Content-Type: text/plain; charset="iso-8859-1"
> 
>> _______________________________________
>> From: Abel Lopez <alopg...@gmail.com>
>> Sent: Monday, October 10, 2016 9:57 PM
>> To: Adam Kijak
>> Cc: openstack-operators
>> Subject: Re: [Openstack-operators] [openstack-operators][ceph][nova] How do 
>> you handle Nova on Ceph?
>> 
>> Have you thought about dedicated pools for cinder/nova and a separate pool 
>> for glance, and any other uses you might have?
>> You need to setup secrets on kvm, but you can have cinder creating volumes 
>> from glance images quickly in different pools
> 
> We already have separate pool for images, volumes and instances. 
> Separate pools doesn't really split the failure domain though.
> Also AFAIK you can't set up multiple pools for instances in nova.conf, right?
> 
> 
> 
> ------------------------------
> 
> Message: 4
> Date: Wed, 12 Oct 2016 09:00:11 -0500
> From: Matt Riedemann <mrie...@linux.vnet.ibm.com>
> To: "openstack-operators@lists.openstack.org"
>       <openstack-operators@lists.openstack.org>
> Subject: [Openstack-operators] [nova] Does anyone use the
>       os-diagnostics API?
> Message-ID: <5dae5c89-b682-15c7-11c6-d9a548107...@linux.vnet.ibm.com>
> Content-Type: text/plain; charset=utf-8; format=flowed
> 
> The current form of the nova os-diagnostics API is hypervisor-specific, 
> which makes it pretty unusable in any generic way, which is why Tempest 
> doesn't test it.
> 
> Way back when the v3 API was a thing for 2 minutes there was work done 
> to standardize the diagnostics information across virt drivers in nova. 
> The only thing is we haven't exposed that out of the REST API yet, but 
> there is a spec proposing to do that now:
> 
> https://review.openstack.org/#/c/357884/
> 
> This is an admin-only API so we're trying to keep an end user point of 
> view out of discussing it. For example, the disk details don't have any 
> unique identifier. We could add one, but would it be useful to an admin?
> 
> This API is really supposed to be for debug, but the question I have for 
> this list is does anyone actually use the existing os-diagnostics API? 
> And if so, how do you use it, and what information is most useful? If 
> you are using it, please review the spec and provide any input on what's 
> proposed for outputs.
> 
> -- 
> 
> Thanks,
> 
> Matt Riedemann
> 
> 
> 
> 
> ------------------------------
> 
> Message: 5
> Date: Wed, 12 Oct 2016 14:17:50 +0000
> From: Ulrich Kleber <ulrich.kle...@huawei.com>
> To: "openstack-operators@lists.openstack.org"
>       <openstack-operators@lists.openstack.org>
> Subject: [Openstack-operators] OPNFV delivered its new Colorado
>       release
> Message-ID: <884BFBB6F562F44F91BF83F77C24E9972E7EB5F0@lhreml507-mbx>
> Content-Type: text/plain; charset="iso-8859-1"
> 
> Hi,
> I didn't see an official announcement, so I like to point you to the new 
> release of OPNFV.
> https://www.opnfv.org/news-faq/press-release/2016/09/open-source-nfv-project-delivers-third-platform-release-introduces-0
> OPNFV is an open source project and one of the most important users of 
> OpenStack in the Telecom/NFV area. Maybe it is interesting for your work.
> Feel free to contact me or meet during the Barcelona summit at the session of 
> the OpenStack Operators Telecom/NFV Functional Team 
> (https://www.openstack.org/summit/barcelona-2016/summit-schedule/events/16768/openstack-operators-telecomnfv-functional-team).
> Cheers,
> Uli
> 
> 
> Ulrich KLEBER
> Chief Architect Cloud Platform
> European Research Center
> IT R&D Division
> [huawei_logo]
> Riesstra?e 25
> 80992 M?nchen
> Mobile: +49 (0)173 4636144
> Mobile (China): +86 13005480404
> 
> 
> 
> -------------- next part --------------
> An HTML attachment was scrubbed...
> URL: 
> <http://lists.openstack.org/pipermail/openstack-operators/attachments/20161012/7dcc7cae/attachment-0001.html>
> -------------- next part --------------
> A non-text attachment was scrubbed...
> Name: image001.jpg
> Type: image/jpeg
> Size: 6737 bytes
> Desc: image001.jpg
> URL: 
> <http://lists.openstack.org/pipermail/openstack-operators/attachments/20161012/7dcc7cae/attachment-0001.jpg>
> 
> ------------------------------
> 
> Message: 6
> Date: Wed, 12 Oct 2016 14:35:54 +0000
> From: Tim Bell <tim.b...@cern.ch>
> To: Matt Riedemann <mrie...@linux.vnet.ibm.com>
> Cc: "openstack-operators@lists.openstack.org"
>       <openstack-operators@lists.openstack.org>
> Subject: Re: [Openstack-operators] [nova] Does anyone use the
>       os-diagnostics API?
> Message-ID: <248c8965-1ece-4ca3-9b88-a7c75cf8b...@cern.ch>
> Content-Type: text/plain; charset="utf-8"
> 
> 
>> On 12 Oct 2016, at 07:00, Matt Riedemann <mrie...@linux.vnet.ibm.com> wrote:
>> 
>> The current form of the nova os-diagnostics API is hypervisor-specific, 
>> which makes it pretty unusable in any generic way, which is why Tempest 
>> doesn't test it.
>> 
>> Way back when the v3 API was a thing for 2 minutes there was work done to 
>> standardize the diagnostics information across virt drivers in nova. The 
>> only thing is we haven't exposed that out of the REST API yet, but there is 
>> a spec proposing to do that now:
>> 
>> https://review.openstack.org/#/c/357884/
>> 
>> This is an admin-only API so we're trying to keep an end user point of view 
>> out of discussing it. For example, the disk details don't have any unique 
>> identifier. We could add one, but would it be useful to an admin?
>> 
>> This API is really supposed to be for debug, but the question I have for 
>> this list is does anyone actually use the existing os-diagnostics API? And 
>> if so, how do you use it, and what information is most useful? If you are 
>> using it, please review the spec and provide any input on what's proposed 
>> for outputs.
>> 
> 
> Matt,
> 
> Thanks for asking. We?ve used the API in the past as a way of getting the 
> usage data out of Nova. We had problems running ceilometer at scale and this 
> was a way of retrieving the data for our accounting reports. We created a 
> special policy configuration to allow authorised users query this data 
> without full admin rights.
> 
> From the look of the new spec, it would be fairly straightforward to adapt 
> the process to use the new format as all the CPU utilisation data is there.
> 
> Tim
> 
>> -- 
>> 
>> Thanks,
>> 
>> Matt Riedemann
>> 
>> 
>> _______________________________________________
>> OpenStack-operators mailing list
>> OpenStack-operators@lists.openstack.org
>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
> 
> 
> ------------------------------
> 
> Message: 7
> Date: Wed, 12 Oct 2016 08:44:14 -0600
> From: Joe Topjian <j...@topjian.net>
> To: "openstack-operators@lists.openstack.org"
>       <openstack-operators@lists.openstack.org>
> Subject: Re: [Openstack-operators] [nova] Does anyone use the
>       os-diagnostics  API?
> Message-ID:
>       <CA+y7hvg9V8sa-esRGK-mW=OGhiDv5+P7ak=u4zgzppmwzdu...@mail.gmail.com>
> Content-Type: text/plain; charset="utf-8"
> 
> Hi Matt, Tim,
> 
> Thanks for asking. We?ve used the API in the past as a way of getting the
>> usage data out of Nova. We had problems running ceilometer at scale and
>> this was a way of retrieving the data for our accounting reports. We
>> created a special policy configuration to allow authorised users query this
>> data without full admin rights.
>> 
> 
> We do this as well.
> 
> 
>> From the look of the new spec, it would be fairly straightforward to adapt
>> the process to use the new format as all the CPU utilisation data is there.
>> 
> 
> I agree.
> -------------- next part --------------
> An HTML attachment was scrubbed...
> URL: 
> <http://lists.openstack.org/pipermail/openstack-operators/attachments/20161012/dda817a4/attachment-0001.html>
> 
> ------------------------------
> 
> Message: 8
> Date: Wed, 12 Oct 2016 13:02:11 -0400
> From: Jay Pipes <jaypi...@gmail.com>
> To: openstack-operators@lists.openstack.org
> Subject: Re: [Openstack-operators] OPNFV delivered its new Colorado
>       release
> Message-ID: <fe2e53bb-6ed2-c1af-bbee-97c3220a3...@gmail.com>
> Content-Type: text/plain; charset=windows-1252; format=flowed
> 
> On 10/12/2016 10:17 AM, Ulrich Kleber wrote:
>> Hi,
>> 
>> I didn?t see an official announcement, so I like to point you to the new
>> release of OPNFV.
>> 
>> https://www.opnfv.org/news-faq/press-release/2016/09/open-source-nfv-project-delivers-third-platform-release-introduces-0
>> 
>> OPNFV is an open source project and one of the most important users of
>> OpenStack in the Telecom/NFV area. Maybe it is interesting for your work.
> 
> Hi Ulrich,
> 
> I'm hoping you can explain to me what exactly OPNFV is producing in its 
> releases. I've been through a number of the Jira items linked in the 
> press release above and simply cannot tell what is being actually 
> delivered by OPNFV versus what is just something that is in an OpenStack 
> component or deployment.
> 
> A good example of this is the IPV6 project's Jira item here:
> 
> https://jira.opnfv.org/browse/IPVSIX-37
> 
> Which has the title of "Auto-installation of both underlay IPv6 and 
> overlay IPv6". The issue is marked as "Fixed" in Colorado 1.0. However, 
> I can't tell what code was produced in OPNFV that delivers the 
> auto-installation of both an underlay IPv6 and an overlay IPv6.
> 
> In short, I'm confused about what OPNFV is producing and hope to get 
> some insights from you.
> 
> Best,
> -jay
> 
> 
> 
> ------------------------------
> 
> Message: 9
> Date: Wed, 12 Oct 2016 11:21:27 -0600
> From: Curtis <serverasc...@gmail.com>
> To: "openstack-operators@lists.openstack.org"
>       <openstack-operators@lists.openstack.org>
> Subject: [Openstack-operators] glance, nova backed by NFS
> Message-ID:
>       <CAJ_JamBDSb1q9zz2oBwHRBaee5-V0tZJoRd=8yyrpnbfxy4...@mail.gmail.com>
> Content-Type: text/plain; charset=UTF-8
> 
> Hi All,
> 
> I've never used NFS with OpenStack before. But I am now with a small
> lab deployment with a few compute nodes.
> 
> Is there anything special I should do with NFS and glance and nova? I
> remember there was an issue way back when of images being deleted b/c
> certain components weren't aware they are on NFS. I'm guessing that
> has changed but just wanted to check if there is anything specific I
> should be doing configuration-wise.
> 
> I can't seem to find many examples of NFS usage...so feel free to
> point me to any documentation, blog posts, etc. I may have just missed
> it.
> 
> Thanks,
> Curtis.
> 
> 
> 
> ------------------------------
> 
> Message: 10
> Date: Wed, 12 Oct 2016 17:58:47 +0000
> From: "Kris G. Lindgren" <klindg...@godaddy.com>
> To: Curtis <serverasc...@gmail.com>,
>       "openstack-operators@lists.openstack.org"
>       <openstack-operators@lists.openstack.org>
> Subject: Re: [Openstack-operators] glance, nova backed by NFS
> Message-ID: <94226dbf-0d6f-4585-9341-37e193c5f...@godaddy.com>
> Content-Type: text/plain; charset="utf-8"
> 
> We don?t use shared storage at all.  But I do remember what you are talking 
> about.  The issue is that compute nodes weren?t aware they were on shared 
> storage, and would nuke the backing mage from shared storage, after all vm?s 
> on *that* compute node had stopped using it. Not after all vm?s had stopped 
> using it.
> 
> https://bugs.launchpad.net/nova/+bug/1620341 - Looks like some code to 
> address that concern has landed  but only in trunk maybe mitaka.  Any stable 
> releases don?t appear to be shared backing image safe.
> 
> You might be able to get around this by setting the compute image manager 
> task to not run.  But the issue with that will be one missed compute node, 
> and everyone will have a bad day.
> 
> ___________________________________________________________________
> Kris Lindgren
> Senior Linux Systems Engineer
> GoDaddy
> 
> On 10/12/16, 11:21 AM, "Curtis" <serverasc...@gmail.com> wrote:
> 
>    Hi All,
> 
>    I've never used NFS with OpenStack before. But I am now with a small
>    lab deployment with a few compute nodes.
> 
>    Is there anything special I should do with NFS and glance and nova? I
>    remember there was an issue way back when of images being deleted b/c
>    certain components weren't aware they are on NFS. I'm guessing that
>    has changed but just wanted to check if there is anything specific I
>    should be doing configuration-wise.
> 
>    I can't seem to find many examples of NFS usage...so feel free to
>    point me to any documentation, blog posts, etc. I may have just missed
>    it.
> 
>    Thanks,
>    Curtis.
> 
>    _______________________________________________
>    OpenStack-operators mailing list
>    OpenStack-operators@lists.openstack.org
>    http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
> 
> 
> 
> ------------------------------
> 
> Message: 11
> Date: Wed, 12 Oct 2016 17:59:13 +0000
> From: Tobias Sch?n <tobias.sc...@fiberdata.se>
> To: Curtis <serverasc...@gmail.com>,
>       "openstack-operators@lists.openstack.org"
>       <openstack-operators@lists.openstack.org>
> Subject: Re: [Openstack-operators] glance, nova backed by NFS
> Message-ID: <58f74b23f1254f5886df90183092a...@elara.ad.fiberdata.se>
> Content-Type: text/plain; charset="iso-8859-1"
> 
> Hi,
> 
> We have an environment with glance and cinder using NFS.
> It's important that they have the correct rights. The shares should be owned 
> by nova on compute if mounted up on /var/lib/nova/instances
> And the same for nova and glance on the controller..
> 
> It's important that you map the glance and nova up in fstab.
> 
> The cinder one is controlled by the nfsdriver.
> 
> We are running rhelosp6, Openstack Juno.
> 
> This parameter is used:
> nfs_shares_config=/etc/cinder/shares-nfs.conf in the /etc/cinder/cinder.conf 
> file and then we have specified the share in /etc/cinder/shares-nfs.conf.
> 
> chmod 0640 /etc/cinder/shares-nfs.conf
> 
> setsebool -P virt_use_nfs on
> This one is important to make it work with SELinux
> 
> How up to date this is actually I don't know tbh, but it was up to date as of 
> redhat documentation when we deployed it around 1.5y ago.
> 
> //Tobias
> 
> -----Ursprungligt meddelande-----
> Fr?n: Curtis [mailto:serverasc...@gmail.com] 
> Skickat: den 12 oktober 2016 19:21
> Till: openstack-operators@lists.openstack.org
> ?mne: [Openstack-operators] glance, nova backed by NFS
> 
> Hi All,
> 
> I've never used NFS with OpenStack before. But I am now with a small lab 
> deployment with a few compute nodes.
> 
> Is there anything special I should do with NFS and glance and nova? I 
> remember there was an issue way back when of images being deleted b/c certain 
> components weren't aware they are on NFS. I'm guessing that has changed but 
> just wanted to check if there is anything specific I should be doing 
> configuration-wise.
> 
> I can't seem to find many examples of NFS usage...so feel free to point me to 
> any documentation, blog posts, etc. I may have just missed it.
> 
> Thanks,
> Curtis.
> 
> _______________________________________________
> OpenStack-operators mailing list
> OpenStack-operators@lists.openstack.org
> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
> 
> 
> 
> ------------------------------
> 
> Message: 12
> Date: Wed, 12 Oct 2016 18:06:31 +0000
> From: "Kris G. Lindgren" <klindg...@godaddy.com>
> To: Tobias Sch?n <tobias.sc...@fiberdata.se>, Curtis
>       <serverasc...@gmail.com>, "openstack-operators@lists.openstack.org"
>       <openstack-operators@lists.openstack.org>
> Subject: Re: [Openstack-operators] glance, nova backed by NFS
> Message-ID: <88ae471e-81cd-4e72-935d-4390c05f5...@godaddy.com>
> Content-Type: text/plain; charset="utf-8"
> 
> Tobias does bring up something that we have ran into before.
> 
> With NFSv3 user mapping is done by ID, so you need to ensure that all of your 
> servers use the same UID for nova/glance.  If you are using 
> packages/automation that do useradd?s  without the same userid its *VERY* 
> easy to have mismatched username/uid?s across multiple boxes.
> 
> NFSv4, iirc, sends the username and the nfs server does the translation of 
> the name to uid, so it should not have this issue.  But we have been bit by 
> that more than once on nfsv3.
> 
> 
> ___________________________________________________________________
> Kris Lindgren
> Senior Linux Systems Engineer
> GoDaddy
> 
> On 10/12/16, 11:59 AM, "Tobias Sch?n" <tobias.sc...@fiberdata.se> wrote:
> 
>    Hi,
> 
>    We have an environment with glance and cinder using NFS.
>    It's important that they have the correct rights. The shares should be 
> owned by nova on compute if mounted up on /var/lib/nova/instances
>    And the same for nova and glance on the controller..
> 
>    It's important that you map the glance and nova up in fstab.
> 
>    The cinder one is controlled by the nfsdriver.
> 
>    We are running rhelosp6, Openstack Juno.
> 
>    This parameter is used:
>    nfs_shares_config=/etc/cinder/shares-nfs.conf in the 
> /etc/cinder/cinder.conf file and then we have specified the share in 
> /etc/cinder/shares-nfs.conf.
> 
>    chmod 0640 /etc/cinder/shares-nfs.conf
> 
>    setsebool -P virt_use_nfs on
>    This one is important to make it work with SELinux
> 
>    How up to date this is actually I don't know tbh, but it was up to date as 
> of redhat documentation when we deployed it around 1.5y ago.
> 
>    //Tobias
> 
>    -----Ursprungligt meddelande-----
>    Fr?n: Curtis [mailto:serverasc...@gmail.com] 
>    Skickat: den 12 oktober 2016 19:21
>    Till: openstack-operators@lists.openstack.org
>    ?mne: [Openstack-operators] glance, nova backed by NFS
> 
>    Hi All,
> 
>    I've never used NFS with OpenStack before. But I am now with a small lab 
> deployment with a few compute nodes.
> 
>    Is there anything special I should do with NFS and glance and nova? I 
> remember there was an issue way back when of images being deleted b/c certain 
> components weren't aware they are on NFS. I'm guessing that has changed but 
> just wanted to check if there is anything specific I should be doing 
> configuration-wise.
> 
>    I can't seem to find many examples of NFS usage...so feel free to point me 
> to any documentation, blog posts, etc. I may have just missed it.
> 
>    Thanks,
>    Curtis.
> 
>    _______________________________________________
>    OpenStack-operators mailing list
>    OpenStack-operators@lists.openstack.org
>    http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
> 
>    _______________________________________________
>    OpenStack-operators mailing list
>    OpenStack-operators@lists.openstack.org
>    http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
> 
> 
> 
> ------------------------------
> 
> Message: 13
> Date: Wed, 12 Oct 2016 12:18:40 -0600
> From: Curtis <serverasc...@gmail.com>
> To: "Kris G. Lindgren" <klindg...@godaddy.com>
> Cc: "openstack-operators@lists.openstack.org"
>       <openstack-operators@lists.openstack.org>
> Subject: Re: [Openstack-operators] glance, nova backed by NFS
> Message-ID:
>       <caj_jamdhhbm7apewdo1hmefm7yeb3rt7x_cojxycrp3jhvo...@mail.gmail.com>
> Content-Type: text/plain; charset=UTF-8
> 
> On Wed, Oct 12, 2016 at 11:58 AM, Kris G. Lindgren
> <klindg...@godaddy.com> wrote:
>> We don?t use shared storage at all.  But I do remember what you are talking 
>> about.  The issue is that compute nodes weren?t aware they were on shared 
>> storage, and would nuke the backing mage from shared storage, after all vm?s 
>> on *that* compute node had stopped using it. Not after all vm?s had stopped 
>> using it.
>> 
>> https://bugs.launchpad.net/nova/+bug/1620341 - Looks like some code to 
>> address that concern has landed  but only in trunk maybe mitaka.  Any stable 
>> releases don?t appear to be shared backing image safe.
>> 
>> You might be able to get around this by setting the compute image manager 
>> task to not run.  But the issue with that will be one missed compute node, 
>> and everyone will have a bad day.
> 
> Cool, thanks Kris. Exactly what I was talking about. I'm on Mitaka,
> and I will look into that bugfix. I guess I need to test this lol.
> 
> Thanks,
> Curtis.
> 
>> 
>> ___________________________________________________________________
>> Kris Lindgren
>> Senior Linux Systems Engineer
>> GoDaddy
>> 
>> On 10/12/16, 11:21 AM, "Curtis" <serverasc...@gmail.com> wrote:
>> 
>>    Hi All,
>> 
>>    I've never used NFS with OpenStack before. But I am now with a small
>>    lab deployment with a few compute nodes.
>> 
>>    Is there anything special I should do with NFS and glance and nova? I
>>    remember there was an issue way back when of images being deleted b/c
>>    certain components weren't aware they are on NFS. I'm guessing that
>>    has changed but just wanted to check if there is anything specific I
>>    should be doing configuration-wise.
>> 
>>    I can't seem to find many examples of NFS usage...so feel free to
>>    point me to any documentation, blog posts, etc. I may have just missed
>>    it.
>> 
>>    Thanks,
>>    Curtis.
>> 
>>    _______________________________________________
>>    OpenStack-operators mailing list
>>    OpenStack-operators@lists.openstack.org
>>    http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>> 
>> 
> 
> 
> 
> -- 
> Blog: serverascode.com
> 
> 
> 
> ------------------------------
> 
> Message: 14
> Date: Wed, 12 Oct 2016 11:34:39 -0700
> From: James Penick <jpen...@gmail.com>
> To: Curtis <serverasc...@gmail.com>
> Cc: "openstack-operators@lists.openstack.org"
>       <openstack-operators@lists.openstack.org>
> Subject: Re: [Openstack-operators] glance, nova backed by NFS
> Message-ID:
>       <camomh-6y5h_2etguy_2_uoz+sq8poulb9vskbwwckovb8qd...@mail.gmail.com>
> Content-Type: text/plain; charset="utf-8"
> 
> Are you backing both glance and nova-compute with NFS? If you're only
> putting the glance store on NFS you don't need any special changes. It'll
> Just Work.
> 
> On Wed, Oct 12, 2016 at 11:18 AM, Curtis <serverasc...@gmail.com> wrote:
> 
>> On Wed, Oct 12, 2016 at 11:58 AM, Kris G. Lindgren
>> <klindg...@godaddy.com> wrote:
>>> We don?t use shared storage at all.  But I do remember what you are
>> talking about.  The issue is that compute nodes weren?t aware they were on
>> shared storage, and would nuke the backing mage from shared storage, after
>> all vm?s on *that* compute node had stopped using it. Not after all vm?s
>> had stopped using it.
>>> 
>>> https://bugs.launchpad.net/nova/+bug/1620341 - Looks like some code to
>> address that concern has landed  but only in trunk maybe mitaka.  Any
>> stable releases don?t appear to be shared backing image safe.
>>> 
>>> You might be able to get around this by setting the compute image
>> manager task to not run.  But the issue with that will be one missed
>> compute node, and everyone will have a bad day.
>> 
>> Cool, thanks Kris. Exactly what I was talking about. I'm on Mitaka,
>> and I will look into that bugfix. I guess I need to test this lol.
>> 
>> Thanks,
>> Curtis.
>> 
>>> 
>>> ___________________________________________________________________
>>> Kris Lindgren
>>> Senior Linux Systems Engineer
>>> GoDaddy
>>> 
>>> On 10/12/16, 11:21 AM, "Curtis" <serverasc...@gmail.com> wrote:
>>> 
>>>    Hi All,
>>> 
>>>    I've never used NFS with OpenStack before. But I am now with a small
>>>    lab deployment with a few compute nodes.
>>> 
>>>    Is there anything special I should do with NFS and glance and nova? I
>>>    remember there was an issue way back when of images being deleted b/c
>>>    certain components weren't aware they are on NFS. I'm guessing that
>>>    has changed but just wanted to check if there is anything specific I
>>>    should be doing configuration-wise.
>>> 
>>>    I can't seem to find many examples of NFS usage...so feel free to
>>>    point me to any documentation, blog posts, etc. I may have just
>> missed
>>>    it.
>>> 
>>>    Thanks,
>>>    Curtis.
>>> 
>>>    _______________________________________________
>>>    OpenStack-operators mailing list
>>>    OpenStack-operators@lists.openstack.org
>>>    http://lists.openstack.org/cgi-bin/mailman/listinfo/
>> openstack-operators
>>> 
>>> 
>> 
>> 
>> 
>> --
>> Blog: serverascode.com
>> 
>> _______________________________________________
>> OpenStack-operators mailing list
>> OpenStack-operators@lists.openstack.org
>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>> 
> -------------- next part --------------
> An HTML attachment was scrubbed...
> URL: 
> <http://lists.openstack.org/pipermail/openstack-operators/attachments/20161012/d894399e/attachment-0001.html>
> 
> ------------------------------
> 
> Message: 15
> Date: Wed, 12 Oct 2016 12:49:40 -0600
> From: Curtis <serverasc...@gmail.com>
> To: James Penick <jpen...@gmail.com>
> Cc: "openstack-operators@lists.openstack.org"
>       <openstack-operators@lists.openstack.org>
> Subject: Re: [Openstack-operators] glance, nova backed by NFS
> Message-ID:
>       <CAJ_JamBYkB54yKS=V8a_b+FWKSuUtMEJFqm==mzarj738rx...@mail.gmail.com>
> Content-Type: text/plain; charset=UTF-8
> 
> On Wed, Oct 12, 2016 at 12:34 PM, James Penick <jpen...@gmail.com> wrote:
>> Are you backing both glance and nova-compute with NFS? If you're only
>> putting the glance store on NFS you don't need any special changes. It'll
>> Just Work.
> 
> I've got both glance and nova backed by NFS. Haven't put up cinder
> yet, but that will also be NFS backed. I just have very limited
> storage on the compute hosts, basically just enough for the operating
> system; this is just a small but permanent lab deployment. Good to
> hear that Glance will Just Work. :) Thanks!
> 
> Thanks,
> Curtis.
> 
>> 
>> On Wed, Oct 12, 2016 at 11:18 AM, Curtis <serverasc...@gmail.com> wrote:
>>> 
>>> On Wed, Oct 12, 2016 at 11:58 AM, Kris G. Lindgren
>>> <klindg...@godaddy.com> wrote:
>>>> We don?t use shared storage at all.  But I do remember what you are
>>>> talking about.  The issue is that compute nodes weren?t aware they were on
>>>> shared storage, and would nuke the backing mage from shared storage, after
>>>> all vm?s on *that* compute node had stopped using it. Not after all vm?s 
>>>> had
>>>> stopped using it.
>>>> 
>>>> https://bugs.launchpad.net/nova/+bug/1620341 - Looks like some code to
>>>> address that concern has landed  but only in trunk maybe mitaka.  Any 
>>>> stable
>>>> releases don?t appear to be shared backing image safe.
>>>> 
>>>> You might be able to get around this by setting the compute image
>>>> manager task to not run.  But the issue with that will be one missed 
>>>> compute
>>>> node, and everyone will have a bad day.
>>> 
>>> Cool, thanks Kris. Exactly what I was talking about. I'm on Mitaka,
>>> and I will look into that bugfix. I guess I need to test this lol.
>>> 
>>> Thanks,
>>> Curtis.
>>> 
>>>> 
>>>> ___________________________________________________________________
>>>> Kris Lindgren
>>>> Senior Linux Systems Engineer
>>>> GoDaddy
>>>> 
>>>> On 10/12/16, 11:21 AM, "Curtis" <serverasc...@gmail.com> wrote:
>>>> 
>>>>    Hi All,
>>>> 
>>>>    I've never used NFS with OpenStack before. But I am now with a small
>>>>    lab deployment with a few compute nodes.
>>>> 
>>>>    Is there anything special I should do with NFS and glance and nova?
>>>> I
>>>>    remember there was an issue way back when of images being deleted
>>>> b/c
>>>>    certain components weren't aware they are on NFS. I'm guessing that
>>>>    has changed but just wanted to check if there is anything specific I
>>>>    should be doing configuration-wise.
>>>> 
>>>>    I can't seem to find many examples of NFS usage...so feel free to
>>>>    point me to any documentation, blog posts, etc. I may have just
>>>> missed
>>>>    it.
>>>> 
>>>>    Thanks,
>>>>    Curtis.
>>>> 
>>>>    _______________________________________________
>>>>    OpenStack-operators mailing list
>>>>    OpenStack-operators@lists.openstack.org
>>>> 
>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>>>> 
>>>> 
>>> 
>>> 
>>> 
>>> --
>>> Blog: serverascode.com
>>> 
>>> _______________________________________________
>>> OpenStack-operators mailing list
>>> OpenStack-operators@lists.openstack.org
>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>> 
>> 
> 
> 
> 
> -- 
> Blog: serverascode.com
> 
> 
> 
> ------------------------------
> 
> Message: 16
> Date: Thu, 13 Oct 2016 08:24:16 +1300
> From: Xav Paice <xavpa...@gmail.com>
> To: Lutz Birkhahn <lutz.birkh...@noris.de>
> Cc: "openstack-operators@lists.openstack.org"
>       <openstack-operators@lists.openstack.org>
> Subject: Re: [Openstack-operators] Ubuntu package for Octavia
> Message-ID:
>       <CAMb5Lvru-USCW=GztsWxigccVeUGszYcYn-txk=r9zquicv...@mail.gmail.com>
> Content-Type: text/plain; charset="utf-8"
> 
> I highly recommend looking in to Giftwrap for that, until there's UCA
> packages.
> 
> The thing missing from the packages that Giftwrap will produce is init
> scripts, config file examples, and the various user and directory setup
> stuff.  That's easy enough to put into config management or a separate
> package if you wanted to.
> 
> On 13 October 2016 at 01:25, Lutz Birkhahn <lutz.birkh...@noris.de> wrote:
> 
>> Has anyone seen Ubuntu packages for Octavia yet?
>> 
>> We?re running Ubuntu 16.04 with Newton, but for whatever reason I can not
>> find any Octavia package?
>> 
>> So far I?ve only found in https://wiki.openstack.org/
>> wiki/Neutron/LBaaS/HowToRun the following:
>> 
>>     Ubuntu Packages Setup: Install octavia with your favorite
>> distribution: ?pip install octavia?
>> 
>> That was not exactly what we would like to do in our production cloud?
>> 
>> Thanks,
>> 
>> /lutz
>> _______________________________________________
>> OpenStack-operators mailing list
>> OpenStack-operators@lists.openstack.org
>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>> 
>> 
> -------------- next part --------------
> An HTML attachment was scrubbed...
> URL: 
> <http://lists.openstack.org/pipermail/openstack-operators/attachments/20161013/bd85871c/attachment-0001.html>
> 
> ------------------------------
> 
> Message: 17
> Date: Wed, 12 Oct 2016 16:02:55 -0400
> From: Warren Wang <war...@wangspeed.com>
> To: Adam Kijak <adam.ki...@corp.ovh.com>
> Cc: openstack-operators <openstack-operators@lists.openstack.org>
> Subject: Re: [Openstack-operators] [openstack-operators][ceph][nova]
>       How do you handle Nova on Ceph?
> Message-ID:
>       <CAARB8+vKoT3WAtde_vst2x4cDAOiv+S4G4TDOVAq1NniR=4...@mail.gmail.com>
> Content-Type: text/plain; charset="utf-8"
> 
> If fault domain is a concern, you can always split the cloud up into 3
> regions, each having a dedicate Ceph cluster. It isn't necessarily going to
> mean more hardware, just logical splits. This is kind of assuming that the
> network doesn't share the same fault domain though.
> 
> Alternatively, you can split the hardware for the Ceph boxes into multiple
> clusters, and use multi backend Cinder to talk to the same set of
> hypervisors to use multiple Ceph clusters. We're doing that to migrate from
> one Ceph cluster to another. You can even mount a volume from each cluster
> into a single instance.
> 
> Keep in mind that you don't really want to shrink a Ceph cluster too much.
> What's "too big"? You should keep growing so that the fault domains aren't
> too small (3 physical rack min), or you guarantee that the entire cluster
> stops if you lose network.
> 
> Just my 2 cents,
> Warren
> 
> On Wed, Oct 12, 2016 at 8:35 AM, Adam Kijak <adam.ki...@corp.ovh.com> wrote:
> 
>>> _______________________________________
>>> From: Abel Lopez <alopg...@gmail.com>
>>> Sent: Monday, October 10, 2016 9:57 PM
>>> To: Adam Kijak
>>> Cc: openstack-operators
>>> Subject: Re: [Openstack-operators] [openstack-operators][ceph][nova]
>> How do you handle Nova on Ceph?
>>> 
>>> Have you thought about dedicated pools for cinder/nova and a separate
>> pool for glance, and any other uses you might have?
>>> You need to setup secrets on kvm, but you can have cinder creating
>> volumes from glance images quickly in different pools
>> 
>> We already have separate pool for images, volumes and instances.
>> Separate pools doesn't really split the failure domain though.
>> Also AFAIK you can't set up multiple pools for instances in nova.conf,
>> right?
>> 
>> _______________________________________________
>> OpenStack-operators mailing list
>> OpenStack-operators@lists.openstack.org
>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>> 
> -------------- next part --------------
> An HTML attachment was scrubbed...
> URL: 
> <http://lists.openstack.org/pipermail/openstack-operators/attachments/20161012/656c5638/attachment-0001.html>
> 
> ------------------------------
> 
> Message: 18
> Date: Wed, 12 Oct 2016 13:46:01 -0700
> From: Clint Byrum <cl...@fewbar.com>
> To: openstack-operators <openstack-operators@lists.openstack.org>
> Subject: Re: [Openstack-operators] [openstack-operators][ceph][nova]
>       How do  you handle Nova on Ceph?
> Message-ID: <1476304977-sup-4...@fewbar.com>
> Content-Type: text/plain; charset=UTF-8
> 
> Excerpts from Adam Kijak's message of 2016-10-12 12:23:41 +0000:
>>> ________________________________________
>>> From: Xav Paice <xavpa...@gmail.com>
>>> Sent: Monday, October 10, 2016 8:41 PM
>>> To: openstack-operators@lists.openstack.org
>>> Subject: Re: [Openstack-operators] [openstack-operators][ceph][nova] How do 
>>> you handle Nova on Ceph?
>>> 
>>> On Mon, 2016-10-10 at 13:29 +0000, Adam Kijak wrote:
>>>> Hello,
>>>> 
>>>> We use a Ceph cluster for Nova (Glance and Cinder as well) and over
>>>> time,
>>>> more and more data is stored there. We can't keep the cluster so big
>>>> because of
>>>> Ceph's limitations. Sooner or later it needs to be closed for adding
>>>> new
>>>> instances, images and volumes. Not to mention it's a big failure
>>>> domain.
>>> 
>>> I'm really keen to hear more about those limitations.
>> 
>> Basically it's all related to the failure domain ("blast radius") and risk 
>> management.
>> Bigger Ceph cluster means more users.
> 
> Are these risks well documented? Since Ceph is specifically designed
> _not_ to have the kind of large blast radius that one might see with
> say, a centralized SAN, I'm curious to hear what events trigger
> cluster-wide blasts.
> 
>> Growing the Ceph cluster temporary slows it down, so many users will be 
>> affected.
> 
> One might say that a Ceph cluster that can't be grown without the users
> noticing is an over-subscribed Ceph cluster. My understanding is that
> one is always advised to provision a certain amount of cluster capacity
> for growing and replicating to replaced drives.
> 
>> There are bugs in Ceph which can cause data corruption. It's rare, but when 
>> it happens 
>> it can affect many (maybe all) users of the Ceph cluster.
>> 
> 
> :(
> 
> 
> 
> ------------------------------
> 
> Message: 19
> Date: Thu, 13 Oct 2016 13:37:58 +1100
> From: Blair Bethwaite <blair.bethwa...@gmail.com>
> To: "openstack-oper." <openstack-operators@lists.openstack.org>
> Subject: [Openstack-operators] Disable console for an instance
> Message-ID:
>       <ca+z5dsykjc6z4e+xojv_a-ukbv+bx-+bt2mxdyp3c2e-bjb...@mail.gmail.com>
> Content-Type: text/plain; charset="utf-8"
> 
> Hi all,
> 
> Does anyone know whether there is a way to disable the novnc console on a
> per instance basis?
> 
> Cheers,
> Blair
> -------------- next part --------------
> An HTML attachment was scrubbed...
> URL: 
> <http://lists.openstack.org/pipermail/openstack-operators/attachments/20161013/afafbdc9/attachment-0001.html>
> 
> ------------------------------
> 
> Message: 20
> Date: Thu, 13 Oct 2016 06:12:59 +0000
> From: "Juvonen, Tomi (Nokia - FI/Espoo)" <tomi.juvo...@nokia.com>
> To: "OpenStack-operators@lists.openstack.org"
>       <OpenStack-operators@lists.openstack.org>
> Subject: [Openstack-operators] host maintenance
> Message-ID:
>       
> <am4pr07mb15694e5c03f9d1e9c255957b85...@am4pr07mb1569.eurprd07.prod.outlook.com>
>       
> Content-Type: text/plain; charset="us-ascii"
> 
> Hi,
> 
> Had the session in Austin summit for the maintenance:
> https://etherpad.openstack.org/p/AUS-ops-Nova-maint
> 
> Now the discussion have gotten to a point that should start prototyping a 
> service hosting the maintenance. For maintenance Nova could have a link to 
> this new service, but no functionality for the maintenance should be placed 
> in Nova project. Was working to have this, but now looking better to have the 
> prototype first:
> https://review.openstack.org/310510/
> 
>> From the discussion over above review, the new service might have 
>> maintenance API connection point that links to host by utilizing "hostid" 
>> used in Nova and then there should be "tenant _id" specific end point to get 
>> what needed by project. Something like:
> http://maintenancethingy/maintenance/{hostid}
> http://maintenancethingy/maintenance/{hostid}/{tenant_id}
> This will ensure tenant will not know details about host, but can get needed 
> information about maintenance effecting to his instances.
> 
> In Telco/NFV side we have OPNFV Doctor project that sets the requirements for 
> this from that direction. I am personally interested in that part, but to 
> have this to serve all operator requirements, it is best to bring this here.
> 
> This could be further discussed in Barcelona and should get other people 
> interested to help starting with this. Any suggestion for the Ops session?
> 
> Looking forward,
> Tomi
> 
> 
> -------------- next part --------------
> An HTML attachment was scrubbed...
> URL: 
> <http://lists.openstack.org/pipermail/openstack-operators/attachments/20161013/9f01d2ce/attachment-0001.html>
> 
> ------------------------------
> 
> Message: 21
> Date: Thu, 13 Oct 2016 15:19:51 +0800
> From: Tom Fifield <t...@openstack.org>
> To: OpenStack Operators <openstack-operators@lists.openstack.org>
> Subject: [Openstack-operators] Ops@Barcelona - Call for Moderators
> Message-ID: <ca8e7f82-b11e-ab3f-0d7c-9cc26719e...@openstack.org>
> Content-Type: text/plain; charset=utf-8; format=flowed
> 
> Hello all,
> 
> The Ops design summit sessions are now listed on the schedule!
> 
> 
> https://www.openstack.org/summit/barcelona-2016/summit-schedule/global-search?t=Ops+Summit%3A
> 
> Please tick them and set up your summit app :)
> 
> 
> We are still looking for moderators for the following sessions:
> 
> * OpenStack on Containers
> * Containers on OpenStack
> * Migration to OpenStack
> * Fleet Management
> * Feedback to PWG
> * Neutron pain points
> * Config Mgmt
> * HAProy, MySQL, Rabbit Tuning
> * Swift
> * Horizon
> * OpenStack CLI
> * Baremetal Deploy
> * OsOps
> * CI/CD workflows
> * Alt Deployment tech
> * ControlPlane Design(multi region)
> * Docs
> 
> 
> ==> If you are interested in moderating a session, please
> 
> * write your name in its etherpad (multiple moderators OK!)
> 
> ==> I'll be honest, I have no idea what some of the sessions are 
> supposed to be, so also:
> 
> * write a short description for the session so the agenda can be updated
> 
> 
> For those of you who want to know what it takes check out the 
> Moderator's Guide: 
> https://wiki.openstack.org/wiki/Operations/Meetups#Moderators_Guide & 
> ask questions - we're here to help!
> 
> 
> 
> Regards,
> 
> 
> Tom, on behalf of the Ops Meetups Team
> https://wiki.openstack.org/wiki/Ops_Meetups_Team
> 
> 
> 
> ------------------------------
> 
> _______________________________________________
> OpenStack-operators mailing list
> OpenStack-operators@lists.openstack.org
> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
> 
> 
> End of OpenStack-operators Digest, Vol 72, Issue 11
> ***************************************************

_______________________________________________
OpenStack-operators mailing list
OpenStack-operators@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators

Reply via email to