[Bug 1852221] Re: ovs-vswitchd needs to be forced to reconfigure after adding protocols to bridges

2020-09-18 Thread Xav Paice
re-reading this, the issue I was seeing was that the protocol wasn't negotiated - did not need to restart ovs to get the 'good' test. Apologies for the noise, it does actually look like this is also fixed in 2.13.0-0ubuntu1 and possibly could be updated for openvswitch (Ubuntu). -- You received

[Bug 1852221] Re: ovs-vswitchd needs to be forced to reconfigure after adding protocols to bridges

2020-09-18 Thread Xav Paice
Seeing this in Focal, openvswitch version 2.13.0-0ubuntu1 -- You received this bug notification because you are a member of Ubuntu Bugs, which is subscribed to Ubuntu. https://bugs.launchpad.net/bugs/1852221 Title: ovs-vswitchd needs to be forced to reconfigure after adding protocols to

[Bug 1734204] Re: Insufficient free host memory pages available to allocate guest RAM with Open vSwitch DPDK in Newton

2020-08-19 Thread Xav Paice
Hi, just wondering if there's any update on the work to get this into Bionic? -- You received this bug notification because you are a member of Ubuntu Bugs, which is subscribed to Ubuntu. https://bugs.launchpad.net/bugs/1734204 Title: Insufficient free host memory pages available to allocate

[Bug 1874344] [NEW] php 7.2 failure on Bionic install of icingaweb2

2020-04-22 Thread Xav Paice
Public bug reported: When I use the icingaweb2 package 2.7.2-1.bionic, I hit https://github.com/Icinga/icingaweb2/issues/3459. If I use 2.7.3-1.bionic from the upstream repo https://packages.icinga.com/ubuntu this is fixed - can we get a refresh? Reproducer: follow

[Bug 1078213] Re: logs are not logrotated

2019-06-18 Thread Xav Paice
** Tags added: canonical-bootstack ** Changed in: juju (Ubuntu) Status: Triaged => New ** Package changed: juju (Ubuntu) => juju -- You received this bug notification because you are a member of Ubuntu Bugs, which is subscribed to Ubuntu. https://bugs.launchpad.net/bugs/1078213 Title:

[Bug 1828617] Re: Hosts randomly 'losing' disks, breaking ceph-osd service enumeration

2019-06-04 Thread Xav Paice
The pvscan issue is likely something different, just wanted to make sure folks are aware of it for completeness. The logs /var/log/ceph/ceph-volume-systemd.log and ceph-volume.log are empty. -- You received this bug notification because you are a member of Ubuntu Bugs, which is subscribed to

[Bug 1828617] Re: Hosts randomly 'losing' disks, breaking ceph-osd service enumeration

2019-06-04 Thread Xav Paice
Let me word that last comment differently. I went to the host and installed the PPA update, then rebooted. When the box booted up, the PV which hosts the wal LVs wasn't listed in lsblk or 'pvs' or lvs. I then ran pvscan --cache, which brought the LVs back online, but not the OSDs, so I

[Bug 1828617] Re: Hosts randomly 'losing' disks, breaking ceph-osd service enumeration

2019-06-03 Thread Xav Paice
After installing that PPA update and rebooting, the PV for the wal didn't come online till I ran pvscan --cache. Seems a second reboot didn't do that though, might have been a red herring from prior attempts. Unfortunately, the OSDs didn't seem to come online in exactly the same way after

[Bug 1828617] Re: Hosts randomly 'losing' disks, breaking ceph-osd service enumeration

2019-05-29 Thread Xav Paice
Thanks, will do. FWIW, the symlinks are in place before reboot. -- You received this bug notification because you are a member of Ubuntu Bugs, which is subscribed to Ubuntu. https://bugs.launchpad.net/bugs/1828617 Title: Hosts randomly 'losing' disks, breaking ceph-osd service enumeration

[Bug 1828617] Re: Hosts randomly 'losing' disks, breaking ceph-osd service enumeration

2019-05-28 Thread Xav Paice
journalctl --no-pager -lu systemd-udevd.service >/tmp/1828617-1.out Hostname obfusticated lsblk: NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT loop0

[Bug 1828617] Re: Hosts randomly 'losing' disks, breaking ceph-osd service enumeration

2019-05-28 Thread Xav Paice
udevadm info -e >/tmp/1828617-2.out ~# ls -l /var/lib/ceph/osd/ceph* -rw--- 1 ceph ceph 69 May 21 08:44 /var/lib/ceph/osd/ceph.client.osd-upgrade.keyring /var/lib/ceph/osd/ceph-11: total 24 lrwxrwxrwx 1 ceph ceph 93 May 28 22:12 block ->

[Bug 1828617] Re: Hosts randomly 'losing' disks, breaking ceph-osd service enumeration

2019-05-28 Thread Xav Paice
Charm is cs:ceph-osd-284 Ceph version is 12.2.11-0ubuntu0.18.04.2 The udev rules are created by curtin during the maas install. Here's an example udev rule: cat bcache4.rules # Written by curtin SUBSYSTEM=="block", ACTION=="add|change",

[Bug 1828617] Re: Hosts randomly 'losing' disks, breaking ceph-osd service enumeration

2019-05-22 Thread Xav Paice
Just one update, if I change the perms of the symlink made (chown -h) the OSD will actually start. After rebooting, however, I found that the links I had made had gone again and the whole process needed repeating in order to start the OSD. -- You received this bug notification because you are a

[Bug 1828617] Re: Hosts randomly 'losing' disks, breaking ceph-osd service enumeration

2019-05-22 Thread Xav Paice
Added field-critical, there's a cloud deploy ongoing where I currently can't reboot any hosts, nor get some of the OSDs back from a host I rebooted, until we have a workaround. -- You received this bug notification because you are a member of Ubuntu Bugs, which is subscribed to Ubuntu.

[Bug 1828617] Re: Hosts randomly 'losing' disks, breaking ceph-osd service enumeration

2019-05-22 Thread Xav Paice
I'm seeing this in a slightly different manner, on Bionic/Queens. We have LVMs encrypted (thanks Vault), and rebooting a host results in at least one OSD not returning fairly consistently. The LVs appear in the list, however the difference between a working and a non-working OSD is the lack of

[Bug 1828617] Re: Hosts randomly 'losing' disks, breaking ceph-osd service enumeration

2019-05-22 Thread Xav Paice
** Tags added: canonical-bootstack -- You received this bug notification because you are a member of Ubuntu Bugs, which is subscribed to Ubuntu. https://bugs.launchpad.net/bugs/1828617 Title: Hosts randomly 'losing' disks, breaking ceph-osd service enumeration To manage notifications about

[Bug 1784342] Re: AttributeError: 'Subnet' object has no attribute '_obj_network_id'

2019-05-02 Thread Xav Paice
subscribed field-high, added Ubuntu Neutron package, since this has occurred in multiple production sites. ** Also affects: neutron (Ubuntu) Importance: Undecided Status: New -- You received this bug notification because you are a member of Ubuntu Bugs, which is subscribed to Ubuntu.

[Bug 1827159] Re: check_all_disks includes squashfs /snap/* which are 100%

2019-04-30 Thread Xav Paice
** Merge proposal linked: https://code.launchpad.net/~xavpaice/nagios-charm/+git/nagios-charm/+merge/366740 -- You received this bug notification because you are a member of Ubuntu Bugs, which is subscribed to Ubuntu. https://bugs.launchpad.net/bugs/1827159 Title: check_all_disks includes

[Bug 1827159] [NEW] check_all_disks includes squashfs /snap/* which are 100%

2019-04-30 Thread Xav Paice
Public bug reported: When using nagios to monitor the Nagios host itself, if the host is not a container, the template for checking the disk space on the Nagios host does not exclude any snap filesystems. This means we get a Critical report if any snap is installed. This can be changed by

[Bug 1820789] [NEW] Removing a hypervisor doesn't delete it entirely

2019-03-18 Thread Xav Paice
Public bug reported: When removing a host (because it got rebuilt, for example), we use: openstack compute service delete I expected that to remove the hostname cleanly from the database (or at least mark it as deleted) so that the hostname can be re-used. This isn't the case, the host

[Bug 1809454] Re: [SRU] nova rbd auth fallback uses cinder user with libvirt secret

2019-02-10 Thread Xav Paice
How do we go about getting this moving forward from cloud-archive :queens-proposed to stable so we can run this in production? -- You received this bug notification because you are a member of Ubuntu Bugs, which is subscribed to Ubuntu. https://bugs.launchpad.net/bugs/1809454 Title: [SRU]

[Bug 1809454] Re: [SRU] nova rbd auth fallback uses cinder user with libvirt secret

2018-12-22 Thread Xav Paice
** Tags added: canonical-bootstack -- You received this bug notification because you are a member of Ubuntu Bugs, which is subscribed to Ubuntu. https://bugs.launchpad.net/bugs/1809454 Title: [SRU] nova rbd auth fallback uses cinder user with libvirt secret To manage notifications about this

[Bug 1802226] Re: upgrade to 13.0.1-0ubuntu3~cloud0 caused loss of css

2018-11-07 Thread Xav Paice
fwiw, setting Horizon to run with debug appears to allow things to work OK, but of course we don't want to leave it that way. -- You received this bug notification because you are a member of Ubuntu Bugs, which is subscribed to Ubuntu. https://bugs.launchpad.net/bugs/1802226 Title: upgrade to

[Bug 1802226] [NEW] upgrade to 13.0.1-0ubuntu3~cloud0 caused loss of css

2018-11-07 Thread Xav Paice
Public bug reported: Using Queens on Xenial. We updated the packages to the current versions: ~$ apt-cache policy openstack-dashboard-ubuntu-theme openstack-dashboard-ubuntu-theme: Installed: 3:13.0.1-0ubuntu3~cloud0 Candidate: 3:13.0.1-0ubuntu3~cloud0 Version table: ***

[Bug 1452641] Re: Static Ceph mon IP addresses in connection_info can prevent VM startup

2018-09-20 Thread Xav Paice
Just a clarification on the process to 'move' ceph-mon units. I added ceph mons to the cluster, and removed the old ones - in this case it was a 'juju add-unit' and 'juju remove-unit' but any process to achieve the same thing would have the same result - the mons are now all on different

[Bug 1737866] Re: Too many open files when large number of routers on a host

2018-08-16 Thread Xav Paice
Subscribed field-high because we have an active environment (more?) that are are affected by this using Xenial/Ocata, and we really need that SRU released. -- You received this bug notification because you are a member of Ubuntu Bugs, which is subscribed to Ubuntu.

[Bug 1737866] Re: Too many open files when large number of routers on a host

2018-08-14 Thread Xav Paice
Any update on when we might land an SRU for Xenial? -- You received this bug notification because you are a member of Ubuntu Bugs, which is subscribed to Ubuntu. https://bugs.launchpad.net/bugs/1737866 Title: Too many open files when large number of routers on a host To manage notifications

[Bug 1744062] Re: [SRU] L3 HA: multiple agents are active at the same time

2018-07-12 Thread Xav Paice
Subscribed field-high. This is affecting production environments. -- You received this bug notification because you are a member of Ubuntu Bugs, which is subscribed to Ubuntu. https://bugs.launchpad.net/bugs/1744062 Title: [SRU] L3 HA: multiple agents are active at the same time To manage

[Bug 1731595] Re: L3 HA: multiple agents are active at the same time

2018-07-03 Thread Xav Paice
Corey, as far as I'm aware there isn't a bug open for the keepalived package (for Xenial at least). Are you suggesting that we open a bug for a backport to the current cloudarchive package? -- You received this bug notification because you are a member of Ubuntu Bugs, which is subscribed to

[Bug 1731595] Re: L3 HA: multiple agents are active at the same time

2018-06-29 Thread Xav Paice
Comment for the folks that are noticing this as 'fix released' but still affected - see https://github.com/acassen/keepalived/commit/e90a633c34fbe6ebbb891aa98bf29ce579b8b45c for the rest of this fix, we need keepalived to be at least 1.4.0 in order to have this commit. -- You received this bug

[Bug 1777070] Re: firefox plugin libwidevinecdm.so crashes due to apparmor denial

2018-06-17 Thread Xav Paice
Thanks! I won't claim to understand what that change did, but adding the two lines as requested does seem to resolve the issue. I opened up Netflix and was able to watch, without the crash, and there wasn't any new entries in syslog. -- You received this bug notification because you are a

[Bug 1777070] [NEW] firefox plugin libwidevinecdm.so crashes due to apparmor denial

2018-06-15 Thread Xav Paice
Public bug reported: Ubuntu 18.04, Firefox 60.0.1+build2-0ubuntu0.18.04.1 Running firefix, then going to netflix.com and attempting to play a movie. The widevinecdm plugin crashes, the following is found in syslog: Jun 15 19:13:22 xplt kernel: [301351.553043] audit: type=1400

[Bug 1452641] Re: Static Ceph mon IP addresses in connection_info can prevent VM startup

2018-06-11 Thread Xav Paice
FWIW, in the cloud we saw this, migrating the (stopped) instance also updated the connection info - it was just that migrating hundreds of instances wasn't practical. -- You received this bug notification because you are a member of Ubuntu Bugs, which is subscribed to Ubuntu.

[Bug 1313539] Re: [DisplayPort] monitor shows black screen and "no input signal" after turning the monitor off and on manually

2018-05-14 Thread Xav Paice
FWIW, seeing this using any desktop environment, and at the login screen. Have tried, i3, awesomewm, Gnome and Unity. -- You received this bug notification because you are a member of Ubuntu Bugs, which is subscribed to Ubuntu. https://bugs.launchpad.net/bugs/1313539 Title: [DisplayPort]

[Bug 1313539] Re: [DisplayPort] monitor shows black screen and "no input signal" after turning the monitor off and on manually

2018-05-14 Thread Xav Paice
Seeing this on Bionic also, with 2 external screens and the built in laptop display. product: UX303UA vendor: ASUSTeK COMPUTER INC. product: HD Graphics 520 configuration: driver=i915 latency=0 I've attached my Xorg.log if that's helpful. On first boot, or after unplugging the

[Bug 1313539] Re: [DisplayPort] monitor shows black screen and "no input signal" after turning the monitor off and on manually

2018-05-14 Thread Xav Paice
It really didn't like me trying to attach the log file. Here's a pastebin: https://paste.ubuntu.com/p/mm4cwkGv4z/ -- You received this bug notification because you are a member of Ubuntu Bugs, which is subscribed to Ubuntu. https://bugs.launchpad.net/bugs/1313539 Title: [DisplayPort] monitor

[Bug 1770040] Re: lbaas load balancer does not forward traffic unless agent restarted

2018-05-14 Thread Xav Paice
This was reproduced with a heat template, but just running the steps at the start of the case from horizon are enough. Note that neutron- gateway was deployed with aa-profile-mode set to complain, not the default setting. Changing this to 'disable' seems to have fixed the problem, more testing

[Bug 1770040] Re: lbaas load balancer does not forward traffic unless agent restarted

2018-05-13 Thread Xav Paice
Apparmor is in 'complain' mode, the logs show the same entries but allowed rather than denied. Worth trying that change first, then installing -proposed if that makes no difference. This is a production site after all. -- You received this bug notification because you are a member of Ubuntu

[Bug 1770040] Re: lbaas load balancer does not forward traffic unless agent restarted

2018-05-10 Thread Xav Paice
Please note that this affects customers as follows; - customer creates a lbaas, no backends come up - we restart the service, and backends come to life - customer creates another lbaas, the running one is fine but the new one has no backends - we restart... etc This means for every new load

[Bug 1770040] Re: lbaas load balancer does not forward traffic unless agent restarted

2018-05-10 Thread Xav Paice
Due to customer impact, have subscribed field-high. -- You received this bug notification because you are a member of Ubuntu Bugs, which is subscribed to Ubuntu. https://bugs.launchpad.net/bugs/1770040 Title: lbaas load balancer does not forward traffic unless agent restarted To manage

[Bug 1724173] Re: bcache makes the whole io system hang after long run time

2018-02-14 Thread Xav Paice
We're also seeing this with 4.4.0-111-generic (on Trusty), and a very similar hardware profile. The boxes in question are running Swift with a large (millions) number of objects all approx 32k in size. I'm currently fio'ing in a test environment to try to reproduce this away from production. --

[Bug 1724173] Re: bcache makes the whole io system hang after long run time

2018-02-14 Thread Xav Paice
** Tags added: canonical-bootstack -- You received this bug notification because you are a member of Ubuntu Bugs, which is subscribed to Ubuntu. https://bugs.launchpad.net/bugs/1724173 Title: bcache makes the whole io system hang after long run time To manage notifications about this bug go

[Bug 1731595] Re: L3 HA: multiple agents are active at the same time

2017-12-12 Thread Xav Paice
We have installed the Ocata -proposed package, however the situation is this: - there's 464 routers configured, on 3 Neutron gateway hosts, using l3-ha, and each router is scheduled to all 3 hosts. - we installed the package because were in a situation with a current incident with multiple l3

[Bug 1731595] Re: L3 HA: multiple agents are active at the same time

2017-12-11 Thread Xav Paice
Please note, we now have a client affected by this running Mitaka as well. -- You received this bug notification because you are a member of Ubuntu Bugs, which is subscribed to Ubuntu. https://bugs.launchpad.net/bugs/1731595 Title: L3 HA: multiple agents are active at the same time To manage

[Bug 1623658] Re: livestatus socket permission

2017-09-05 Thread Xav Paice
** Changed in: nagios-charm Status: Fix Committed => Fix Released -- You received this bug notification because you are a member of Ubuntu Bugs, which is subscribed to Ubuntu. https://bugs.launchpad.net/bugs/1623658 Title: livestatus socket permission To manage notifications about

[Bug 1623658] Re: livestatus socket permission

2017-08-21 Thread Xav Paice
https://code.launchpad.net/~xavpaice/nagios-charm/+git/nagios- charm/+merge/329344 ** Merge proposal linked: https://code.launchpad.net/~xavpaice/nagios-charm/+git/nagios-charm/+merge/329344 ** Changed in: nagios-charm Status: In Progress => Fix Committed -- You received this bug

[Bug 1623658] Re: livestatus socket permission

2017-08-07 Thread Xav Paice
** Changed in: nagios-charm Status: New => In Progress -- You received this bug notification because you are a member of Ubuntu Bugs, which is subscribed to Ubuntu. https://bugs.launchpad.net/bugs/1623658 Title: livestatus socket permission To manage notifications about this bug go

[Bug 1623658] Re: livestatus socket permission

2017-08-07 Thread Xav Paice
https://code.launchpad.net/~xavpaice/nagios-charm/+git/nagios- charm/+merge/328677 -- You received this bug notification because you are a member of Ubuntu Bugs, which is subscribed to Ubuntu. https://bugs.launchpad.net/bugs/1623658 Title: livestatus socket permission To manage notifications

[Bug 1623658] Re: livestatus socket permission

2017-08-07 Thread Xav Paice
In https://git.launchpad.net/nagios-charm/tree/hooks/install the Nagios charm creates the mklivestatus directory and sets perms. We will need to change this to add +x. -- You received this bug notification because you are a member of Ubuntu Bugs, which is subscribed to Ubuntu.

[Bug 1702595] Re: Upgrade neutron-plugin-openvswitch-agent package causes nova-compute to fall over

2017-07-05 Thread Xav Paice
apologies for the vile wrapping. For those with access, https://pastebin.canonical.com/192695/ might be easier to read. -- You received this bug notification because you are a member of Ubuntu Bugs, which is subscribed to Ubuntu. https://bugs.launchpad.net/bugs/1702595 Title: Upgrade

[Bug 1702595] [NEW] Upgrade neutron-plugin-openvswitch-agent package causes nova-compute to fall over

2017-07-05 Thread Xav Paice
Public bug reported: On upgrading neutron on a compute node, instances on that node wound up losing some of their network plumbing via openvswitch. This cloud: Mitaka, xenial, openvswitch with gre. 2017-07-05 16:17:52, the following auto-upgrade occurred: neutron- common:amd64

[Bug 1623658] Re: livestatus socket permission

2017-06-29 Thread Xav Paice
** Package changed: nagios (Juju Charms Collection) => check-mk (Ubuntu) ** Also affects: nagios-charm Importance: Undecided Status: New -- You received this bug notification because you are a member of Ubuntu Bugs, which is subscribed to Ubuntu.

[Bug 1679823] Re: bond0: Invalid MTU 9000 requested, hw max 1500 with kernel 4.10 (or 4.8.0-49, xenial-hwe)

2017-05-12 Thread Xav Paice
** Tags added: canonical-bootstack -- You received this bug notification because you are a member of Ubuntu Bugs, which is subscribed to Ubuntu. https://bugs.launchpad.net/bugs/1679823 Title: bond0: Invalid MTU 9000 requested, hw max 1500 with kernel 4.10 (or 4.8.0-49, xenial-hwe) To

[Bug 1588391] Re: ceilometer charm creates world-readable /etc/ceilometer/ceilometer.conf, exposing credentials

2017-02-08 Thread Xav Paice
** Tags added: canonical-bootstack -- You received this bug notification because you are a member of Ubuntu Bugs, which is subscribed to Ubuntu. https://bugs.launchpad.net/bugs/1588391 Title: ceilometer charm creates world-readable /etc/ceilometer/ceilometer.conf, exposing credentials To

[Bug 1403152] Re: unregister_netdevice: waiting for lo to become free. Usage count

2016-10-24 Thread Xav Paice
>From the logs it looks like the patch is now a part of https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1627730 which hit 4.4.0-46.67~14.04.1 (proposed) on 22nd Oct? -- You received this bug notification because you are a member of Ubuntu Bugs, which is subscribed to Ubuntu.

[Bug 1474667] Re: log dir permissions are incorrect for use with swift

2015-10-01 Thread Xav Paice
comment #8 was pretty clear that this isn't regarded as something that needs fixing. If that's still the case, this should be closed as wontfix. I don't know another way round it though - and if changing the package is the wrong approach I would like to know the right approach so we can take the

[Bug 1474667] Re: log dir permissions are incorrect

2015-07-21 Thread Xav Paice
I see your point, and agree in principal, but the effect of this is that we cannot use the Ubuntu Cloud Archive until Swift is changed to use the Ceilometer code in some other way. That's a pretty significant change, and a diversion from the approach described by the devs of Ceilometer in their

[Bug 1474667] Re: log dir permissions are incorrect

2015-07-15 Thread Xav Paice
I'm happy to submit a patch if someone could please point me at how? I looked at https://wiki.ubuntu.com/ServerTeam/OpenStack#Submitting_a_Patch but get bzr: ERROR: development focus https://api.launchpad.net/1.0/ceilometer/liberty has no branch in return. No doubt some newbie problem here, a

[Bug 1474667] [NEW] log dir permissions are incorrect

2015-07-14 Thread Xav Paice
Public bug reported: In ceilometer-common.postinst, permissions for the dir /var/log/ceilometer are set to 750. In http://docs.openstack.org/developer/ceilometer/install/manual.html it's stated to Note Please make sure that ceilometer’s logging directory (if it’s configured) is read and write

[Bug 1474667] [NEW] log dir permissions are incorrect

2015-07-14 Thread Xav Paice
Public bug reported: In ceilometer-common.postinst, permissions for the dir /var/log/ceilometer are set to 750. In http://docs.openstack.org/developer/ceilometer/install/manual.html it's stated to Note Please make sure that ceilometer’s logging directory (if it’s configured) is read and write

[Bug 1389239] Re: apparmor is uninstalled when deploying icehouse nova-compute on Precise

2014-11-04 Thread Xav Paice
*** This bug is a duplicate of bug 1387251 *** https://bugs.launchpad.net/bugs/1387251 ** This bug has been marked a duplicate of bug 1387251 apparmor conflict with precise cloud archive -- You received this bug notification because you are a member of Ubuntu Server Team, which is

[Bug 1389239] Re: apparmor is uninstalled when deploying icehouse nova-compute on Precise

2014-11-04 Thread Xav Paice
*** This bug is a duplicate of bug 1387251 *** https://bugs.launchpad.net/bugs/1387251 ** This bug has been marked a duplicate of bug 1387251 apparmor conflict with precise cloud archive -- You received this bug notification because you are a member of Ubuntu Bugs, which is subscribed to