[Ubuntu-ha] [Bug 1890314] Re: RFE HAProxy 2.2 for groovy

2020-08-25 Thread Rafael David Tinoco
*** This bug is a duplicate of bug 1892871 ***
https://bugs.launchpad.net/bugs/1892871

** This bug has been marked a duplicate of bug 1892871
   haproxy needs merge with debian experimental 2.2 (LTS version)

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to haproxy in Ubuntu.
https://bugs.launchpad.net/bugs/1890314

Title:
  RFE HAProxy 2.2 for groovy

Status in haproxy package in Ubuntu:
  Triaged

Bug description:
  Is there any chance of getting HAProxy 2.2.x (LTS) into the 20.10
  groovy release?

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/haproxy/+bug/1890314/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1892871] Re: haproxy needs merge with debian experimental 2.2 (LTS version)

2020-08-25 Thread Rafael David Tinoco
This bug was fixed in the package haproxy - 2.2.2-1

---
haproxy (2.2.2-1) experimental; urgency=medium

  * New upstream version.
- BUG/MAJOR: dns: don't treat Authority records as an error
- BUG/MAJOR: dns: fix null pointer dereference in
 snr_update_srv_status

 -- Vincent Bernat   Sat, 01 Aug 2020 17:06:42 +0200

haproxy (2.2.1-1) experimental; urgency=medium

  * New upstream version.
- BUG/MAJOR: tasks: don't requeue global tasks into the local
 queue
- BUG/MAJOR: dns: Make the do-resolve action thread-safe

 -- Vincent Bernat   Thu, 23 Jul 2020 13:39:14 +0200

haproxy (2.2.0-1) experimental; urgency=medium

  * New upstream version.
  * Upload to experimental
  * Update d/watch to look for 2.2 stable releases
  * d/gbp.conf: set branch names for 2.2
  * d/patches: refresh patches

 -- Vincent Bernat   Tue, 14 Jul 2020 16:53:23 +0200

haproxy (2.1.7-1) experimental; urgency=medium

  * New upstream version.

 -- Vincent Bernat   Fri, 12 Jun 2020 07:50:48 +0200

haproxy (2.1.5-1) experimental; urgency=medium

  * New upstream version.
- BUG/MAJOR: mux-fcgi: Stop sending loop if FCGI stream is blocked for
 any reason
- Revert "BUG/MINOR: connection: always send address-less LOCAL PROXY
  connections"
- Revert "BUG/MINOR: connection: make sure to correctly tag local
  PROXY connections"

 -- Vincent Bernat   Mon, 01 Jun 2020 08:52:56 +0200

haproxy (2.1.4-1) experimental; urgency=medium

  * New upstream version.
- BUG/CRITICAL: hpack: never index a header into the headroom after
wrapping
- BUG/MAJOR: http-ana: Always abort the request when a tarpit is
 triggered
- BUG/MAJOR: list: fix invalid element address calculation
- BUG/MAJOR: proxy_protocol: Properly validate TLV lengths
  * d/control: fix maintainer address. Closes: #93.

 -- Vincent Bernat   Sun, 12 Apr 2020 13:29:54 +0200

haproxy (2.1.3-3) experimental; urgency=medium

  * d/copryight: document OpenSSL exception. Closes: #951782.
  * d/haproxy.cfg: use "ssl-min-ver" to set minimum version.
  * d/patches: fix an overflow in HTTP/2 header handling.
Fix CVE-2020-11100.

 -- Vincent Bernat   Wed, 01 Apr 2020 21:18:57 +0200

haproxy (2.1.3-2) experimental; urgency=medium

  * d/dconv: use Python 3 to build the documentation.
Closes: #948296, #950435.
  * d/dconv: replace cgi.escape by html.escape. Closes: #951416.

 -- Vincent Bernat   Wed, 19 Feb 2020 07:53:53 +0100

haproxy (2.1.3-1) experimental; urgency=medium

  * New upstream version.
- BUG/MAJOR: hashes: fix the signedness of the hash inputs
- BUG/MAJOR: memory: Don't forget to unlock the rwlock if the pool is
 empty.

 -- Vincent Bernat   Mon, 20 Jan 2020 06:53:23 +0100

haproxy (2.1.2-1) experimental; urgency=medium

  * New upstream version 2.1.2.
- BUG/MAJOR: task: add a new TASK_SHARED_WQ flag to fix foreign requeuing
  * d/logrotate.conf: use rsyslog helper instead of SysV init script.
Closes: #946973.

 -- Vincent Bernat   Fri, 20 Dec 2019 08:20:33 +0100

haproxy (2.1.1-1) experimental; urgency=medium

  * New upstream version 2.1.1.
- BUG/MAJOR: dns: add minimalist error processing on the Rx path

 -- Vincent Bernat   Sat, 14 Dec 2019 11:20:32 +0100

haproxy (2.1.0-2) experimental; urgency=medium

  * Link against libatomic on riscv64

 -- Apollon Oikonomopoulos   Fri, 29 Nov 2019
14:03:49 +0200

haproxy (2.1.0-1) experimental; urgency=medium

  * New upstream version 2.1.0
  * Upload to experimental
  * Update d/watch to look for 2.1 stable releases
  * d/gbp.conf: set branch names for 2.1
  * Bump Standards-Version to 4.4.1; no changes needed
  * Bump dh compat level to 12
+ B-D on debhelper-compat and remove debian/compat
+ Override dh_installsystemd with the same args as dh_installinit
+ Add ${misc:Pre-Depends} to haproxy's Pre-Depends

 -- Apollon Oikonomopoulos   Wed, 27 Nov 2019
23:30:30 +0200

** Changed in: haproxy (Ubuntu)
   Status: In Progress => Fix Released

** CVE added: https://cve.mitre.org/cgi-bin/cvename.cgi?name=2020-11100

** Changed in: haproxy (Ubuntu)
 Assignee: Rafael David Tinoco (rafaeldtinoco) => (unassigned)

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to haproxy in Ubuntu.
https://bugs.launchpad.net/bugs/1892871

Title:
  haproxy needs merge with debian experimental 2.2 (LTS version)

Status in haproxy package in Ubuntu:
  Fix Released

Bug description:
  Currently Debian experimental is at "2025-Q2 (LTS)" which is the LTS
  version for HAPROXY. We should merge this version sooner than later in
  the LTS cycles.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/haproxy/+bug/1892871/+subscriptions

___
Mailing list: https:/

[Ubuntu-ha] [Bug 1892871] [NEW] haproxy needs merge with debian experimental 2.2 (LTS version)

2020-08-25 Thread Rafael David Tinoco
Public bug reported:

Currently Debian experimental is at "2025-Q2 (LTS)" which is the LTS
version for HAPROXY. We should merge this version sooner than later in
the LTS cycles.

** Affects: haproxy (Ubuntu)
 Importance: Medium
 Assignee: Rafael David Tinoco (rafaeldtinoco)
 Status: Fix Released

** Changed in: haproxy (Ubuntu)
   Status: New => In Progress

** Changed in: haproxy (Ubuntu)
   Importance: Undecided => Medium

** Changed in: haproxy (Ubuntu)
 Assignee: (unassigned) => Rafael David Tinoco (rafaeldtinoco)

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to haproxy in Ubuntu.
https://bugs.launchpad.net/bugs/1892871

Title:
  haproxy needs merge with debian experimental 2.2 (LTS version)

Status in haproxy package in Ubuntu:
  Fix Released

Bug description:
  Currently Debian experimental is at "2025-Q2 (LTS)" which is the LTS
  version for HAPROXY. We should merge this version sooner than later in
  the LTS cycles.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/haproxy/+bug/1892871/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1890491] Re: A pacemaker node fails monitor (probe) and stop /start operations on a resource because it returns "rc=189

2020-08-20 Thread Rafael David Tinoco
Thanks for taking care of this Jorge. Let me know whenever you have a
fix ready for this.

Cheers!

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1890491

Title:
  A pacemaker node fails monitor (probe) and stop /start operations on a
  resource because it returns "rc=189

Status in pacemaker package in Ubuntu:
  Fix Released
Status in pacemaker source package in Bionic:
  In Progress
Status in pacemaker source package in Focal:
  Fix Released
Status in pacemaker source package in Groovy:
  Fix Released

Bug description:
  Cause: Pacemaker implicitly ordered all stops needed on a Pacemaker
  Remote node before the stop of the node's Pacemaker Remote connection,
  including stops that were implied by fencing of the node. Also,
  Pacemaker scheduled actions on Pacemaker Remote nodes with a failed
  connection so that the actions could be done once the connection is
  recovered, even if the connection wasn't being recovered (for example,
  if the node was shutting down when the failure occurred).

  Consequence: If a Pacemaker Remote node needed to be fenced while it
  was in the process of shutting down, once the fencing completed
  pacemaker scheduled probes on the node. The probes fail because the
  connection is not actually active. Due to the failed probe, a stop is
  scheduled which also fails, leading to fencing of the node again, and
  the situation repeats itself indefinitely.

  Fix: Pacemaker Remote connection stops are no longer ordered after
  implied stops, and actions are not scheduled on Pacemaker Remote nodes
  when the connection is failed and not being started again.

  Result: A Pacemaker Remote node that needs to be fenced while it is in
  the process of shutting down is fenced once, without repeating
  indefinitely.

  The fix seems to be fixed in pacemaker-1.1.21-1.el7

  Related to https://bugzilla.redhat.com/show_bug.cgi?id=1704870

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1890491/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1890185] Re: scheduler: support/backport on-fail="demote" recovery policy for promoted resources

2020-08-10 Thread Rafael David Tinoco
 Forwarded Message 
Subject: [ClusterLabs] Coming in Pacemaker 2.0.5: on-fail=demote / 
no-quorum-policy=demote
Date: Mon, 10 Aug 2020 11:47:24 -0500
From: Ken Gaillot 
Reply-To: Cluster Labs - All topics related to open-source clustering welcomed 

Organization: Red Hat
To: Cluster Labs - All topics related to open-source clustering welcomed 


Hi all,

Looking ahead to the Pacemaker 2.0.5 release expected at the end of
this year, here is a new feature already in the master branch.

When configuring resource operations, Pacemaker lets you set an "on-
fail" policy to specify whether to restart the resource, fence the
node, etc., if the operation fails. With 2.0.5, a new possible value
will be "demote", which will mean "demote this resource but do not
fully restart it".

"Demote" will be a valid value only for promote actions, and for
recurring monitors with "role" set to "Master".

Once the resource is demoted, it will be eligible for promotion again,
so if the promotion scores have not changed, a promote on the same node
may be attempted. If this is not desired, the agent can change the
promotion scores either in the failed monitor or the demote.

The intended use case is an application where a successful demote assures a 
well-functioning service, and a full restart would be
unnecessarily heavyweight. A large database might be an example.

Similarly, Pacemaker offers the cluster-wide "no-quorum-policy" option
to specify what happens to resources when quorum is lost (the default
being to stop them). With 2.0.5, "demote" will be a possible value here
as well, and will mean "demote all promotable resources and stop all
other resources".

The intended use case is an application that cannot cause any harm
after being demoted, and may be useful in a demoted role even if there
is no quorum. A database that operates read-only when demoted and
doesn't depend on any non-promotable resources might be an example.

Happy clustering :)
-- 
Ken Gaillot 

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1890185

Title:
  scheduler: support/backport on-fail="demote" recovery policy for
  promoted resources

Status in pacemaker package in Ubuntu:
  Triaged
Status in pacemaker source package in Bionic:
  Won't Fix
Status in pacemaker source package in Eoan:
  Won't Fix
Status in pacemaker source package in Focal:
  Triaged

Bug description:
  There is a request for the feature:

  Feature: scheduler: new on-fail="demote" recovery policy for promoted…

  to be backported to Ubuntu LTS versions.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1890185/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1890314] Re: RFE HAProxy 2.2 for groovy

2020-08-07 Thread Rafael David Tinoco
Hello Michael,

Thanks for pointing it out. Actually, it was a nice headsup for us. We
were missing the fact that 2.2.x is LTS so I'm marking this as a TODO
for this month in 20.10 cycle.

Will keep this bug updated about it!

** Tags removed: server-triage-discuss

** Changed in: haproxy (Ubuntu)
 Assignee: (unassigned) => Rafael David Tinoco (rafaeldtinoco)

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to haproxy in Ubuntu.
https://bugs.launchpad.net/bugs/1890314

Title:
  RFE HAProxy 2.2 for groovy

Status in haproxy package in Ubuntu:
  Triaged

Bug description:
  Is there any chance of getting HAProxy 2.2.x (LTS) into the 20.10
  groovy release?

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/haproxy/+bug/1890314/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1890314] Re: RFE HAProxy 2.2 for groovy

2020-08-07 Thread Rafael David Tinoco
** Changed in: haproxy (Ubuntu)
   Status: New => Triaged

** Changed in: haproxy (Ubuntu)
   Importance: Undecided => Wishlist

** Tags added: server-triage-discuss

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to haproxy in Ubuntu.
https://bugs.launchpad.net/bugs/1890314

Title:
  RFE HAProxy 2.2 for groovy

Status in haproxy package in Ubuntu:
  Triaged

Bug description:
  Is there any chance of getting HAProxy 2.2.x (LTS) into the 20.10
  groovy release?

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/haproxy/+bug/1890314/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1890185] Re: scheduler: support/backport on-fail="demote" recovery policy for promoted resources

2020-08-03 Thread Rafael David Tinoco
This feature is in 2.0.x releases, so SRU for bionic would be out of question:

 pacemaker | 1.1.18-0ubuntu1   | bionic  | source
 pacemaker | 1.1.18-0ubuntu1.1 | bionic-security | source
 pacemaker | 1.1.18-0ubuntu1.2 | bionic-updates  | source
 pacemaker | 2.0.1-4ubuntu2| eoan| source
 pacemaker | 2.0.3-3ubuntu3| focal   | source
 pacemaker | 2.0.4-2ubuntu1| groovy  | source

and Eoan is not LTS so the effort is not worth.

That would make the need for:

To backport

(1)

874f75e0f - Feature: scheduler: new on-fail="demote" recovery policy for 
promoted resources
2f1e2df1f - Feature: xml: add on-fail="demote" option to resources schema

and all related patches to Focal.

(2)

To make sure Groovy catches up with the feature or to backport similar
patches to Groovy.



** Changed in: pacemaker (Ubuntu Bionic)
   Status: New => Won't Fix

** Changed in: pacemaker (Ubuntu Eoan)
   Status: New => Won't Fix

** Changed in: pacemaker (Ubuntu Focal)
   Status: New => Triaged

** Changed in: pacemaker (Ubuntu Focal)
   Importance: Undecided => Medium

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1890185

Title:
  scheduler: support/backport on-fail="demote" recovery policy for
  promoted resources

Status in pacemaker package in Ubuntu:
  Triaged
Status in pacemaker source package in Bionic:
  Won't Fix
Status in pacemaker source package in Eoan:
  Won't Fix
Status in pacemaker source package in Focal:
  Triaged

Bug description:
  There is a request for the feature:

  Feature: scheduler: new on-fail="demote" recovery policy for promoted…

  to be backported to Ubuntu LTS versions.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1890185/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1890185] Re: scheduler: support/backport on-fail="demote" recovery policy for promoted resources

2020-08-03 Thread Rafael David Tinoco
** Also affects: pacemaker (Ubuntu Eoan)
   Importance: Undecided
   Status: New

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1890185

Title:
  scheduler: support/backport on-fail="demote" recovery policy for
  promoted resources

Status in pacemaker package in Ubuntu:
  Triaged
Status in pacemaker source package in Bionic:
  New
Status in pacemaker source package in Eoan:
  New
Status in pacemaker source package in Focal:
  New

Bug description:
  There is a request for the feature:

  Feature: scheduler: new on-fail="demote" recovery policy for promoted…

  to be backported to Ubuntu LTS versions.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1890185/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1890185] [NEW] scheduler: support/backport on-fail="demote" recovery policy for promoted resources

2020-08-03 Thread Rafael David Tinoco
Public bug reported:

There is a request for the feature:

Feature: scheduler: new on-fail="demote" recovery policy for promoted…

to be backported to Ubuntu LTS versions.

** Affects: pacemaker (Ubuntu)
 Importance: Medium
 Status: Triaged

** Affects: pacemaker (Ubuntu Bionic)
 Importance: Undecided
 Status: New

** Affects: pacemaker (Ubuntu Focal)
 Importance: Undecided
 Status: New

** Changed in: pacemaker (Ubuntu)
   Status: New => Triaged

** Changed in: pacemaker (Ubuntu)
   Importance: Undecided => Medium

** Also affects: pacemaker (Ubuntu Focal)
   Importance: Undecided
   Status: New

** Also affects: pacemaker (Ubuntu Bionic)
   Importance: Undecided
   Status: New

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1890185

Title:
  scheduler: support/backport on-fail="demote" recovery policy for
  promoted resources

Status in pacemaker package in Ubuntu:
  Triaged
Status in pacemaker source package in Bionic:
  New
Status in pacemaker source package in Focal:
  New

Bug description:
  There is a request for the feature:

  Feature: scheduler: new on-fail="demote" recovery policy for promoted…

  to be backported to Ubuntu LTS versions.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1890185/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1886546] Re: pacemaker - 2.0.3-3ubuntu3 needs update (2.0.4-2)

2020-07-06 Thread Rafael David Tinoco
** Description changed:

- pacemaker - 2.0.3-3ubuntu3 needs update (2.0.4-2)
+ BUG: https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1886546
+ PPA: https://launchpad.net/~rafaeldtinoco/+archive/ubuntu/lp1886546
+ MERGE:

** Description changed:

  BUG: https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1886546
  PPA: https://launchpad.net/~rafaeldtinoco/+archive/ubuntu/lp1886546
- MERGE:
+ MERGE: https://tinyurl.com/y83wtxny

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1886546

Title:
  pacemaker - 2.0.3-3ubuntu3 needs update (2.0.4-2)

Status in pacemaker package in Ubuntu:
  In Progress

Bug description:
  BUG: https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1886546
  PPA: https://launchpad.net/~rafaeldtinoco/+archive/ubuntu/lp1886546
  MERGE: https://tinyurl.com/y83wtxny

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1886546/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1886546] [NEW] pacemaker - 2.0.3-3ubuntu3 needs update (2.0.4-2)

2020-07-06 Thread Rafael David Tinoco
Public bug reported:

pacemaker - 2.0.3-3ubuntu3 needs update (2.0.4-2)

** Affects: pacemaker (Ubuntu)
 Importance: Medium
 Status: In Progress

** Changed in: pacemaker (Ubuntu)
   Status: New => In Progress

** Changed in: pacemaker (Ubuntu)
   Importance: Undecided => Medium

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1886546

Title:
  pacemaker - 2.0.3-3ubuntu3 needs update (2.0.4-2)

Status in pacemaker package in Ubuntu:
  In Progress

Bug description:
  pacemaker - 2.0.3-3ubuntu3 needs update (2.0.4-2)

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1886546/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1881762] Re: resource timeout not respecting units

2020-06-17 Thread Rafael David Tinoco
** Changed in: pacemaker (Ubuntu)
   Status: New => Triaged

** Changed in: pacemaker (Ubuntu)
 Assignee: (unassigned) => Rafael David Tinoco (rafaeldtinoco)

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1881762

Title:
  resource timeout not respecting units

Status in pacemaker package in Ubuntu:
  Triaged

Bug description:
  While working on pacemaker, i discovered a issue with timeouts

  haproxy_stop_0 on primary 'OCF_TIMEOUT' (198): call=583, status='Timed
  Out', exitreason='', last-rc-change='1970-01-04 17:21:18 -05:00',
  queued=44ms,  exec=176272ms

  this lead me down the path of finding that setting a timeout unit
  value was not doing anything

  primitive haproxy systemd:haproxy \
  op monitor interval=2s \
  op start interval=0s timeout=500s \
  op stop interval=0s timeout=500s \
  meta migration-threshold=2

  primitive haproxy systemd:haproxy \
  op monitor interval=2s \
  op start interval=0s timeout=500 \
  op stop interval=0s timeout=500 \
  meta migration-threshold=2

  the two above configs result in the same behaviour, pacemaker/crm seems to be 
ignoring the "s"
  I file a bug with pacemaker itself
  https://bugs.clusterlabs.org/show_bug.cgi?id=5429

  but this lead to the following responsed, copied from the ticket:

  <https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1881762/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1412438] Re: OCF:pacemaker:o2cb broken in 14.04

2020-05-21 Thread Rafael David Tinoco
*** This bug is a duplicate of bug 1869643 ***
https://bugs.launchpad.net/bugs/1869643

Thank you for taking the time to report this bug. In an effort to keep an
up-to-date and valid list of bugs to work on, I have reviewed this report to
verify it still requires effort and occurs on an Ubuntu release in standard
support.

This is and old behavior that might still be present in current
resource-agents package agents and is being drive at:

https://bugs.launchpad.net/ubuntu/+source/resource-agents/+bug/1869643

I'm marking this as duplicate so I can keep track of same issues for the
recent Ubuntu versions in an easier way.

** Also affects: ocfs2-tools (Ubuntu Trusty)
   Importance: Undecided
   Status: New

** Also affects: ocfs2-tools (Ubuntu Precise)
   Importance: Undecided
   Status: New

** This bug has been marked a duplicate of bug 1869643
   {fence,resource}-agents: PATH for all binaries called by all agents should 
be resolvable

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to ocfs2-tools in Ubuntu.
https://bugs.launchpad.net/bugs/1412438

Title:
  OCF:pacemaker:o2cb broken in 14.04

Status in ocfs2-tools package in Ubuntu:
  Confirmed
Status in ocfs2-tools source package in Precise:
  New
Status in ocfs2-tools source package in Trusty:
  New

Bug description:
  The pacemaker resource agent ocf:pacemaker:o2cb requires

  /usr/sbin/ocfs2_controld.pcmk

  in order to execute correctly. This binary used to be in ocsf2-tools-
  pacemaker in earlier releases (e.g. 12.04). However, this is omitted
  from 14.04. As a result, resource agent is broken and pacemaker cannot
  be used to manage ocfs2 file systems

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/ocfs2-tools/+bug/1412438/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1584629] Re: Failed to start LSB: Load O2CB cluster services at system boot.

2020-05-21 Thread Rafael David Tinoco
Thank you for taking the time to report this bug. In an effort to keep an
up-to-date and valid list of bugs to work on, I have reviewed this report to
verify it still requires effort and occurs on an Ubuntu release in standard
support, and it does not.

It is unfortunate that we were unable to resolve this defect, however there
appears to be no further action possible at this time. I am therefore moving the
bug to 'Fix Released' to current development release as the issue does not apply
to it any longer.

If you disagree or have new information, we would be grateful if you could
please open a new bug mentioning new affected versions with a possible
reproducer in the bug description.


** Changed in: ocfs2-tools (Ubuntu Trusty)
   Status: New => Triaged

** No longer affects: ocfs2-tools (Ubuntu Xenial)

** No longer affects: ocfs2-tools (Ubuntu Yakkety)

** Changed in: ocfs2-tools (Ubuntu)
   Status: Triaged => Fix Released

** Changed in: ocfs2-tools (Ubuntu Trusty)
   Importance: Undecided => Low

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to ocfs2-tools in Ubuntu.
https://bugs.launchpad.net/bugs/1584629

Title:
  Failed to start LSB: Load O2CB cluster services at system boot.

Status in ocfs2-tools package in Ubuntu:
  Fix Released
Status in ocfs2-tools source package in Trusty:
  Triaged

Bug description:
  Ubuntu 16.04.

  Sometimes (not every boot) o2cb failed to start:

  systemctl status o2cb
  ● o2cb.service - LSB: Load O2CB cluster services at system boot.
     Loaded: loaded (/etc/init.d/o2cb; bad; vendor preset: enabled)
     Active: failed (Result: exit-code) since Пн 2016-05-23 11:46:43 SAMT; 2min 
12s ago
   Docs: man:systemd-sysv-generator(8)
    Process: 1526 ExecStart=/etc/init.d/o2cb start (code=exited, 
status=1/FAILURE)

  май 23 11:46:43 inetgw1 systemd[1]: Starting LSB: Load O2CB cluster services 
at system boot
  май 23 11:46:43 inetgw1 o2cb[1526]: Loading filesystem "configfs": OK
  май 23 11:46:43 inetgw1 o2cb[1526]: Mounting configfs filesystem at 
/sys/kernel/config: mount: configfs is already
  май 23 11:46:43 inetgw1 o2cb[1526]:configfs is already mounted on 
/sys/kernel/config
  май 23 11:46:43 inetgw1 o2cb[1526]: Unable to mount configfs filesystem
  май 23 11:46:43 inetgw1 o2cb[1526]: Failed
  май 23 11:46:43 inetgw1 systemd[1]: o2cb.service: Control process exited, 
code=exited status=1
  май 23 11:46:43 inetgw1 systemd[1]: Failed to start LSB: Load O2CB cluster 
services at system boot..
  май 23 11:46:43 inetgw1 systemd[1]: o2cb.service: Unit entered failed state.
  май 23 11:46:43 inetgw1 systemd[1]: o2cb.service: Failed with result 
'exit-code'.

  next try is successful:
  systemctl status o2cb
  ● o2cb.service - LSB: Load O2CB cluster services at system boot.
     Loaded: loaded (/etc/init.d/o2cb; bad; vendor preset: enabled)
     Active: active (exited) since Пн 2016-05-23 11:49:07 SAMT; 1s ago
   Docs: man:systemd-sysv-generator(8)
    Process: 2101 ExecStart=/etc/init.d/o2cb start (code=exited, 
status=0/SUCCESS)

  май 23 11:49:07 inetgw1 systemd[1]: Starting LSB: Load O2CB cluster services 
at system boot
  май 23 11:49:07 inetgw1 o2cb[2101]: Loading stack plugin "o2cb": OK
  май 23 11:49:07 inetgw1 o2cb[2101]: Loading filesystem "ocfs2_dlmfs": OK
  май 23 11:49:07 inetgw1 o2cb[2101]: Mounting ocfs2_dlmfs filesystem at /dlm: 
OK
  май 23 11:49:07 inetgw1 o2cb[2101]: Setting cluster stack "o2cb": OK
  май 23 11:49:07 inetgw1 o2cb[2101]: Starting O2CB cluster inetgw: OK
  май 23 11:49:07 inetgw1 systemd[1]: Started LSB: Load O2CB cluster services 
at system boot..

  I guess this is startup dependency problem.

  Thank you!

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/ocfs2-tools/+bug/1584629/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1449046] Re: needing non existed file

2020-05-21 Thread Rafael David Tinoco
*** This bug is a duplicate of bug 1869643 ***
https://bugs.launchpad.net/bugs/1869643

** This bug is no longer a duplicate of bug 1412438
   OCF:pacemaker:o2cb broken in 14.04
** This bug has been marked a duplicate of bug 1869643
   {fence,resource}-agents: PATH for all binaries called by all agents should 
be resolvable

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to ocfs2-tools in Ubuntu.
https://bugs.launchpad.net/bugs/1449046

Title:
  needing non existed file

Status in ocfs2-tools package in Ubuntu:
  Confirmed

Bug description:
  $ lsb_release -rd
  Description:  Ubuntu 14.04.2 LTS
  Release:  14.04

  $ sudo apt-cache policy ocfs2-tools
  [sudo] password for termant: 
  ocfs2-tools:
Installed: 1.6.4-3ubuntu1
Candidate: 1.6.4-3ubuntu1
Version table:
   *** 1.6.4-3ubuntu1 0
  500 http://fi.archive.ubuntu.com/ubuntu/ trusty/main amd64 Packages
  100 /var/lib/dpkg/status


  I have installed the following packages:

  apt-get install drbd8-utils corosync pacemaker dlm clvm ocfs2-tools


  Then in my pacemaker configuration (using crm) I have defined

  primitive p-drbd ocf:linbit:drbd \
  params drbd_resource="md2" \
  op monitor interval="50" role="Master" timeout="30" \
  op monitor interval="60" role="Slave" timeout="30" \
  op start interval="0" timeout="240" \
  op stop interval="0" timeout="100"
  ms ms-drbd p-drbd \
  meta master-max="2" clone-max="2" notify="true" interleave="true"

  primitive p-dlm ocf:pacemaker:controld \
  op monitor interval="120" timeout="30" \
  op start interval="0" timeout="90" \
  op stop interval="0" timeout="100"

  primitive p-clvm ocf:lvm2:clvmd \
 params daemon_timeout="30" \
 op monitor interval="60" timeout="30" \
 op start interval="0" timeout="90" \
 op stop interval="0" timeout="100"

  primitive p-o2cb ocf:pacemaker:o2cb \
  op monitor interval="120" timeout="30" \
  op start interval="0" timeout="90" \
  op stop interval="0" timeout="100"

  group g-lock p-dlm p-clvm p-o2cb
  clone c-lock g-lock \
  meta globally-unique="false" interleave="true"
  colocation col-drbd-lock inf: c-lock ms-drbd:Master
  order ord-drbd-lock inf: ms-drbd:promote c-lock


  When I started my cluster with two nodes, output of the crm_mon tool
  was following:

  Last updated: Mon Apr 27 16:25:46 2015
  Last change: Mon Apr 27 16:14:07 2015 via cibadmin on node1
  Stack: corosync
  Current DC: node2 (2) - partition with quorum
  Version: 1.1.10-42f2063
  2 Nodes configured
  10 Resources configured

  
  Online: [ node1 node2 ]

  p-ipmi-stonith-node1 (stonith:fence_ipmilan):Started node2
  p-ipmi-stonith-node2   (stonith:fence_ipmilan):Started node1
   Master/Slave Set: ms-drbd [p-drbd]
   Masters: [ node1 node2 ]

  Failed actions:
  p-o2cb_monitor_0 (node=node1, call=275, rc=5, status=complete, 
last-rc-change
  =Mon Apr 27 16:25:45 2015
  , queued=18ms, exec=0ms
  ): not installed
  p-o2cb_monitor_0 (node=node2, call=66, rc=5, status=complete, 
last-rc-chang
  e=Mon Apr 27 16:25:45 2015
  , queued=21ms, exec=0ms
  ): not installed


  What I expected was:

  Online: [ node1 node2 ]

  Master/Slave Set: ms-drbd [p-drbd]
   Masters: [ node2 node1 ]
  Clone Set: c-lock [g-lock]
   Started: [ node1 node2 ]


  /var/log/syslog indicates that file /usr/sbin/ocfs2_controld.pcmk is
  missing (ERROR: Setup problem: couldn't find command:
  /usr/sbin/ocfs2_controld.pcmk). Then I searched a package which
  contains this file but I couldn't find a match for Ubuntu 14.04. But,
  there had been a file ocfs2-tools-pacemaker which includes this file,
  which is now missing.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/ocfs2-tools/+bug/1449046/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1879106] Re: Sync ocfs2-tools 1.8.6-3 (main) from Debian unstable (main)

2020-05-21 Thread Rafael David Tinoco
This bug was fixed in the package ocfs2-tools - 1.8.6-5
Sponsored for Logan Rosen (logan)

---
ocfs2-tools (1.8.6-5) unstable; urgency=medium

  * debian/control: limit build to linux architectures

 -- Valentin Vidic   Thu, 21 May 2020 14:10:56 +0200

ocfs2-tools (1.8.6-4) unstable; urgency=medium

  * debian/patches: fix o2image segfault on s390x
  * debian/lintian-overrides: ignore manpage warning
  * debian/tests: update test output
  * debian/rules: move startup scripts to /usr/libexec

 -- Valentin Vidic   Thu, 21 May 2020 13:35:08 +0200

ocfs2-tools (1.8.6-3) unstable; urgency=medium

  * debian/salsa-ci.yml: enable CI
  * debian/control: specify only supported architectures (LP: #1745155)
  * debian/salsa-ci.yml: disable blhc warnings
  * debian/control: update Standards-Version to 4.5.0
  * debian/control: use debhelper v13
  * debian/control: update Maintainer mailing list address
  * debian/NEWS: fix file name

 -- Valentin Vidic   Wed, 13 May 2020 21:31:59 +0200

** Changed in: ocfs2-tools (Ubuntu)
   Status: Triaged => Fix Released

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to ocfs2-tools in Ubuntu.
https://bugs.launchpad.net/bugs/1879106

Title:
  Sync ocfs2-tools 1.8.6-3 (main) from Debian unstable (main)

Status in ocfs2-tools package in Ubuntu:
  Fix Released

Bug description:
  Please sync ocfs2-tools 1.8.6-3 (main) from Debian unstable (main)

  Explanation of the Ubuntu delta and why it can be dropped:
* debian/control: specify only supported architectures (LP: #1745155)
  The same change was made in Debian.

  Changelog entries since current groovy version 1.8.6-2ubuntu1:

  ocfs2-tools (1.8.6-3) unstable; urgency=medium

* debian/salsa-ci.yml: enable CI
* debian/control: specify only supported architectures (LP: #1745155)
* debian/salsa-ci.yml: disable blhc warnings
* debian/control: update Standards-Version to 4.5.0
* debian/control: use debhelper v13
* debian/control: update Maintainer mailing list address
* debian/NEWS: fix file name

   -- Valentin Vidic   Wed, 13 May 2020 21:31:59
  +0200

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/ocfs2-tools/+bug/1879106/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 246613] Re: ocfs2 startup link prior to open-iscsi

2020-05-21 Thread Rafael David Tinoco
Thank you for taking the time to report this bug. In an effort to keep an
up-to-date and valid list of bugs to work on, I have reviewed this report
to verify it still requires effort and occurs on an Ubuntu release in
standard support, and it does not.

It is unfortunate that we were unable to resolve this defect, however
there appears to be no further action possible at this time. I am
therefore moving the bug to 'Fix Released' to current development
release as the issue does not apply to it any longer.

If you disagree or have new information, we would be grateful if
you could please open a new bug mentioning new affected versions with
a possible reproducer in the bug description.


** Changed in: ocfs2-tools (Ubuntu)
   Status: Confirmed => Fix Released

** Also affects: ocfs2-tools (Ubuntu Precise)
   Importance: Undecided
   Status: New

** Changed in: ocfs2-tools (Ubuntu Precise)
   Status: New => Won't Fix

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to ocfs2-tools in Ubuntu.
https://bugs.launchpad.net/bugs/246613

Title:
  ocfs2 startup link prior to open-iscsi

Status in ocfs2-tools package in Ubuntu:
  Fix Released
Status in ocfs2-tools source package in Precise:
  Won't Fix

Bug description:
  Binary package hint: ocfs2-tools

  The ocfs2-tools package creates a startup link at /etc/rcS.d/S65ocfs2.
  However, the ocfs2 filesystem is (as I understand it) intended for use
  as a network shared file system.  A common way to accomplish this is
  through the use of iSCSI.  The startup link for open-iscsi is created
  at /etc/rc2.d/S20open-iscsi which means that when ocfs2 starts the
  iSCSI layer is not yet in place and thus the volume(s) are not
  mounted.  Moving the ocfs2 startup link to /etc/rc2.d corrects the
  issue.  I'm not sure if the link for ocfs2 or open-iscsi should be the
  one to move.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/ocfs2-tools/+bug/246613/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1018671] Re: mismatch between ocfs2-tools version and kernel

2020-05-21 Thread Rafael David Tinoco
Thank you for taking the time to report this bug. In an effort to keep an
up-to-date and valid list of bugs to work on, I have reviewed this report to
verify it still requires effort and occurs on an Ubuntu release in standard
support, and it does not.

It is unfortunate that we were unable to resolve this defect, however there
appears to be no further action possible at this time. I am therefore moving the
bug to 'Fix Released' to current development release as the issue does not apply
to it any longer.

If you disagree or have new information, we would be grateful if you could
please open a new bug mentioning new affected versions with a possible
reproducer in the bug description.


** Also affects: ocfs2-tools (Ubuntu Precise)
   Importance: Undecided
   Status: New

** Changed in: ocfs2-tools (Ubuntu Precise)
   Status: New => Won't Fix

** Changed in: ocfs2-tools (Ubuntu)
   Status: Confirmed => Fix Released

** Changed in: ocfs2-tools (Ubuntu)
   Importance: High => Undecided

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to ocfs2-tools in Ubuntu.
https://bugs.launchpad.net/bugs/1018671

Title:
  mismatch between ocfs2-tools version and kernel

Status in ocfs2-tools package in Ubuntu:
  Fix Released
Status in ocfs2-tools source package in Precise:
  Won't Fix

Bug description:
  In at least 11.10 (likely 12.04 as well, however...), the ocfs2-tools
  is behind what the 3.0 kernel is creating on disk.  This can be seen
  trivially by attempting to examine the locks on an ocfs2 filesystem:

  ===# debugfs.ocfs2 -n -R "fs_locks" /dev/dm-1
  Debug string proto 3 found, but 2 is the highest I understand.

  The 1.8 tree in the ocfs2-tools git repo includes support for proto 3.
  That set of tools should probably be distributed with 11.10 and up, so
  that the tools match what the kernel is providing.

  Not having the tools match up to the kernel makes things very
  difficult, like trying to debug locking issues.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/ocfs2-tools/+bug/1018671/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 474215] Re: mountall-net tries to mount ocfs2 before o2cb is started

2020-05-21 Thread Rafael David Tinoco
Thank you for taking the time to report this bug. In an effort to keep an
up-to-date and valid list of bugs to work on, I have reviewed this report to
verify it still requires effort and occurs on an Ubuntu release in standard
support, and it does not.

It is unfortunate that we were unable to resolve this defect, however there
appears to be no further action possible at this time. I am therefore moving the
bug to 'Fix Released' to current development release as the issue does not apply
to it any longer.

If you disagree or have new information, we would be grateful if you could
please open a new bug mentioning new affected versions with a possible
reproducer in the bug description.


** Also affects: ocfs2-tools (Ubuntu Trusty)
   Importance: Undecided
   Status: New

** Also affects: ocfs2-tools (Ubuntu Precise)
   Importance: Undecided
   Status: New

** Changed in: ocfs2-tools (Ubuntu Trusty)
   Status: New => Won't Fix

** Changed in: ocfs2-tools (Ubuntu Precise)
   Status: New => Won't Fix

** Changed in: ocfs2-tools (Ubuntu Trusty)
   Status: Won't Fix => Triaged

** Changed in: ocfs2-tools (Ubuntu)
   Status: Confirmed => Fix Released

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to ocfs2-tools in Ubuntu.
https://bugs.launchpad.net/bugs/474215

Title:
  mountall-net tries to mount ocfs2 before o2cb is started

Status in ocfs2-tools package in Ubuntu:
  Fix Released
Status in ocfs2-tools source package in Precise:
  Won't Fix
Status in ocfs2-tools source package in Trusty:
  Triaged

Bug description:
  Binary package hint: ocfs2-tools

  Upstart tries to mount an ocfs2 filesystem defined in fstab before 
/etc/init.d/o2cb has been run, the mount
  therefore fails.

  As a workaround I run '/etc/init.d/o2cb start' in mountall-net.conf.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/ocfs2-tools/+bug/474215/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1879106] Re: Sync ocfs2-tools 1.8.6-3 (main) from Debian unstable (main)

2020-05-19 Thread Rafael David Tinoco
Hello Logan,

I haven't started yet the work of merging and syncing HA packages for
20.10, but will do very soon. Added this to my queue!

Thank you!

** Changed in: ocfs2-tools (Ubuntu)
   Status: New => Triaged

** Changed in: ocfs2-tools (Ubuntu)
 Assignee: (unassigned) => Rafael David Tinoco (rafaeldtinoco)

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to ocfs2-tools in Ubuntu.
https://bugs.launchpad.net/bugs/1879106

Title:
  Sync ocfs2-tools 1.8.6-3 (main) from Debian unstable (main)

Status in ocfs2-tools package in Ubuntu:
  Triaged

Bug description:
  Please sync ocfs2-tools 1.8.6-3 (main) from Debian unstable (main)

  Explanation of the Ubuntu delta and why it can be dropped:
* debian/control: specify only supported architectures (LP: #1745155)
  The same change was made in Debian.

  Changelog entries since current groovy version 1.8.6-2ubuntu1:

  ocfs2-tools (1.8.6-3) unstable; urgency=medium

* debian/salsa-ci.yml: enable CI
* debian/control: specify only supported architectures (LP: #1745155)
* debian/salsa-ci.yml: disable blhc warnings
* debian/control: update Standards-Version to 4.5.0
* debian/control: use debhelper v13
* debian/control: update Maintainer mailing list address
* debian/NEWS: fix file name

   -- Valentin Vidic   Wed, 13 May 2020 21:31:59
  +0200

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/ocfs2-tools/+bug/1879106/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1815101] Re: [master] Restarting systemd-networkd breaks keepalived, heartbeat, corosync, pacemaker (interface aliases are restarted)

2020-05-14 Thread Rafael David Tinoco
TL;DR TODO SUMMARY:

- netplan change to support KeepConfiguration= for systemd-networkd backend 
(Groovy)
- backport this change: netplan for Ubuntu Focal  (SRU)
- backport this change: netplan for Ubuntu Eoan   (SRU, WontFix due to EOL ?)
- backport this change: netplan for Ubuntu Bionic (SRU)
- backport this change: netplan for Ubuntu Xenial (SRU, WontFix ?)

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to keepalived in Ubuntu.
https://bugs.launchpad.net/bugs/1815101

Title:
  [master] Restarting systemd-networkd breaks keepalived, heartbeat,
  corosync, pacemaker (interface aliases are restarted)

Status in netplan:
  Confirmed
Status in heartbeat package in Ubuntu:
  Won't Fix
Status in keepalived package in Ubuntu:
  In Progress
Status in systemd package in Ubuntu:
  In Progress
Status in keepalived source package in Xenial:
  Confirmed
Status in systemd source package in Xenial:
  Confirmed
Status in keepalived source package in Bionic:
  Confirmed
Status in systemd source package in Bionic:
  Confirmed
Status in systemd source package in Disco:
  Won't Fix
Status in systemd source package in Eoan:
  Fix Released
Status in keepalived source package in Focal:
  Confirmed
Status in systemd source package in Focal:
  Fix Released

Bug description:
  [impact]

  - ALL related HA software has a small problem if interfaces are being
  managed by systemd-networkd: nic restarts/reconfigs are always going
  to wipe all interfaces aliases when HA software is not expecting it to
  (no coordination between them.

  - keepalived, smb ctdb, pacemaker, all suffer from this. Pacemaker is
  smarter in this case because it has a service monitor that will
  restart the virtual IP resource, in affected node & nic, before
  considering a real failure, but other HA service might consider a real
  failure when it is not.

  [test case]

  - comment #14 is a full test case: to have 3 node pacemaker, in that
  example, and cause a networkd service restart: it will trigger a
  failure for the virtual IP resource monitor.

  - other example is given in the original description for keepalived.
  both suffer from the same issue (and other HA softwares as well).

  [regression potential]

  - this backports KeepConfiguration parameter, which adds some
  significant complexity to networkd's configuration and behavior, which
  could lead to regressions in correctly configuring the network at
  networkd start, or incorrectly maintaining configuration at networkd
  restart, or losing network state at networkd stop.

  - Any regressions are most likely to occur during networkd start,
  restart, or stop, and most likely to involve missing or incorrect ip
  address(es).

  - the change is based in upstream patches adding the exact feature we
  needed to fix this issue & it will be integrated with a netplan change
  to add the needed stanza to systemd nic configuration file
  (KeepConfiguration=)

  [other info]

  original description:
  ---

  Configure netplan for interfaces, for example (a working config with
  IP addresses obfuscated)

  network:
  ethernets:
  eth0:
  addresses: [192.168.0.5/24]
  dhcp4: false
  nameservers:
    search: [blah.com, other.blah.com, hq.blah.com, cust.blah.com, 
phone.blah.com]
    addresses: [10.22.11.1]
  eth2:
  addresses:
    - 12.13.14.18/29
    - 12.13.14.19/29
  gateway4: 12.13.14.17
  dhcp4: false
  nameservers:
    search: [blah.com, other.blah.com, hq.blah.com, cust.blah.com, 
phone.blah.com]
    addresses: [10.22.11.1]
  eth3:
  addresses: [10.22.11.6/24]
  dhcp4: false
  nameservers:
    search: [blah.com, other.blah.com, hq.blah.com, cust.blah.com, 
phone.blah.com]
    addresses: [10.22.11.1]
  eth4:
  addresses: [10.22.14.6/24]
  dhcp4: false
  nameservers:
    search: [blah.com, other.blah.com, hq.blah.com, cust.blah.com, 
phone.blah.com]
    addresses: [10.22.11.1]
  eth7:
  addresses: [9.5.17.34/29]
  dhcp4: false
  optional: true
  nameservers:
    search: [blah.com, other.blah.com, hq.blah.com, cust.blah.com, 
phone.blah.com]
    addresses: [10.22.11.1]
  version: 2

  Configure keepalived (again, a working config with IP addresses
  obfuscated)

  global_defs   # Block id
  {
  notification_email {
  sysadm...@blah.com
  }
  notification_email_from keepali...@system3.hq.blah.com
  smtp_server 10.22.11.7 # IP
  smtp_connect_timeout 30  # integer, seconds
  router_id system3  # string identifying the machine,
   # (doesn't 

[Ubuntu-ha] [Bug 1815101] Re: [master] Restarting systemd-networkd breaks keepalived, heartbeat, corosync, pacemaker (interface aliases are restarted)

2020-05-14 Thread Rafael David Tinoco
@napsty: the "workaround" (from your blog) is actually to use:

- ifupdown/bridge-utils/vlan/resolvconf for network setup   OR
- use systemd-networkd DIRECTLY with the KeepConfiguration= option in .network 
file

Just highlighting it here.

@ddstreet, you said you would try to come up with the netplan change for
KeepConfiguration. Did you have time to check on this ? (just checking).

Cheers o/

** Changed in: keepalived (Ubuntu)
 Assignee: Rafael David Tinoco (rafaeldtinoco) => (unassigned)

** Changed in: keepalived (Ubuntu Xenial)
     Assignee: Rafael David Tinoco (rafaeldtinoco) => (unassigned)

** Changed in: keepalived (Ubuntu Bionic)
 Assignee: Rafael David Tinoco (rafaeldtinoco) => (unassigned)

** Changed in: keepalived (Ubuntu Disco)
     Assignee: Rafael David Tinoco (rafaeldtinoco) => (unassigned)

** Changed in: keepalived (Ubuntu Eoan)
 Assignee: Rafael David Tinoco (rafaeldtinoco) => (unassigned)

** Changed in: systemd (Ubuntu)
     Assignee: Rafael David Tinoco (rafaeldtinoco) => (unassigned)

** Changed in: systemd (Ubuntu Xenial)
 Assignee: Rafael David Tinoco (rafaeldtinoco) => (unassigned)

** Changed in: systemd (Ubuntu Bionic)
     Assignee: Rafael David Tinoco (rafaeldtinoco) => (unassigned)

** Changed in: systemd (Ubuntu Disco)
 Assignee: Rafael David Tinoco (rafaeldtinoco) => (unassigned)

** Changed in: systemd (Ubuntu Eoan)
 Assignee: Rafael David Tinoco (rafaeldtinoco) => (unassigned)

** Changed in: netplan
 Assignee: Rafael David Tinoco (rafaeldtinoco) => (unassigned)

** No longer affects: keepalived (Ubuntu Eoan)

** No longer affects: keepalived (Ubuntu Disco)

** Also affects: heartbeat (Ubuntu Focal)
   Importance: Undecided
   Status: New

** Also affects: keepalived (Ubuntu Focal)
   Importance: Undecided
   Status: New

** Also affects: systemd (Ubuntu Focal)
   Importance: Undecided
   Status: New

** Changed in: systemd (Ubuntu Focal)
   Status: New => Fix Released

** Changed in: keepalived (Ubuntu Focal)
   Status: New => Confirmed

** No longer affects: heartbeat (Ubuntu Focal)

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to keepalived in Ubuntu.
https://bugs.launchpad.net/bugs/1815101

Title:
  [master] Restarting systemd-networkd breaks keepalived, heartbeat,
  corosync, pacemaker (interface aliases are restarted)

Status in netplan:
  Confirmed
Status in heartbeat package in Ubuntu:
  Won't Fix
Status in keepalived package in Ubuntu:
  In Progress
Status in systemd package in Ubuntu:
  In Progress
Status in keepalived source package in Xenial:
  Confirmed
Status in systemd source package in Xenial:
  Confirmed
Status in keepalived source package in Bionic:
  Confirmed
Status in systemd source package in Bionic:
  Confirmed
Status in systemd source package in Disco:
  Won't Fix
Status in systemd source package in Eoan:
  Fix Released
Status in keepalived source package in Focal:
  Confirmed
Status in systemd source package in Focal:
  Fix Released

Bug description:
  [impact]

  - ALL related HA software has a small problem if interfaces are being
  managed by systemd-networkd: nic restarts/reconfigs are always going
  to wipe all interfaces aliases when HA software is not expecting it to
  (no coordination between them.

  - keepalived, smb ctdb, pacemaker, all suffer from this. Pacemaker is
  smarter in this case because it has a service monitor that will
  restart the virtual IP resource, in affected node & nic, before
  considering a real failure, but other HA service might consider a real
  failure when it is not.

  [test case]

  - comment #14 is a full test case: to have 3 node pacemaker, in that
  example, and cause a networkd service restart: it will trigger a
  failure for the virtual IP resource monitor.

  - other example is given in the original description for keepalived.
  both suffer from the same issue (and other HA softwares as well).

  [regression potential]

  - this backports KeepConfiguration parameter, which adds some
  significant complexity to networkd's configuration and behavior, which
  could lead to regressions in correctly configuring the network at
  networkd start, or incorrectly maintaining configuration at networkd
  restart, or losing network state at networkd stop.

  - Any regressions are most likely to occur during networkd start,
  restart, or stop, and most likely to involve missing or incorrect ip
  address(es).

  - the change is based in upstream patches adding the exact feature we
  needed to fix this issue & it will be integrated with a netplan change
  to add the needed stanza to systemd nic configuration file
  (KeepConfiguration=)

  [other info]

  original description:
  ---

  Configure netplan for interfaces, for example (a working config with
  IP addresses obfuscated)

  network:
  ethernets:
  eth0:
   

[Ubuntu-ha] [Bug 1874719] Re: Focal deploy creates a 'node1' node

2020-04-24 Thread Rafael David Tinoco
** Changed in: pacemaker (Ubuntu)
   Status: New => Invalid

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1874719

Title:
  Focal deploy creates a 'node1' node

Status in OpenStack hacluster charm:
  New
Status in pacemaker package in Ubuntu:
  Invalid

Bug description:
  Testing of masakari on focal zaza tests failed because the test checks
  that all pacemaker nodes are online. This check failed due the
  appearance of a new node called 'node1' which was marked as offline. I
  don't know where that node came from or what is supposed to represent
  but it seems like an unwanted change in behaviour.

To manage notifications about this bug go to:
https://bugs.launchpad.net/charm-hacluster/+bug/1874719/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1874719] Re: Focal deploy creates a 'node1' node

2020-04-24 Thread Rafael David Tinoco
Here you will find the autopkgtests for pacemaker:

http://autopkgtest.ubuntu.com/packages/p/pacemaker/focal/amd64

the installation ends with a corosync node online:

Cluster Summary:
  * Stack: corosync
  * Current DC: node1 (version 2.0.3-4b1f869f0f) - partition with quorum
  * Last updated: Fri Apr 17 22:55:34 2020
  * Last change:  Fri Apr 17 22:55:22 2020 by hacluster via crmd on node1
  * 1 node configured
  * 0 resource instances configured

Node List:
  * Online: [ node1 ]

Active Resources:
  * No active resources

pacemakerPASS
pacemakerPASS

So, for some reason, this localhost corosync installation made the
service to be unavailable in your cause. I wonder if you have more logs
so we can check what happened for it not to be online.

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1874719

Title:
  Focal deploy creates a 'node1' node

Status in OpenStack hacluster charm:
  New
Status in pacemaker package in Ubuntu:
  New

Bug description:
  Testing of masakari on focal zaza tests failed because the test checks
  that all pacemaker nodes are online. This check failed due the
  appearance of a new node called 'node1' which was marked as offline. I
  don't know where that node came from or what is supposed to represent
  but it seems like an unwanted change in behaviour.

To manage notifications about this bug go to:
https://bugs.launchpad.net/charm-hacluster/+bug/1874719/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1874719] Re: Focal deploy creates a 'node1' node

2020-04-24 Thread Rafael David Tinoco
** Changed in: pacemaker (Ubuntu)
 Assignee: (unassigned) => Rafael David Tinoco (rafaeldtinoco)

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1874719

Title:
  Focal deploy creates a 'node1' node

Status in OpenStack hacluster charm:
  New
Status in pacemaker package in Ubuntu:
  New

Bug description:
  Testing of masakari on focal zaza tests failed because the test checks
  that all pacemaker nodes are online. This check failed due the
  appearance of a new node called 'node1' which was marked as offline. I
  don't know where that node came from or what is supposed to represent
  but it seems like an unwanted change in behaviour.

To manage notifications about this bug go to:
https://bugs.launchpad.net/charm-hacluster/+bug/1874719/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1866119] Re: [bionic] fence_scsi not working properly with Pacemaker 1.1.18-2ubuntu1.1

2020-04-16 Thread Rafael David Tinoco
Okay, I had verified this from the day in it landed in -proposed. It is
working as expected (https://discourse.ubuntu.com/t/ubuntu-high-
availability-corosync-pacemaker-shared-disk-environments/). I'm marking
this as verification-done as it has stayed in -proposed for sometime now
and no bad feedback was given from those who were asked to test it.

** Tags removed: block-proposed-bionic verification-needed 
verification-needed-bionic
** Tags added: verification-done verification-done-bionic

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1866119

Title:
  [bionic] fence_scsi not working properly with Pacemaker
  1.1.18-2ubuntu1.1

Status in pacemaker package in Ubuntu:
  Fix Released
Status in pacemaker source package in Bionic:
  Fix Committed

Bug description:
  OBS: This bug was originally into LP: #1865523 but it was split.

   SRU: pacemaker

  [Impact]

   * fence_scsi is not currently working in a share disk environment

   * all clusters relying in fence_scsi and/or fence_scsi + watchdog
  won't be able to start the fencing agents OR, in worst case scenarios,
  the fence_scsi agent might start but won't make scsi reservations in
  the shared scsi disk.

   * this bug is taking care of pacemaker 1.1.18 issues with fence_scsi,
  since the later was fixed at LP: #1865523.

  [Test Case]

   * having a 3-node setup, nodes called "clubionic01, clubionic02,
  clubionic03", with a shared scsi disk (fully supporting persistent
  reservations) /dev/sda, with corosync and pacemaker operational and
  running, one might try:

  rafaeldtinoco@clubionic01:~$ crm configure
  crm(live)configure# property stonith-enabled=on
  crm(live)configure# property stonith-action=off
  crm(live)configure# property no-quorum-policy=stop
  crm(live)configure# property have-watchdog=true
  crm(live)configure# commit
  crm(live)configure# end
  crm(live)# end

  rafaeldtinoco@clubionic01:~$ crm configure primitive fence_clubionic \
  stonith:fence_scsi params \
  pcmk_host_list="clubionic01 clubionic02 clubionic03" \
  devices="/dev/sda" \
  meta provides=unfencing

  And see the following errors:

  Failed Actions:
  * fence_clubionic_start_0 on clubionic02 'unknown error' (1): call=6, 
status=Error, exitreason='',
  last-rc-change='Wed Mar  4 19:53:12 2020', queued=0ms, exec=1105ms
  * fence_clubionic_start_0 on clubionic03 'unknown error' (1): call=6, 
status=Error, exitreason='',
  last-rc-change='Wed Mar  4 19:53:13 2020', queued=0ms, exec=1109ms
  * fence_clubionic_start_0 on clubionic01 'unknown error' (1): call=6, 
status=Error, exitreason='',
  last-rc-change='Wed Mar  4 19:53:11 2020', queued=0ms, exec=1108ms

  and corosync.log will show:

  warning: unpack_rsc_op_failure: Processing failed op start for
  fence_clubionic on clubionic01: unknown error (1)

  [Regression Potential]

   * LP: #1865523 shows fence_scsi fully operational after SRU for that
  bug is done.

   * LP: #1865523 used pacemaker 1.1.19 (vanilla) in order to fix
  fence_scsi.

   * There are changes to: cluster resource manager daemon, local
  resource manager daemon and police engine. From all the changes, the
  police engine fix is the biggest, but still not big for a SRU. This
  could cause police engine, thus cluster decisions, to mal function.

   * All patches are based in upstream fixes made right after
  Pacemaker-1.1.18, used by Ubuntu Bionic and were tested with
  fence_scsi to make sure it fixed the issues.

  [Other Info]

   * Original Description:

  Trying to setup a cluster with an iscsi shared disk, using fence_scsi
  as the fencing mechanism, I realized that fence_scsi is not working in
  Ubuntu Bionic. I first thought it was related to Azure environment
  (LP: #1864419), where I was trying this environment, but then, trying
  locally, I figured out that somehow pacemaker 1.1.18 is not fencing
  the shared scsi disk properly.

  Note: I was able to "backport" vanilla 1.1.19 from upstream and
  fence_scsi worked. I have then tried 1.1.18 without all quilt patches
  and it didnt work as well. I think that bisecting 1.1.18 <-> 1.1.19
  might tell us which commit has fixed the behaviour needed by the
  fence_scsi agent.

  (k)rafaeldtinoco@clubionic01:~$ crm conf show
  node 1: clubionic01.private
  node 2: clubionic02.private
  node 3: clubionic03.private
  primitive fence_clubionic stonith:fence_scsi \
  params pcmk_host_list="10.250.3.10 10.250.3.11 10.250.3.12" 
devices="/dev/sda" \
  meta provides=unfencing
  property cib-bootstrap-options: \
  have-watchdog=false \
  dc-version=1.1.18-2b07d5c5a9 \
  cluster-infrastructure=corosync \
  cluster-name=clubionic \
  stonith-enabled=on \
  stonith-action=off \
  no-quorum-policy=stop \
  symmetric-cluster=true

  

  

[Ubuntu-ha] [Bug 1871353] Re: after split brain detection heartbeart service stops unexpectedly

2020-04-09 Thread Rafael David Tinoco
** Changed in: heartbeat (Ubuntu)
   Status: New => Triaged

** Changed in: heartbeat (Ubuntu)
   Importance: Undecided => Medium

** Also affects: heartbeat (Ubuntu Xenial)
   Importance: Undecided
   Status: New

** Changed in: heartbeat (Ubuntu Xenial)
   Status: New => Triaged

** Changed in: heartbeat (Ubuntu Xenial)
   Importance: Undecided => Medium

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to heartbeat in Ubuntu.
https://bugs.launchpad.net/bugs/1871353

Title:
  after split brain detection heartbeart service stops unexpectedly

Status in heartbeat package in Ubuntu:
  Triaged
Status in heartbeat source package in Xenial:
  Triaged

Bug description:
  Ubuntu Release
  --
  Description:  Ubuntu 16.04.6 LTS
  Release:  16.04

  
  heartbeat Package
  -
  heartbeat:
Installed: 1:3.0.6-2
Candidate: 1:3.0.6-2
Version table:
   *** 1:3.0.6-2 500
  500 http://mirror.hetzner.de/ubuntu/packages xenial/main amd64 
Packages
  100 /var/lib/dpkg/status

  
  Scenario Description
  
  When heartbeat detects a split brain scenario a restart is triggered by the 
heartbeat service itself.

  
  Expectation
  ---
  The heartbeat service should be running after a split brain scenario was 
detected.

  
  Obervation
  --
  The heartbeat service was no longer running after the split brain scenario.

  
  Further Investigation
  -
  systemd detects the restart and executes the ExecStop action.
  This behaviour is documented at 
https://www.freedesktop.org/software/systemd/man/systemd.service.html#ExecStop=

  This problem most likely arises because of the automatically generated
  systemctl service file (converted from the init.d script by systemd-
  sysv-generator).

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/heartbeat/+bug/1871353/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1828228] Re: corosync fails to start in unprivileged containers - autopkgtest failure

2020-04-06 Thread Rafael David Tinoco
** Changed in: pcs (Ubuntu)
 Assignee: Rafael David Tinoco (rafaeldtinoco) => (unassigned)

** Changed in: pacemaker (Ubuntu)
 Assignee: Rafael David Tinoco (rafaeldtinoco) => (unassigned)

** Changed in: corosync (Ubuntu)
 Assignee: Rafael David Tinoco (rafaeldtinoco) => (unassigned)

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to corosync in Ubuntu.
https://bugs.launchpad.net/bugs/1828228

Title:
  corosync fails to start in unprivileged containers - autopkgtest
  failure

Status in Auto Package Testing:
  Invalid
Status in corosync package in Ubuntu:
  Fix Released
Status in pacemaker package in Ubuntu:
  Fix Released
Status in pcs package in Ubuntu:
  Fix Released

Bug description:
  Currently pacemaker v2 fails to start in armhf containers (and by
  extension corosync too).

  I found that it is reproducible locally, and that I had to bump a few
  limits to get it going.

  Specifically I did:

  1) bump memlock limits
  2) bump rmem_max limits

  = 1) Bump memlock limits =

  I have no idea, which one of these finally worked, and/or is
  sufficient. A bit of a whack-a-mole.

  cat >>/etc/security/limits.conf </etc/systemd/system/snap.lxd.daemon.service.d/override.conf <&1 | grep sockop
  [pid   447] setsockopt(12, SOL_SOCKET, SO_RCVBUF, [8388608], 4) = 0
  [pid   447] getsockopt(12, SOL_SOCKET, SO_RCVBUF, [425984], [4]) = 0
  [pid   447] setsockopt(12, SOL_SOCKET, SO_RCVBUFFORCE, [8388608], 4) = -1 
EPERM (Operation not permitted)

  Bumped mem_max using:
  sudo sysctl -w net.core.wmem_max=8388608
  sudo sysctl -w net.core.rmem_max=8388608

  (Not sure if the desired sized depends on the machine/container I am
  running on)

  
  Can we check the values for above things on our armhf containers and/or bump 
them? or like can we mark pacemaker v2.0 autopkgtest as ignored on armhf?

To manage notifications about this bug go to:
https://bugs.launchpad.net/auto-package-testing/+bug/1828228/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1870235] Re: [focal] pacemaker v2.0.3 last upstream fixes

2020-04-06 Thread Rafael David Tinoco
PPA successfully built pacemaker fixes:

https://launchpad.net/~ubuntu-server-ha/+archive/ubuntu/staging

and the merge request has been +1.

** Description changed:

- This bug backport/cherrypick fixes released after 2.0.3 release.
  
- The correct patches are going to be defined within this bug comments.
+ [Impact]
+ 
+ * Idea here is latest release stabilization with backports of fixes
+ released after the current version.
+ 
+ * Since this is a LTS, it is worth applying "straightforward" fixes
+ before the final release.
+ 
+ [Test Case]
+ 
+ * https://discourse.ubuntu.com/t/ubuntu-high-availability-corosync-
+ pacemaker-shared-disk-environments/
+ 
+ [Regression Potential]
+ 
+ Areas of potential regression: stonith, pengine, crmservice:
+ 
+ * stonith: I have regression tests and they look good.
+ 
+ * pengine: could not identify bad decisions when fencing resources but
+ haven't explored in detail as there are many possible combinations and
+ decisions that pengine can take -> changes are very minimal here.
+ 
+ * crmservice: pacemaker tests don't show regressions, resource agents
+ start/stop/probe/migrate correctly afaik so the fix looks good.
+ 
+ [Other Info]
+ 
+ * This bug backport/cherrypick fixes released after 2.0.3 release.
+ 
+ * The correct patches are going to be defined within this bug comments.

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1870235

Title:
  [focal] pacemaker v2.0.3 last upstream fixes

Status in pacemaker package in Ubuntu:
  In Progress

Bug description:
  
  [Impact]

  * Idea here is latest release stabilization with backports of fixes
  released after the current version.

  * Since this is a LTS, it is worth applying "straightforward" fixes
  before the final release.

  [Test Case]

  * https://discourse.ubuntu.com/t/ubuntu-high-availability-corosync-
  pacemaker-shared-disk-environments/

  [Regression Potential]

  Areas of potential regression: stonith, pengine, crmservice:

  * stonith: I have regression tests and they look good.

  * pengine: could not identify bad decisions when fencing resources but
  haven't explored in detail as there are many possible combinations and
  decisions that pengine can take -> changes are very minimal here.

  * crmservice: pacemaker tests don't show regressions, resource agents
  start/stop/probe/migrate correctly afaik so the fix looks good.

  [Other Info]

  * This bug backport/cherrypick fixes released after 2.0.3 release.

  * The correct patches are going to be defined within this bug
  comments.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1870235/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1869751] Re: [focal] pacemaker FTBFS because of deprecated ftime()

2020-04-06 Thread Rafael David Tinoco
Thanks @doko,

This was fixed by:

# ftime causes deprecation errors now, but the clock_gettime replacement is
# disabled by default in 2.0.3. Enable it.
CPPFLAGS = -UPCMK_TIME_EMERGENCY_CGT

in

pacemaker (2.0.3-3ubuntu2) focal; urgency=medium

  * Forcibly switch from ftime to clock_gettime, since building with ftime now
results in deprecation errors

 -- William Grant   Sat, 04 Apr 2020 19:15:34 +1100

I'll rebase my merge request.

** Merge proposal unlinked:
   
https://code.launchpad.net/~rafaeldtinoco/ubuntu/+source/pacemaker/+git/pacemaker/+merge/381684

** Changed in: pacemaker (Ubuntu)
 Assignee: Rafael David Tinoco (rafaeldtinoco) => (unassigned)

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1869751

Title:
  [focal] pacemaker FTBFS because of deprecated ftime()

Status in pacemaker package in Ubuntu:
  Fix Committed

Bug description:
  https://people.canonical.com/~doko/ftbfs-report/test-rebuild-20200327
  -focal-focal.html

  shows that pacemaker started to be FTBFS because of:

  """
  gcc -DHAVE_CONFIG_H -I. -I../../include  -DSUPPORT_REMOTE -I../../include 
-I../../include -I../../libltdl -I../../libltdl -DPCMK_TIME_EMERGENCY_CGT 
-Wdate-time -D_FORTIFY_SOURCE=2 -I/usr/include/glib-2.0 
-I/usr/lib/i386-linux-gnu/glib-2.0/include -I/usr/include/libxml2 
-I/usr/include/heartbeat -I/usr/include/dbus-1.0 
-I/usr/lib/i386-linux-gnu/dbus-1.0/include -fPIE -g -O2 
-fdebug-prefix-map=/<>=. -fstack-protector-strong -Wformat 
-Werror=format-security   -ggdb  -fgnu89-inline -Wall -Waggregate-return 
-Wbad-function-cast -Wcast-align -Wdeclaration-after-statement -Wendif-labels 
-Wfloat-equal -Wformat-security -Wmissing-prototypes -Wmissing-declarations 
-Wnested-externs -Wno-long-long -Wno-strict-aliasing -Wpointer-arith 
-Wwrite-strings -Wunused-but-set-variable -Wformat=2 -Wformat-nonliteral 
-fstack-protector-strong -Werror -c -o pacemaker_remoted-pacemaker-execd.o 
`test -f 'pacemaker-execd.c' || echo './'`pacemaker-execd.c
  execd_commands.c: In function ‘stonith_recurring_op_helper’:
  execd_commands.c:257:5: error: ‘ftime’ is deprecated 
[-Werror=deprecated-declarations]
257 | ftime(>t_queue);
| ^
  In file included from execd_commands.c:23:
  /usr/include/i386-linux-gnu/sys/timeb.h:39:12: note: declared here
 39 | extern int ftime (struct timeb *__timebuf)
|^
  execd_commands.c: In function ‘schedule_lrmd_cmd’:
  execd_commands.c:389:5: error: ‘ftime’ is deprecated 
[-Werror=deprecated-declarations]
389 | ftime(>t_queue);
| ^
  """

  And man page shows:

  
  SYNOPSIS
 #include 

 int ftime(struct timeb *tp);

  DESCRIPTION
 NOTE: This function is deprecated, and will be removed in a future 
version of the GNU C library.  Use clock_gettime(2) instead.

  
  I'll fix this together with other fixes, opening this bug to track the issue.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1869751/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1870235] Re: [focal] pacemaker v2.0.3 last upstream fixes

2020-04-06 Thread Rafael David Tinoco
** Description changed:

- Together with fix for:
- 
- https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1869751 (FTBFS)
- 
- This bug will also backport/cherrypick fixes released after 2.0.3
- release.
+ This bug backport/cherrypick fixes released after 2.0.3 release.
  
  The correct patches are going to be defined within this bug comments.

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1870235

Title:
  [focal] pacemaker v2.0.3 last upstream fixes

Status in pacemaker package in Ubuntu:
  In Progress

Bug description:
  This bug backport/cherrypick fixes released after 2.0.3 release.

  The correct patches are going to be defined within this bug comments.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1870235/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1870235] Re: [focal] pacemaker v2.0.3 last upstream fixes

2020-04-03 Thread Rafael David Tinoco
There is an on-going merge request for this bug.

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1870235

Title:
  [focal] pacemaker v2.0.3 last upstream fixes

Status in pacemaker package in Ubuntu:
  In Progress

Bug description:
  Together with fix for:

  https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1869751
  (FTBFS)

  This bug will also backport/cherrypick fixes released after 2.0.3
  release.

  The correct patches are going to be defined within this bug comments.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1870235/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1869751] Re: [focal] pacemaker FTBFS because of deprecated ftime()

2020-04-03 Thread Rafael David Tinoco
Check:

https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1870235

for the fix (same merge request).

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1869751

Title:
  [focal] pacemaker FTBFS because of deprecated ftime()

Status in pacemaker package in Ubuntu:
  Confirmed

Bug description:
  https://people.canonical.com/~doko/ftbfs-report/test-rebuild-20200327
  -focal-focal.html

  shows that pacemaker started to be FTBFS because of:

  """
  gcc -DHAVE_CONFIG_H -I. -I../../include  -DSUPPORT_REMOTE -I../../include 
-I../../include -I../../libltdl -I../../libltdl -DPCMK_TIME_EMERGENCY_CGT 
-Wdate-time -D_FORTIFY_SOURCE=2 -I/usr/include/glib-2.0 
-I/usr/lib/i386-linux-gnu/glib-2.0/include -I/usr/include/libxml2 
-I/usr/include/heartbeat -I/usr/include/dbus-1.0 
-I/usr/lib/i386-linux-gnu/dbus-1.0/include -fPIE -g -O2 
-fdebug-prefix-map=/<>=. -fstack-protector-strong -Wformat 
-Werror=format-security   -ggdb  -fgnu89-inline -Wall -Waggregate-return 
-Wbad-function-cast -Wcast-align -Wdeclaration-after-statement -Wendif-labels 
-Wfloat-equal -Wformat-security -Wmissing-prototypes -Wmissing-declarations 
-Wnested-externs -Wno-long-long -Wno-strict-aliasing -Wpointer-arith 
-Wwrite-strings -Wunused-but-set-variable -Wformat=2 -Wformat-nonliteral 
-fstack-protector-strong -Werror -c -o pacemaker_remoted-pacemaker-execd.o 
`test -f 'pacemaker-execd.c' || echo './'`pacemaker-execd.c
  execd_commands.c: In function ‘stonith_recurring_op_helper’:
  execd_commands.c:257:5: error: ‘ftime’ is deprecated 
[-Werror=deprecated-declarations]
257 | ftime(>t_queue);
| ^
  In file included from execd_commands.c:23:
  /usr/include/i386-linux-gnu/sys/timeb.h:39:12: note: declared here
 39 | extern int ftime (struct timeb *__timebuf)
|^
  execd_commands.c: In function ‘schedule_lrmd_cmd’:
  execd_commands.c:389:5: error: ‘ftime’ is deprecated 
[-Werror=deprecated-declarations]
389 | ftime(>t_queue);
| ^
  """

  And man page shows:

  
  SYNOPSIS
 #include 

 int ftime(struct timeb *tp);

  DESCRIPTION
 NOTE: This function is deprecated, and will be removed in a future 
version of the GNU C library.  Use clock_gettime(2) instead.

  
  I'll fix this together with other fixes, opening this bug to track the issue.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1869751/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1870235] Re: [focal] pacemaker v2.0.3 last upstream fixes

2020-04-03 Thread Rafael David Tinoco
Cherry-picks that seem to make sense after 2.0.3 release are:

Fix: tools: Fix definition of curses_indented_printf.
Fix: iso8601: Fix crm_time_parse_offset() to parse offset with plus sign.
Log: libcrmcommon: correctly raise detail log line length
Low: libcrmservice: handle child wait errors appropriately
Log: libcrmservice: improve messages when waiting for child process
Refactor: libcrmservice: isolate SIGCHLD handling
Fix: libstonithd, tools: Fix arguments to stonith-event.
Fix: libpengine: Options should be unsigned int, not long.
Fix: libstonithd: Some validate arguments should be non-const.
Fix: scheduler: make sure cluster-wide maintenance-mode=true overrides 
per-resource settings
Fix: tools: Correct the crm_mon man page.

There are too many fixes but mostly related to in-development
refactorings so it's hard to distinguish if a fix is fixing something
that exists already in v2.0.3.

I preferred to be conservative and pick the ones that made more sense -
core parts - and were solved right after the release (and of course
didn't introduce new features and major code refactoring).

A list of all commits after v2.0.3 can be seen at:

https://paste.ubuntu.com/p/pGzWc2qZpB/

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1870235

Title:
  [focal] pacemaker v2.0.3 last upstream fixes

Status in pacemaker package in Ubuntu:
  In Progress

Bug description:
  Together with fix for:

  https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1869751
  (FTBFS)

  This bug will also backport/cherrypick fixes released after 2.0.3
  release.

  The correct patches are going to be defined within this bug comments.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1870235/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1870235] Re: [focal] pacemaker v2.0.3 last upstream fixes

2020-04-03 Thread Rafael David Tinoco
** Description changed:

  Together with fix for:
  
  https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1869751 (FTBFS)
  
  This bug will also backport/cherrypick fixes released after 2.0.3
- release:
- 
- c1bc0a7f4 Merge pull request #2023 from kgaillot/fixes
- f00d1120f Merge pull request #2005 from clumens/tag-fixes
- 5c543bc73 Fix: xml: Add a new version of the tags schema.
- d9109fb83 Fix: tools: Apply various fixes to the crm_diff man page.
- 269aea44b Merge pull request #2017 from kgaillot/fixes
- ad532cfff Low: fencer,libstonithd: fix Y2038 issue in fence history
- ef2037126 Doc: libcrmcommon: fix doc string mistake
- f0be039f9 Merge pull request #2015 from clumens/more-blanks
- e2215c722 Fix: tools: One more blank line related fix for crm_mon.
- ef1208197 Merge pull request #2010 from kgaillot/fixes
- 4c997e61a Merge pull request #2006 from HideoYamauchi/trac4454-fix
- 7cd3ecfd3 Low: crm: Fix argument order in declaration
- 78db2071e Merge pull request #1999 from clumens/crm_mon-fixes
- 5e3b8cdba Merge pull request #2004 from jnpkrn/build-gcc10-s390x
- 4267f7fc4 Doc: Pacemaker Development: crm_str easy to forget, like its prime 
use
- 9acf1f9b3 Build: fix compilation -Werror compilation issue with GCC 10 with 
s390x
- 0625eb1a2 Merge pull request #1996 from kgaillot/fixes
- 902e79f43 Feature: Rework how a prefix for bans is given in crm_mon.
- 4860c572c Fix: tools: Move print_neg_location_prefix into options.
- 398611256 Fix: tools: Use formatted output for more errors in crm_mon.
- a290911ef Merge pull request #1989 from kgaillot/fixes
- 7599e6d98 Build: configure: fix syntax error
- 68bd2b4da Merge pull request #1953 from 
jnpkrn/build-fix-inkscape-1.0-beta-compat
- df6c286d9 Merge pull request #1985 from kgaillot/rhbz1760669
- cdf84f849 Merge pull request #1984 from kgaillot/fixes
- 38ef24516 Merge pull request #1979 from kgaillot/fixes
- 403c5b1f8 Merge pull request #1978 from jnpkrn/gcc10-fno-common-fix
- ea5b06fae Merge pull request #1976 from jnpkrn/fix-process-will-not-die
- e0ca2a0b7 Fix: libpacemaker: Fix setting options when calling the history opt.
- 3fa025298 Fix: libpacemaker: Fix handling of some operation return values.
- 529ff2851 Merge pull request #1971 from clumens/crm_mon-fixes
- 5b98dd71c Fix: tools: Re-enable CGI output from crm_mon.
- ed6972a74 Fix: tools: Maintain a list of extra HTML headers.
- c2e5d2d02 Fix: tools: Correct sec vs. msec discrepancy in stonith_admin.
- bc6f54495 Merge pull request #1968 from kgaillot/fixes
- f11fedd40 Refactor: libcrmcommon,libpacemaker,tools: convert 
pcmk__output_new() to standard return codes
- 264afd724 Merge pull request #1958 from kgaillot/fixes
- aa779a942 Merge pull request #1960 from yoshfuji/fix-iso8601-parser-20191217
- 33e28dcad Merge pull request #1955 from kgaillot/fixes
- d671faa22 Merge pull request #1954 from 
aleksei-burlakov/fix-libpe_status-maintenance-mode
- 47ecd21b9 Build: fix unability to build with Inkscape 1.0 beta version(s)
- 4f5207a28 Fix: tools: Correct the crm_mon man page.
+ release.
  
  The correct patches are going to be defined within this bug comments.

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1870235

Title:
  [focal] pacemaker v2.0.3 last upstream fixes

Status in pacemaker package in Ubuntu:
  In Progress

Bug description:
  Together with fix for:

  https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1869751
  (FTBFS)

  This bug will also backport/cherrypick fixes released after 2.0.3
  release.

  The correct patches are going to be defined within this bug comments.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1870235/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1638210] Re: saidar segmentation fault

2020-04-02 Thread Rafael David Tinoco
There are other comments saying that the issue does not seem to be
solved:

https://github.com/libstatgrab/libstatgrab/issues/102#issuecomment-509715203

I'm marking this as incomplete since we still need a dump to do anything
further.


** Also affects: libstatgrab (Ubuntu Eoan)
   Importance: Undecided
   Status: New

** Also affects: libstatgrab (Ubuntu Xenial)
   Importance: Undecided
   Status: New

** Also affects: libstatgrab (Ubuntu Bionic)
   Importance: Undecided
   Status: New

** Changed in: libstatgrab (Ubuntu)
   Status: Confirmed => Fix Released

** Changed in: libstatgrab (Ubuntu Xenial)
   Status: New => Incomplete

** Changed in: libstatgrab (Ubuntu Bionic)
   Status: New => Incomplete

** Changed in: libstatgrab (Ubuntu Eoan)
   Status: New => Incomplete

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to libstatgrab in Ubuntu.
https://bugs.launchpad.net/bugs/1638210

Title:
  saidar segmentation fault

Status in libstatgrab package in Ubuntu:
  Fix Released
Status in libstatgrab source package in Xenial:
  Incomplete
Status in libstatgrab source package in Bionic:
  Incomplete
Status in libstatgrab source package in Eoan:
  Incomplete
Status in libstatgrab package in Debian:
  Fix Released

Bug description:
  Erreur de segmentation (core dumped)

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/libstatgrab/+bug/1638210/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1869751] Re: [focal] pacemaker FTBFS because of deprecated ftime()

2020-04-02 Thread Rafael David Tinoco
While some may think the fix for this is to suppress deprecated warning
OR to "fix" ftime()... the actual fix for this is to start using
MONOTONIC clock for the related functions:

(c)rafaeldtinoco@clufocaldev:~/.../sources/ubuntu/pacemaker$ git diff HEAD
diff --git a/configure.ac b/configure.ac
index a4c8ae93..b7edea57 100644
--- a/configure.ac
+++ b/configure.ac
@@ -981,7 +981,7 @@ AC_CHECK_DECLS([CLOCK_MONOTONIC], [], [], [[
 # will only be applied in 2.0.3 release and will become opt-out since
 # (which hopefully explains the name of the macro as it will start to
 # make more sense then, and the continuity is important)
-CPPFLAGS="-DPCMK_TIME_EMERGENCY_CGT $CPPFLAGS"
+# CPPFLAGS="-DPCMK_TIME_EMERGENCY_CGT $CPPFLAGS"


 dnl 

This solves the building issue.

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1869751

Title:
  [focal] pacemaker FTBFS because of deprecated ftime()

Status in pacemaker package in Ubuntu:
  Confirmed

Bug description:
  https://people.canonical.com/~doko/ftbfs-report/test-rebuild-20200327
  -focal-focal.html

  shows that pacemaker started to be FTBFS because of:

  """
  gcc -DHAVE_CONFIG_H -I. -I../../include  -DSUPPORT_REMOTE -I../../include 
-I../../include -I../../libltdl -I../../libltdl -DPCMK_TIME_EMERGENCY_CGT 
-Wdate-time -D_FORTIFY_SOURCE=2 -I/usr/include/glib-2.0 
-I/usr/lib/i386-linux-gnu/glib-2.0/include -I/usr/include/libxml2 
-I/usr/include/heartbeat -I/usr/include/dbus-1.0 
-I/usr/lib/i386-linux-gnu/dbus-1.0/include -fPIE -g -O2 
-fdebug-prefix-map=/<>=. -fstack-protector-strong -Wformat 
-Werror=format-security   -ggdb  -fgnu89-inline -Wall -Waggregate-return 
-Wbad-function-cast -Wcast-align -Wdeclaration-after-statement -Wendif-labels 
-Wfloat-equal -Wformat-security -Wmissing-prototypes -Wmissing-declarations 
-Wnested-externs -Wno-long-long -Wno-strict-aliasing -Wpointer-arith 
-Wwrite-strings -Wunused-but-set-variable -Wformat=2 -Wformat-nonliteral 
-fstack-protector-strong -Werror -c -o pacemaker_remoted-pacemaker-execd.o 
`test -f 'pacemaker-execd.c' || echo './'`pacemaker-execd.c
  execd_commands.c: In function ‘stonith_recurring_op_helper’:
  execd_commands.c:257:5: error: ‘ftime’ is deprecated 
[-Werror=deprecated-declarations]
257 | ftime(>t_queue);
| ^
  In file included from execd_commands.c:23:
  /usr/include/i386-linux-gnu/sys/timeb.h:39:12: note: declared here
 39 | extern int ftime (struct timeb *__timebuf)
|^
  execd_commands.c: In function ‘schedule_lrmd_cmd’:
  execd_commands.c:389:5: error: ‘ftime’ is deprecated 
[-Werror=deprecated-declarations]
389 | ftime(>t_queue);
| ^
  """

  And man page shows:

  
  SYNOPSIS
 #include 

 int ftime(struct timeb *tp);

  DESCRIPTION
 NOTE: This function is deprecated, and will be removed in a future 
version of the GNU C library.  Use clock_gettime(2) instead.

  
  I'll fix this together with other fixes, opening this bug to track the issue.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1869751/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1869751] Re: [focal] pacemaker FTBFS because of deprecated ftime()

2020-04-02 Thread Rafael David Tinoco
commit 4b8b84cce1fd57eec1f47ca44780d60c148b399d
Author: Jan Pokorný 
Date:   Fri Nov 15 16:06:57 2019 +0100

Build: restore buildability in the face of obsolete ftime(3)

Since the usage of ftime(3) is purely optional and since
clock_gettime(3) is mandated with POSIX 2001, we can simply
look at whether CLOCK_MONOTONIC is defined to be used as an
identifier for the particular clock (kind exactly suitable
for this context).  But due to being late in the release cycle,
such a change is kept as opt-in (see configure.ac comment for
details), and for compatibility stability concerns[*], also
dropping some old surrounding cruft is delayed.

In this form, constitutes first step out of two to restore
out-of-the-box buildability with recent enough glibc, again,
refer to configure.ac comment.

References:

https://sourceware.org/git/?p=glibc.git;a=commit;h=2b5fea833bcd0f651579afd16ed7842770ecbae1

https://src.fedoraproject.org/rpms/glibc/c/ebf75398f06dd27357d8a5321e8e5959633b8182?branch=master
(for a Fedora Rawhide follow-the-upstream update that led to this
discovery)

[*] in case you opt-in (as described), CLOCK_MONOTONIC gets detected
in time.h positively but it starts choking for whatever reason in
the actual build or even in run-time, you can rescind that,
or you can shortcut any checking and refrain from any time period
measurements altogher with something like:

  env \
ac_cv_header_sys_timeb_h=no
ac_cv_have_decl_CLOCK_MONOTONIC=no \
./configure

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1869751

Title:
  [focal] pacemaker FTBFS because of deprecated ftime()

Status in pacemaker package in Ubuntu:
  Confirmed

Bug description:
  https://people.canonical.com/~doko/ftbfs-report/test-rebuild-20200327
  -focal-focal.html

  shows that pacemaker started to be FTBFS because of:

  """
  gcc -DHAVE_CONFIG_H -I. -I../../include  -DSUPPORT_REMOTE -I../../include 
-I../../include -I../../libltdl -I../../libltdl -DPCMK_TIME_EMERGENCY_CGT 
-Wdate-time -D_FORTIFY_SOURCE=2 -I/usr/include/glib-2.0 
-I/usr/lib/i386-linux-gnu/glib-2.0/include -I/usr/include/libxml2 
-I/usr/include/heartbeat -I/usr/include/dbus-1.0 
-I/usr/lib/i386-linux-gnu/dbus-1.0/include -fPIE -g -O2 
-fdebug-prefix-map=/<>=. -fstack-protector-strong -Wformat 
-Werror=format-security   -ggdb  -fgnu89-inline -Wall -Waggregate-return 
-Wbad-function-cast -Wcast-align -Wdeclaration-after-statement -Wendif-labels 
-Wfloat-equal -Wformat-security -Wmissing-prototypes -Wmissing-declarations 
-Wnested-externs -Wno-long-long -Wno-strict-aliasing -Wpointer-arith 
-Wwrite-strings -Wunused-but-set-variable -Wformat=2 -Wformat-nonliteral 
-fstack-protector-strong -Werror -c -o pacemaker_remoted-pacemaker-execd.o 
`test -f 'pacemaker-execd.c' || echo './'`pacemaker-execd.c
  execd_commands.c: In function ‘stonith_recurring_op_helper’:
  execd_commands.c:257:5: error: ‘ftime’ is deprecated 
[-Werror=deprecated-declarations]
257 | ftime(>t_queue);
| ^
  In file included from execd_commands.c:23:
  /usr/include/i386-linux-gnu/sys/timeb.h:39:12: note: declared here
 39 | extern int ftime (struct timeb *__timebuf)
|^
  execd_commands.c: In function ‘schedule_lrmd_cmd’:
  execd_commands.c:389:5: error: ‘ftime’ is deprecated 
[-Werror=deprecated-declarations]
389 | ftime(>t_queue);
| ^
  """

  And man page shows:

  
  SYNOPSIS
 #include 

 int ftime(struct timeb *tp);

  DESCRIPTION
 NOTE: This function is deprecated, and will be removed in a future 
version of the GNU C library.  Use clock_gettime(2) instead.

  
  I'll fix this together with other fixes, opening this bug to track the issue.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1869751/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1870235] [NEW] [focal] pacemaker v2.0.3 last upstream fixes

2020-04-01 Thread Rafael David Tinoco
Public bug reported:

Together with fix for:

https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1869751 (FTBFS)

This bug will also backport/cherrypick fixes released after 2.0.3
release:

c1bc0a7f4 Merge pull request #2023 from kgaillot/fixes
f00d1120f Merge pull request #2005 from clumens/tag-fixes
5c543bc73 Fix: xml: Add a new version of the tags schema.
d9109fb83 Fix: tools: Apply various fixes to the crm_diff man page.
269aea44b Merge pull request #2017 from kgaillot/fixes
ad532cfff Low: fencer,libstonithd: fix Y2038 issue in fence history
ef2037126 Doc: libcrmcommon: fix doc string mistake
f0be039f9 Merge pull request #2015 from clumens/more-blanks
e2215c722 Fix: tools: One more blank line related fix for crm_mon.
ef1208197 Merge pull request #2010 from kgaillot/fixes
4c997e61a Merge pull request #2006 from HideoYamauchi/trac4454-fix
7cd3ecfd3 Low: crm: Fix argument order in declaration
78db2071e Merge pull request #1999 from clumens/crm_mon-fixes
5e3b8cdba Merge pull request #2004 from jnpkrn/build-gcc10-s390x
4267f7fc4 Doc: Pacemaker Development: crm_str easy to forget, like its prime use
9acf1f9b3 Build: fix compilation -Werror compilation issue with GCC 10 with 
s390x
0625eb1a2 Merge pull request #1996 from kgaillot/fixes
902e79f43 Feature: Rework how a prefix for bans is given in crm_mon.
4860c572c Fix: tools: Move print_neg_location_prefix into options.
398611256 Fix: tools: Use formatted output for more errors in crm_mon.
a290911ef Merge pull request #1989 from kgaillot/fixes
7599e6d98 Build: configure: fix syntax error
68bd2b4da Merge pull request #1953 from 
jnpkrn/build-fix-inkscape-1.0-beta-compat
df6c286d9 Merge pull request #1985 from kgaillot/rhbz1760669
cdf84f849 Merge pull request #1984 from kgaillot/fixes
38ef24516 Merge pull request #1979 from kgaillot/fixes
403c5b1f8 Merge pull request #1978 from jnpkrn/gcc10-fno-common-fix
ea5b06fae Merge pull request #1976 from jnpkrn/fix-process-will-not-die
e0ca2a0b7 Fix: libpacemaker: Fix setting options when calling the history opt.
3fa025298 Fix: libpacemaker: Fix handling of some operation return values.
529ff2851 Merge pull request #1971 from clumens/crm_mon-fixes
5b98dd71c Fix: tools: Re-enable CGI output from crm_mon.
ed6972a74 Fix: tools: Maintain a list of extra HTML headers.
c2e5d2d02 Fix: tools: Correct sec vs. msec discrepancy in stonith_admin.
bc6f54495 Merge pull request #1968 from kgaillot/fixes
f11fedd40 Refactor: libcrmcommon,libpacemaker,tools: convert pcmk__output_new() 
to standard return codes
264afd724 Merge pull request #1958 from kgaillot/fixes
aa779a942 Merge pull request #1960 from yoshfuji/fix-iso8601-parser-20191217
33e28dcad Merge pull request #1955 from kgaillot/fixes
d671faa22 Merge pull request #1954 from 
aleksei-burlakov/fix-libpe_status-maintenance-mode
47ecd21b9 Build: fix unability to build with Inkscape 1.0 beta version(s)
4f5207a28 Fix: tools: Correct the crm_mon man page.

The correct patches are going to be defined within this bug comments.

** Affects: pacemaker (Ubuntu)
 Importance: High
 Assignee: Rafael David Tinoco (rafaeldtinoco)
 Status: In Progress

** Changed in: pacemaker (Ubuntu)
   Status: New => In Progress

** Changed in: pacemaker (Ubuntu)
   Importance: Undecided => High

** Changed in: pacemaker (Ubuntu)
 Assignee: (unassigned) => Rafael David Tinoco (rafaeldtinoco)

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1870235

Title:
  [focal] pacemaker v2.0.3 last upstream fixes

Status in pacemaker package in Ubuntu:
  In Progress

Bug description:
  Together with fix for:

  https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1869751
  (FTBFS)

  This bug will also backport/cherrypick fixes released after 2.0.3
  release:

  c1bc0a7f4 Merge pull request #2023 from kgaillot/fixes
  f00d1120f Merge pull request #2005 from clumens/tag-fixes
  5c543bc73 Fix: xml: Add a new version of the tags schema.
  d9109fb83 Fix: tools: Apply various fixes to the crm_diff man page.
  269aea44b Merge pull request #2017 from kgaillot/fixes
  ad532cfff Low: fencer,libstonithd: fix Y2038 issue in fence history
  ef2037126 Doc: libcrmcommon: fix doc string mistake
  f0be039f9 Merge pull request #2015 from clumens/more-blanks
  e2215c722 Fix: tools: One more blank line related fix for crm_mon.
  ef1208197 Merge pull request #2010 from kgaillot/fixes
  4c997e61a Merge pull request #2006 from HideoYamauchi/trac4454-fix
  7cd3ecfd3 Low: crm: Fix argument order in declaration
  78db2071e Merge pull request #1999 from clumens/crm_mon-fixes
  5e3b8cdba Merge pull request #2004 from jnpkrn/build-gcc10-s390x
  4267f7fc4 Doc: Pacemaker Development: crm_str easy to forget, like its prime 
use
  9acf1f9b3 Build: fix compilation -Werror compilation issue with GCC 10 with 
s390x
  0625eb1a2 Merge pull request #1996 from kgaillot/fixes
  902e79f43 Featur

[Ubuntu-ha] [Bug 1828223] Re: pacemaker v2 appears to be missing log directory

2020-04-01 Thread Rafael David Tinoco
Thank you for taking the time to report this bug. In an effort to keep an
up-to-date and valid list of bugs to work on, I have reviewed this report
to verify it still requires effort and occurs on an Ubuntu release in
standard support, and it does not.

It is unfortunate that we were unable to resolve this defect, however
there appears to be no further action possible at this time. I am
therefore moving the bug to 'Incomplete'. If you disagree or have
new information, we would be grateful if you could please add a comment 
stating why and then change the status of the bug to 'New'.

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1828223

Title:
  pacemaker v2 appears to be missing log directory

Status in pacemaker package in Ubuntu:
  Fix Released
Status in pacemaker source package in Disco:
  Won't Fix
Status in pacemaker source package in Eoan:
  Fix Released
Status in pacemaker package in Debian:
  Fix Released

Bug description:
  # journalctl -u pacemaker | grep 'logging to'
  May 08 12:30:49 nice-mako pacemakerd[325]:  error: Directory 
'/var/log/pacemaker' does not exist: logging to 
'/var/log/pacemaker/pacemaker.log' is disabled
  May 08 12:30:49 nice-mako pacemaker-fenced[329]:  error: Directory 
'/var/log/pacemaker' does not exist: logging to 
'/var/log/pacemaker/pacemaker.log' is disabled
  May 08 12:30:49 nice-mako pacemaker-based[328]:  error: Directory 
'/var/log/pacemaker' does not exist: logging to 
'/var/log/pacemaker/pacemaker.log' is disabled
  May 08 12:30:49 nice-mako pacemaker-schedulerd[332]:  error: Directory 
'/var/log/pacemaker' does not exist: logging to 
'/var/log/pacemaker/pacemaker.log' is disabled
  May 08 12:30:49 nice-mako pacemaker-attrd[331]:  error: Directory 
'/var/log/pacemaker' does not exist: logging to 
'/var/log/pacemaker/pacemaker.log' is disabled
  May 08 12:30:49 nice-mako pacemaker-execd[330]:  error: Directory 
'/var/log/pacemaker' does not exist: logging to 
'/var/log/pacemaker/pacemaker.log' is disabled
  May 08 12:30:49 nice-mako pacemaker-controld[333]:  error: Directory 
'/var/log/pacemaker' does not exist: logging to 
'/var/log/pacemaker/pacemaker.log' is disabled

  I guess pacemaker v2 should either be configured to log to journal by
  default, or that directory needs to be created? or configuration
  changed to log to /var/log/pacemaker.log?

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1828223/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1828223] Re: pacemaker v2 appears to be missing log directory

2020-04-01 Thread Rafael David Tinoco
As Paride mentioned, this is already fixed in Eoan and Focal and was an
issue for Disco.

** Also affects: pacemaker (Ubuntu Focal)
   Importance: Medium
 Assignee: Rafael David Tinoco (rafaeldtinoco)
   Status: Confirmed

** Also affects: pacemaker (Ubuntu Eoan)
   Importance: Undecided
   Status: New

** Also affects: pacemaker (Ubuntu Disco)
   Importance: Undecided
   Status: New

** Changed in: pacemaker (Ubuntu Eoan)
   Status: New => Fix Released

** No longer affects: pacemaker (Ubuntu Focal)

** Changed in: pacemaker (Ubuntu)
   Status: Confirmed => Triaged

** Changed in: pacemaker (Ubuntu)
   Status: Triaged => Won't Fix

** Changed in: pacemaker (Ubuntu Disco)
   Status: New => Won't Fix

** Changed in: pacemaker (Ubuntu)
   Status: Won't Fix => Fix Released

** Changed in: pacemaker (Ubuntu)
 Assignee: Rafael David Tinoco (rafaeldtinoco) => (unassigned)

** Changed in: pacemaker (Ubuntu)
   Importance: Medium => Undecided

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1828223

Title:
  pacemaker v2 appears to be missing log directory

Status in pacemaker package in Ubuntu:
  Fix Released
Status in pacemaker source package in Disco:
  Won't Fix
Status in pacemaker source package in Eoan:
  Fix Released
Status in pacemaker package in Debian:
  Fix Released

Bug description:
  # journalctl -u pacemaker | grep 'logging to'
  May 08 12:30:49 nice-mako pacemakerd[325]:  error: Directory 
'/var/log/pacemaker' does not exist: logging to 
'/var/log/pacemaker/pacemaker.log' is disabled
  May 08 12:30:49 nice-mako pacemaker-fenced[329]:  error: Directory 
'/var/log/pacemaker' does not exist: logging to 
'/var/log/pacemaker/pacemaker.log' is disabled
  May 08 12:30:49 nice-mako pacemaker-based[328]:  error: Directory 
'/var/log/pacemaker' does not exist: logging to 
'/var/log/pacemaker/pacemaker.log' is disabled
  May 08 12:30:49 nice-mako pacemaker-schedulerd[332]:  error: Directory 
'/var/log/pacemaker' does not exist: logging to 
'/var/log/pacemaker/pacemaker.log' is disabled
  May 08 12:30:49 nice-mako pacemaker-attrd[331]:  error: Directory 
'/var/log/pacemaker' does not exist: logging to 
'/var/log/pacemaker/pacemaker.log' is disabled
  May 08 12:30:49 nice-mako pacemaker-execd[330]:  error: Directory 
'/var/log/pacemaker' does not exist: logging to 
'/var/log/pacemaker/pacemaker.log' is disabled
  May 08 12:30:49 nice-mako pacemaker-controld[333]:  error: Directory 
'/var/log/pacemaker' does not exist: logging to 
'/var/log/pacemaker/pacemaker.log' is disabled

  I guess pacemaker v2 should either be configured to log to journal by
  default, or that directory needs to be created? or configuration
  changed to log to /var/log/pacemaker.log?

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1828223/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1825992] Re: Upgrade to version 2.0 to satisfy pcs dependency

2020-04-01 Thread Rafael David Tinoco
Thank you for taking the time to report this bug. In an effort to keep
an up-to-date and valid list of bugs to work on, I have reviewed this
report to verify it still requires effort and occurs on an Ubuntu
release in standard support, and it does not.

I have marked Disco was won't fix as it was EOL in this last January
2020.

Currently, for focal we have:

- PCS: 0.10.4-2ubuntu1 (Debian: 0.10.4-2, Upstream: 0.10.5)
- COROSYNC: 3.0.3-2ubuntu1 (Debian: 3.0.3-2 +fixes, Upstream: v3.0.3)
- PACEMAKER: 2.0.3-3ubuntu1 (Debian: 2.0.3-3, Upstream: 2.0.3)

AND we're still targeting crmsh tool as the "official" way of setting up
PACEMAKER. We are targeting PCS to replace CRMSH in a few releases from
now (crmsh is currently back in [main] while PCS is still in
[universe]).

With that said I'm considering this bug as fix released since Eoan.


** No longer affects: pacemaker (Ubuntu Focal)

** Changed in: pacemaker (Ubuntu)
   Importance: Medium => Undecided

** Changed in: pacemaker (Ubuntu Eoan)
   Importance: Medium => Undecided

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1825992

Title:
  Upgrade to version 2.0 to satisfy pcs dependency

Status in pacemaker package in Ubuntu:
  Fix Released
Status in pacemaker source package in Disco:
  Won't Fix
Status in pacemaker source package in Eoan:
  Fix Released

Bug description:
  Is there a way we could upgrade the version to 2.0? The pcs package
  requires a version of pacemaker that is greater than or equal to 2.0,
  and there is already a debian version packaged for v2. Installing the
  Ubuntu package for pcs will delete the pacemaker package as there is
  no version of pacemaker that is greater than 2.0.

  I can help with build/testing and packaging for 2.0 based on the
  existing debian deb if needed.

  https://pkgs.org/download/pacemaker

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1825992/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1825992] Re: Upgrade to version 2.0 to satisfy pcs dependency

2020-04-01 Thread Rafael David Tinoco
** Also affects: pacemaker (Ubuntu Focal)
   Importance: Medium
 Assignee: Rafael David Tinoco (rafaeldtinoco)
   Status: Confirmed

** Changed in: pacemaker (Ubuntu Disco)
   Status: Confirmed => Won't Fix

** Changed in: pacemaker (Ubuntu Disco)
 Assignee: Rafael David Tinoco (rafaeldtinoco) => (unassigned)

** Changed in: pacemaker (Ubuntu Eoan)
   Status: Confirmed => Fix Released

** Changed in: pacemaker (Ubuntu Focal)
   Status: Confirmed => Fix Released

** Changed in: pacemaker (Ubuntu Focal)
     Assignee: Rafael David Tinoco (rafaeldtinoco) => (unassigned)

** Changed in: pacemaker (Ubuntu Eoan)
 Assignee: Rafael David Tinoco (rafaeldtinoco) => (unassigned)

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1825992

Title:
  Upgrade to version 2.0 to satisfy pcs dependency

Status in pacemaker package in Ubuntu:
  Fix Released
Status in pacemaker source package in Disco:
  Won't Fix
Status in pacemaker source package in Eoan:
  Fix Released

Bug description:
  Is there a way we could upgrade the version to 2.0? The pcs package
  requires a version of pacemaker that is greater than or equal to 2.0,
  and there is already a debian version packaged for v2. Installing the
  Ubuntu package for pcs will delete the pacemaker package as there is
  no version of pacemaker that is greater than 2.0.

  I can help with build/testing and packaging for 2.0 based on the
  existing debian deb if needed.

  https://pkgs.org/download/pacemaker

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1825992/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1627083] Re: ipt_CLUSTERIP is deprecated and it will removed soon, use xt_cluster instead

2020-04-01 Thread Rafael David Tinoco
commit 92c49b6f2847546f3f938b10a2a97021774f0be3
Author: Jan Pokorný 
Date:   Wed Dec 4 14:36:59 2019 +0100

IPaddr2: ipt_CLUSTERIP "iptables" extension not "nft" backend
compatible

Reference:
https://lists.clusterlabs.org/pipermail/users/2019-December/026674.html
(thread also sketches a future ambition for a [presumably, to revert
the habit of a functional overloading] separate agent to use
"xt_cluster" extension/cluster match).

Signed-off-by: Jan Pokorný 

---

It is a well known upstream decision and it has been recently documented
in v4.5.0 of "resource-agents".

The following resource-agent description is about to get added to Focal
when FFe:

https://bugs.launchpad.net/ubuntu/+source/resource-agents/+bug/1866383

is accepted and merge is finished.

** Also affects: resource-agents (Ubuntu)
   Importance: Undecided
   Status: New

** No longer affects: strongswan (Ubuntu)

** No longer affects: pacemaker (Ubuntu)

** Changed in: resource-agents (Ubuntu)
   Importance: Undecided => Wishlist

** Changed in: resource-agents (Ubuntu)
   Status: New => Confirmed

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1627083

Title:
  ipt_CLUSTERIP is deprecated and it will removed soon, use xt_cluster
  instead

Status in resource-agents package in Ubuntu:
  Confirmed

Bug description:
  pacemaker still uses iptable's "CLUSTERIP" -- and dmesg shows a
  deprecation warning:

  [   15.027333] ipt_CLUSTERIP: ClusterIP Version 0.8 loaded successfully
  [   15.027464] ipt_CLUSTERIP: ipt_CLUSTERIP is deprecated and it will removed 
soon, use xt_cluster instead

  ~# iptables -L
  Chain INPUT (policy ACCEPT)
  target prot opt source   destination 
  CLUSTERIP  all  --  anywhere proxy.charite.de CLUSTERIP 
hashmode=sourceip-sourceport clustermac=EF:EE:6B:F9:7B:67 total_nodes=4 
local_node=2 hash_init=0

  ProblemType: Bug
  DistroRelease: Ubuntu 16.04
  Package: pacemaker 1.1.14-2ubuntu1.1
  ProcVersionSignature: Ubuntu 4.4.0-38.57-generic 4.4.19
  Uname: Linux 4.4.0-38-generic x86_64
  ApportVersion: 2.20.1-0ubuntu2.1
  Architecture: amd64
  Date: Fri Sep 23 17:26:01 2016
  InstallationDate: Installed on 2014-08-19 (766 days ago)
  InstallationMedia: Ubuntu-Server 14.04.1 LTS "Trusty Tahr" - Release amd64 
(20140722.3)
  SourcePackage: pacemaker
  UpgradeStatus: Upgraded to xenial on 2016-09-22 (1 days ago)

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/resource-agents/+bug/1627083/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1490727] Re: "Invalid IPC credentials" after corosync, pacemaker service restarts

2020-04-01 Thread Rafael David Tinoco
For pacemaker, this is a duplicate of:

https://bugs.launchpad.net/bugs/1439649

So, instead of marking it as duplicate, since it had charm work, I'm
removing pacemaker from the bug and leaving this comment:

https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1439649/comments/20

"""
>From Corosync 2.4.1 Release Notes:

This release contains fix for one regression and few more smaller fixes.

"""
During 2.3.6 development the bug which is causing pacemaker to not work after 
corosync configuration file is reloaded happened. Solution is ether to use this 
fixed version (recommended) or as a quick workaround (for users who wants to 
stay on 2.3.6 or 2.4.0) is to create file pacemaker (file name can be 
arbitrary) in /etc/corosync/uidgid.d directory with following content (you can 
also put same stanza into /etc/corosync/corosync.conf):

uidgid {
gid: haclient
}
"""

Anyone relying in Trusty or Xenial corosync:

 corosync | 2.3.3-1ubuntu1 | trusty
 corosync | 2.3.3-1ubuntu4 | trusty-updates
 corosync | 2.3.5-3ubuntu1 | xenial
 corosync | 2.3.5-3ubuntu2.3 | xenial-security
 corosync | 2.3.5-3ubuntu2.3 | xenial-updates

should apply the mitigation above, like discovered previously by
commenters of this bug.

Note: Trusty is already EOS so I'm marking it as "won't fix".

Xenial should include the mitigation in a SRU.
"""

** No longer affects: pacemaker (Ubuntu)

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1490727

Title:
  "Invalid IPC credentials" after corosync, pacemaker service restarts

Status in Landscape Server:
  Fix Released
Status in Landscape Server 15.07 series:
  Fix Released
Status in Landscape Server cisco-odl series:
  Fix Released
Status in hacluster package in Juju Charms Collection:
  Fix Released

Bug description:
  Followup from 
https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1439649,
  see there comments #14, #15 as it _maybe_ related to missing uidgid ACLs for
  hacluster:haclient (as apparently presented by pacemaker).

  FYI you can find relevant IPC resources with:
  $ find /run/shm -user hacluster -group haclient -ls

To manage notifications about this bug go to:
https://bugs.launchpad.net/landscape/+bug/1490727/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1439649] Re: Pacemaker unable to communicate with corosync on restart under lxc

2020-04-01 Thread Rafael David Tinoco
** Also affects: pacemaker (Ubuntu Xenial)
   Importance: Undecided
   Status: New

** Also affects: pacemaker (Ubuntu Trusty)
   Importance: Undecided
   Status: New

** Changed in: pacemaker (Ubuntu)
   Status: Confirmed => Fix Released

** Also affects: pacemaker (Ubuntu Bionic)
   Importance: Undecided
   Status: New

** Changed in: pacemaker (Ubuntu Bionic)
   Status: New => Fix Released

** Changed in: pacemaker (Ubuntu Xenial)
   Status: New => Confirmed

** Changed in: pacemaker (Ubuntu Trusty)
   Status: New => Confirmed

** Changed in: pacemaker (Ubuntu Xenial)
   Importance: Undecided => High

** Changed in: pacemaker (Ubuntu Trusty)
   Importance: Undecided => Medium

** Changed in: pacemaker (Ubuntu Xenial)
   Importance: High => Medium

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1439649

Title:
  Pacemaker unable to communicate with corosync on restart under lxc

Status in pacemaker package in Ubuntu:
  Fix Released
Status in pacemaker source package in Trusty:
  Won't Fix
Status in pacemaker source package in Xenial:
  Confirmed
Status in pacemaker source package in Bionic:
  Fix Released

Bug description:
  We've seen this a few times with three node clusters, all running in
  LXC containers; pacemaker fails to restart correctly as it can't
  communicate with corosync, resulting in a down cluster.  Rebooting the
  containers resolves the issue, so suspect some sort of bad state
  either in corosync or pacemaker.

  Apr  2 11:41:32 juju-machine-4-lxc-4 pacemakerd[1033741]:   notice: 
mcp_read_config: Configured corosync to accept connections from group 115: 
Library error (2)
  Apr  2 11:41:32 juju-machine-4-lxc-4 pacemakerd[1033741]:   notice: main: 
Starting Pacemaker 1.1.10 (Build: 42f2063):  generated-manpages agent-manpages 
ncurses libqb-logging libqb-ipc lha-fencing upstart nagios  heartbeat 
corosync-native snmp libesmtp
  Apr  2 11:41:32 juju-machine-4-lxc-4 pacemakerd[1033741]:   notice: 
cluster_connect_quorum: Quorum acquired
  Apr  2 11:41:32 juju-machine-4-lxc-4 pacemakerd[1033741]:   notice: 
corosync_node_name: Unable to get node name for nodeid 1000
  Apr  2 11:41:32 juju-machine-4-lxc-4 pacemakerd[1033741]:   notice: 
corosync_node_name: Unable to get node name for nodeid 1001
  Apr  2 11:41:32 juju-machine-4-lxc-4 pacemakerd[1033741]:   notice: 
corosync_node_name: Unable to get node name for nodeid 1003
  Apr  2 11:41:32 juju-machine-4-lxc-4 pacemakerd[1033741]:   notice: 
corosync_node_name: Unable to get node name for nodeid 1001
  Apr  2 11:41:32 juju-machine-4-lxc-4 pacemakerd[1033741]:   notice: 
get_node_name: Defaulting to uname -n for the local corosync node name
  Apr  2 11:41:32 juju-machine-4-lxc-4 pacemakerd[1033741]:   notice: 
crm_update_peer_state: pcmk_quorum_notification: Node 
juju-machine-4-lxc-4[1001] - state is now member (was (null))
  Apr  2 11:41:32 juju-machine-4-lxc-4 pacemakerd[1033741]:   notice: 
corosync_node_name: Unable to get node name for nodeid 1003
  Apr  2 11:41:32 juju-machine-4-lxc-4 pacemakerd[1033741]:   notice: 
crm_update_peer_state: pcmk_quorum_notification: Node (null)[1003] - state is 
now member (was (null))
  Apr  2 11:41:32 juju-machine-4-lxc-4 crmd[1033748]:   notice: main: CRM Git 
Version: 42f2063
  Apr  2 11:41:32 juju-machine-4-lxc-4 stonith-ng[1033744]:   notice: 
crm_cluster_connect: Connecting to cluster infrastructure: corosync
  Apr  2 11:41:32 juju-machine-4-lxc-4 stonith-ng[1033744]:   notice: 
corosync_node_name: Unable to get node name for nodeid 1001
  Apr  2 11:41:32 juju-machine-4-lxc-4 stonith-ng[1033744]:   notice: 
get_node_name: Defaulting to uname -n for the local corosync node name
  Apr  2 11:41:32 juju-machine-4-lxc-4 attrd[1033746]:   notice: 
crm_cluster_connect: Connecting to cluster infrastructure: corosync
  Apr  2 11:41:32 juju-machine-4-lxc-4 corosync[1033732]:  [MAIN  ] Denied 
connection attempt from 109:115
  Apr  2 11:41:32 juju-machine-4-lxc-4 corosync[1033732]:  [QB] Invalid IPC 
credentials (1033732-1033746).
  Apr  2 11:41:32 juju-machine-4-lxc-4 attrd[1033746]:error: 
cluster_connect_cpg: Could not connect to the Cluster Process Group API: 11
  Apr  2 11:41:32 juju-machine-4-lxc-4 attrd[1033746]:error: main: HA 
Signon failed
  Apr  2 11:41:32 juju-machine-4-lxc-4 attrd[1033746]:error: main: Aborting 
startup
  Apr  2 11:41:32 juju-machine-4-lxc-4 pacemakerd[1033741]:error: 
pcmk_child_exit: Child process attrd (1033746) exited: Network is down (100)
  Apr  2 11:41:32 juju-machine-4-lxc-4 pacemakerd[1033741]:  warning: 
pcmk_child_exit: Pacemaker child process attrd no longer wishes to be 
respawned. Shutting ourselves down.
  Apr  2 11:41:32 juju-machine-4-lxc-4 pacemakerd[1033741]:   notice: 
pcmk_shutdown_worker: Shuting down Pacemaker
  Apr  2 11:41:32 juju-machine-4-lxc-4 pacemakerd[1033741]:   notice: 

[Ubuntu-ha] [Bug 1870069] Re: pacemaker ftbfs in focal

2020-04-01 Thread Rafael David Tinoco
*** This bug is a duplicate of bug 1869751 ***
https://bugs.launchpad.net/bugs/1869751

** This bug has been marked a duplicate of bug 1869751
   [focal] pacemaker FTBFS because of deprecated ftime()

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1870069

Title:
  pacemaker ftbfs in focal

Status in pacemaker package in Ubuntu:
  Confirmed

Bug description:
  seen in the second focal test rebuild
  
https://launchpadlibrarian.net/471775801/buildlog_ubuntu-focal-amd64.pacemaker_2.0.3-3ubuntu1_BUILDING.txt.gz

  execd_commands.c: In function ‘stonith_recurring_op_helper’:
  execd_commands.c:257:5: error: ‘ftime’ is deprecated 
[-Werror=deprecated-declarations]
257 | ftime(>t_queue);
| ^
  In file included from execd_commands.c:23:
  /usr/include/x86_64-linux-gnu/sys/timeb.h:39:12: note: declared here
 39 | extern int ftime (struct timeb *__timebuf)
|^
  execd_commands.c: In function ‘schedule_lrmd_cmd’:
  execd_commands.c:389:5: error: ‘ftime’ is deprecated 
[-Werror=deprecated-declarations]
389 | ftime(>t_queue);
| ^
  In file included from execd_commands.c:23:
  /usr/include/x86_64-linux-gnu/sys/timeb.h:39:12: note: declared here
 39 | extern int ftime (struct timeb *__timebuf)
|^
  execd_commands.c: In function ‘time_diff_ms’:
  execd_commands.c:480:9: error: ‘ftime’ is deprecated 
[-Werror=deprecated-declarations]
480 | ftime(_now);
| ^
  In file included from execd_commands.c:23:
  /usr/include/x86_64-linux-gnu/sys/timeb.h:39:12: note: declared here
 39 | extern int ftime (struct timeb *__timebuf)
|^
  execd_commands.c: In function ‘action_complete’:
  execd_commands.c:896:9: error: ‘ftime’ is deprecated 
[-Werror=deprecated-declarations]
896 | ftime(>t_rcchange);
| ^
  In file included from execd_commands.c:23:
  /usr/include/x86_64-linux-gnu/sys/timeb.h:39:12: note: declared here
 39 | extern int ftime (struct timeb *__timebuf)
|^
  execd_commands.c: In function ‘lrmd_rsc_execute’:
  execd_commands.c:1428:13: error: ‘ftime’ is deprecated 
[-Werror=deprecated-declarations]
   1428 | ftime(>t_first_run);
| ^
  In file included from execd_commands.c:23:
  /usr/include/x86_64-linux-gnu/sys/timeb.h:39:12: note: declared here
 39 | extern int ftime (struct timeb *__timebuf)
|^
  execd_commands.c:1430:9: error: ‘ftime’ is deprecated 
[-Werror=deprecated-declarations]
   1430 | ftime(>t_run);
| ^
  In file included from execd_commands.c:23:
  /usr/include/x86_64-linux-gnu/sys/timeb.h:39:12: note: declared here
 39 | extern int ftime (struct timeb *__timebuf)
|^

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1870069/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1869751] [NEW] [focal] pacemaker FTBFS because of deprecated ftime()

2020-03-30 Thread Rafael David Tinoco
Public bug reported:

https://people.canonical.com/~doko/ftbfs-report/test-rebuild-20200327
-focal-focal.html

shows that pacemaker started to be FTBFS because of:

"""
gcc -DHAVE_CONFIG_H -I. -I../../include  -DSUPPORT_REMOTE -I../../include 
-I../../include -I../../libltdl -I../../libltdl -DPCMK_TIME_EMERGENCY_CGT 
-Wdate-time -D_FORTIFY_SOURCE=2 -I/usr/include/glib-2.0 
-I/usr/lib/i386-linux-gnu/glib-2.0/include -I/usr/include/libxml2 
-I/usr/include/heartbeat -I/usr/include/dbus-1.0 
-I/usr/lib/i386-linux-gnu/dbus-1.0/include -fPIE -g -O2 
-fdebug-prefix-map=/<>=. -fstack-protector-strong -Wformat 
-Werror=format-security   -ggdb  -fgnu89-inline -Wall -Waggregate-return 
-Wbad-function-cast -Wcast-align -Wdeclaration-after-statement -Wendif-labels 
-Wfloat-equal -Wformat-security -Wmissing-prototypes -Wmissing-declarations 
-Wnested-externs -Wno-long-long -Wno-strict-aliasing -Wpointer-arith 
-Wwrite-strings -Wunused-but-set-variable -Wformat=2 -Wformat-nonliteral 
-fstack-protector-strong -Werror -c -o pacemaker_remoted-pacemaker-execd.o 
`test -f 'pacemaker-execd.c' || echo './'`pacemaker-execd.c
execd_commands.c: In function ‘stonith_recurring_op_helper’:
execd_commands.c:257:5: error: ‘ftime’ is deprecated 
[-Werror=deprecated-declarations]
  257 | ftime(>t_queue);
  | ^
In file included from execd_commands.c:23:
/usr/include/i386-linux-gnu/sys/timeb.h:39:12: note: declared here
   39 | extern int ftime (struct timeb *__timebuf)
  |^
execd_commands.c: In function ‘schedule_lrmd_cmd’:
execd_commands.c:389:5: error: ‘ftime’ is deprecated 
[-Werror=deprecated-declarations]
  389 | ftime(>t_queue);
  | ^
"""

And man page shows:


SYNOPSIS
   #include 

   int ftime(struct timeb *tp);

DESCRIPTION
   NOTE: This function is deprecated, and will be removed in a future 
version of the GNU C library.  Use clock_gettime(2) instead.


I'll fix this together with other fixes, opening this bug to track the issue.

** Affects: pacemaker (Ubuntu)
     Importance: High
 Assignee: Rafael David Tinoco (rafaeldtinoco)
 Status: Confirmed

** Changed in: pacemaker (Ubuntu)
   Status: New => Won't Fix

** Changed in: pacemaker (Ubuntu)
   Status: Won't Fix => Confirmed

** Changed in: pacemaker (Ubuntu)
   Importance: Undecided => High

** Changed in: pacemaker (Ubuntu)
 Assignee: (unassigned) => Rafael David Tinoco (rafaeldtinoco)

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1869751

Title:
  [focal] pacemaker FTBFS because of deprecated ftime()

Status in pacemaker package in Ubuntu:
  Confirmed

Bug description:
  https://people.canonical.com/~doko/ftbfs-report/test-rebuild-20200327
  -focal-focal.html

  shows that pacemaker started to be FTBFS because of:

  """
  gcc -DHAVE_CONFIG_H -I. -I../../include  -DSUPPORT_REMOTE -I../../include 
-I../../include -I../../libltdl -I../../libltdl -DPCMK_TIME_EMERGENCY_CGT 
-Wdate-time -D_FORTIFY_SOURCE=2 -I/usr/include/glib-2.0 
-I/usr/lib/i386-linux-gnu/glib-2.0/include -I/usr/include/libxml2 
-I/usr/include/heartbeat -I/usr/include/dbus-1.0 
-I/usr/lib/i386-linux-gnu/dbus-1.0/include -fPIE -g -O2 
-fdebug-prefix-map=/<>=. -fstack-protector-strong -Wformat 
-Werror=format-security   -ggdb  -fgnu89-inline -Wall -Waggregate-return 
-Wbad-function-cast -Wcast-align -Wdeclaration-after-statement -Wendif-labels 
-Wfloat-equal -Wformat-security -Wmissing-prototypes -Wmissing-declarations 
-Wnested-externs -Wno-long-long -Wno-strict-aliasing -Wpointer-arith 
-Wwrite-strings -Wunused-but-set-variable -Wformat=2 -Wformat-nonliteral 
-fstack-protector-strong -Werror -c -o pacemaker_remoted-pacemaker-execd.o 
`test -f 'pacemaker-execd.c' || echo './'`pacemaker-execd.c
  execd_commands.c: In function ‘stonith_recurring_op_helper’:
  execd_commands.c:257:5: error: ‘ftime’ is deprecated 
[-Werror=deprecated-declarations]
257 | ftime(>t_queue);
| ^
  In file included from execd_commands.c:23:
  /usr/include/i386-linux-gnu/sys/timeb.h:39:12: note: declared here
 39 | extern int ftime (struct timeb *__timebuf)
|^
  execd_commands.c: In function ‘schedule_lrmd_cmd’:
  execd_commands.c:389:5: error: ‘ftime’ is deprecated 
[-Werror=deprecated-declarations]
389 | ftime(>t_queue);
| ^
  """

  And man page shows:

  
  SYNOPSIS
 #include 

 int ftime(struct timeb *tp);

  DESCRIPTION
 NOTE: This function is deprecated, and will be removed in a future 
version of the GNU C library.  Use clock_gettime(2) instead.

  
  I'll fix this together with other fixes, opening this bug to track the issue.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+sourc

[Ubuntu-ha] [Bug 1425431] Re: Pacemaker [SIGSEGV - Segmentation violation] using with Heartbeat 3.x

2020-03-29 Thread Rafael David Tinoco
Thank you for taking the time to report this bug. In an effort to keep an
up-to-date and valid list of bugs to work on, I have reviewed this report
to verify it still requires effort and occurs on an Ubuntu release in
standard support, and it does not.

Judging by the existing comments, other related bugs and the pointed
upstream discussion thread, it appears that the wrong lrmd was being
set after upgrade and that led existing cluster to confusion when
connecting to the local resource manager daemon.

It is unfortunate that we were unable to resolve this defect, however
there appears to be no further action possible at this time. I am
therefore moving the bug to 'Incomplete'. If you disagree or have
new information, we would be grateful if you could please add a comment
stating why and then change the status of the bug to 'New'.

** Also affects: pacemaker (Ubuntu Trusty)
   Importance: Undecided
   Status: New

** Changed in: pacemaker (Ubuntu Trusty)
   Status: New => Triaged

** Changed in: pacemaker (Ubuntu)
   Status: Confirmed => Fix Released

** Changed in: pacemaker (Ubuntu Trusty)
   Status: Triaged => Incomplete

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1425431

Title:
  Pacemaker [SIGSEGV - Segmentation violation] using with Heartbeat 3.x

Status in pacemaker package in Ubuntu:
  Fix Released
Status in pacemaker source package in Trusty:
  Incomplete

Bug description:
  According to:
  http://oss.clusterlabs.org/pipermail/pacemaker/2013-November/020152.html

  and my test case using:

  Pacemaker - Version: 1.1.10+git20130802-1ubuntu2.3 - Architecture: armhf
  Hearbeat - Version: 1:3.0.5-3.2 - Architecture: armhf
  Cluster-Glue- Version: 1.0.11+hg2754-1.1build1 - Architecture: armhf

  
  I got (into syslog):

   ccm: [4356]: info: Hostname: 
   ccm: [4356]: info: Break tie for 2 nodes cluster
   ccm: [4356]: WARN: ccm_state_joined: received message with unknown cookie, 
just dropping
   ccm: [4356]: info: client (pid=4361) removed from ccm
   heartbeat: [4269]: WARN: Managed /usr/lib/heartbeat/crmd process 4361 killed 
by signal 11 [SIGSEGV - Segmentation violation].
   heartbeat: [4269]: ERROR: Client /usr/lib/heartbeat/crmd (pid=4361) killed 
by signal 11.
   heartbeat: [4269]: ERROR: Respawning client "/usr/lib/heartbeat/crmd":
   heartbeat: [4269]: info: Starting child client "/usr/lib/heartbeat/crmd" 
(109,114)
   heartbeat: [4427]: info: Starting "/usr/lib/heartbeat/crmd" as uid 109  gid 
114 (pid 4427)
   heartbeat: [4269]: info: the send queue length from heartbeat to client crmd 
is set to 1024
   heartbeat: [4269]: info: killing /usr/lib/heartbeat/crmd process group 4427 
with signal 15
   heartbeat: [4269]: info: Core process 4273 exited. 7 remaining
   heartbeat: [4269]: info: Core process 4274 exited. 6 remaining
   heartbeat: [4269]: info: Core process 4275 exited. 5 remaining
   heartbeat: [4269]: info: Core process 4276 exited. 4 remaining
   heartbeat: [4269]: info: Core process 4277 exited. 3 remaining
   heartbeat: [4269]: info: Core process 4278 exited. 2 remaining

  then node always reboot by crmd after some time.
  Using Heartbeat without Pacemaker (crm no into ha.cf) no problem occur.

  Note that /usr/lib/heartbeat/crmd are soft links to:
  lrwxrwxrwx  1 root root 18 Feb  5 19:51 /usr/lib/heartbeat/attrd -> 
../pacemaker/attrd
  lrwxrwxrwx  1 root root 16 Feb  5 19:51 /usr/lib/heartbeat/cib -> 
../pacemaker/cib
  lrwxrwxrwx  1 root root 17 Feb  5 19:51 /usr/lib/heartbeat/crmd -> 
../pacemaker/crmd
  lrwxrwxrwx  1 root root 20 Feb  5 19:51 /usr/lib/heartbeat/pengine -> 
../pacemaker/pengine
  lrwxrwxrwx  1 root root 21 Feb  5 19:51 /usr/lib/heartbeat/stonithd -> 
../pacemaker/stonithd

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1425431/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1412831] Re: crm_mon crashed with SIGABRT in g_return_if_fail_warning()

2020-03-29 Thread Rafael David Tinoco
** Also affects: pacemaker (Ubuntu Trusty)
   Importance: Undecided
   Status: New

** Changed in: pacemaker (Ubuntu Trusty)
   Status: New => Triaged

** Changed in: pacemaker (Ubuntu)
   Status: Confirmed => Fix Released

** Changed in: pacemaker (Ubuntu)
   Importance: Medium => Undecided

** Changed in: pacemaker (Ubuntu Trusty)
   Importance: Undecided => Medium

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1412831

Title:
  crm_mon crashed with SIGABRT in g_return_if_fail_warning()

Status in pacemaker package in Ubuntu:
  Fix Released
Status in pacemaker source package in Trusty:
  Triaged

Bug description:
  $ sudo crm_mon -1fA
  will generate a crash report.

  $ sudo crm_mon -1fA
  Last updated: Tue Jan 20 15:02:21 2015
  Last change: Tue Jan 20 14:42:22 2015 via crmd on juju-nobuto-machine-46
  Stack: corosync
  Current DC: juju-nobuto-machine-45 (1002) - partition with quorum
  Version: 1.1.10-42f2063
  3 Nodes configured
  4 Resources configured

  
  Online: [ juju-nobuto-machine-44 juju-nobuto-machine-45 
juju-nobuto-machine-46 ]

   Clone Set: cl_ping [ping]
   Started: [ juju-nobuto-machine-44 juju-nobuto-machine-45 
juju-nobuto-machine-46 ]
   Resource Group: grp_percona_cluster
   res_mysql_vip  (ocf::heartbeat:IPaddr2):   Started 
juju-nobuto-machine-44 

  Node Attributes:
  * Node juju-nobuto-machine-44:
  + pingd : 100   crm_glib_handler: Cannot 
wait on forked child 13386: No child processes (10)
  + pingd : 100   crm_glib_handler: Cannot 
wait on forked child 13388: No child processes (10)
  + pingd : 100   
  * Node juju-nobuto-machine-45:
  + pingd : 100   crm_glib_handler: Cannot 
wait on forked child 13390: No child processes (10)
  + pingd : 100   crm_glib_handler: Cannot 
wait on forked child 13392: No child processes (10)
  + pingd : 100   
  * Node juju-nobuto-machine-46:
  + pingd : 100   crm_glib_handler: Cannot 
wait on forked child 13394: No child processes (10)
  + pingd : 100   crm_glib_handler: Cannot 
wait on forked child 13396: No child processes (10)
  + pingd : 100   

  Migration summary:
  * Node juju-nobuto-machine-44: 
  * Node juju-nobuto-machine-46: 
  * Node juju-nobuto-machine-45: 

  $ sudo crm configure show
  node $id="1000" juju-nobuto-machine-44
  node $id="1001" juju-nobuto-machine-46
  node $id="1002" juju-nobuto-machine-45
  primitive ping ocf:pacemaker:ping \
  params host_list="10.5.0.1" multiplier="100" \
  op monitor interval="5s"
  primitive res_mysql_vip ocf:heartbeat:IPaddr2 \
  params ip="10.5.100.1" cidr_netmask="16" nic="eth0"
  group grp_percona_cluster res_mysql_vip
  clone cl_ping ping \
  meta interleave="true"
  location Ping-res_mysql_vip res_mysql_vip \
  rule $id="Ping-res_mysql_vip-rule" -inf: pingd lte 0
  property $id="cib-bootstrap-options" \
  dc-version="1.1.10-42f2063" \
  cluster-infrastructure="corosync" \
  no-quorum-policy="stop" \
  stonith-enabled="false" \
  last-lrm-refresh="1421764942"
  rsc_defaults $id="rsc-options" \
  resource-stickiness="100"

  ProblemType: Crash
  DistroRelease: Ubuntu 14.04
  Package: pacemaker-cli-utils 1.1.10+git20130802-1ubuntu2.1
  ProcVersionSignature: Ubuntu 3.13.0-40.69-generic 3.13.11.10
  Uname: Linux 3.13.0-40-generic x86_64
  ApportVersion: 2.14.1-0ubuntu3.6
  Architecture: amd64
  CrashCounter: 1
  Date: Tue Jan 20 14:54:20 2015
  Ec2AMI: ami-000f
  Ec2AMIManifest: FIXME
  Ec2AvailabilityZone: nova
  Ec2InstanceType: m1.small
  Ec2Kernel: aki-0002
  Ec2Ramdisk: ari-0002
  ExecutablePath: /usr/sbin/crm_mon
  ExecutableTimestamp: 1411498637
  ProcCmdline: crm_mon -1fA
  ProcCwd: /home/ubuntu
  ProcEnviron:
   TERM=screen
   PATH=(custom, no user)
   LANG=en_US.UTF-8
   SHELL=/bin/bash
  Signal: 6
  SourcePackage: pacemaker
  StacktraceTop:
   g_return_if_fail_warning (log_domain=log_domain@entry=0x7fab566acb4e "GLib", 
pretty_function=pretty_function@entry=0x7fab566ba7e0 <__FUNCTION__.11709> 
"g_strsplit", expression=expression@entry=0x7fab56704fca "string != NULL") at 
/build/buildd/glib2.0-2.40.2/./glib/gmessages.c:1080
   g_strsplit (string=0x0, delimiter=, max_tokens=) at /build/buildd/glib2.0-2.40.2/./glib/gstrfuncs.c:2269
   print_attr_msg (node=0x7fab58af09e0, rsc_list=, 
attrname=0x7fab58afb5f0 "pingd", attrvalue=0x7fab58afb5d0 "100") at 
crm_mon.c:989
   print_node_attribute (name=0x7fab58afb5f0, node_data=0x7fab58af09e0) at 
crm_mon.c:1042
   g_list_foreach (list=, func=0x7fab5834c030 
, 

[Ubuntu-ha] [Bug 1322899] Re: pacemaker init script links aren't created on upgrade

2020-03-29 Thread Rafael David Tinoco
*** This bug is a duplicate of bug 1052449 ***
https://bugs.launchpad.net/bugs/1052449

** This bug has been marked a duplicate of bug 1052449
   corosync hangs due to missing pacemaker shutdown scripts

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1322899

Title:
  pacemaker init script links aren't created on upgrade

Status in pacemaker package in Ubuntu:
  Triaged

Bug description:
  When upgrading my pacemaker/corosync machines from ubuntu 12.04 to
  14.04, update-rc.d for pacemaker is not run, so the new pacemaker
  init.d script is never executed on system startup, causing
  corosync/pacemaker HA system to not start.

  When adding the new pacemaker init.d script, update-rc.d should be
  run, so pacemaker init.d is run to start pacemaker as needed.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1322899/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1227580] Re: Monitoring in Master state doesn't always work

2020-03-29 Thread Rafael David Tinoco
Thank you for taking the time to report this bug. In an effort to keep an
up-to-date and valid list of bugs to work on, I have reviewed this report
to verify it still requires effort and occurs on an Ubuntu release in
standard support, and it does not.

It is unfortunate that we were unable to resolve this defect, however
there appears to be no further action possible at this time. I am
therefore moving the bug to 'Incomplete'. If you disagree or have
new information, we would be grateful if you could please add a comment 
stating why and then change the status of the bug to 'New'.

** Changed in: pacemaker (Ubuntu Precise)
   Status: New => Incomplete

** Changed in: pacemaker (Ubuntu)
   Status: New => Fix Released

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1227580

Title:
  Monitoring in Master state doesn't always work

Status in pacemaker package in Ubuntu:
  Fix Released
Status in pacemaker source package in Precise:
  Incomplete

Bug description:
  Hi
  There is a bug that master/slave resources are not always monitored. It is 
already fixed in pacemaker trunk, but not in libpengine3 1.1.6-2ubuntu3. I 
patched lib/pengine/unpack.c

  
  1391c1391,1395
  <   } else if(safe_str_eq(task, CRMD_ACTION_START)) {
  ---
  >   } else if(safe_str_eq(task, CRMD_ACTION_START) || 
  >   safe_str_eq(task, CRMD_ACTION_MIGRATED) || 
  >   safe_str_eq(task, CRMD_ACTION_PROMOTE) || 
  >   safe_str_eq(task, CRMD_ACTION_DEMOTE)) {
  > 

  With this monitoring also works if resource has been promoted. This
  bug can cause much trouble as a failed service wont be recognized.

  Best regards

  Achim

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1227580/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1227580] Re: Monitoring in Master state doesn't always work

2020-03-29 Thread Rafael David Tinoco
The mentioned patch is the following one:

commit fc03be02bf3a045babfe8233cbc99227da71d024
Author: David Vossel 
Date:   Fri Jun 29 11:24:52 2012 -0500

High: pengine: cl#5072 - Fixes monitor op stopping after rsc
promotion.

and was supposed to be applied to Precise.

** Also affects: pacemaker (Ubuntu Precise)
   Importance: Undecided
   Status: New

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1227580

Title:
  Monitoring in Master state doesn't always work

Status in pacemaker package in Ubuntu:
  Fix Released
Status in pacemaker source package in Precise:
  Incomplete

Bug description:
  Hi
  There is a bug that master/slave resources are not always monitored. It is 
already fixed in pacemaker trunk, but not in libpengine3 1.1.6-2ubuntu3. I 
patched lib/pengine/unpack.c

  
  1391c1391,1395
  <   } else if(safe_str_eq(task, CRMD_ACTION_START)) {
  ---
  >   } else if(safe_str_eq(task, CRMD_ACTION_START) || 
  >   safe_str_eq(task, CRMD_ACTION_MIGRATED) || 
  >   safe_str_eq(task, CRMD_ACTION_PROMOTE) || 
  >   safe_str_eq(task, CRMD_ACTION_DEMOTE)) {
  > 

  With this monitoring also works if resource has been promoted. This
  bug can cause much trouble as a failed service wont be recognized.

  Best regards

  Achim

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1227580/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1052449] Re: corosync hangs due to missing pacemaker shutdown scripts

2020-03-29 Thread Rafael David Tinoco
I'm marking Precise and Trusty as affected.

The suggested fix is currently implemented by Xenial already:

mcp/pacemaker.in:

...
# Required-Start:   $network corosync
# Should-Start: $syslog
# Required-Stop:$network corosync
...


** Also affects: pacemaker (Ubuntu Precise)
   Importance: Undecided
   Status: New

** Also affects: pacemaker (Ubuntu Xenial)
   Importance: Undecided
   Status: New

** Also affects: pacemaker (Ubuntu Trusty)
   Importance: Undecided
   Status: New

** Changed in: pacemaker (Ubuntu Xenial)
   Status: New => Fix Released

** Changed in: pacemaker (Ubuntu Trusty)
   Status: New => Confirmed

** Changed in: pacemaker (Ubuntu Precise)
   Status: New => Confirmed

** Changed in: pacemaker (Ubuntu)
   Status: Triaged => Fix Released

** Changed in: pacemaker (Ubuntu)
   Importance: High => Undecided

** Changed in: pacemaker (Ubuntu Trusty)
   Status: Confirmed => Triaged

** Changed in: pacemaker (Ubuntu Precise)
   Status: Confirmed => Triaged

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1052449

Title:
  corosync hangs due to missing pacemaker shutdown scripts

Status in pacemaker package in Ubuntu:
  Fix Released
Status in pacemaker source package in Precise:
  Triaged
Status in pacemaker source package in Trusty:
  Triaged
Status in pacemaker source package in Xenial:
  Fix Released

Bug description:
  The pacemaker package installs the right init script but doesn't link
  it to the according runlevels. If corosync is activated and started on
  the system this leads to a hanging shutdown / reboot because corosync
  only ends if pacemaker is stopped beforehand. In addition to this the
  pacemaker daemon has to start after corosync but has to stop before
  corosync.

  A possible solution would be to link the init script accordingly an
  enable it throug /etc/default/

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1052449/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1049037] Re: Pacemaker 1.1.6 VirtualDomain OCF wrong timeout handle on shutdown

2020-03-29 Thread Rafael David Tinoco
Fix was already included in current pacemaker version by patch:

commit fcfe6fe522138343e4138248829926700fac213e
Author: Yan Gao 
Date:   Mon Oct 17 11:19:01 2011 +0800

Medium: crmd: Send out all of the meta parameters to lrmd for stop
actions

So I'm flagging this as "Fix Released" as it does not affect current
supported versions of pacemaker.

** Changed in: pacemaker (Ubuntu)
   Status: Confirmed => Fix Released

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1049037

Title:
  Pacemaker 1.1.6 VirtualDomain OCF wrong timeout handle on shutdown

Status in pacemaker package in Ubuntu:
  Fix Released

Bug description:
  Pacemaker 1.1.6 included in current LTS has a bug which causes
  VirtualDomain to fail upon VM shutdown.

  Full thread with explaination here:
  http://www.gossamer-threads.com/lists/linuxha/pacemaker/78969

  it would be great if either the patch is backported or Pacemaker
  upgraded into LTS.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1049037/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 955164] Re: PostgreSQL resource agent has wrong default paths

2020-03-29 Thread Rafael David Tinoco
Thank you for taking the time to report this bug. In an effort to keep an
up-to-date and valid list of bugs to work on, I have reviewed this report
to verify it still requires effort and occurs on an Ubuntu release in
standard support, and it does not.

It is unfortunate that we were unable to resolve this defect, however
there appears to be no further action possible at this time. I am
therefore moving the bug to 'Incomplete'. If you disagree or have
new information, we would be grateful if you could please add a comment 
stating why and then change the status of the bug to 'New'.

Observation:

The correct affected package for this bug is resource-agents:

(k)rafaeldtinoco@clufocal01:~$ dpkg -L resource-agents | grep -i pgsql
/usr/lib/ocf/resource.d/heartbeat/pgsql

and I'm opening a generic bug to check ALL resource agents paths
existence.

** Also affects: resource-agents (Ubuntu)
   Importance: Undecided
   Status: New

** No longer affects: pacemaker (Ubuntu)

** Also affects: resource-agents (Ubuntu Precise)
   Importance: Undecided
   Status: New

** Changed in: resource-agents (Ubuntu Precise)
   Status: New => Incomplete

** Changed in: resource-agents (Ubuntu)
   Status: New => Invalid

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/955164

Title:
  PostgreSQL resource agent has wrong default paths

Status in resource-agents package in Ubuntu:
  Invalid
Status in resource-agents source package in Precise:
  Incomplete

Bug description:
  The pgsql resource agent has default paths which don't match those
  used by the Ubuntu postgresql package

  In the resource agent:
   * The default configuration file is /etc/postgresql/9.1/main/postgresql.conf
   * The default loction of pg_ctl is /usr/lib/postgresql/9.1/bin/pg_ctl
   * The default data directory is given as /var/lib/pgsql/data

  This causes Pacemaker to be unable to start the postgresql server, and the 
following errors in /var/log/syslog:
  Mar 14 14:51:38 OB1 pgsql[3533]: ERROR: Configuration file 
/var/lib/pgsql/data/postgresql.conf doesn't exist
  Mar 14 14:51:38 OB1 pgsql[3533]: ERROR: Setup problem: couldn't find command: 
/usr/bin/pg_ctl

  The problem is in the file /usr/lib/ocf/resource.d/heartbeat/pgsql

  The following patch fixes these problems:
  =
  --- pgsql.old 2012-03-14 15:17:09.123507205 +
  +++ pgsql 2012-03-14 15:19:37.521641615 +
  @@ -45,13 +45,13 @@
   }
   
   # Defaults
  -OCF_RESKEY_pgctl_default=/usr/bin/pg_ctl
  +OCF_RESKEY_pgctl_default=/usr/lib/postgresql/9.1/bin/pg_ctl
   OCF_RESKEY_psql_default=/usr/bin/psql
  -OCF_RESKEY_pgdata_default=/var/lib/pgsql/data
  +OCF_RESKEY_pgdata_default=/var/lib/postgresql/9.1/main
   OCF_RESKEY_pgdba_default=postgres
   OCF_RESKEY_pghost_default=""
   OCF_RESKEY_pgport_default=5432
  -OCF_RESKEY_config_default=""
  +OCF_RESKEY_config_default=/etc/postgresql/9.1/main/postgresql.conf
   OCF_RESKEY_start_opt_default=""
   OCF_RESKEY_pgdb_default=template1
   OCF_RESKEY_logfile_default=/dev/null
  =

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/resource-agents/+bug/955164/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1869622] Re: [focal] corosync v3.0.3 last upstream fixes

2020-03-29 Thread Rafael David Tinoco
** Changed in: corosync (Ubuntu)
   Status: Confirmed => In Progress

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to corosync in Ubuntu.
https://bugs.launchpad.net/bugs/1869622

Title:
  [focal] corosync v3.0.3 last upstream fixes

Status in corosync package in Ubuntu:
  In Progress

Bug description:
  Together with fixes for:

  https://bugs.launchpad.net/ubuntu/+source/corosync/+bug/1437359
  https://bugs.launchpad.net/ubuntu/+source/corosync/+bug/1677684

  This bug will also backport/cherrypick the following post release
  fixes:

  ->  c631951e icmap: icmap_init_r() leaks if trie_create() fails = fix
  ->  ca320bea votequorum: set wfa status only on startup = fix
  ->  5f543465 quorumtool: exit on invalid expected votes = fix
  ->  0c16442f votequorum: Change check of expected_votes = fix
  ->  8ce65bf9 votequorum: Reflect runtime change of 2Node to WFA = fix
  ->  89b0d62f stats: Check return code of stats_map_get  = fix
  ->  56ee8503 quorumtool: Assert copied string length= assert
  ->  1fb095b0 notifyd: Check cmap_track_add result   = assert
  ->  8ff7760c cmapctl: Free bin_value on error   = fix
  ->  35c312f8 votequorum: Assert copied strings length   = assert
  ->  29109683 totemknet: Assert strcpy length assert = assert
  ->  0c118d8f totemknet: Check result of fcntl O_NONBLOCK call   = assert
  ->  a24cbad5 totemconfig: Initialize warnings variable  = build 
fix
  ->  74eed54a sync: Assert sync_callbacks.name length= assert
  ->  380b744e totemknet: Don't mix corosync and knet error codes = fix
  ->  624b6a47 stats: Assert value_len when value is needed   = assert
  ->  09f6d34a logconfig: Remove double free of value = fix
  ->  efe48120 totemconfig: Free leaks found by coverity  = fix

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/corosync/+bug/1869622/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1677684] Re: /usr/bin/corosync-blackbox: 34: /usr/bin/corosync-blackbox: qb-blackbox: not found

2020-03-29 Thread Rafael David Tinoco
BUG: https://bugs.launchpad.net/ubuntu/+source/corosync/+bug/1869622
PPA: https://launchpad.net/~ubuntu-server-ha/+archive/ubuntu/staging

** Changed in: corosync (Ubuntu Focal)
   Status: Confirmed => In Progress

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to corosync in Ubuntu.
https://bugs.launchpad.net/bugs/1677684

Title:
  /usr/bin/corosync-blackbox: 34: /usr/bin/corosync-blackbox: qb-
  blackbox: not found

Status in corosync package in Ubuntu:
  In Progress
Status in corosync source package in Trusty:
  Won't Fix
Status in corosync source package in Xenial:
  Confirmed
Status in corosync source package in Zesty:
  Won't Fix
Status in corosync source package in Bionic:
  Confirmed
Status in corosync source package in Disco:
  Won't Fix
Status in corosync source package in Eoan:
  Confirmed
Status in corosync source package in Focal:
  In Progress

Bug description:
  [Environment]

  Ubuntu Xenial 16.04
  Amd64

  [Test Case]

  1) sudo apt-get install corosync
  2) sudo corosync-blackbox.

  root@juju-niedbalski-xenial-machine-5:/home/ubuntu# dpkg -L corosync |grep 
black
  /usr/bin/corosync-blackbox

  Expected results: corosync-blackbox runs OK.

  Current results:

  $ sudo corosync-blackbox
  /usr/bin/corosync-blackbox: 34: /usr/bin/corosync-blackbox: qb-blackbox: not 
found

  [Impact]

   * Cannot run corosync-blackbox

  [Regression Potential]

  * None identified.

  [Fix]
  Make the package dependant of libqb-dev

  root@juju-niedbalski-xenial-machine-5:/home/ubuntu# dpkg -L libqb-dev | grep 
qb-bl
  /usr/sbin/qb-blackbox

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/corosync/+bug/1677684/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1869622] Re: [focal] corosync v3.0.3 last upstream fixes

2020-03-29 Thread Rafael David Tinoco
For the 2 fixes:

LP: #1437359
LP: #1677684

I have submitted to upstream the following merge request:

https://salsa.debian.org/ha-team/corosync/-/merge_requests/2

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to corosync in Ubuntu.
https://bugs.launchpad.net/bugs/1869622

Title:
  [focal] corosync v3.0.3 last upstream fixes

Status in corosync package in Ubuntu:
  Confirmed

Bug description:
  Together with fixes for:

  https://bugs.launchpad.net/ubuntu/+source/corosync/+bug/1437359
  https://bugs.launchpad.net/ubuntu/+source/corosync/+bug/1677684

  This bug will also backport/cherrypick the following post release
  fixes:

  ->  c631951e icmap: icmap_init_r() leaks if trie_create() fails = fix
  ->  ca320bea votequorum: set wfa status only on startup = fix
  ->  5f543465 quorumtool: exit on invalid expected votes = fix
  ->  0c16442f votequorum: Change check of expected_votes = fix
  ->  8ce65bf9 votequorum: Reflect runtime change of 2Node to WFA = fix
  ->  89b0d62f stats: Check return code of stats_map_get  = fix
  ->  56ee8503 quorumtool: Assert copied string length= assert
  ->  1fb095b0 notifyd: Check cmap_track_add result   = assert
  ->  8ff7760c cmapctl: Free bin_value on error   = fix
  ->  35c312f8 votequorum: Assert copied strings length   = assert
  ->  29109683 totemknet: Assert strcpy length assert = assert
  ->  0c118d8f totemknet: Check result of fcntl O_NONBLOCK call   = assert
  ->  a24cbad5 totemconfig: Initialize warnings variable  = build 
fix
  ->  74eed54a sync: Assert sync_callbacks.name length= assert
  ->  380b744e totemknet: Don't mix corosync and knet error codes = fix
  ->  624b6a47 stats: Assert value_len when value is needed   = assert
  ->  09f6d34a logconfig: Remove double free of value = fix
  ->  efe48120 totemconfig: Free leaks found by coverity  = fix

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/corosync/+bug/1869622/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1869622] Re: [focal] corosync v3.0.3 last upstream fixes

2020-03-29 Thread Rafael David Tinoco
** Description changed:

  Together with fixes for:
  
  https://bugs.launchpad.net/ubuntu/+source/corosync/+bug/1437359
  https://bugs.launchpad.net/ubuntu/+source/corosync/+bug/1677684
  
  This bug will also backport/cherrypick the following post release fixes:
  
- ->  c631951e icmap: icmap_init_r() leaks if trie_create() fails = coverity fix
- ->  ca320bea votequorum: set wfa status only on startup = fix for corner case
- ->  5f543465 quorumtool: exit on invalid expected votes = fix
- ->  0c16442f votequorum: Change check of expected_votes = fix
- ->  8ce65bf9 votequorum: Reflect runtime change of 2Node to WFA = fix
- ->  89b0d62f stats: Check return code of stats_map_get = fix
- ->  56ee8503 quorumtool: Assert copied string length = assert (better support)
- ->  1fb095b0 notifyd: Check cmap_track_add result = assert (better support)
- ->  8ff7760c cmapctl: Free bin_value on error (fix)
- ->  35c312f8 votequorum: Assert copied strings length = assert (better 
support)
- ->  29109683 totemknet: Assert strcpy length assert (better support)
- ->  0c118d8f totemknet: Check result of fcntl O_NONBLOCK call
- ->  a24cbad5 totemconfig: Initialize warnings variable = build fix
- ->  74eed54a sync: Assert sync_callbacks.name length = assert (better support)
- ->  380b744e totemknet: Don't mix corosync and knet error codes = fix
- ->  624b6a47 stats: Assert value_len when value is needed = assert (bet 
support)
- ->  09f6d34a logconfig: Remove double free of value = fix
- ->  efe48120 totemconfig: Free leaks found by coverity = fix
+ ->  c631951e icmap: icmap_init_r() leaks if trie_create() fails = fix
+ ->  ca320bea votequorum: set wfa status only on startup = fix
+ ->  5f543465 quorumtool: exit on invalid expected votes = fix
+ ->  0c16442f votequorum: Change check of expected_votes = fix
+ ->  8ce65bf9 votequorum: Reflect runtime change of 2Node to WFA = fix
+ ->  89b0d62f stats: Check return code of stats_map_get  = fix
+ ->  56ee8503 quorumtool: Assert copied string length= assert
+ ->  1fb095b0 notifyd: Check cmap_track_add result   = assert
+ ->  8ff7760c cmapctl: Free bin_value on error   = fix
+ ->  35c312f8 votequorum: Assert copied strings length   = assert
+ ->  29109683 totemknet: Assert strcpy length assert = assert
+ ->  0c118d8f totemknet: Check result of fcntl O_NONBLOCK call   = assert
+ ->  a24cbad5 totemconfig: Initialize warnings variable  = build 
fix
+ ->  74eed54a sync: Assert sync_callbacks.name length= assert
+ ->  380b744e totemknet: Don't mix corosync and knet error codes = fix
+ ->  624b6a47 stats: Assert value_len when value is needed   = assert
+ ->  09f6d34a logconfig: Remove double free of value = fix
+ ->  efe48120 totemconfig: Free leaks found by coverity  = fix

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to corosync in Ubuntu.
https://bugs.launchpad.net/bugs/1869622

Title:
  [focal] corosync v3.0.3 last upstream fixes

Status in corosync package in Ubuntu:
  Confirmed

Bug description:
  Together with fixes for:

  https://bugs.launchpad.net/ubuntu/+source/corosync/+bug/1437359
  https://bugs.launchpad.net/ubuntu/+source/corosync/+bug/1677684

  This bug will also backport/cherrypick the following post release
  fixes:

  ->  c631951e icmap: icmap_init_r() leaks if trie_create() fails = fix
  ->  ca320bea votequorum: set wfa status only on startup = fix
  ->  5f543465 quorumtool: exit on invalid expected votes = fix
  ->  0c16442f votequorum: Change check of expected_votes = fix
  ->  8ce65bf9 votequorum: Reflect runtime change of 2Node to WFA = fix
  ->  89b0d62f stats: Check return code of stats_map_get  = fix
  ->  56ee8503 quorumtool: Assert copied string length= assert
  ->  1fb095b0 notifyd: Check cmap_track_add result   = assert
  ->  8ff7760c cmapctl: Free bin_value on error   = fix
  ->  35c312f8 votequorum: Assert copied strings length   = assert
  ->  29109683 totemknet: Assert strcpy length assert = assert
  ->  0c118d8f totemknet: Check result of fcntl O_NONBLOCK call   = assert
  ->  a24cbad5 totemconfig: Initialize warnings variable  = build 
fix
  ->  74eed54a sync: Assert sync_callbacks.name length= assert
  ->  380b744e totemknet: Don't mix corosync and knet error codes = fix
  ->  624b6a47 stats: Assert value_len when value is needed   = assert
  ->  09f6d34a logconfig: Remove double free of value = fix
  ->  efe48120 totemconfig: Free leaks found by coverity  = fix

To manage notifications about this bug go to:

[Ubuntu-ha] [Bug 1869622] [NEW] [focal] corosync v3.0.3 last upstream fixes

2020-03-29 Thread Rafael David Tinoco
Public bug reported:

Apart from fixes for:

https://bugs.launchpad.net/ubuntu/+source/corosync/+bug/1437359
https://bugs.launchpad.net/ubuntu/+source/corosync/+bug/1677684

This bug will also backport/cherrypick the following post release fixes:

->  c631951e icmap: icmap_init_r() leaks if trie_create() fails = coverity fix
->  ca320bea votequorum: set wfa status only on startup = fix for corner case
->  5f543465 quorumtool: exit on invalid expected votes = fix
->  0c16442f votequorum: Change check of expected_votes = fix
->  8ce65bf9 votequorum: Reflect runtime change of 2Node to WFA = fix
->  89b0d62f stats: Check return code of stats_map_get = fix
->  56ee8503 quorumtool: Assert copied string length = assert (better support)
->  1fb095b0 notifyd: Check cmap_track_add result = assert (better support)
->  8ff7760c cmapctl: Free bin_value on error (fix)
->  35c312f8 votequorum: Assert copied strings length = assert (better support)
->  29109683 totemknet: Assert strcpy length assert (better support)
->  0c118d8f totemknet: Check result of fcntl O_NONBLOCK call
->  a24cbad5 totemconfig: Initialize warnings variable = build fix
->  74eed54a sync: Assert sync_callbacks.name length = assert (better support)
->  380b744e totemknet: Don't mix corosync and knet error codes = fix
->  624b6a47 stats: Assert value_len when value is needed = assert (bet support)
->  09f6d34a logconfig: Remove double free of value = fix
->  efe48120 totemconfig: Free leaks found by coverity = fix

** Affects: corosync (Ubuntu)
     Importance: High
 Assignee: Rafael David Tinoco (rafaeldtinoco)
 Status: Confirmed

** Changed in: corosync (Ubuntu)
   Status: New => Confirmed

** Changed in: corosync (Ubuntu)
   Importance: Undecided => High

** Changed in: corosync (Ubuntu)
 Assignee: (unassigned) => Rafael David Tinoco (rafaeldtinoco)

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to corosync in Ubuntu.
https://bugs.launchpad.net/bugs/1869622

Title:
  [focal] corosync v3.0.3 last upstream fixes

Status in corosync package in Ubuntu:
  Confirmed

Bug description:
  Apart from fixes for:

  https://bugs.launchpad.net/ubuntu/+source/corosync/+bug/1437359
  https://bugs.launchpad.net/ubuntu/+source/corosync/+bug/1677684

  This bug will also backport/cherrypick the following post release
  fixes:

  ->  c631951e icmap: icmap_init_r() leaks if trie_create() fails = coverity fix
  ->  ca320bea votequorum: set wfa status only on startup = fix for corner case
  ->  5f543465 quorumtool: exit on invalid expected votes = fix
  ->  0c16442f votequorum: Change check of expected_votes = fix
  ->  8ce65bf9 votequorum: Reflect runtime change of 2Node to WFA = fix
  ->  89b0d62f stats: Check return code of stats_map_get = fix
  ->  56ee8503 quorumtool: Assert copied string length = assert (better support)
  ->  1fb095b0 notifyd: Check cmap_track_add result = assert (better support)
  ->  8ff7760c cmapctl: Free bin_value on error (fix)
  ->  35c312f8 votequorum: Assert copied strings length = assert (better 
support)
  ->  29109683 totemknet: Assert strcpy length assert (better support)
  ->  0c118d8f totemknet: Check result of fcntl O_NONBLOCK call
  ->  a24cbad5 totemconfig: Initialize warnings variable = build fix
  ->  74eed54a sync: Assert sync_callbacks.name length = assert (better support)
  ->  380b744e totemknet: Don't mix corosync and knet error codes = fix
  ->  624b6a47 stats: Assert value_len when value is needed = assert (bet 
support)
  ->  09f6d34a logconfig: Remove double free of value = fix
  ->  efe48120 totemconfig: Free leaks found by coverity = fix

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/corosync/+bug/1869622/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1437359] Re: A PIDFILE is double-defined for the corosync-notifyd init script

2020-03-23 Thread Rafael David Tinoco
** Tags added: server-triage-discuss

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to corosync in Ubuntu.
https://bugs.launchpad.net/bugs/1437359

Title:
  A PIDFILE is double-defined for the corosync-notifyd init script

Status in corosync package in Ubuntu:
  In Progress
Status in corosync source package in Trusty:
  Won't Fix
Status in corosync source package in Xenial:
  Triaged
Status in corosync source package in Bionic:
  Triaged
Status in corosync source package in Disco:
  Won't Fix
Status in corosync source package in Eoan:
  Triaged
Status in corosync source package in Focal:
  In Progress

Bug description:
  A /etc/init.d/corosync-notifyd contains two definitions for the PIDFILE:
  > PIDFILE=/var/run/$NAME.pid
  > SCRIPTNAME=/etc/init.d/$NAME
  > PIDFILE=/var/run/corosync.pid

  The first one is correct and the second one is wrong as it refers to
  the corosync service's pidfile instead

  The corosync package version is 2.3.3-1ubuntu1

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/corosync/+bug/1437359/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1340172] Re: [OSCI] Add mamange/unmanage blocks to corosync init scripts

2020-03-19 Thread Rafael David Tinoco
Thank you for taking the time to report this bug. In an effort to keep an
up-to-date and valid list of bugs to work on, I have reviewed this report
to verify it still requires effort and occurs on an Ubuntu release in
standard support, and it does not.

Unfortunately this bug is old for us to consider having an sysv-compatible
option to put the cluster into maintenance mode. Corosync is also being managed,
for sometime now, by systemd so this particular change wouldn't make much sense.
IMHO this type of cluster change should be done by the cluster administrator in
specific tools like crmsh, pcs, and not in an init script.

It is unfortunate that we were unable to resolve this defect, however
there appears to be no further action possible at this time. I am
therefore moving the bug to 'Incomplete'. If you disagree or have
new information, we would be grateful if you could please add a comment 
stating why and then change the status of the bug to 'New'.


** Also affects: corosync (Ubuntu Precise)
   Importance: Undecided
   Status: New

** Changed in: corosync (Ubuntu Precise)
   Status: New => Incomplete

** Changed in: corosync (Ubuntu)
   Status: Triaged => Won't Fix

** Changed in: corosync (Ubuntu)
 Assignee: Rafael David Tinoco (rafaeldtinoco) => (unassigned)

** Changed in: corosync (Ubuntu Precise)
   Status: Incomplete => Won't Fix

** Changed in: corosync (Ubuntu)
   Importance: High => Undecided

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to corosync in Ubuntu.
https://bugs.launchpad.net/bugs/1340172

Title:
  [OSCI] Add mamange/unmanage blocks to corosync init scripts

Status in Fuel for OpenStack:
  Fix Released
Status in corosync package in Ubuntu:
  Won't Fix
Status in corosync source package in Precise:
  Won't Fix

Bug description:
  Please add manage and unmanage fucntions to corosync init scripts'
  restart actions for both Centos and Ubuntu

  
  Here is patch for centos

  
  --- corosync-1.4.6/init/generic.in2013-05-29 18:33:27.0 +0400
  +++ corosync-1.4.6.new/init/generic.in2014-07-10 16:34:27.127641857 
+0400
  @@ -70,6 +70,50 @@
   pidof -c -o $$ -o $PPID -o %PPID "${1##*/}"
   }
   
  +pcmk_maintance() {
  +  which cibadmin 1>/dev/null 2>/dev/null
  +
  +  if [ "${?}" -gt "0" ]; then
  +return 0
  +  fi
  +
  +  pcmk_retry="30"
  +  pcmk_maintance="${1}"
  +  pcmk_sleep="1"
  +
  +  if [ "${pcmk_maintance}" = "" ]; then
  +pcmk_maintance="true"
  +  fi
  +
  +  while [ "${pcmk_retry}" -gt "0" ]; do
  +cat <
  +  
  +
  +  
  +
  +  
  +
  +  
  +
  +  
  +
  +  
  +
  +EOF
  +
  +if [ "${?}" -eq "0" ]; then
  +  break
  +fi
  +
  +echo "Pacemaker connection failed! Retries left: ${pcmk_retry}"
  +pcmk_retry="$((pcmk_retry - 1))"
  +sleep "${pcmk_sleep}"
  +
  +  done
  +}
  +
   cluster_disabled_at_boot()
   {
  if grep -q nocluster /proc/cmdline && \
  @@ -150,8 +194,13 @@
   
   restart()
   {
  + status $prog
  + if [ $? -eq 0 ]; then
  + pcmk_maintance true
  + fi
stop
start
  + pcmk_maintance false
   }
   
   rtrn=0
  ===

  
  And patch for ubuntu

  
  ===
  --- debian/corosync.init  2011-08-24 11:27:27.0 +0400
  +++ debian.new/corosync.init  2014-07-10 16:47:12.823624121 +0400
  @@ -39,6 +39,50 @@
   # Depend on lsb-base (>= 3.0-6) to ensure that this file is present.
   . /lib/lsb/init-functions
   
  +pcmk_maintance() {
  +  which cibadmin 1>/dev/null 2>/dev/null
  +
  +  if [ "${?}" -gt "0" ]; then
  +return 0
  +  fi
  +
  +  pcmk_retry="30"
  +  pcmk_maintance="${1}"
  +  pcmk_sleep="1"
  +
  +  if [ "${pcmk_maintance}" = "" ]; then
  +pcmk_maintance="true"
  +  fi
  +
  +  while [ "${pcmk_retry}" -gt "0" ]; do
  +cat <
  +  
  +
  +  
  +
  +  
  +
  +  
  +
  +  
  +
  +  
  +
  +EOF
  +
  +if [ "${?}" -eq "0" ]; then
  +  break
  +fi
  +
  +echo "Pacemaker connection failed! Retries left: ${pcmk_retry}"
  +pcmk_retry="$((pcmk_retry - 1))"
  +sleep "${pcmk_sleep}"
  +
  +  done
  +}
  +
   #
   # Function that starts the daemon/service
   #
  @@ -95,6 +139,10 @@
;;
 restart|force-reload)
log_daemon_msg "Restarting $DESC" &

[Ubuntu-ha] [Bug 1677684] Re: /usr/bin/corosync-blackbox: 34: /usr/bin/corosync-blackbox: qb-blackbox: not found

2020-03-19 Thread Rafael David Tinoco
This issue still exists and should be fixed. I'm putting together with
some other SRUs so all of them are done at once. Thanks @niedbalski for
bringing up this issue. I'll fix Ubuntu Focal for now and try to get
along with the needed SRUs.

** Changed in: corosync (Ubuntu Trusty)
 Assignee: Jorge Niedbalski (niedbalski) => (unassigned)

** Changed in: corosync (Ubuntu)
 Assignee: Jorge Niedbalski (niedbalski) => (unassigned)

** Changed in: corosync (Ubuntu Zesty)
 Assignee: Jorge Niedbalski (niedbalski) => (unassigned)

** Changed in: corosync (Ubuntu Xenial)
 Assignee: Jorge Niedbalski (niedbalski) => (unassigned)

** Changed in: corosync (Ubuntu)
   Importance: Medium => Undecided

** Changed in: corosync (Ubuntu Zesty)
   Importance: Medium => Undecided

** Changed in: corosync (Ubuntu Trusty)
   Importance: Medium => Undecided

** Changed in: corosync (Ubuntu Xenial)
   Importance: Medium => Undecided

** Also affects: corosync (Ubuntu Bionic)
   Importance: Undecided
   Status: New

** Also affects: corosync (Ubuntu Eoan)
   Importance: Undecided
   Status: New

** Also affects: corosync (Ubuntu Focal)
   Importance: Undecided
   Status: Incomplete

** Also affects: corosync (Ubuntu Disco)
   Importance: Undecided
   Status: New

** Changed in: corosync (Ubuntu Focal)
   Status: Incomplete => Confirmed

** Changed in: corosync (Ubuntu Eoan)
   Status: New => Confirmed

** Changed in: corosync (Ubuntu Disco)
   Status: New => Won't Fix

** Changed in: corosync (Ubuntu Bionic)
   Status: New => Confirmed

** Changed in: corosync (Ubuntu Zesty)
   Status: Incomplete => Won't Fix

** Changed in: corosync (Ubuntu Xenial)
   Status: Incomplete => Confirmed

** Changed in: corosync (Ubuntu Trusty)
   Status: Incomplete => Won't Fix

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to corosync in Ubuntu.
https://bugs.launchpad.net/bugs/1677684

Title:
  /usr/bin/corosync-blackbox: 34: /usr/bin/corosync-blackbox: qb-
  blackbox: not found

Status in corosync package in Ubuntu:
  Confirmed
Status in corosync source package in Trusty:
  Won't Fix
Status in corosync source package in Xenial:
  Confirmed
Status in corosync source package in Zesty:
  Won't Fix
Status in corosync source package in Bionic:
  Confirmed
Status in corosync source package in Disco:
  Won't Fix
Status in corosync source package in Eoan:
  Confirmed
Status in corosync source package in Focal:
  Confirmed

Bug description:
  [Environment]

  Ubuntu Xenial 16.04
  Amd64

  [Test Case]

  1) sudo apt-get install corosync
  2) sudo corosync-blackbox.

  root@juju-niedbalski-xenial-machine-5:/home/ubuntu# dpkg -L corosync |grep 
black
  /usr/bin/corosync-blackbox

  Expected results: corosync-blackbox runs OK.

  Current results:

  $ sudo corosync-blackbox
  /usr/bin/corosync-blackbox: 34: /usr/bin/corosync-blackbox: qb-blackbox: not 
found

  [Impact]

   * Cannot run corosync-blackbox

  [Regression Potential]

  * None identified.

  [Fix]
  Make the package dependant of libqb-dev

  root@juju-niedbalski-xenial-machine-5:/home/ubuntu# dpkg -L libqb-dev | grep 
qb-bl
  /usr/sbin/qb-blackbox

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/corosync/+bug/1677684/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1437359] Re: A PIDFILE is double-defined for the corosync-notifyd init script

2020-03-19 Thread Rafael David Tinoco
(c)rafaeldtinoco@clusterdev:~/.../sources/ubuntu/corosync$ git diff HEAD
diff --git a/debian/corosync-notifyd.init b/debian/corosync-notifyd.init
index c908618..837e48a 100644
--- a/debian/corosync-notifyd.init
+++ b/debian/corosync-notifyd.init
@@ -21,7 +21,6 @@ NAME=corosync-notifyd
 DAEMON=/usr/sbin/$NAME
 PIDFILE=/var/run/$NAME.pid
 SCRIPTNAME=/etc/init.d/$NAME
-PIDFILE=/var/run/corosync.pid
 RARUNDIR=/var/run/resource-agents

 # Exit if the package is not installed

If anyone faces this issue, just remove the second PIDFILE as it appears
to be a leftover. I'll fix this and point this out to Debian as it also
faces the same issue. NEVERTHELESS, remember that this issue would only
exist if you are using the "sysv" scripts OR the sysv-generator (making
systemd to control the service through sysv scripts).

All others are probably relying in systemd to start/stop/monitor
corosync daemon (/lib/systemd/system/corosync.service).

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to corosync in Ubuntu.
https://bugs.launchpad.net/bugs/1437359

Title:
  A PIDFILE is double-defined for the corosync-notifyd init script

Status in corosync package in Ubuntu:
  In Progress
Status in corosync source package in Trusty:
  Won't Fix
Status in corosync source package in Xenial:
  Triaged
Status in corosync source package in Bionic:
  Triaged
Status in corosync source package in Disco:
  Won't Fix
Status in corosync source package in Eoan:
  Triaged
Status in corosync source package in Focal:
  In Progress

Bug description:
  A /etc/init.d/corosync-notifyd contains two definitions for the PIDFILE:
  > PIDFILE=/var/run/$NAME.pid
  > SCRIPTNAME=/etc/init.d/$NAME
  > PIDFILE=/var/run/corosync.pid

  The first one is correct and the second one is wrong as it refers to
  the corosync service's pidfile instead

  The corosync package version is 2.3.3-1ubuntu1

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/corosync/+bug/1437359/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1654403] Re: Race condition in hacluster charm that leaves pacemaker down

2020-03-19 Thread Rafael David Tinoco
** Also affects: corosync (Ubuntu Xenial)
   Importance: Undecided
   Status: New

** Changed in: corosync (Ubuntu Xenial)
   Status: New => Incomplete

** Changed in: corosync (Ubuntu)
   Status: Incomplete => Fix Released

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to corosync in Ubuntu.
https://bugs.launchpad.net/bugs/1654403

Title:
  Race condition in hacluster charm that leaves pacemaker down

Status in OpenStack hacluster charm:
  Fix Released
Status in corosync package in Ubuntu:
  Fix Released
Status in corosync source package in Xenial:
  Incomplete
Status in hacluster package in Juju Charms Collection:
  Invalid

Bug description:
  Symptom: one or more hacluster nodes are left in an executing state.
  Observing the process list on the affected nodes the command 'crm node list' 
is in an infinite loop and pacemaker is not started. On nodes that complete the 
crm node list and other crm commands pacemaker is started.

  See the artefacts from this run:
  
https://openstack-ci-reports.ubuntu.com/artifacts/test_charm_pipeline/openstack/charm-percona-cluster/417131/1/1873/index.html

  Hypothesis: There is a race that leads to crm node list being executed
  before pacemaker is started. It is also possible that something causes
  pacemaker to fail to start.

  Suggest a check for pacemaker heath before any crm commands are run.

To manage notifications about this bug go to:
https://bugs.launchpad.net/charm-hacluster/+bug/1654403/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1437359] Re: A PIDFILE is double-defined for the corosync-notifyd init script

2020-03-19 Thread Rafael David Tinoco
Okay, this is a very simple fix but it is tricky... mainly because ..
possibly 99% of the users of this package are using systemd and the
corosync service unit file... which does not face this issue. I'm not
entirely sure a SRU is the right thing to do on all affected Ubuntu
versions (Xenial, Bionic, Eoan). Will discuss this with the server team.

Meanwhile, fixing Focal seems appropriate just because its not released
yet ... :\ but I'll keep this "on the hand" until I have more fixes so I
do the SRU all at once.

** Changed in: corosync (Ubuntu Focal)
   Status: Triaged => In Progress

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to corosync in Ubuntu.
https://bugs.launchpad.net/bugs/1437359

Title:
  A PIDFILE is double-defined for the corosync-notifyd init script

Status in corosync package in Ubuntu:
  In Progress
Status in corosync source package in Trusty:
  Won't Fix
Status in corosync source package in Xenial:
  Triaged
Status in corosync source package in Bionic:
  Triaged
Status in corosync source package in Disco:
  Won't Fix
Status in corosync source package in Eoan:
  Triaged
Status in corosync source package in Focal:
  In Progress

Bug description:
  A /etc/init.d/corosync-notifyd contains two definitions for the PIDFILE:
  > PIDFILE=/var/run/$NAME.pid
  > SCRIPTNAME=/etc/init.d/$NAME
  > PIDFILE=/var/run/corosync.pid

  The first one is correct and the second one is wrong as it refers to
  the corosync service's pidfile instead

  The corosync package version is 2.3.3-1ubuntu1

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/corosync/+bug/1437359/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1437359] Re: A PIDFILE is double-defined for the corosync-notifyd init script

2020-03-19 Thread Rafael David Tinoco
** Also affects: corosync (Ubuntu Disco)
   Importance: Undecided
   Status: New

** Also affects: corosync (Ubuntu Focal)
   Importance: Undecided
 Assignee: Rafael David Tinoco (rafaeldtinoco)
   Status: Triaged

** Also affects: corosync (Ubuntu Trusty)
   Importance: Undecided
   Status: New

** Also affects: corosync (Ubuntu Bionic)
   Importance: Undecided
   Status: New

** Also affects: corosync (Ubuntu Xenial)
   Importance: Undecided
   Status: New

** Also affects: corosync (Ubuntu Eoan)
   Importance: Undecided
   Status: New

** Changed in: corosync (Ubuntu Focal)
 Assignee: Rafael David Tinoco (rafaeldtinoco) => (unassigned)

** Changed in: corosync (Ubuntu Disco)
   Status: New => Won't Fix

** Changed in: corosync (Ubuntu Trusty)
   Status: New => Won't Fix

** Changed in: corosync (Ubuntu Eoan)
   Status: New => Triaged

** Changed in: corosync (Ubuntu Xenial)
   Status: New => Triaged

** Changed in: corosync (Ubuntu Bionic)
   Status: New => Triaged

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to corosync in Ubuntu.
https://bugs.launchpad.net/bugs/1437359

Title:
  A PIDFILE is double-defined for the corosync-notifyd init script

Status in corosync package in Ubuntu:
  In Progress
Status in corosync source package in Trusty:
  Won't Fix
Status in corosync source package in Xenial:
  Triaged
Status in corosync source package in Bionic:
  Triaged
Status in corosync source package in Disco:
  Won't Fix
Status in corosync source package in Eoan:
  Triaged
Status in corosync source package in Focal:
  In Progress

Bug description:
  A /etc/init.d/corosync-notifyd contains two definitions for the PIDFILE:
  > PIDFILE=/var/run/$NAME.pid
  > SCRIPTNAME=/etc/init.d/$NAME
  > PIDFILE=/var/run/corosync.pid

  The first one is correct and the second one is wrong as it refers to
  the corosync service's pidfile instead

  The corosync package version is 2.3.3-1ubuntu1

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/corosync/+bug/1437359/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1586876] Re: Corosync report "Started" itself too early

2020-03-19 Thread Rafael David Tinoco
>From upstream documentation:

"""
Pacemaker used to obtain membership and quorum from a custom Corosync plugin. 
This plugin also had the capability to start Pacemaker automatically when 
Corosync was started. Neither behavior is possible with Corosync 2.0 and beyond 
as support for plugins was removed.

Instead, Pacemaker must be started as a separate job/initscript. Also, since 
Pacemaker made use of the plugin for message routing, a node using the plugin 
(Corosync prior to 2.0) cannot talk to one that isn’t (Corosync 2.0+).
Rolling upgrades between these versions are therefore not possible and an 
alternate strategy must be used.
"""

showing that since Ubuntu Trusty this detection behavior is not
supported any longer. Nowadays, we start both services separately and
using systemd.

Corosync starts with a simple one-node only (localhost) ring configured:

(c)rafaeldtinoco@clusterdev:~$ systemctl status corosync
● corosync.service - Corosync Cluster Engine
 Loaded: loaded (/lib/systemd/system/corosync.service; enabled; vendor 
preset: enabled)
 Active: active (running) since Thu 2020-03-19 20:16:49 UTC; 45min ago
   Docs: man:corosync
 man:corosync.conf
 man:corosync_overview
   Main PID: 851 (corosync)
  Tasks: 9 (limit: 23186)
 Memory: 125.9M
 CGroup: /system.slice/corosync.service
 └─851 /usr/sbin/corosync -f

(c)rafaeldtinoco@clusterdev:~$ sudo corosync-quorumtool
Quorum information
--
Date: Thu Mar 19 21:02:21 2020
Quorum provider:  corosync_votequorum
Nodes:1
Node ID:  1
Ring ID:  1.5
Quorate:  Yes

Votequorum information
--
Expected votes:   1
Highest expected: 1
Total votes:  1
Quorum:   1
Flags:Quorate

Membership information
--
Nodeid  Votes Name
 1  1 node1 (local)

AND systemd is responsible to guarantee the synchronicity needed.



>From pacemaker service unit:

...
After=corosync.service
Requires=corosync.service

...

# If you want Corosync to stop whenever Pacemaker is stopped,
# uncomment the next line too:
#
# ExecStopPost=/bin/sh -c 'pidof pacemaker-controld || killall -TERM corosync'

...

# Pacemaker will restart along with Corosync if Corosync is stopped while
# Pacemaker is running.
# In this case, if you want to be fenced always (if you do not want to restart)
# uncomment ExecStopPost below.
#
# ExecStopPost=/bin/sh -c 'pidof corosync || \
#  /usr/bin/systemctl --no-block stop pacemaker' 

you have different options to control behavior for start/stop and
restart accordingly with corosync status.


** Changed in: corosync (Ubuntu Focal)
   Status: Triaged => Fix Released

** Changed in: corosync (Ubuntu Eoan)
   Status: Triaged => Fix Released

** Changed in: corosync (Ubuntu Bionic)
   Status: Triaged => Fix Released

** Changed in: corosync (Ubuntu Xenial)
   Status: Triaged => Won't Fix

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to corosync in Ubuntu.
https://bugs.launchpad.net/bugs/1586876

Title:
  Corosync report "Started" itself too early

Status in corosync package in Ubuntu:
  Fix Released
Status in corosync source package in Trusty:
  Won't Fix
Status in corosync source package in Xenial:
  Won't Fix
Status in corosync source package in Bionic:
  Fix Released
Status in corosync source package in Disco:
  Won't Fix
Status in corosync source package in Eoan:
  Fix Released
Status in corosync source package in Focal:
  Fix Released

Bug description:
  Problem description:
  currently, we have no service state check after start-stop-daemon in 
do_start(),
  it might lead to an error if corosync report itself started too early,
  pacemaker might think it is a 'heartbeat' backended, which is not we desired,
  we should check if corosync is "really" started, then report its state,

  syslog with wrong state:
  May 24 19:53:50 myhost corosync[1018]:   [MAIN  ] Corosync Cluster Engine 
('1.4.2'): started and ready to provide service.
  May 24 19:53:50 myhost corosync[1018]:   [MAIN  ] Corosync built-in features: 
nss
  May 24 19:53:50 myhost corosync[1018]:   [MAIN  ] Successfully read main 
configuration file '/etc/corosync/corosync.conf'.
  May 24 19:53:50 myhost corosync[1018]:   [TOTEM ] Initializing transport 
(UDP/IP Unicast).
  May 24 19:53:50 myhost corosync[1018]:   [TOTEM ] Initializing 
transmit/receive security: libtomcrypt SOBER128/SHA1HMAC (mode 0).
  May 24 19:53:50 myhost pacemakerd: [1094]: info: Invoked: pacemakerd
  May 24 19:53:50 myhost pacemakerd: [1094]: info: crm_log_init_worker: Changed 
active directory to /var/lib/heartbeat/cores/root
  May 24 19:53:50 myhost pacemakerd: [1094]: info: get_cluster_type: Assuming a 
'heartbeat' based cluster
  May 24 19:53:50 myhost pacemakerd: [1094]: info: read_config: Reading 
configure for stack: 

[Ubuntu-ha] [Bug 1239734] Re: corosync sends first time users on a hunt, simple fix

2020-03-19 Thread Rafael David Tinoco
Currently corosync is started by default with a single node ring formed
with localhost:

(c)rafaeldtinoco@clusterdev:~/.../sources/ubuntu/corosync$ systemctl status 
corosync
● corosync.service - Corosync Cluster Engine
 Loaded: loaded (/lib/systemd/system/corosync.service; enabled; vendor 
preset: enabled)
 Active: active (running) since Thu 2020-03-19 20:16:49 UTC; 2min 10s ago
   Docs: man:corosync
 man:corosync.conf
 man:corosync_overview
   Main PID: 851 (corosync)
  Tasks: 9 (limit: 23186)
 Memory: 125.7M
 CGroup: /system.slice/corosync.service
 └─851 /usr/sbin/corosync -f

(c)rafaeldtinoco@clusterdev:~/.../sources/ubuntu/corosync$ sudo 
corosync-quorumtool
Quorum information
--
Date: Thu Mar 19 20:19:37 2020
Quorum provider:  corosync_votequorum
Nodes:1
Node ID:  1
Ring ID:  1.5
Quorate:  Yes

Votequorum information
--
Expected votes:   1
Highest expected: 1
Total votes:  1
Quorum:   1
Flags:Quorate

Membership information
--
Nodeid  Votes Name
 1  1 node1 (local)

Because we try to make things "easier" (no failure on startups because
the lack of configuration files, for example). Of course having a HA
cluster is something beyond the scope of this bug and it might require
an experienced system administrator to be able to make corosync to work
in a meaningful way.

Nevertheless, for any future bug readers, may I recommend the following
reading:

https://discourse.ubuntu.com/t/ubuntu-high-availability-shared-scsi-
disk-environments/14874

in order to get a simple, yet supported, cluster configuration in a
shared scsi environment.

Thanks for reporting this.

** Changed in: corosync (Ubuntu)
   Status: Triaged => Fix Released

** Changed in: corosync (Ubuntu)
     Assignee: Rafael David Tinoco (rafaeldtinoco) => (unassigned)

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to corosync in Ubuntu.
https://bugs.launchpad.net/bugs/1239734

Title:
  corosync sends first time users on a hunt, simple fix

Status in corosync package in Ubuntu:
  Fix Released

Bug description:
  Corosync's init script properly fails to start by default to protect
  us. I get that, but the wild goose chase that a new user ends up on is
  a UX bug and in at least one case, I saw a user who simply changed
  distros because he couldn't figure it out claiming Ubuntu wasn't ready
  for business. (Yeah, I know that's nonsense but his complaint on
  spiceworks forums is among the top links on google for 'ubuntu
  corosync'. )

  The puppet daemon does the same thing, but it has a bit of logic that
  when someone tries to start it and this file overrides them, it simply
  tells the user and tells them where to change the variable.

  Here is the code they use in their init.

  start_puppet_agent() {
  if is_true "$START" ; then
  start-stop-daemon --start --quiet --pidfile $PIDFILE \
--startas $DAEMON -- $NAME $DAEMON_OPTS
  else
  echo ""
  echo "puppet not configured to start, please edit /etc/default/puppet 
to enable"
  fi
  }

  Here is the equivalent code in corosync's init.

  do_start()
  {
  # Return
  #   0 if daemon has been started
  #   1 if daemon was already running
  #   2 if daemon could not be started
  start-stop-daemon --start --quiet --exec $DAEMON --test > /dev/null \
  || return 1
  start-stop-daemon --start --quiet --exec $DAEMON -- $DAEMON_ARGS \
  || return 2
  # Add code here, if necessary, that waits for the process to be ready
  # to handle requests from services started subsequently which depend
  # on this one.  As a last resort, sleep for some time.
  pidof corosync > $PIDFILE
  }

  I won't bother combining it as I find each item I try that the devs
  usually find my solution rather overly complicated. I'm sure you can
  handle it better than I.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/corosync/+bug/1239734/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1586876] Re: Corosync report "Started" itself too early

2020-03-19 Thread Rafael David Tinoco
** Also affects: corosync (Ubuntu Eoan)
   Importance: Undecided
   Status: New

** Also affects: corosync (Ubuntu Disco)
   Importance: Undecided
   Status: New

** Also affects: corosync (Ubuntu Xenial)
   Importance: Undecided
   Status: New

** Also affects: corosync (Ubuntu Bionic)
   Importance: Undecided
   Status: New

** Also affects: corosync (Ubuntu Trusty)
   Importance: Undecided
   Status: New

** Also affects: corosync (Ubuntu Focal)
   Importance: Medium
 Assignee: guessi (guessi)
   Status: In Progress

** Changed in: corosync (Ubuntu Focal)
 Assignee: guessi (guessi) => (unassigned)

** Changed in: corosync (Ubuntu Focal)
   Importance: Medium => Undecided

** Changed in: corosync (Ubuntu Trusty)
   Status: New => Won't Fix

** Changed in: corosync (Ubuntu Disco)
   Status: New => Won't Fix

** Changed in: corosync (Ubuntu Focal)
   Status: In Progress => Triaged

** Changed in: corosync (Ubuntu Eoan)
   Status: New => Triaged

** Changed in: corosync (Ubuntu Xenial)
   Status: New => Triaged

** Changed in: corosync (Ubuntu Bionic)
   Status: New => Triaged

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to corosync in Ubuntu.
https://bugs.launchpad.net/bugs/1586876

Title:
  Corosync report "Started" itself too early

Status in corosync package in Ubuntu:
  Triaged
Status in corosync source package in Trusty:
  Won't Fix
Status in corosync source package in Xenial:
  Triaged
Status in corosync source package in Bionic:
  Triaged
Status in corosync source package in Disco:
  Won't Fix
Status in corosync source package in Eoan:
  Triaged
Status in corosync source package in Focal:
  Triaged

Bug description:
  Problem description:
  currently, we have no service state check after start-stop-daemon in 
do_start(),
  it might lead to an error if corosync report itself started too early,
  pacemaker might think it is a 'heartbeat' backended, which is not we desired,
  we should check if corosync is "really" started, then report its state,

  syslog with wrong state:
  May 24 19:53:50 myhost corosync[1018]:   [MAIN  ] Corosync Cluster Engine 
('1.4.2'): started and ready to provide service.
  May 24 19:53:50 myhost corosync[1018]:   [MAIN  ] Corosync built-in features: 
nss
  May 24 19:53:50 myhost corosync[1018]:   [MAIN  ] Successfully read main 
configuration file '/etc/corosync/corosync.conf'.
  May 24 19:53:50 myhost corosync[1018]:   [TOTEM ] Initializing transport 
(UDP/IP Unicast).
  May 24 19:53:50 myhost corosync[1018]:   [TOTEM ] Initializing 
transmit/receive security: libtomcrypt SOBER128/SHA1HMAC (mode 0).
  May 24 19:53:50 myhost pacemakerd: [1094]: info: Invoked: pacemakerd
  May 24 19:53:50 myhost pacemakerd: [1094]: info: crm_log_init_worker: Changed 
active directory to /var/lib/heartbeat/cores/root
  May 24 19:53:50 myhost pacemakerd: [1094]: info: get_cluster_type: Assuming a 
'heartbeat' based cluster
  May 24 19:53:50 myhost pacemakerd: [1094]: info: read_config: Reading 
configure for stack: heartbeat

  expected result:
  May 24 21:45:02 myhost corosync[1021]:   [MAIN  ] Completed service 
synchronization, ready to provide service.
  May 24 21:45:02 myhost pacemakerd: [1106]: info: Invoked: pacemakerd
  May 24 21:45:02 myhost pacemakerd: [1106]: info: crm_log_init_worker: Changed 
active directory to /var/lib/heartbeat/cores/root
  May 24 21:45:02 myhost pacemakerd: [1106]: info: config_find_next: Processing 
additional service options...
  May 24 21:45:02 myhost pacemakerd: [1106]: info: get_config_opt: Found 
'pacemaker' for option: name
  May 24 21:45:02 myhost pacemakerd: [1106]: info: get_config_opt: Found '1' 
for option: ver
  May 24 21:45:02 myhost pacemakerd: [1106]: info: get_cluster_type: Detected 
an active 'classic openais (with plugin)' cluster

  please note the order of following two lines:
  * corosync: [MAIN  ] Completed service synchronization, ready to provide 
service.
  * pacemakerd: info: get_cluster_type: ...

  affected versions:
  ALL (precise, trusty, vivid, wily, xenial, yakkety)

  upstream solution: wait_for_ipc()
  https://github.com/corosync/corosync/blob/master/init/corosync.in#L84-L99

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/corosync/+bug/1586876/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1019833] Re: problem wtih the configuration of the cluster.conf file. when

2020-03-19 Thread Rafael David Tinoco
** Also affects: corosync (Ubuntu Trusty)
   Importance: Undecided
   Status: New

** Changed in: corosync (Ubuntu Trusty)
   Status: New => Incomplete

** Changed in: corosync (Ubuntu)
   Status: Incomplete => Fix Released

** Changed in: corosync (Ubuntu)
 Assignee: Joao (nobrefr) => (unassigned)

** Changed in: corosync (Ubuntu)
   Importance: Medium => Undecided

** Changed in: corosync (Ubuntu)
   Status: Fix Released => Invalid

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to corosync in Ubuntu.
https://bugs.launchpad.net/bugs/1019833

Title:
   problem wtih the configuration of the cluster.conf file. when

Status in corosync package in Ubuntu:
  Invalid
Status in corosync source package in Trusty:
  Incomplete

Bug description:
  Hello,

   I'm trying to install and configure a small corosync cluster  with two
   nodes, on OS Ubuntu 12.04 LTS.
   There is a problem wtih the configuration of the cluster.conf file. when
   I try to validate it I get
   the error message: extra element rm in interleave.

   command used to validate: ccs_config_validate.

   Both nodes are working. - Used command clustat and cman_tool nodes

   I send a copy of  cluster.conf files of both nodes.

   Corosync version: 1.4.2

   Thanks in advance,

   João Francisco

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/corosync/+bug/1019833/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 637127] Re: TTL is one - prevents multicast routing

2020-03-19 Thread Rafael David Tinoco
Correct flag for this bug is: Fix Released

per upstream fix:

commit d3b983953d43dd17162be04de405d223fb21cd26
Author: Angus Salkeld 
Date:   Wed Nov 24 14:35:56 2010 +1100

Add totem/interface/ttl config option.

This adds a per-interface config option to
adjust the TTL.

Signed-off-by: Angus Salkeld 
Reviewed-by: Steven Dake 

** Changed in: corosync (Ubuntu)
   Status: Incomplete => Fix Released

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to corosync in Ubuntu.
https://bugs.launchpad.net/bugs/637127

Title:
  TTL is one - prevents multicast routing

Status in corosync package in Ubuntu:
  Fix Released
Status in corosync package in Fedora:
  Won't Fix

Bug description:
  Binary package hint: corosync

  corosync uses a TTL of 1 in its multicast packets - restricting its
  use to a local segment and making it unusable in a multicast routed
  environment (such as with routed virtual machines).

  ProblemType: Bug
  DistroRelease: Ubuntu 10.04
  Package: corosync 1.2.0-0ubuntu1
  ProcVersionSignature: Ubuntu 2.6.32-22.36-generic 2.6.32.11+drm33.2
  Uname: Linux 2.6.32-22-generic i686
  Architecture: i386
  Date: Mon Sep 13 13:17:59 2010
  ProcEnviron:
   PATH=(custom, no user)
   LANG=en_GB
   SHELL=/bin/bash
  SourcePackage: corosync

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/corosync/+bug/637127/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1251298] Re: Failed to sign on to LRMd with Heartbeat/Pacemaker

2020-03-19 Thread Rafael David Tinoco
Thank you for taking the time to report this bug. In an effort to keep an
up-to-date and valid list of bugs to work on, I have reviewed this report
to verify it still requires effort and occurs on an Ubuntu release in
standard support, and it does not.

Judging by the existing comments, and the pointed upstream discussion thread, it
appears that the wrong lrmd was being set after upgrade and that led existing
cluster to confusion when connecting to the local resource manager daemon.

It is unfortunate that we were unable to resolve this defect, however
there appears to be no further action possible at this time. I am
therefore moving the bug to 'Incomplete'. If you disagree or have
new information, we would be grateful if you could please add a comment
stating why and then change the status of the bug to 'New'.

** Changed in: cluster-glue (Ubuntu)
   Status: Confirmed => Incomplete

** Also affects: cluster-glue (Ubuntu Trusty)
   Importance: Undecided
   Status: New

** Changed in: cluster-glue (Ubuntu Trusty)
   Status: New => Incomplete

** Changed in: cluster-glue (Ubuntu)
   Status: Incomplete => Fix Released

** Changed in: cluster-glue (Ubuntu)
 Assignee: Rafael David Tinoco (rafaeldtinoco) => (unassigned)

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to cluster-glue in Ubuntu.
https://bugs.launchpad.net/bugs/1251298

Title:
  Failed to sign on to LRMd with Heartbeat/Pacemaker

Status in cluster-glue package in Ubuntu:
  Fix Released
Status in cluster-glue source package in Trusty:
  Incomplete

Bug description:
  I'm running a 2 node heartbeat/pacemaker cluster, which was working fine with 
Ubuntu 13.04
  After upgrading from Ubuntu 13.04 to Ubuntu 13.10, Heartbeat/Pacemaker keeps 
restarting the system due to sign on errors of lrmd and heartbeat tries to 
recover.

  As one system is already on ubuntu 13.10 and one system still running
  13.04, I've tried it without the second node, which leads to the same
  behavior, which occurs before any cluster communication happens.

  Syslog:
  Nov 14 15:53:06 wolverine crmd[2464]:  warning: do_lrm_control: Failed to 
sign on to the LRM 1 (30 max) times
  Nov 14 15:53:06 wolverine crmd[2464]:   notice: crmd_client_status_callback: 
Status update: Client wolverine.domain.tld/crmd now has status [join] (DC=false)
  Nov 14 15:53:06 wolverine crmd[2464]:   notice: crmd_client_status_callback: 
Status update: Client wolverine.domain.tld/crmd now has status [online] 
(DC=false)
  Nov 14 15:53:06 wolverine crmd[2464]:  warning: do_lrm_control: Failed to 
sign on to the LRM 2 (30 max) times
  Nov 14 15:53:06 wolverine crmd[2464]:  warning: do_lrm_control: Failed to 
sign on to the LRM 3 (30 max) times
  Nov 14 15:53:07 wolverine stonith-ng[2462]:   notice: setup_cib: Watching for 
stonith topology changes
  Nov 14 15:53:07 wolverine stonith-ng[2462]:   notice: unpack_config: On loss 
of CCM Quorum: Ignore
  Nov 14 15:53:08 wolverine crmd[2464]:  warning: do_lrm_control: Failed to 
sign on to the LRM 4 (30 max) times
  Nov 14 15:53:10 wolverine crmd[2464]:  warning: do_lrm_control: Failed to 
sign on to the LRM 5 (30 max) times
  Nov 14 15:53:12 wolverine crmd[2464]:  warning: do_lrm_control: Failed to 
sign on to the LRM 6 (30 max) times
  Nov 14 15:53:14 wolverine crmd[2464]:  warning: do_lrm_control: Failed to 
sign on to the LRM 7 (30 max) times
  Nov 14 15:53:16 wolverine crmd[2464]:  warning: do_lrm_control: Failed to 
sign on to the LRM 8 (30 max) times
  Nov 14 15:53:18 wolverine crmd[2464]:  warning: do_lrm_control: Failed to 
sign on to the LRM 9 (30 max) times
  Nov 14 15:53:20 wolverine crmd[2464]:  warning: do_lrm_control: Failed to 
sign on to the LRM 10 (30 max) times
  Nov 14 15:53:22 wolverine crmd[2464]:  warning: do_lrm_control: Failed to 
sign on to the LRM 11 (30 max) times
  Nov 14 15:53:24 wolverine crmd[2464]:  warning: do_lrm_control: Failed to 
sign on to the LRM 12 (30 max) times
  Nov 14 15:53:26 wolverine crmd[2464]:  warning: do_lrm_control: Failed to 
sign on to the LRM 13 (30 max) times
  Nov 14 15:53:28 wolverine crmd[2464]:  warning: do_lrm_control: Failed to 
sign on to the LRM 14 (30 max) times
  Nov 14 15:53:30 wolverine crmd[2464]:  warning: do_lrm_control: Failed to 
sign on to the LRM 15 (30 max) times
  Nov 14 15:53:32 wolverine crmd[2464]:  warning: do_lrm_control: Failed to 
sign on to the LRM 16 (30 max) times
  Nov 14 15:53:34 wolverine crmd[2464]:  warning: do_lrm_control: Failed to 
sign on to the LRM 17 (30 max) times
  Nov 14 15:53:36 wolverine crmd[2464]:  warning: do_lrm_control: Failed to 
sign on to the LRM 18 (30 max) times
  Nov 14 15:53:38 wolverine crmd[2464]:  warning: do_lrm_control: Failed to 
sign on to the LRM 19 (30 max) times
  Nov 14 15:53:40 wolverine crmd[2464]:  warning: do_lrm_control: Failed to 
sign on to the LRM 20 (30 max) times
  Nov 14 15:53:42 wolverine crmd[2464]:  warning: do_lrm_con

[Ubuntu-ha] [Bug 1866119] Re: [bionic] fence_scsi not working properly with Pacemaker 1.1.18-2ubuntu1.1

2020-03-06 Thread Rafael David Tinoco
** Merge proposal unlinked:
   
https://code.launchpad.net/~rafaeldtinoco/ubuntu/+source/pacemaker/+git/pacemaker/+merge/380336

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1866119

Title:
  [bionic] fence_scsi not working properly with Pacemaker
  1.1.18-2ubuntu1.1

Status in pacemaker package in Ubuntu:
  Fix Released
Status in pacemaker source package in Bionic:
  In Progress

Bug description:
  OBS: This bug was originally into LP: #1865523 but it was split.

   SRU: pacemaker

  [Impact]

   * fence_scsi is not currently working in a share disk environment

   * all clusters relying in fence_scsi and/or fence_scsi + watchdog
  won't be able to start the fencing agents OR, in worst case scenarios,
  the fence_scsi agent might start but won't make scsi reservations in
  the shared scsi disk.

   * this bug is taking care of pacemaker 1.1.18 issues with fence_scsi,
  since the later was fixed at LP: #1865523.

  [Test Case]

   * having a 3-node setup, nodes called "clubionic01, clubionic02,
  clubionic03", with a shared scsi disk (fully supporting persistent
  reservations) /dev/sda, with corosync and pacemaker operational and
  running, one might try:

  rafaeldtinoco@clubionic01:~$ crm configure
  crm(live)configure# property stonith-enabled=on
  crm(live)configure# property stonith-action=off
  crm(live)configure# property no-quorum-policy=stop
  crm(live)configure# property have-watchdog=true
  crm(live)configure# commit
  crm(live)configure# end
  crm(live)# end

  rafaeldtinoco@clubionic01:~$ crm configure primitive fence_clubionic \
  stonith:fence_scsi params \
  pcmk_host_list="clubionic01 clubionic02 clubionic03" \
  devices="/dev/sda" \
  meta provides=unfencing

  And see the following errors:

  Failed Actions:
  * fence_clubionic_start_0 on clubionic02 'unknown error' (1): call=6, 
status=Error, exitreason='',
  last-rc-change='Wed Mar  4 19:53:12 2020', queued=0ms, exec=1105ms
  * fence_clubionic_start_0 on clubionic03 'unknown error' (1): call=6, 
status=Error, exitreason='',
  last-rc-change='Wed Mar  4 19:53:13 2020', queued=0ms, exec=1109ms
  * fence_clubionic_start_0 on clubionic01 'unknown error' (1): call=6, 
status=Error, exitreason='',
  last-rc-change='Wed Mar  4 19:53:11 2020', queued=0ms, exec=1108ms

  and corosync.log will show:

  warning: unpack_rsc_op_failure: Processing failed op start for
  fence_clubionic on clubionic01: unknown error (1)

  [Regression Potential]

   * LP: #1865523 shows fence_scsi fully operational after SRU for that
  bug is done.

   * LP: #1865523 used pacemaker 1.1.19 (vanilla) in order to fix
  fence_scsi.

   * There are changes to: cluster resource manager daemon, local
  resource manager daemon and police engine. From all the changes, the
  police engine fix is the biggest, but still not big for a SRU. This
  could cause police engine, thus cluster decisions, to mal function.

   * All patches are based in upstream fixes made right after
  Pacemaker-1.1.18, used by Ubuntu Bionic and were tested with
  fence_scsi to make sure it fixed the issues.

  [Other Info]

   * Original Description:

  Trying to setup a cluster with an iscsi shared disk, using fence_scsi
  as the fencing mechanism, I realized that fence_scsi is not working in
  Ubuntu Bionic. I first thought it was related to Azure environment
  (LP: #1864419), where I was trying this environment, but then, trying
  locally, I figured out that somehow pacemaker 1.1.18 is not fencing
  the shared scsi disk properly.

  Note: I was able to "backport" vanilla 1.1.19 from upstream and
  fence_scsi worked. I have then tried 1.1.18 without all quilt patches
  and it didnt work as well. I think that bisecting 1.1.18 <-> 1.1.19
  might tell us which commit has fixed the behaviour needed by the
  fence_scsi agent.

  (k)rafaeldtinoco@clubionic01:~$ crm conf show
  node 1: clubionic01.private
  node 2: clubionic02.private
  node 3: clubionic03.private
  primitive fence_clubionic stonith:fence_scsi \
  params pcmk_host_list="10.250.3.10 10.250.3.11 10.250.3.12" 
devices="/dev/sda" \
  meta provides=unfencing
  property cib-bootstrap-options: \
  have-watchdog=false \
  dc-version=1.1.18-2b07d5c5a9 \
  cluster-infrastructure=corosync \
  cluster-name=clubionic \
  stonith-enabled=on \
  stonith-action=off \
  no-quorum-policy=stop \
  symmetric-cluster=true

  

  (k)rafaeldtinoco@clubionic02:~$ sudo crm_mon -1
  Stack: corosync
  Current DC: clubionic01.private (version 1.1.18-2b07d5c5a9) - partition with 
quorum
  Last updated: Mon Mar 2 15:55:30 2020
  Last change: Mon Mar 2 15:45:33 2020 by root via cibadmin on 
clubionic01.private

  3 nodes configured
  1 resource configured

  Online: [ clubionic01.private clubionic02.private clubionic03.private
  ]

[Ubuntu-ha] [Bug 1866119] Re: [bionic] fence_scsi not working properly with Pacemaker 1.1.18-2ubuntu1.1

2020-03-06 Thread Rafael David Tinoco
** Description changed:

  OBS: This bug was originally into LP: #1865523 but it was split.
  
   SRU: pacemaker
  
  [Impact]
  
-  * fence_scsi is not currently working in a share disk environment
+  * fence_scsi is not currently working in a share disk environment
  
-  * all clusters relying in fence_scsi and/or fence_scsi + watchdog won't
+  * all clusters relying in fence_scsi and/or fence_scsi + watchdog won't
  be able to start the fencing agents OR, in worst case scenarios, the
  fence_scsi agent might start but won't make scsi reservations in the
  shared scsi disk.
  
-  * this bug is taking care of pacemaker 1.1.18 issues with fence_scsi,
+  * this bug is taking care of pacemaker 1.1.18 issues with fence_scsi,
  since the later was fixed at LP: #1865523.
  
  [Test Case]
  
-  * having a 3-node setup, nodes called "clubionic01, clubionic02,
+  * having a 3-node setup, nodes called "clubionic01, clubionic02,
  clubionic03", with a shared scsi disk (fully supporting persistent
  reservations) /dev/sda, with corosync and pacemaker operational and
  running, one might try:
  
  rafaeldtinoco@clubionic01:~$ crm configure
  crm(live)configure# property stonith-enabled=on
  crm(live)configure# property stonith-action=off
  crm(live)configure# property no-quorum-policy=stop
  crm(live)configure# property have-watchdog=true
  crm(live)configure# commit
  crm(live)configure# end
  crm(live)# end
  
  rafaeldtinoco@clubionic01:~$ crm configure primitive fence_clubionic \
- stonith:fence_scsi params \
- pcmk_host_list="clubionic01 clubionic02 clubionic03" \
- devices="/dev/sda" \
- meta provides=unfencing
+ stonith:fence_scsi params \
+ pcmk_host_list="clubionic01 clubionic02 clubionic03" \
+ devices="/dev/sda" \
+ meta provides=unfencing
  
  And see the following errors:
  
  Failed Actions:
  * fence_clubionic_start_0 on clubionic02 'unknown error' (1): call=6, 
status=Error, exitreason='',
- last-rc-change='Wed Mar  4 19:53:12 2020', queued=0ms, exec=1105ms
+ last-rc-change='Wed Mar  4 19:53:12 2020', queued=0ms, exec=1105ms
  * fence_clubionic_start_0 on clubionic03 'unknown error' (1): call=6, 
status=Error, exitreason='',
- last-rc-change='Wed Mar  4 19:53:13 2020', queued=0ms, exec=1109ms
+ last-rc-change='Wed Mar  4 19:53:13 2020', queued=0ms, exec=1109ms
  * fence_clubionic_start_0 on clubionic01 'unknown error' (1): call=6, 
status=Error, exitreason='',
- last-rc-change='Wed Mar  4 19:53:11 2020', queued=0ms, exec=1108ms
+ last-rc-change='Wed Mar  4 19:53:11 2020', queued=0ms, exec=1108ms
  
  and corosync.log will show:
  
  warning: unpack_rsc_op_failure: Processing failed op start for
  fence_clubionic on clubionic01: unknown error (1)
  
  [Regression Potential]
  
-  * LP: #1865523 shows fence_scsi fully operational after SRU for that
+  * LP: #1865523 shows fence_scsi fully operational after SRU for that
  bug is done.
  
-  * LP: #1865523 used pacemaker 1.1.19 (vanilla) in order to fix
+  * LP: #1865523 used pacemaker 1.1.19 (vanilla) in order to fix
  fence_scsi.
  
-  * TODO
+  * There are changes to: cluster resource manager daemon, local resource
+ manager daemon and police engine. From all the changes, the police
+ engine fix is the biggest, but still not big for a SRU. This could cause
+ police engine, thus cluster decisions, to mal function.
+ 
+  * All patches are based in upstream fixes made right after
+ Pacemaker-1.1.18, used by Ubuntu Bionic and were tested with fence_scsi
+ to make sure it fixed the issues.
  
  [Other Info]
  
-  * Original Description:
+  * Original Description:
  
  Trying to setup a cluster with an iscsi shared disk, using fence_scsi as
  the fencing mechanism, I realized that fence_scsi is not working in
  Ubuntu Bionic. I first thought it was related to Azure environment (LP:
  #1864419), where I was trying this environment, but then, trying
  locally, I figured out that somehow pacemaker 1.1.18 is not fencing the
  shared scsi disk properly.
  
  Note: I was able to "backport" vanilla 1.1.19 from upstream and
  fence_scsi worked. I have then tried 1.1.18 without all quilt patches
  and it didnt work as well. I think that bisecting 1.1.18 <-> 1.1.19
  might tell us which commit has fixed the behaviour needed by the
  fence_scsi agent.
  
  (k)rafaeldtinoco@clubionic01:~$ crm conf show
  node 1: clubionic01.private
  node 2: clubionic02.private
  node 3: clubionic03.private
  primitive fence_clubionic stonith:fence_scsi \
  params pcmk_host_list="10.250.3.10 10.250.3.11 10.250.3.12" 
devices="/dev/sda" \
  meta provides=unfencing
  property cib-bootstrap-options: \
  have-watchdog=false \
  dc-version=1.1.18-2b07d5c5a9 \
  cluster-infrastructure=corosync \
  cluster-name=clubionic \
  stonith-enabled=on \
  stonith-action=off \
  no-quorum-policy=stop \
  symmetric-cluster=true
  
  
  
  

[Ubuntu-ha] [Bug 1866119] Re: [bionic] fence_scsi not working properly with 1.1.18-2ubuntu1.1

2020-03-05 Thread Rafael David Tinoco
SUMMARY:

One using packages from:

https://launchpad.net/~ubuntu-server-ha/+archive/ubuntu/staging

in Ubuntu Bionic will be workedaround temporarily (until fixes are
released) for the bugs:

https://bugs.launchpad.net/ubuntu/+source/fence-agents/+bug/1865523
https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1866119

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1866119

Title:
  [bionic] fence_scsi not working properly with Pacemaker
  1.1.18-2ubuntu1.1

Status in pacemaker package in Ubuntu:
  Fix Released
Status in pacemaker source package in Bionic:
  In Progress

Bug description:
  OBS: This bug was originally into LP: #1865523 but it was split.

   SRU: pacemaker

  [Impact]

   * fence_scsi is not currently working in a share disk environment

   * all clusters relying in fence_scsi and/or fence_scsi + watchdog
  won't be able to start the fencing agents OR, in worst case scenarios,
  the fence_scsi agent might start but won't make scsi reservations in
  the shared scsi disk.

   * this bug is taking care of pacemaker 1.1.18 issues with fence_scsi,
  since the later was fixed at LP: #1865523.

  [Test Case]

   * having a 3-node setup, nodes called "clubionic01, clubionic02,
  clubionic03", with a shared scsi disk (fully supporting persistent
  reservations) /dev/sda, with corosync and pacemaker operational and
  running, one might try:

  rafaeldtinoco@clubionic01:~$ crm configure
  crm(live)configure# property stonith-enabled=on
  crm(live)configure# property stonith-action=off
  crm(live)configure# property no-quorum-policy=stop
  crm(live)configure# property have-watchdog=true
  crm(live)configure# commit
  crm(live)configure# end
  crm(live)# end

  rafaeldtinoco@clubionic01:~$ crm configure primitive fence_clubionic \
  stonith:fence_scsi params \
  pcmk_host_list="clubionic01 clubionic02 clubionic03" \
  devices="/dev/sda" \
  meta provides=unfencing

  And see the following errors:

  Failed Actions:
  * fence_clubionic_start_0 on clubionic02 'unknown error' (1): call=6, 
status=Error, exitreason='',
  last-rc-change='Wed Mar  4 19:53:12 2020', queued=0ms, exec=1105ms
  * fence_clubionic_start_0 on clubionic03 'unknown error' (1): call=6, 
status=Error, exitreason='',
  last-rc-change='Wed Mar  4 19:53:13 2020', queued=0ms, exec=1109ms
  * fence_clubionic_start_0 on clubionic01 'unknown error' (1): call=6, 
status=Error, exitreason='',
  last-rc-change='Wed Mar  4 19:53:11 2020', queued=0ms, exec=1108ms

  and corosync.log will show:

  warning: unpack_rsc_op_failure: Processing failed op start for
  fence_clubionic on clubionic01: unknown error (1)

  [Regression Potential]

   * LP: #1865523 shows fence_scsi fully operational after SRU for that
  bug is done.

   * LP: #1865523 used pacemaker 1.1.19 (vanilla) in order to fix
  fence_scsi.

   * TODO

  [Other Info]

   * Original Description:

  Trying to setup a cluster with an iscsi shared disk, using fence_scsi
  as the fencing mechanism, I realized that fence_scsi is not working in
  Ubuntu Bionic. I first thought it was related to Azure environment
  (LP: #1864419), where I was trying this environment, but then, trying
  locally, I figured out that somehow pacemaker 1.1.18 is not fencing
  the shared scsi disk properly.

  Note: I was able to "backport" vanilla 1.1.19 from upstream and
  fence_scsi worked. I have then tried 1.1.18 without all quilt patches
  and it didnt work as well. I think that bisecting 1.1.18 <-> 1.1.19
  might tell us which commit has fixed the behaviour needed by the
  fence_scsi agent.

  (k)rafaeldtinoco@clubionic01:~$ crm conf show
  node 1: clubionic01.private
  node 2: clubionic02.private
  node 3: clubionic03.private
  primitive fence_clubionic stonith:fence_scsi \
  params pcmk_host_list="10.250.3.10 10.250.3.11 10.250.3.12" 
devices="/dev/sda" \
  meta provides=unfencing
  property cib-bootstrap-options: \
  have-watchdog=false \
  dc-version=1.1.18-2b07d5c5a9 \
  cluster-infrastructure=corosync \
  cluster-name=clubionic \
  stonith-enabled=on \
  stonith-action=off \
  no-quorum-policy=stop \
  symmetric-cluster=true

  

  (k)rafaeldtinoco@clubionic02:~$ sudo crm_mon -1
  Stack: corosync
  Current DC: clubionic01.private (version 1.1.18-2b07d5c5a9) - partition with 
quorum
  Last updated: Mon Mar 2 15:55:30 2020
  Last change: Mon Mar 2 15:45:33 2020 by root via cibadmin on 
clubionic01.private

  3 nodes configured
  1 resource configured

  Online: [ clubionic01.private clubionic02.private clubionic03.private
  ]

  Active resources:

   fence_clubionic (stonith:fence_scsi): Started clubionic01.private

  

  (k)rafaeldtinoco@clubionic02:~$ sudo sg_persist --in --read-keys 
--device=/dev/sda
    LIO-ORG cluster.bionic. 4.0
    Peripheral device 

[Ubuntu-ha] [Bug 1866119] Re: [bionic] fence_scsi not working properly with Pacemaker 1.1.18-2ubuntu1.1

2020-03-05 Thread Rafael David Tinoco
** Summary changed:

- [bionic] fence_scsi not working properly with 1.1.18-2ubuntu1.1
+ [bionic] fence_scsi not working properly with Pacemaker 1.1.18-2ubuntu1.1

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1866119

Title:
  [bionic] fence_scsi not working properly with Pacemaker
  1.1.18-2ubuntu1.1

Status in pacemaker package in Ubuntu:
  Fix Released
Status in pacemaker source package in Bionic:
  In Progress

Bug description:
  OBS: This bug was originally into LP: #1865523 but it was split.

   SRU: pacemaker

  [Impact]

   * fence_scsi is not currently working in a share disk environment

   * all clusters relying in fence_scsi and/or fence_scsi + watchdog
  won't be able to start the fencing agents OR, in worst case scenarios,
  the fence_scsi agent might start but won't make scsi reservations in
  the shared scsi disk.

   * this bug is taking care of pacemaker 1.1.18 issues with fence_scsi,
  since the later was fixed at LP: #1865523.

  [Test Case]

   * having a 3-node setup, nodes called "clubionic01, clubionic02,
  clubionic03", with a shared scsi disk (fully supporting persistent
  reservations) /dev/sda, with corosync and pacemaker operational and
  running, one might try:

  rafaeldtinoco@clubionic01:~$ crm configure
  crm(live)configure# property stonith-enabled=on
  crm(live)configure# property stonith-action=off
  crm(live)configure# property no-quorum-policy=stop
  crm(live)configure# property have-watchdog=true
  crm(live)configure# commit
  crm(live)configure# end
  crm(live)# end

  rafaeldtinoco@clubionic01:~$ crm configure primitive fence_clubionic \
  stonith:fence_scsi params \
  pcmk_host_list="clubionic01 clubionic02 clubionic03" \
  devices="/dev/sda" \
  meta provides=unfencing

  And see the following errors:

  Failed Actions:
  * fence_clubionic_start_0 on clubionic02 'unknown error' (1): call=6, 
status=Error, exitreason='',
  last-rc-change='Wed Mar  4 19:53:12 2020', queued=0ms, exec=1105ms
  * fence_clubionic_start_0 on clubionic03 'unknown error' (1): call=6, 
status=Error, exitreason='',
  last-rc-change='Wed Mar  4 19:53:13 2020', queued=0ms, exec=1109ms
  * fence_clubionic_start_0 on clubionic01 'unknown error' (1): call=6, 
status=Error, exitreason='',
  last-rc-change='Wed Mar  4 19:53:11 2020', queued=0ms, exec=1108ms

  and corosync.log will show:

  warning: unpack_rsc_op_failure: Processing failed op start for
  fence_clubionic on clubionic01: unknown error (1)

  [Regression Potential]

   * LP: #1865523 shows fence_scsi fully operational after SRU for that
  bug is done.

   * LP: #1865523 used pacemaker 1.1.19 (vanilla) in order to fix
  fence_scsi.

   * TODO

  [Other Info]

   * Original Description:

  Trying to setup a cluster with an iscsi shared disk, using fence_scsi
  as the fencing mechanism, I realized that fence_scsi is not working in
  Ubuntu Bionic. I first thought it was related to Azure environment
  (LP: #1864419), where I was trying this environment, but then, trying
  locally, I figured out that somehow pacemaker 1.1.18 is not fencing
  the shared scsi disk properly.

  Note: I was able to "backport" vanilla 1.1.19 from upstream and
  fence_scsi worked. I have then tried 1.1.18 without all quilt patches
  and it didnt work as well. I think that bisecting 1.1.18 <-> 1.1.19
  might tell us which commit has fixed the behaviour needed by the
  fence_scsi agent.

  (k)rafaeldtinoco@clubionic01:~$ crm conf show
  node 1: clubionic01.private
  node 2: clubionic02.private
  node 3: clubionic03.private
  primitive fence_clubionic stonith:fence_scsi \
  params pcmk_host_list="10.250.3.10 10.250.3.11 10.250.3.12" 
devices="/dev/sda" \
  meta provides=unfencing
  property cib-bootstrap-options: \
  have-watchdog=false \
  dc-version=1.1.18-2b07d5c5a9 \
  cluster-infrastructure=corosync \
  cluster-name=clubionic \
  stonith-enabled=on \
  stonith-action=off \
  no-quorum-policy=stop \
  symmetric-cluster=true

  

  (k)rafaeldtinoco@clubionic02:~$ sudo crm_mon -1
  Stack: corosync
  Current DC: clubionic01.private (version 1.1.18-2b07d5c5a9) - partition with 
quorum
  Last updated: Mon Mar 2 15:55:30 2020
  Last change: Mon Mar 2 15:45:33 2020 by root via cibadmin on 
clubionic01.private

  3 nodes configured
  1 resource configured

  Online: [ clubionic01.private clubionic02.private clubionic03.private
  ]

  Active resources:

   fence_clubionic (stonith:fence_scsi): Started clubionic01.private

  

  (k)rafaeldtinoco@clubionic02:~$ sudo sg_persist --in --read-keys 
--device=/dev/sda
    LIO-ORG cluster.bionic. 4.0
    Peripheral device type: disk
    PR generation=0x0, there are NO registered reservation keys

  (k)rafaeldtinoco@clubionic02:~$ sudo sg_persist -r /dev/sda
    LIO-ORG 

[Ubuntu-ha] [Bug 1866119] Re: [bionic] fence_scsi not working properly with 1.1.18-2ubuntu1.1

2020-03-04 Thread Rafael David Tinoco
** Description changed:

+ OBS: This bug was originally into LP: #1865523 but it was split.
+ 
+  SRU: pacemaker
+ 
+ [Impact]
+ 
+  * fence_scsi is not currently working in a share disk environment
+ 
+  * all clusters relying in fence_scsi and/or fence_scsi + watchdog won't
+ be able to start the fencing agents OR, in worst case scenarios, the
+ fence_scsi agent might start but won't make scsi reservations in the
+ shared scsi disk.
+ 
+  * this bug is taking care of pacemaker 1.1.18 issues with fence_scsi,
+ since the later was fixed at LP: #1865523.
+ 
+ [Test Case]
+ 
+  * having a 3-node setup, nodes called "clubionic01, clubionic02,
+ clubionic03", with a shared scsi disk (fully supporting persistent
+ reservations) /dev/sda, with corosync and pacemaker operational and
+ running, one might try:
+ 
+ rafaeldtinoco@clubionic01:~$ crm configure
+ crm(live)configure# property stonith-enabled=on
+ crm(live)configure# property stonith-action=off
+ crm(live)configure# property no-quorum-policy=stop
+ crm(live)configure# property have-watchdog=true
+ crm(live)configure# commit
+ crm(live)configure# end
+ crm(live)# end
+ 
+ rafaeldtinoco@clubionic01:~$ crm configure primitive fence_clubionic \
+ stonith:fence_scsi params \
+ pcmk_host_list="clubionic01 clubionic02 clubionic03" \
+ devices="/dev/sda" \
+ meta provides=unfencing
+ 
+ And see the following errors:
+ 
+ Failed Actions:
+ * fence_clubionic_start_0 on clubionic02 'unknown error' (1): call=6, 
status=Error, exitreason='',
+ last-rc-change='Wed Mar  4 19:53:12 2020', queued=0ms, exec=1105ms
+ * fence_clubionic_start_0 on clubionic03 'unknown error' (1): call=6, 
status=Error, exitreason='',
+ last-rc-change='Wed Mar  4 19:53:13 2020', queued=0ms, exec=1109ms
+ * fence_clubionic_start_0 on clubionic01 'unknown error' (1): call=6, 
status=Error, exitreason='',
+ last-rc-change='Wed Mar  4 19:53:11 2020', queued=0ms, exec=1108ms
+ 
+ and corosync.log will show:
+ 
+ warning: unpack_rsc_op_failure: Processing failed op start for
+ fence_clubionic on clubionic01: unknown error (1)
+ 
+ [Regression Potential]
+ 
+  * LP: #1865523 shows fence_scsi fully operational after SRU for that
+ bug is done.
+ 
+  * LP: #1865523 used pacemaker 1.1.19 (vanilla) in order to fix
+ fence_scsi.
+ 
+  * TODO
+ 
+ [Other Info]
+ 
+  * Original Description:
+ 
  Trying to setup a cluster with an iscsi shared disk, using fence_scsi as
  the fencing mechanism, I realized that fence_scsi is not working in
  Ubuntu Bionic. I first thought it was related to Azure environment (LP:
  #1864419), where I was trying this environment, but then, trying
  locally, I figured out that somehow pacemaker 1.1.18 is not fencing the
  shared scsi disk properly.
  
  Note: I was able to "backport" vanilla 1.1.19 from upstream and
  fence_scsi worked. I have then tried 1.1.18 without all quilt patches
  and it didnt work as well. I think that bisecting 1.1.18 <-> 1.1.19
  might tell us which commit has fixed the behaviour needed by the
  fence_scsi agent.
  
  (k)rafaeldtinoco@clubionic01:~$ crm conf show
  node 1: clubionic01.private
  node 2: clubionic02.private
  node 3: clubionic03.private
  primitive fence_clubionic stonith:fence_scsi \
- params pcmk_host_list="10.250.3.10 10.250.3.11 10.250.3.12" 
devices="/dev/sda" \
- meta provides=unfencing
+ params pcmk_host_list="10.250.3.10 10.250.3.11 10.250.3.12" 
devices="/dev/sda" \
+ meta provides=unfencing
  property cib-bootstrap-options: \
- have-watchdog=false \
- dc-version=1.1.18-2b07d5c5a9 \
- cluster-infrastructure=corosync \
- cluster-name=clubionic \
- stonith-enabled=on \
- stonith-action=off \
- no-quorum-policy=stop \
- symmetric-cluster=true
+ have-watchdog=false \
+ dc-version=1.1.18-2b07d5c5a9 \
+ cluster-infrastructure=corosync \
+ cluster-name=clubionic \
+ stonith-enabled=on \
+ stonith-action=off \
+ no-quorum-policy=stop \
+ symmetric-cluster=true
  
  
  
  (k)rafaeldtinoco@clubionic02:~$ sudo crm_mon -1
  Stack: corosync
  Current DC: clubionic01.private (version 1.1.18-2b07d5c5a9) - partition with 
quorum
  Last updated: Mon Mar 2 15:55:30 2020
  Last change: Mon Mar 2 15:45:33 2020 by root via cibadmin on 
clubionic01.private
  
  3 nodes configured
  1 resource configured
  
  Online: [ clubionic01.private clubionic02.private clubionic03.private ]
  
  Active resources:
  
-  fence_clubionic (stonith:fence_scsi): Started clubionic01.private
+  fence_clubionic (stonith:fence_scsi): Started clubionic01.private
  
  
  
  (k)rafaeldtinoco@clubionic02:~$ sudo sg_persist --in --read-keys 
--device=/dev/sda
-   LIO-ORG cluster.bionic. 4.0
-   Peripheral device type: disk
-   PR generation=0x0, there are NO registered reservation keys
+   LIO-ORG cluster.bionic. 4.0
+   Peripheral device type: disk
+   PR generation=0x0, there 

[Ubuntu-ha] [Bug 1865523] Re: [bionic] fence_scsi not working properly with 1.1.18-2ubuntu1.1

2020-03-04 Thread Rafael David Tinoco
** No longer affects: pacemaker (Ubuntu)

** No longer affects: pacemaker (Ubuntu Bionic)

** Description changed:

+ OBS: I have split this bug into 2 bugs:
+  - fence-agents (this) and pacemaker (LP: #1866119)
+ 
   SRU: fence-agents
  
  [Impact]
  
   * fence_scsi is not currently working in a share disk environment
  
   * all clusters relying in fence_scsi and/or fence_scsi + watchdog won't
  be able to start the fencing agents OR, in worst case scenarios, the
  fence_scsi agent might start but won't make scsi reservations in the
  shared scsi disk.
  
  [Test Case]
  
   * having a 3-node setup, nodes called "clubionic01, clubionic02,
  clubionic03", with a shared scsi disk (fully supporting persistent
  reservations) /dev/sda, one might try the following command:
  
  sudo fence_scsi --verbose -n clubionic01 -d /dev/sda -k 3abe -o off
  
  from nodes "clubionic02 or clubionic03" and check if the reservation
  worked:
  
  (k)rafaeldtinoco@clubionic02:~$ sudo sg_persist --in --read-keys 
--device=/dev/sda
    LIO-ORG   cluster.bionic.   4.0
    Peripheral device type: disk
    PR generation=0x0, there are NO registered reservation keys
  
  (k)rafaeldtinoco@clubionic02:~$ sudo sg_persist -r /dev/sda
    LIO-ORG   cluster.bionic.   4.0
    Peripheral device type: disk
    PR generation=0x0, there is NO reservation held
  
   * having a 3-node setup, nodes called "clubionic01, clubionic02,
  clubionic03", with a shared scsi disk (fully supporting persistent
  reservations) /dev/sda, with corosync and pacemaker operational and
  running, one might try:
  
  rafaeldtinoco@clubionic01:~$ crm configure
  crm(live)configure# property stonith-enabled=on
  crm(live)configure# property stonith-action=off
  crm(live)configure# property no-quorum-policy=stop
  crm(live)configure# property have-watchdog=true
  crm(live)configure# property symmetric-cluster=true
  crm(live)configure# commit
  crm(live)configure# end
  crm(live)# end
  
  rafaeldtinoco@clubionic01:~$ crm configure primitive fence_clubionic \
  stonith:fence_scsi params \
  pcmk_host_list="clubionic01 clubionic02 clubionic03" \
  devices="/dev/sda" \
  meta provides=unfencing
  
  And see that crm_mon won't show fence_clubionic resource operational.
  
  [Regression Potential]
  
-  * Comments #3 and #4 show this new version fully working.
-  
-  * This fix has a potential of breaking other "nowadays working" fencing 
agent. If that happens, I suggest that ones affected revert previous to 
previous package AND open a bug against either pacemaker and/or fence-agents.
+  * Comments #3 and #4 show this new version fully working.
+ 
+  * This fix has a potential of breaking other "nowadays working" fencing
+ agent. If that happens, I suggest that ones affected revert previous to
+ previous package AND open a bug against either pacemaker and/or fence-
+ agents.
  
   * Judging by this issue, it is very likely that any Ubuntu user that
  have tried using fence_scsi has probably migrated to a newer version
  because fence_scsi agent is broken since its release.
  
   * The way I fixed fence_scsi was this:
  
  I packaged pacemaker in latest 1.1.X version and kept it "vanilla" so I
  could bisect fence-agents. At that moment I realized that bisecting was
  going to be hard because there were multiple issues, not only one. I
  backported the latest fence-agents together with Pacemaker 1.1.19-0 and
  saw that it worked.
  
  From then on, I bisected the following intervals:
  
  4.3.0 .. 4.4.0 (eoan - working)
  4.2.0 .. 4.3.0
  4.1.0 .. 4.2.0
  4.0.25 .. 4.1.0 (bionic - not working)
  
  In each of those intervals I discovered issues. For example, Using 4.3.0
  I faced problems so I had to backport fixes that were in between 4.4.0
  and 4.3.0. Then, backporting 4.2.0, I faced issues so I had to backport
  fixes from the 4.3.0 <-> 4.2.0 interval. I did this until I was at
  4.0.25 version, current Bionic fence-agents version.
  
  [Other Info]
  
   * Original Description:
  
  Trying to setup a cluster with an iscsi shared disk, using fence_scsi as
  the fencing mechanism, I realized that fence_scsi is not working in
  Ubuntu Bionic. I first thought it was related to Azure environment (LP:
  #1864419), where I was trying this environment, but then, trying
  locally, I figured out that somehow pacemaker 1.1.18 is not fencing the
  shared scsi disk properly.
  
  Note: I was able to "backport" vanilla 1.1.19 from upstream and
  fence_scsi worked. I have then tried 1.1.18 without all quilt patches
  and it didnt work as well. I think that bisecting 1.1.18 <-> 1.1.19
  might tell us which commit has fixed the behaviour needed by the
  fence_scsi agent.
  
  (k)rafaeldtinoco@clubionic01:~$ crm conf show
  node 1: clubionic01.private
  node 2: clubionic02.private
  node 3: clubionic03.private
  primitive fence_clubionic stonith:fence_scsi \
  params pcmk_host_list="10.250.3.10 10.250.3.11 10.250.3.12" 
devices="/dev/sda" \
    

[Ubuntu-ha] [Bug 1866119] [NEW] [bionic] fence_scsi not working properly with 1.1.18-2ubuntu1.1

2020-03-04 Thread Rafael David Tinoco
Public bug reported:

OBS: This bug was originally into LP: #1865523 but it was split.

 SRU: pacemaker

[Impact]

 * fence_scsi is not currently working in a share disk environment

 * all clusters relying in fence_scsi and/or fence_scsi + watchdog won't
be able to start the fencing agents OR, in worst case scenarios, the
fence_scsi agent might start but won't make scsi reservations in the
shared scsi disk.

 * this bug is taking care of pacemaker 1.1.18 issues with fence_scsi,
since the later was fixed at LP: #1865523.

[Test Case]

 * having a 3-node setup, nodes called "clubionic01, clubionic02,
clubionic03", with a shared scsi disk (fully supporting persistent
reservations) /dev/sda, with corosync and pacemaker operational and
running, one might try:

rafaeldtinoco@clubionic01:~$ crm configure
crm(live)configure# property stonith-enabled=on
crm(live)configure# property stonith-action=off
crm(live)configure# property no-quorum-policy=stop
crm(live)configure# property have-watchdog=true
crm(live)configure# commit
crm(live)configure# end
crm(live)# end

rafaeldtinoco@clubionic01:~$ crm configure primitive fence_clubionic \
stonith:fence_scsi params \
pcmk_host_list="clubionic01 clubionic02 clubionic03" \
devices="/dev/sda" \
meta provides=unfencing

And see the following errors:

Failed Actions:
* fence_clubionic_start_0 on clubionic02 'unknown error' (1): call=6, 
status=Error, exitreason='',
last-rc-change='Wed Mar  4 19:53:12 2020', queued=0ms, exec=1105ms
* fence_clubionic_start_0 on clubionic03 'unknown error' (1): call=6, 
status=Error, exitreason='',
last-rc-change='Wed Mar  4 19:53:13 2020', queued=0ms, exec=1109ms
* fence_clubionic_start_0 on clubionic01 'unknown error' (1): call=6, 
status=Error, exitreason='',
last-rc-change='Wed Mar  4 19:53:11 2020', queued=0ms, exec=1108ms

and corosync.log will show:

warning: unpack_rsc_op_failure: Processing failed op start for
fence_clubionic on clubionic01: unknown error (1)

[Regression Potential]

 * LP: #1865523 shows fence_scsi fully operational after SRU for that
bug is done.

 * LP: #1865523 used pacemaker 1.1.19 (vanilla) in order to fix
fence_scsi.

 * TODO

[Other Info]

 * Original Description:

Trying to setup a cluster with an iscsi shared disk, using fence_scsi as
the fencing mechanism, I realized that fence_scsi is not working in
Ubuntu Bionic. I first thought it was related to Azure environment (LP:
#1864419), where I was trying this environment, but then, trying
locally, I figured out that somehow pacemaker 1.1.18 is not fencing the
shared scsi disk properly.

Note: I was able to "backport" vanilla 1.1.19 from upstream and
fence_scsi worked. I have then tried 1.1.18 without all quilt patches
and it didnt work as well. I think that bisecting 1.1.18 <-> 1.1.19
might tell us which commit has fixed the behaviour needed by the
fence_scsi agent.

(k)rafaeldtinoco@clubionic01:~$ crm conf show
node 1: clubionic01.private
node 2: clubionic02.private
node 3: clubionic03.private
primitive fence_clubionic stonith:fence_scsi \
params pcmk_host_list="10.250.3.10 10.250.3.11 10.250.3.12" 
devices="/dev/sda" \
meta provides=unfencing
property cib-bootstrap-options: \
have-watchdog=false \
dc-version=1.1.18-2b07d5c5a9 \
cluster-infrastructure=corosync \
cluster-name=clubionic \
stonith-enabled=on \
stonith-action=off \
no-quorum-policy=stop \
symmetric-cluster=true



(k)rafaeldtinoco@clubionic02:~$ sudo crm_mon -1
Stack: corosync
Current DC: clubionic01.private (version 1.1.18-2b07d5c5a9) - partition with 
quorum
Last updated: Mon Mar 2 15:55:30 2020
Last change: Mon Mar 2 15:45:33 2020 by root via cibadmin on clubionic01.private

3 nodes configured
1 resource configured

Online: [ clubionic01.private clubionic02.private clubionic03.private ]

Active resources:

 fence_clubionic (stonith:fence_scsi): Started clubionic01.private



(k)rafaeldtinoco@clubionic02:~$ sudo sg_persist --in --read-keys 
--device=/dev/sda
  LIO-ORG cluster.bionic. 4.0
  Peripheral device type: disk
  PR generation=0x0, there are NO registered reservation keys

(k)rafaeldtinoco@clubionic02:~$ sudo sg_persist -r /dev/sda
  LIO-ORG cluster.bionic. 4.0
  Peripheral device type: disk
  PR generation=0x0, there is NO reservation held

** Affects: pacemaker (Ubuntu)
 Importance: Undecided
 Status: Fix Released

** Affects: pacemaker (Ubuntu Bionic)
 Importance: Undecided
 Status: In Progress

** Also affects: pacemaker (Ubuntu Bionic)
   Importance: Undecided
   Status: New

** Changed in: pacemaker (Ubuntu)
   Status: New => Fix Released

** Changed in: pacemaker (Ubuntu Bionic)
   Status: New => Confirmed

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1866119

Title:
  [bionic] fence_scsi not working 

[Ubuntu-ha] [Bug 1865523] Re: [bionic] fence_scsi not working properly with 1.1.18-2ubuntu1.1

2020-03-04 Thread Rafael David Tinoco
** Description changed:

+  SRU: fence-agents
+ 
+ [Impact]
+ 
+  * fence_scsi is not currently working in a share disk environment
+ 
+  * all clusters relying in fence_scsi and/or fence_scsi + watchdog won't
+ be able to start the fencing agents OR, in worst case scenarios, the
+ fence_scsi agent might start but won't make scsi reservations in the
+ shared scsi disk.
+ 
+ [Test Case]
+ 
+  * having a 3-node setup, nodes called "clubionic01, clubionic02,
+ clubionic03", with a shared scsi disk (fully supporting persistent
+ reservations) /dev/sda, one might try the following command:
+ 
+ sudo fence_scsi --verbose -n clubionic01 -d /dev/sda -k 3abe -o off
+ 
+ from nodes "clubionic02 or clubionic03" and check if the reservation
+ worked:
+ 
+ (k)rafaeldtinoco@clubionic02:~$ sudo sg_persist --in --read-keys 
--device=/dev/sda
+   LIO-ORG   cluster.bionic.   4.0
+   Peripheral device type: disk
+   PR generation=0x0, there are NO registered reservation keys
+ 
+ (k)rafaeldtinoco@clubionic02:~$ sudo sg_persist -r /dev/sda
+   LIO-ORG   cluster.bionic.   4.0
+   Peripheral device type: disk
+   PR generation=0x0, there is NO reservation held
+ 
+  * having a 3-node setup, nodes called "clubionic01, clubionic02,
+ clubionic03", with a shared scsi disk (fully supporting persistent
+ reservations) /dev/sda, with corosync and pacemaker operational and
+ running, one might try:
+ 
+ rafaeldtinoco@clubionic01:~$ crm configure
+ crm(live)configure# property stonith-enabled=on
+ crm(live)configure# property stonith-action=off
+ crm(live)configure# property no-quorum-policy=stop
+ crm(live)configure# property have-watchdog=true
+ crm(live)configure# property symmetric-cluster=true
+ crm(live)configure# commit
+ crm(live)configure# end
+ crm(live)# end
+ 
+ rafaeldtinoco@clubionic01:~$ crm configure primitive fence_clubionic \
+ stonith:fence_scsi params \
+ pcmk_host_list="clubionic01 clubionic02 clubionic03" \
+ devices="/dev/sda" \
+ meta provides=unfencing
+ 
+ And see that crm_mon won't show fence_clubionic resource operational.
+ 
+ [Regression Potential]
+ 
+  * Judging by this issue, it is very likely that any Ubuntu user that
+ have tried using fence_scsi has probably migrated to a newer version
+ because fence_scsi agent is broken since its release.
+ 
+  * The way I fixed fence_scsi was this:
+ 
+ I packaged pacemaker in latest 1.1.X version and kept it "vanilla" so I
+ could bisect fence-agents. At that moment I realized that bisecting was
+ going to be hard because there were multiple issues, not only one. I
+ backported the latest fence-agents together with Pacemaker 1.1.19-0 and
+ saw that it worked.
+ 
+ From then on, I bisected the following intervals:
+ 
+ 4.3.0 .. 4.4.0 (eoan - working)
+ 4.2.0 .. 4.3.0
+ 4.1.0 .. 4.2.0
+ 4.0.25 .. 4.1.0 (bionic - not working)
+ 
+ In each of those intervals I discovered issues. For example, Using 4.3.0
+ I faced problems so I had to backport fixes that were in between 4.4.0
+ and 4.3.0. Then, backporting 4.2.0, I faced issues so I had to backport
+ fixes from the 4.3.0 <-> 4.2.0 interval. I did this until I was at
+ 4.0.25 version, current Bionic fence-agents version.
+ 
+ [Other Info]
+  
+  * Original Description:
+ 
  Trying to setup a cluster with an iscsi shared disk, using fence_scsi as
  the fencing mechanism, I realized that fence_scsi is not working in
  Ubuntu Bionic. I first thought it was related to Azure environment (LP:
  #1864419), where I was trying this environment, but then, trying
  locally, I figured out that somehow pacemaker 1.1.18 is not fencing the
  shared scsi disk properly.
  
  Note: I was able to "backport" vanilla 1.1.19 from upstream and
  fence_scsi worked. I have then tried 1.1.18 without all quilt patches
  and it didnt work as well. I think that bisecting 1.1.18 <-> 1.1.19
  might tell us which commit has fixed the behaviour needed by the
  fence_scsi agent.
  
  (k)rafaeldtinoco@clubionic01:~$ crm conf show
  node 1: clubionic01.private
  node 2: clubionic02.private
  node 3: clubionic03.private
  primitive fence_clubionic stonith:fence_scsi \
- params pcmk_host_list="10.250.3.10 10.250.3.11 10.250.3.12" 
devices="/dev/sda" \
- meta provides=unfencing
+ params pcmk_host_list="10.250.3.10 10.250.3.11 10.250.3.12" 
devices="/dev/sda" \
+ meta provides=unfencing
  property cib-bootstrap-options: \
- have-watchdog=false \
- dc-version=1.1.18-2b07d5c5a9 \
- cluster-infrastructure=corosync \
- cluster-name=clubionic \
- stonith-enabled=on \
- stonith-action=off \
- no-quorum-policy=stop \
- symmetric-cluster=true
+ have-watchdog=false \
+ dc-version=1.1.18-2b07d5c5a9 \
+ cluster-infrastructure=corosync \
+ cluster-name=clubionic \
+ stonith-enabled=on \
+ stonith-action=off \
+ no-quorum-policy=stop \
+ symmetric-cluster=true
  
  
  
  

[Ubuntu-ha] [Bug 1865523] Re: [bionic] fence_scsi not working properly with 1.1.18-2ubuntu1.1

2020-03-03 Thread Rafael David Tinoco
# For Ubuntu Bionic:

Okay, after bisecting fence-scsi and monitoring all its functions I was
able to isolate the patches that I need to take to bionic to make it
compatible with existing version and, at the same time, operational:

Note: all tests were conducted with Pacemaker v1.1.19-0ubuntu1 and this
is not the default in Ubuntu Bionic. I have maintained "vanilla"
Pacemaker v1.1.19 in order to better isolate all fixes for fence-agents.
Now I'm able to create a fixed fence-agent package for Ubuntu Bionic AND
fix Pacemaker.

# Ubuntu Bionic SRU: Fence Agents v4.0.25 PLUS the following
fixes/commits ordered by date:

commit 81b8370844f5aecaee5e7178d82670c70399d824
Author: Oyvind Albrigtsen 
Date:   Mon Jul 24 14:12:15 2017

fence_scsi: add FIPS support

commit eae9d029b7073e7eb8c7ba4df9ec19b755a8f603
Author: Oyvind Albrigtsen 
Date:   Wed Sep 27 12:26:38 2017

fix for ignored options

commit c6f29a653114523e9ac3644aed958b4bb43f3b41
Author: Oyvind Albrigtsen 
Date:   Wed Sep 27 12:42:39 2017

Maintain ABI compatibility for external agents

commit 746fd55b061aa28b27aac5a1bb38714a95812592
Author: Reid Wahl 
Date:   Fri Apr 6 18:31:30 2018

Low: fence_scsi: Remove period from cmd string

commit bec154345d2291c9051c16277de9054387dc9707
Author: Oyvind Albrigtsen 
Date:   Thu Apr 19 11:30:53 2018

fence_scsi: fix plug-parameter and keep support for nodename to
avoid regressions

commit 335aca4e54e4ec46b9b5d86ef30a7d9348e6a216
Author: Valentin Vidic 
Date:   Wed May 23 12:51:23 2018

fence_scsi: fix python3 encoding error #206

commit f77297b654586bf539e78957f26cae1d22c6f081
Author: Oyvind Albrigtsen 
Date:   Fri Nov 2 08:24:56 2018

fence_scsi: fix incorrect SCSI key when node ID is 10 or higher

  The last four digits of the SCSI key will be zero padded digit
between -0009.

commit 1c4a64ca803831b44c96c75022abe5bb8713cd1a
Author: Oyvind Albrigtsen 
Date:   Wed May 22 08:13:34 2019

fence_scsi: detect node ID using new format, and fallback to old format
before failing

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1865523

Title:
  [bionic] fence_scsi not working properly with 1.1.18-2ubuntu1.1

Status in fence-agents package in Ubuntu:
  Fix Released
Status in pacemaker package in Ubuntu:
  Fix Released
Status in fence-agents source package in Bionic:
  Confirmed
Status in pacemaker source package in Bionic:
  Confirmed
Status in fence-agents source package in Disco:
  Confirmed
Status in fence-agents source package in Eoan:
  Fix Released
Status in fence-agents source package in Focal:
  Fix Released

Bug description:
  Trying to setup a cluster with an iscsi shared disk, using fence_scsi
  as the fencing mechanism, I realized that fence_scsi is not working in
  Ubuntu Bionic. I first thought it was related to Azure environment
  (LP: #1864419), where I was trying this environment, but then, trying
  locally, I figured out that somehow pacemaker 1.1.18 is not fencing
  the shared scsi disk properly.

  Note: I was able to "backport" vanilla 1.1.19 from upstream and
  fence_scsi worked. I have then tried 1.1.18 without all quilt patches
  and it didnt work as well. I think that bisecting 1.1.18 <-> 1.1.19
  might tell us which commit has fixed the behaviour needed by the
  fence_scsi agent.

  (k)rafaeldtinoco@clubionic01:~$ crm conf show
  node 1: clubionic01.private
  node 2: clubionic02.private
  node 3: clubionic03.private
  primitive fence_clubionic stonith:fence_scsi \
  params pcmk_host_list="10.250.3.10 10.250.3.11 10.250.3.12" 
devices="/dev/sda" \
  meta provides=unfencing
  property cib-bootstrap-options: \
  have-watchdog=false \
  dc-version=1.1.18-2b07d5c5a9 \
  cluster-infrastructure=corosync \
  cluster-name=clubionic \
  stonith-enabled=on \
  stonith-action=off \
  no-quorum-policy=stop \
  symmetric-cluster=true

  

  (k)rafaeldtinoco@clubionic02:~$ sudo crm_mon -1
  Stack: corosync
  Current DC: clubionic01.private (version 1.1.18-2b07d5c5a9) - partition with 
quorum
  Last updated: Mon Mar  2 15:55:30 2020
  Last change: Mon Mar  2 15:45:33 2020 by root via cibadmin on 
clubionic01.private

  3 nodes configured
  1 resource configured

  Online: [ clubionic01.private clubionic02.private clubionic03.private
  ]

  Active resources:

   fence_clubionic(stonith:fence_scsi):   Started
  clubionic01.private

  

  (k)rafaeldtinoco@clubionic02:~$ sudo sg_persist --in --read-keys 
--device=/dev/sda
LIO-ORG   cluster.bionic.   4.0
Peripheral device type: disk
PR generation=0x0, there are NO registered reservation keys

  (k)rafaeldtinoco@clubionic02:~$ sudo sg_persist -r /dev/sda
LIO-ORG   cluster.bionic.   4.0
Peripheral device type: disk
PR generation=0x0, there is NO reservation held

To manage notifications 

[Ubuntu-ha] [Bug 1865523] Re: [bionic] fence_scsi not working properly with 1.1.18-2ubuntu1.1

2020-03-03 Thread Rafael David Tinoco
# Demonstration of fence_scsi agent working in Bionic:

(k)rafaeldtinoco@clubionic01:~/.../upstream$ sudo dpkg -i ./*.deb
dpkg: warning: downgrading fence-agents from 4.4.0-2 to 4.0.25-2ubuntu1
(Reading database ... 85434 files and directories currently installed.)
Preparing to unpack .../fence-agents_4.0.25-2ubuntu1_amd64.deb ...
Unpacking fence-agents (4.0.25-2ubuntu1) over (4.4.0-2) ...
Preparing to unpack .../fence-agents_4.4.0-2_amd64.deb ...
Unpacking fence-agents (4.4.0-2) over (4.0.25-2ubuntu1) ...
Setting up fence-agents (4.4.0-2) ...
Processing triggers for man-db (2.8.3-2ubuntu0.1) ...

(k)rafaeldtinoco@clubionic01:~/.../upstream$ sudo sg_persist --in --read-keys 
--device=/dev/sda
  LIO-ORG   cluster.bionic.   4.0
  Peripheral device type: disk
  PR generation=0x0, there are NO registered reservation keys

(k)rafaeldtinoco@clubionic01:~/.../upstream$ systemctl restart pacemaker

(k)rafaeldtinoco@clubionic02:~/.../upstream$ crm_mon -1
Stack: corosync
Current DC: clubionic03.private (version 1.1.19-1.1.19) - partition with quorum
Last updated: Tue Mar  3 21:16:04 2020
Last change: Tue Mar  3 20:25:56 2020 by root via cibadmin on 
clubionic01.private

3 nodes configured
1 resource configured

Online: [ clubionic01.private clubionic02.private clubionic03.private ]

Active resources:

 fence_clubionic(stonith:fence_scsi):   Started
clubionic01.private

(k)rafaeldtinoco@clubionic01:~/.../upstream$ sudo sg_persist --in --read-keys 
--device=/dev/sda
  LIO-ORG   cluster.bionic.   4.0
  Peripheral device type: disk
  PR generation=0x3, 3 registered reservation keys follow:
0x3abe
0x3abe0001
0x3abe0002

(k)rafaeldtinoco@clubionic01:~/.../upstream$ sudo sg_persist -r /dev/sda
  LIO-ORG   cluster.bionic.   4.0
  Peripheral device type: disk
  PR generation=0x3, Reservation follows:
Key=0x3abe0001
scope: LU_SCOPE,  type: Write Exclusive, registrants only

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1865523

Title:
  [bionic] fence_scsi not working properly with 1.1.18-2ubuntu1.1

Status in fence-agents package in Ubuntu:
  Fix Released
Status in pacemaker package in Ubuntu:
  Fix Released
Status in fence-agents source package in Bionic:
  Confirmed
Status in pacemaker source package in Bionic:
  Confirmed
Status in fence-agents source package in Disco:
  Confirmed
Status in fence-agents source package in Eoan:
  Fix Released
Status in fence-agents source package in Focal:
  Fix Released

Bug description:
  Trying to setup a cluster with an iscsi shared disk, using fence_scsi
  as the fencing mechanism, I realized that fence_scsi is not working in
  Ubuntu Bionic. I first thought it was related to Azure environment
  (LP: #1864419), where I was trying this environment, but then, trying
  locally, I figured out that somehow pacemaker 1.1.18 is not fencing
  the shared scsi disk properly.

  Note: I was able to "backport" vanilla 1.1.19 from upstream and
  fence_scsi worked. I have then tried 1.1.18 without all quilt patches
  and it didnt work as well. I think that bisecting 1.1.18 <-> 1.1.19
  might tell us which commit has fixed the behaviour needed by the
  fence_scsi agent.

  (k)rafaeldtinoco@clubionic01:~$ crm conf show
  node 1: clubionic01.private
  node 2: clubionic02.private
  node 3: clubionic03.private
  primitive fence_clubionic stonith:fence_scsi \
  params pcmk_host_list="10.250.3.10 10.250.3.11 10.250.3.12" 
devices="/dev/sda" \
  meta provides=unfencing
  property cib-bootstrap-options: \
  have-watchdog=false \
  dc-version=1.1.18-2b07d5c5a9 \
  cluster-infrastructure=corosync \
  cluster-name=clubionic \
  stonith-enabled=on \
  stonith-action=off \
  no-quorum-policy=stop \
  symmetric-cluster=true

  

  (k)rafaeldtinoco@clubionic02:~$ sudo crm_mon -1
  Stack: corosync
  Current DC: clubionic01.private (version 1.1.18-2b07d5c5a9) - partition with 
quorum
  Last updated: Mon Mar  2 15:55:30 2020
  Last change: Mon Mar  2 15:45:33 2020 by root via cibadmin on 
clubionic01.private

  3 nodes configured
  1 resource configured

  Online: [ clubionic01.private clubionic02.private clubionic03.private
  ]

  Active resources:

   fence_clubionic(stonith:fence_scsi):   Started
  clubionic01.private

  

  (k)rafaeldtinoco@clubionic02:~$ sudo sg_persist --in --read-keys 
--device=/dev/sda
LIO-ORG   cluster.bionic.   4.0
Peripheral device type: disk
PR generation=0x0, there are NO registered reservation keys

  (k)rafaeldtinoco@clubionic02:~$ sudo sg_persist -r /dev/sda
LIO-ORG   cluster.bionic.   4.0
Peripheral device type: disk
PR generation=0x0, there is NO reservation held

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/fence-agents/+bug/1865523/+subscriptions


[Ubuntu-ha] [Bug 1865523] Re: [bionic] fence_scsi not working properly with 1.1.18-2ubuntu1.1

2020-03-03 Thread Rafael David Tinoco
Now I'm going to work with this package and check needed Pacemaker
fixes. After that I'm going to propose both merges together.

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1865523

Title:
  [bionic] fence_scsi not working properly with 1.1.18-2ubuntu1.1

Status in fence-agents package in Ubuntu:
  Fix Released
Status in pacemaker package in Ubuntu:
  Fix Released
Status in fence-agents source package in Bionic:
  Confirmed
Status in pacemaker source package in Bionic:
  Confirmed
Status in fence-agents source package in Disco:
  Confirmed
Status in fence-agents source package in Eoan:
  Fix Released
Status in fence-agents source package in Focal:
  Fix Released

Bug description:
  Trying to setup a cluster with an iscsi shared disk, using fence_scsi
  as the fencing mechanism, I realized that fence_scsi is not working in
  Ubuntu Bionic. I first thought it was related to Azure environment
  (LP: #1864419), where I was trying this environment, but then, trying
  locally, I figured out that somehow pacemaker 1.1.18 is not fencing
  the shared scsi disk properly.

  Note: I was able to "backport" vanilla 1.1.19 from upstream and
  fence_scsi worked. I have then tried 1.1.18 without all quilt patches
  and it didnt work as well. I think that bisecting 1.1.18 <-> 1.1.19
  might tell us which commit has fixed the behaviour needed by the
  fence_scsi agent.

  (k)rafaeldtinoco@clubionic01:~$ crm conf show
  node 1: clubionic01.private
  node 2: clubionic02.private
  node 3: clubionic03.private
  primitive fence_clubionic stonith:fence_scsi \
  params pcmk_host_list="10.250.3.10 10.250.3.11 10.250.3.12" 
devices="/dev/sda" \
  meta provides=unfencing
  property cib-bootstrap-options: \
  have-watchdog=false \
  dc-version=1.1.18-2b07d5c5a9 \
  cluster-infrastructure=corosync \
  cluster-name=clubionic \
  stonith-enabled=on \
  stonith-action=off \
  no-quorum-policy=stop \
  symmetric-cluster=true

  

  (k)rafaeldtinoco@clubionic02:~$ sudo crm_mon -1
  Stack: corosync
  Current DC: clubionic01.private (version 1.1.18-2b07d5c5a9) - partition with 
quorum
  Last updated: Mon Mar  2 15:55:30 2020
  Last change: Mon Mar  2 15:45:33 2020 by root via cibadmin on 
clubionic01.private

  3 nodes configured
  1 resource configured

  Online: [ clubionic01.private clubionic02.private clubionic03.private
  ]

  Active resources:

   fence_clubionic(stonith:fence_scsi):   Started
  clubionic01.private

  

  (k)rafaeldtinoco@clubionic02:~$ sudo sg_persist --in --read-keys 
--device=/dev/sda
LIO-ORG   cluster.bionic.   4.0
Peripheral device type: disk
PR generation=0x0, there are NO registered reservation keys

  (k)rafaeldtinoco@clubionic02:~$ sudo sg_persist -r /dev/sda
LIO-ORG   cluster.bionic.   4.0
Peripheral device type: disk
PR generation=0x0, there is NO reservation held

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/fence-agents/+bug/1865523/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1865523] Re: [bionic] fence_scsi not working properly with 1.1.18-2ubuntu1.1

2020-03-03 Thread Rafael David Tinoco
# Demonstration of fence_scsi fencing a node:

(k)rafaeldtinoco@clubionic03:~/.../upstream$ cat /etc/network/interfaces
auto lo
iface lo inet loopback

auto eth0
iface eth0 inet dhcp

# iscsi
auto eth1
iface eth1 inet static
address 10.250.1.12/24

# private
auto eth2
iface eth2 inet static
address 10.250.3.12/24

# public
auto eth3
iface eth3 inet static
address 10.250.98.12/24

(k)rafaeldtinoco@clubionic03:~/.../upstream$ sudo iptables -A INPUT -i
eth2 -j DROP

(k)rafaeldtinoco@clubionic01:~$ crm_mon -1
Stack: corosync
Current DC: clubionic01.private (version 1.1.19-1.1.19) - partition with quorum
Last updated: Tue Mar  3 21:24:55 2020
Last change: Tue Mar  3 20:25:56 2020 by root via cibadmin on 
clubionic01.private

3 nodes configured
1 resource configured

Online: [ clubionic01.private clubionic02.private ]
OFFLINE: [ clubionic03.private ]

Active resources:

 fence_clubionic(stonith:fence_scsi):   Started
clubionic01.private


(k)rafaeldtinoco@clubionic03:~/.../upstream$ sudo sg_persist --in --read-keys 
--device=/dev/sda
  LIO-ORG   cluster.bionic.   4.0
  Peripheral device type: disk
  PR generation=0x4, 2 registered reservation keys follow:
0x3abe
0x3abe0001

(k)rafaeldtinoco@clubionic03:~/.../upstream$ sudo dd if=/dev/zero of=/dev/sda 
bs=1M count=1
[ 3301.867294] print_req_error: critical nexus error, dev sda, sector 0
[ 3301.868543] Buffer I/O error on dev sda, logical block 0, lost async page 
write
[ 3301.869956] Buffer I/O error on dev sda, logical block 1, lost async page 
write
[ 3301.871430] Buffer I/O error on dev sda, logical block 2, lost async page 
write
[ 3301.872929] Buffer I/O error on dev sda, logical block 3, lost async page 
write
[ 3301.874448] Buffer I/O error on dev sda, logical block 4, lost async page 
write
[ 3301.875963] Buffer I/O error on dev sda, logical block 5, lost async page 
write
[ 3301.877486] Buffer I/O error on dev sda, logical block 6, lost async page 
write
[ 3301.879000] Buffer I/O error on dev sda, logical block 7, lost async page 
write
[ 3301.880481] Buffer I/O error on dev sda, logical block 8, lost async page 
write
[ 3301.882014] Buffer I/O error on dev sda, logical block 9, lost async page 
write
1+0 records in
1+0 records out
1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0227557 s, 46.1 MB/s

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1865523

Title:
  [bionic] fence_scsi not working properly with 1.1.18-2ubuntu1.1

Status in fence-agents package in Ubuntu:
  Fix Released
Status in pacemaker package in Ubuntu:
  Fix Released
Status in fence-agents source package in Bionic:
  Confirmed
Status in pacemaker source package in Bionic:
  Confirmed
Status in fence-agents source package in Disco:
  Confirmed
Status in fence-agents source package in Eoan:
  Fix Released
Status in fence-agents source package in Focal:
  Fix Released

Bug description:
  Trying to setup a cluster with an iscsi shared disk, using fence_scsi
  as the fencing mechanism, I realized that fence_scsi is not working in
  Ubuntu Bionic. I first thought it was related to Azure environment
  (LP: #1864419), where I was trying this environment, but then, trying
  locally, I figured out that somehow pacemaker 1.1.18 is not fencing
  the shared scsi disk properly.

  Note: I was able to "backport" vanilla 1.1.19 from upstream and
  fence_scsi worked. I have then tried 1.1.18 without all quilt patches
  and it didnt work as well. I think that bisecting 1.1.18 <-> 1.1.19
  might tell us which commit has fixed the behaviour needed by the
  fence_scsi agent.

  (k)rafaeldtinoco@clubionic01:~$ crm conf show
  node 1: clubionic01.private
  node 2: clubionic02.private
  node 3: clubionic03.private
  primitive fence_clubionic stonith:fence_scsi \
  params pcmk_host_list="10.250.3.10 10.250.3.11 10.250.3.12" 
devices="/dev/sda" \
  meta provides=unfencing
  property cib-bootstrap-options: \
  have-watchdog=false \
  dc-version=1.1.18-2b07d5c5a9 \
  cluster-infrastructure=corosync \
  cluster-name=clubionic \
  stonith-enabled=on \
  stonith-action=off \
  no-quorum-policy=stop \
  symmetric-cluster=true

  

  (k)rafaeldtinoco@clubionic02:~$ sudo crm_mon -1
  Stack: corosync
  Current DC: clubionic01.private (version 1.1.18-2b07d5c5a9) - partition with 
quorum
  Last updated: Mon Mar  2 15:55:30 2020
  Last change: Mon Mar  2 15:45:33 2020 by root via cibadmin on 
clubionic01.private

  3 nodes configured
  1 resource configured

  Online: [ clubionic01.private clubionic02.private clubionic03.private
  ]

  Active resources:

   fence_clubionic(stonith:fence_scsi):   Started
  clubionic01.private

  

  (k)rafaeldtinoco@clubionic02:~$ sudo sg_persist --in --read-keys 
--device=/dev/sda
LIO-ORG   cluster.bionic.   4.0
Peripheral device type: 

[Ubuntu-ha] [Bug 1828228] Re: corosync fails to start in unprivileged containers - autopkgtest failure

2020-02-20 Thread Rafael David Tinoco
** Merge proposal unlinked:
   
https://code.launchpad.net/~rafaeldtinoco/ubuntu/+source/corosync/+git/corosync/+merge/379581

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to corosync in Ubuntu.
https://bugs.launchpad.net/bugs/1828228

Title:
  corosync fails to start in unprivileged containers - autopkgtest
  failure

Status in Auto Package Testing:
  Invalid
Status in corosync package in Ubuntu:
  Fix Released
Status in pacemaker package in Ubuntu:
  Fix Released
Status in pcs package in Ubuntu:
  In Progress

Bug description:
  Currently pacemaker v2 fails to start in armhf containers (and by
  extension corosync too).

  I found that it is reproducible locally, and that I had to bump a few
  limits to get it going.

  Specifically I did:

  1) bump memlock limits
  2) bump rmem_max limits

  = 1) Bump memlock limits =

  I have no idea, which one of these finally worked, and/or is
  sufficient. A bit of a whack-a-mole.

  cat >>/etc/security/limits.conf 

[Ubuntu-ha] [Bug 1864116] Re: Package: pacemaker (debian: 2.0.3-3, ubuntu: 2.0.1-5ubuntu5) needs merge

2020-02-20 Thread Rafael David Tinoco
** Merge proposal linked:
   
https://code.launchpad.net/~rafaeldtinoco/ubuntu/+source/pacemaker/+git/pacemaker/+merge/379595

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1864116

Title:
  Package: pacemaker (debian: 2.0.3-3, ubuntu: 2.0.1-5ubuntu5) needs
  merge

Status in pacemaker package in Ubuntu:
  In Progress

Bug description:
  Package: pacemaker (debian: 2.0.3-3, ubuntu: 2.0.1-5ubuntu5) needs
  merge

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1864116/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1828228] Re: corosync fails to start in unprivileged containers - autopkgtest failure

2020-02-20 Thread Rafael David Tinoco
** Merge proposal unlinked:
   
https://code.launchpad.net/~rafaeldtinoco/ubuntu/+source/pacemaker/+git/pacemaker/+merge/379595

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to corosync in Ubuntu.
https://bugs.launchpad.net/bugs/1828228

Title:
  corosync fails to start in unprivileged containers - autopkgtest
  failure

Status in Auto Package Testing:
  Invalid
Status in corosync package in Ubuntu:
  Fix Released
Status in pacemaker package in Ubuntu:
  Fix Released
Status in pcs package in Ubuntu:
  In Progress

Bug description:
  Currently pacemaker v2 fails to start in armhf containers (and by
  extension corosync too).

  I found that it is reproducible locally, and that I had to bump a few
  limits to get it going.

  Specifically I did:

  1) bump memlock limits
  2) bump rmem_max limits

  = 1) Bump memlock limits =

  I have no idea, which one of these finally worked, and/or is
  sufficient. A bit of a whack-a-mole.

  cat >>/etc/security/limits.conf 

[Ubuntu-ha] [Bug 1864116] [NEW] Package: pacemaker (debian: 2.0.3-3, ubuntu: 2.0.1-5ubuntu5) needs merge

2020-02-20 Thread Rafael David Tinoco
Public bug reported:

Package: pacemaker (debian: 2.0.3-3, ubuntu: 2.0.1-5ubuntu5) needs merge

** Affects: pacemaker (Ubuntu)
 Importance: Medium
 Assignee: Rafael David Tinoco (rafaeldtinoco)
 Status: In Progress

** Changed in: pacemaker (Ubuntu)
   Status: New => In Progress

** Changed in: pacemaker (Ubuntu)
   Importance: Undecided => Medium

** Changed in: pacemaker (Ubuntu)
 Assignee: (unassigned) => Rafael David Tinoco (rafaeldtinoco)

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1864116

Title:
  Package: pacemaker (debian: 2.0.3-3, ubuntu: 2.0.1-5ubuntu5) needs
  merge

Status in pacemaker package in Ubuntu:
  In Progress

Bug description:
  Package: pacemaker (debian: 2.0.3-3, ubuntu: 2.0.1-5ubuntu5) needs
  merge

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1864116/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1864087] Re: Package: corosync (debian: 3.0.3-2, ubuntu: 3.0.2-1ubuntu2) needs merge

2020-02-20 Thread Rafael David Tinoco
corosync (3.0.3-2ubuntu1) focal; urgency=medium

  * Merge with Debian unstable. Remaining changes:
- Skip autopkgtest for unprivileged containers: (LP #1828228)
  + d/t/control: mark tests as skippable
  + d/t/cfgtool: skip if memlock can't be set to unlimited by uid 0
  + d/t/quorumtool: skip if memlock can't be set to unlimited by uid 0

 -- Rafael David Tinoco   Thu, 20 Feb 2020
17:57:22 +

corosync (3.0.3-2) unstable; urgency=medium

  * [d0a06e5] Separate the autopkgtests and make them generate artifacts
  * [8680d48] Run Salsa CI reprotest with diffoscope
  * [1d89c4f] Recognize the autopkgtest artifacts in Salsa CI
  * [8e09226] New patch: man: move cmap_keys man page from section 8 to 7

 -- Ferenc Wágner   Sat, 04 Jan 2020 14:07:31 +0100

corosync (3.0.3-1) unstable; urgency=medium

  * [d103a33] New upstream release (3.0.3)
  * [e6f6831] Refresh our patches
  * [f1e85a3] Enable nozzle support
  * [19d3dd3] Package the votequorum simulator in corosync-vqsim
  * [8ae3235] Update Standards-Version to 4.4.1 (no changes required)
  * [bfd9560] Advertise Rules-Requires-Root: no

 -- Ferenc Wágner   Sat, 07 Dec 2019 23:02:16 +0100

corosync (3.0.2-1ubuntu2) focal; urgency=medium

  * Rebuild for libsnmp35.

 -- Rafael David Tinoco   Tue, 14 Jan 2020
17:55:39 +

corosync (3.0.2-1ubuntu1) focal; urgency=medium

  * Merge with Debian unstable. Remaining changes:
- Skip autopkgtest for unprivileged containers: (LP #1828228)
  + d/t/control: mark corosync test as skippable
  + d/t/corosync: skip if memlock can't be set to unlimited by root

 -- Rafael David Tinoco   Mon, 04 Nov 2019
17:37:30 -0300

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to corosync in Ubuntu.
https://bugs.launchpad.net/bugs/1864087

Title:
  Package: corosync (debian: 3.0.3-2, ubuntu: 3.0.2-1ubuntu2) needs
  merge

Status in corosync package in Ubuntu:
  In Progress

Bug description:
  Package: corosync (debian: 3.0.3-2, ubuntu: 3.0.2-1ubuntu2) needs
  merge

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/corosync/+bug/1864087/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1864087] [NEW] Package: corosync (debian: 3.0.3-2, ubuntu: 3.0.2-1ubuntu2) needs merge

2020-02-20 Thread Rafael David Tinoco
Public bug reported:

Package: corosync (debian: 3.0.3-2, ubuntu: 3.0.2-1ubuntu2) needs merge

** Affects: corosync (Ubuntu)
 Importance: Medium
 Assignee: Rafael David Tinoco (rafaeldtinoco)
 Status: In Progress

** Changed in: corosync (Ubuntu)
   Status: New => In Progress

** Changed in: corosync (Ubuntu)
   Importance: Undecided => Medium

** Changed in: corosync (Ubuntu)
 Assignee: (unassigned) => Rafael David Tinoco (rafaeldtinoco)

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to corosync in Ubuntu.
https://bugs.launchpad.net/bugs/1864087

Title:
  Package: corosync (debian: 3.0.3-2, ubuntu: 3.0.2-1ubuntu2) needs
  merge

Status in corosync package in Ubuntu:
  In Progress

Bug description:
  Package: corosync (debian: 3.0.3-2, ubuntu: 3.0.2-1ubuntu2) needs
  merge

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/corosync/+bug/1864087/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1863677] [NEW] fence_scsi needs lvm2 package so it should depend on it

2020-02-17 Thread Rafael David Tinoco
Public bug reported:

Playing with fence_scsi I have noticed that one of my images did not
have lvm2 package installed. With that, the following error ocurred:

$ sudo fence_scsi -n clufocal03 --action=status
2020-02-17 19:57:02,324 ERROR: Unable to run /sbin/vgs --noheadings --separator 
: --sort pv_uuid --options vg_attr,pv_name --config 'global { locking_type = 0 
} devices { preferred_names = [ "^/dev/dm" ] }'

when trying to scsi fence one of my nodes.

fence-agents should depend and/or suggest all packages it depends on for
the correct execution of provided agents.

** Affects: fence-agents (Ubuntu)
 Importance: Medium
 Assignee: Rafael David Tinoco (rafaeldtinoco)
 Status: Confirmed

** Changed in: fence-agents (Ubuntu)
   Status: New => Confirmed

** Changed in: fence-agents (Ubuntu)
   Importance: Undecided => Medium

** Changed in: fence-agents (Ubuntu)
 Assignee: (unassigned) => Rafael David Tinoco (rafaeldtinoco)

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to the bug report.
https://bugs.launchpad.net/bugs/1863677

Title:
  fence_scsi needs lvm2 package so it should depend on it

Status in fence-agents package in Ubuntu:
  Confirmed

Bug description:
  Playing with fence_scsi I have noticed that one of my images did not
  have lvm2 package installed. With that, the following error ocurred:

  $ sudo fence_scsi -n clufocal03 --action=status
  2020-02-17 19:57:02,324 ERROR: Unable to run /sbin/vgs --noheadings 
--separator : --sort pv_uuid --options vg_attr,pv_name --config 'global { 
locking_type = 0 } devices { preferred_names = [ "^/dev/dm" ] }'

  when trying to scsi fence one of my nodes.

  fence-agents should depend and/or suggest all packages it depends on
  for the correct execution of provided agents.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/fence-agents/+bug/1863677/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1863437] Re: [focal] pacemaker i386 should drop a few i386 only packages

2020-02-17 Thread Rafael David Tinoco
https://salsa.debian.org/ha-team/pacemaker/merge_requests/1

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1863437

Title:
  [focal] pacemaker i386 should drop a few i386 only packages

Status in pacemaker package in Ubuntu:
  In Progress
Status in pacemaker source package in Focal:
  In Progress

Bug description:
  When executing pacemaker i386 autopkgtests I realized that package
  "resource-agents" wasn't available for i386. When discussing this with
  @vorlon we came into the conclusion that some i386 only cluster
  packages could be removed from the repository, towards the effort of
  having *only the essential* packages available in i386 (to be run
  together with an amd64 host).

  IRC log:

  """
   resource-agents i386 binary package
   the pacemaker binary is /not/ present on i386 in the release pocket
   that may have been an overly aggressive removal
   vorlon: are u keeping pacemaker because of dependencies ?
   yeah, I removed it when I shouldn't have
  (https://launchpad.net/ubuntu/focal/i386/pacemaker)
   rafaeldtinoco: pacemaker-dev is a build-dep of something else we 
need,
  see the referenced germinate output for full details

  https://people.canonical.com/~ubuntu-archive/germinate-
  output/i386.focal/i386+build-depends

   pacemaker-dev is a build-dep of dlm
   and libesmtp-dev is a build-dep of pacemaker, not the other way 
around
   (and dlm is a build-dep of lvm2)
   ah gotcha
   dlm -> corosync -> pacemaker
   so, even though I removed the binary in error from the release 
pocket,
  the right answer is still for pacemaker/i386 binary to go away (leaving only 
the
  -dev and lib packages)

   do you want me to fix that up, or do you want to?
   to fix that we should do like we did to samba ?
   yeah

   looks like the binaries you'll need to drop are pacemaker,
  pacemaker-cli-utils, pacemaker-remote
   and I'll remove those from -proposed right now, so that those don't
  hold up migration

   but I'll hold off on adding the hint until they're dropped from the
  source
   deal, and next hint ill do with proper branch
  """

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1863437/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1863437] Re: [focal] pacemaker i386 should drop a few i386 only packages

2020-02-17 Thread Rafael David Tinoco
Upstream already had the following commit (since @vorlon opened the bug
in January):



commit 543574f18
Author: Ferenc Wágner 
Date:   Thu Jan 9 14:19:19 2020

Omit pacemaker{, -cli-utils, -remote} on Ubuntu/i386

Closes: #948379

diff --git a/debian/rules b/debian/rules
index 1d186d741..04acd7da2 100755
--- a/debian/rules
+++ b/debian/rules
@@ -11,8 +11,14 @@ export DEB_LDFLAGS_MAINT_APPEND=-Wl,-z,defs
 # Avoid useless dependencies in the utilities
 export DEB_LDFLAGS_MAINT_APPEND+=-Wl,--as-needed

+# Ubuntu/i386 shrank into a compatibility layer not carrying the
+# dependencies of some of our binary packages (#948379):
+ifeq ($(shell dpkg-vendor --query vendor)/$(DEB_HOST_ARCH), Ubuntu/i386)
+UBUNTU_EXCLUDES = -Npacemaker -Npacemaker-cli-utils -Npacemaker-remote
+endif
+
 %:
-   dh $@ --with python3
+   dh $@ --with python3 $(UBUNTU_EXCLUDES)

 # autoreconf options taken from autogen.sh
 # without symlink usage (-s) to make --as-needed effective 

I'll propose it to include: -Npacemaeker-resource-agents and we will be
good.



https://bugs.debian.org/cgi-bin/bugreport.cgi?bug=948379

** Bug watch added: Debian Bug tracker #948379
   https://bugs.debian.org/cgi-bin/bugreport.cgi?bug=948379

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1863437

Title:
  [focal] pacemaker i386 should drop a few i386 only packages

Status in pacemaker package in Ubuntu:
  In Progress
Status in pacemaker source package in Focal:
  In Progress

Bug description:
  When executing pacemaker i386 autopkgtests I realized that package
  "resource-agents" wasn't available for i386. When discussing this with
  @vorlon we came into the conclusion that some i386 only cluster
  packages could be removed from the repository, towards the effort of
  having *only the essential* packages available in i386 (to be run
  together with an amd64 host).

  IRC log:

  """
   resource-agents i386 binary package
   the pacemaker binary is /not/ present on i386 in the release pocket
   that may have been an overly aggressive removal
   vorlon: are u keeping pacemaker because of dependencies ?
   yeah, I removed it when I shouldn't have
  (https://launchpad.net/ubuntu/focal/i386/pacemaker)
   rafaeldtinoco: pacemaker-dev is a build-dep of something else we 
need,
  see the referenced germinate output for full details

  https://people.canonical.com/~ubuntu-archive/germinate-
  output/i386.focal/i386+build-depends

   pacemaker-dev is a build-dep of dlm
   and libesmtp-dev is a build-dep of pacemaker, not the other way 
around
   (and dlm is a build-dep of lvm2)
   ah gotcha
   dlm -> corosync -> pacemaker
   so, even though I removed the binary in error from the release 
pocket,
  the right answer is still for pacemaker/i386 binary to go away (leaving only 
the
  -dev and lib packages)

   do you want me to fix that up, or do you want to?
   to fix that we should do like we did to samba ?
   yeah

   looks like the binaries you'll need to drop are pacemaker,
  pacemaker-cli-utils, pacemaker-remote
   and I'll remove those from -proposed right now, so that those don't
  hold up migration

   but I'll hold off on adding the hint until they're dropped from the
  source
   deal, and next hint ill do with proper branch
  """

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1863437/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1863437] Re: [focal] pacemaker i386 should drop a few i386 only packages

2020-02-17 Thread Rafael David Tinoco
Merged and uploaded.

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1863437

Title:
  [focal] pacemaker i386 should drop a few i386 only packages

Status in pacemaker package in Ubuntu:
  In Progress
Status in pacemaker source package in Focal:
  In Progress

Bug description:
  When executing pacemaker i386 autopkgtests I realized that package
  "resource-agents" wasn't available for i386. When discussing this with
  @vorlon we came into the conclusion that some i386 only cluster
  packages could be removed from the repository, towards the effort of
  having *only the essential* packages available in i386 (to be run
  together with an amd64 host).

  IRC log:

  """
   resource-agents i386 binary package
   the pacemaker binary is /not/ present on i386 in the release pocket
   that may have been an overly aggressive removal
   vorlon: are u keeping pacemaker because of dependencies ?
   yeah, I removed it when I shouldn't have
  (https://launchpad.net/ubuntu/focal/i386/pacemaker)
   rafaeldtinoco: pacemaker-dev is a build-dep of something else we 
need,
  see the referenced germinate output for full details

  https://people.canonical.com/~ubuntu-archive/germinate-
  output/i386.focal/i386+build-depends

   pacemaker-dev is a build-dep of dlm
   and libesmtp-dev is a build-dep of pacemaker, not the other way 
around
   (and dlm is a build-dep of lvm2)
   ah gotcha
   dlm -> corosync -> pacemaker
   so, even though I removed the binary in error from the release 
pocket,
  the right answer is still for pacemaker/i386 binary to go away (leaving only 
the
  -dev and lib packages)

   do you want me to fix that up, or do you want to?
   to fix that we should do like we did to samba ?
   yeah

   looks like the binaries you'll need to drop are pacemaker,
  pacemaker-cli-utils, pacemaker-remote
   and I'll remove those from -proposed right now, so that those don't
  hold up migration

   but I'll hold off on adding the hint until they're dropped from the
  source
   deal, and next hint ill do with proper branch
  """

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1863437/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1863437] Re: [focal] pacemaker i386 should drop a few i386 only packages

2020-02-16 Thread Rafael David Tinoco
"fakeroot debian/rules build-arch" gives me:

dpkg-deb: building package 'pacemaker-remote' in 
'../pacemaker-remote_2.0.1-5ubuntu5_i386.deb'.
dpkg-deb: building package 'pacemaker' in 
'../pacemaker_2.0.1-5ubuntu5_i386.deb'.
dpkg-deb: building package 'libcrmcluster29' in 
'../libcrmcluster29_2.0.1-5ubuntu5_i386.deb'.
dpkg-deb: building package 'libcrmservice28' in 
'../libcrmservice28_2.0.1-5ubuntu5_i386.deb'.
dpkg-deb: building package 'libpe-rules26' in 
'../libpe-rules26_2.0.1-5ubuntu5_i386.deb'.
dpkg-deb: building package 'libpengine27' in 
'../libpengine27_2.0.1-5ubuntu5_i386.deb'.
dpkg-deb: building package 'libtransitioner25' in 
'../libtransitioner25_2.0.1-5ubuntu5_i386.deb'.
dpkg-deb: building package 'libpe-status28' in 
'../libpe-status28_2.0.1-5ubuntu5_i386.deb'.
dpkg-deb: building package 'pacemaker-dev' in 
'../pacemaker-dev_2.0.1-5ubuntu5_i386.deb'.
dpkg-deb: building package 'libcib27' in '../libcib27_2.0.1-5ubuntu5_i386.deb'.
dpkg-deb: building package 'liblrmd28' in 
'../liblrmd28_2.0.1-5ubuntu5_i386.deb'.
dpkg-deb: building package 'libcrmcommon34' in 
'../libcrmcommon34_2.0.1-5ubuntu5_i386.deb'.
dpkg-deb: building package 'libstonithd26' in 
'../libstonithd26_2.0.1-5ubuntu5_i386.deb'.
dpkg-deb: building package 'pacemaker-cli-utils' in 
'../pacemaker-cli-utils_2.0.1-5ubuntu5_i386.deb'.

Since it doesn't build pacemaker-doc it does not depend on missing
packages in i386 (pelican, for example).

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1863437

Title:
  [focal] pacemaker i386 should drop a few i386 only packages

Status in pacemaker package in Ubuntu:
  In Progress
Status in pacemaker source package in Focal:
  In Progress

Bug description:
  When executing pacemaker i386 autopkgtests I realized that package
  "resource-agents" wasn't available for i386. When discussing this with
  @vorlon we came into the conclusion that some i386 only cluster
  packages could be removed from the repository, towards the effort of
  having *only the essential* packages available in i386 (to be run
  together with an amd64 host).

  IRC log:

  """
   resource-agents i386 binary package
   the pacemaker binary is /not/ present on i386 in the release pocket
   that may have been an overly aggressive removal
   vorlon: are u keeping pacemaker because of dependencies ?
   yeah, I removed it when I shouldn't have
  (https://launchpad.net/ubuntu/focal/i386/pacemaker)
   rafaeldtinoco: pacemaker-dev is a build-dep of something else we 
need,
  see the referenced germinate output for full details

  https://people.canonical.com/~ubuntu-archive/germinate-
  output/i386.focal/i386+build-depends

   pacemaker-dev is a build-dep of dlm
   and libesmtp-dev is a build-dep of pacemaker, not the other way 
around
   (and dlm is a build-dep of lvm2)
   ah gotcha
   dlm -> corosync -> pacemaker
   so, even though I removed the binary in error from the release 
pocket,
  the right answer is still for pacemaker/i386 binary to go away (leaving only 
the
  -dev and lib packages)

   do you want me to fix that up, or do you want to?
   to fix that we should do like we did to samba ?
   yeah

   looks like the binaries you'll need to drop are pacemaker,
  pacemaker-cli-utils, pacemaker-remote
   and I'll remove those from -proposed right now, so that those don't
  hold up migration

   but I'll hold off on adding the hint until they're dropped from the
  source
   deal, and next hint ill do with proper branch
  """

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/pacemaker/+bug/1863437/+subscriptions

___
Mailing list: https://launchpad.net/~ubuntu-ha
Post to : ubuntu-ha@lists.launchpad.net
Unsubscribe : https://launchpad.net/~ubuntu-ha
More help   : https://help.launchpad.net/ListHelp


[Ubuntu-ha] [Bug 1863437] Re: [focal] pacemaker i386 should drop a few i386 only packages

2020-02-15 Thread Rafael David Tinoco
# i386 build in local environment

dpkg-deb: building package 'libcib27' in '../libcib27_2.0.1-5ubuntu5_i386.deb'.
dpkg-deb: building package 'pacemaker-common' in 
'../pacemaker-common_2.0.1-5ubuntu5_all.deb'.
dpkg-deb: building package 'libcrmcommon34' in 
'../libcrmcommon34_2.0.1-5ubuntu5_i386.deb'.
dpkg-deb: building package 'liblrmd28' in 
'../liblrmd28_2.0.1-5ubuntu5_i386.deb'.
dpkg-deb: building package 'libpe-status28' in 
'../libpe-status28_2.0.1-5ubuntu5_i386.deb'.
dpkg-deb: building package 'libcrmservice-dev' in 
'../libcrmservice-dev_2.0.1-5ubuntu5_all.deb'.
dpkg-deb: building package 'pacemaker-dev' in 
'../pacemaker-dev_2.0.1-5ubuntu5_i386.deb'.
dpkg-deb: building package 'libstonithd26' in 
'../libstonithd26_2.0.1-5ubuntu5_i386.deb'.
dpkg-deb: building package 'libpengine-dev' in 
'../libpengine-dev_2.0.1-5ubuntu5_all.deb'.
dpkg-deb: building package 'libcrmcluster-dev' in 
'../libcrmcluster-dev_2.0.1-5ubuntu5_all.deb'.
dpkg-deb: building package 'libstonithd-dev' in 
'../libstonithd-dev_2.0.1-5ubuntu5_all.deb'.
dpkg-deb: building package 'liblrmd-dev' in 
'../liblrmd-dev_2.0.1-5ubuntu5_all.deb'.
dpkg-deb: building package 'libcrmcommon-dev' in 
'../libcrmcommon-dev_2.0.1-5ubuntu5_all.deb'.
dpkg-deb: building package 'libtransitioner25' in 
'../libtransitioner25_2.0.1-5ubuntu5_i386.deb'.
dpkg-deb: building package 'libpe-rules26' in 
'../libpe-rules26_2.0.1-5ubuntu5_i386.deb'.
dpkg-deb: building package 'libcrmcluster29' in 
'../libcrmcluster29_2.0.1-5ubuntu5_i386.deb'.
dpkg-deb: building package 'libcib-dev' in 
'../libcib-dev_2.0.1-5ubuntu5_all.deb'.
dpkg-deb: building package 'pacemaker-resource-agents' in 
'../pacemaker-resource-agents_2.0.1-5ubuntu5_all.deb'
dpkg-deb: building package 'libpengine27' in 
'../libpengine27_2.0.1-5ubuntu5_i386.deb'.
dpkg-deb: building package 'libcrmservice28' in 
'../libcrmservice28_2.0.1-5ubuntu5_i386.deb'.
 dpkg-genbuildinfo --build=binary
 dpkg-genchanges --build=binary >../pacemaker_2.0.1-5ubuntu5_i386.changes
dpkg-genchanges: warning: package pacemaker in control file but not in files 
list
dpkg-genchanges: warning: package pacemaker-cli-utils in control file but not 
in files list
dpkg-genchanges: warning: package pacemaker-remote in control file but not in 
files list
dpkg-genchanges: info: binary-only upload (no source code included)



I had to remove i386 architecture from pacemaker-doc as well, in order
not to Build-Depend-Indep: on asciidoc, doxygen, graphviz, inkscape and
publican.

Build-Depends-Indep:
 asciidoc [amd64 arm64 armel armhf mips64el mipsel ppc64el ppc64 s390x x32],
 doxygen [amd64 arm64 armel armhf mips64el mipsel ppc64el ppc64 s390x x32],
 graphviz [amd64 arm64 armel armhf mips64el mipsel ppc64el ppc64 s390x x32],
 inkscape [amd64 arm64 armel armhf mips64el mipsel ppc64el ppc64 s390x x32],
 publican [amd64 arm64 armel armhf mips64el mipsel ppc64el ppc64 s390x x32],

The other packages I removed with the option: dh -Nx

# Ubuntu i386 binary compatibility only effort: Disable ceph support
ifeq ($(shell dpkg-vendor --is Ubuntu && echo yes) $(DEB_HOST_ARCH), yes i386)
   BUILD_PACKAGES += \
-Npacemaker \
-Npacemaker-cli-utils \
-Npacemaker-remote
endif

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1863437

Title:
  [focal] pacemaker i386 should drop a few i386 only packages

Status in pacemaker package in Ubuntu:
  In Progress
Status in pacemaker source package in Focal:
  In Progress

Bug description:
  When executing pacemaker i386 autopkgtests I realized that package
  "resource-agents" wasn't available for i386. When discussing this with
  @vorlon we came into the conclusion that some i386 only cluster
  packages could be removed from the repository, towards the effort of
  having *only the essential* packages available in i386 (to be run
  together with an amd64 host).

  IRC log:

  """
   resource-agents i386 binary package
   the pacemaker binary is /not/ present on i386 in the release pocket
   that may have been an overly aggressive removal
   vorlon: are u keeping pacemaker because of dependencies ?
   yeah, I removed it when I shouldn't have
  (https://launchpad.net/ubuntu/focal/i386/pacemaker)
   rafaeldtinoco: pacemaker-dev is a build-dep of something else we 
need,
  see the referenced germinate output for full details

  https://people.canonical.com/~ubuntu-archive/germinate-
  output/i386.focal/i386+build-depends

   pacemaker-dev is a build-dep of dlm
   and libesmtp-dev is a build-dep of pacemaker, not the other way 
around
   (and dlm is a build-dep of lvm2)
   ah gotcha
   dlm -> corosync -> pacemaker
   so, even though I removed the binary in error from the release 
pocket,
  the right answer is still for pacemaker/i386 binary to go away (leaving only 
the
  -dev and lib packages)

   do you want me to fix that up, or do you want to?
   to fix that we 

[Ubuntu-ha] [Bug 1863437] [NEW] [focal] pacemaker i386 should drop a few i386 only packages

2020-02-15 Thread Rafael David Tinoco
Public bug reported:

When executing pacemaker i386 autopkgtests I realized that package
"resource-agents" wasn't available for i386. When discussing this with
@vorlon we came into the conclusion that some i386 only cluster packages
could be removed from the repository, towards the effort of having *only
the essential* packages available in i386 (to be run together with an
amd64 host).

IRC log:

"""
 resource-agents i386 binary package
 the pacemaker binary is /not/ present on i386 in the release pocket
 that may have been an overly aggressive removal
 vorlon: are u keeping pacemaker because of dependencies ?
 yeah, I removed it when I shouldn't have
(https://launchpad.net/ubuntu/focal/i386/pacemaker)
 rafaeldtinoco: pacemaker-dev is a build-dep of something else we need,
see the referenced germinate output for full details

https://people.canonical.com/~ubuntu-archive/germinate-
output/i386.focal/i386+build-depends

 pacemaker-dev is a build-dep of dlm
 and libesmtp-dev is a build-dep of pacemaker, not the other way around
 (and dlm is a build-dep of lvm2)
 ah gotcha
 dlm -> corosync -> pacemaker
 so, even though I removed the binary in error from the release pocket,
the right answer is still for pacemaker/i386 binary to go away (leaving only the
-dev and lib packages)

 do you want me to fix that up, or do you want to?
 to fix that we should do like we did to samba ?
 yeah

 looks like the binaries you'll need to drop are pacemaker,
pacemaker-cli-utils, pacemaker-remote
 and I'll remove those from -proposed right now, so that those don't
hold up migration

 but I'll hold off on adding the hint until they're dropped from the
source
 deal, and next hint ill do with proper branch
"""

** Affects: pacemaker (Ubuntu)
 Importance: Medium
 Assignee: Rafael David Tinoco (rafaeldtinoco)
 Status: In Progress

** Affects: pacemaker (Ubuntu Focal)
 Importance: Medium
 Assignee: Rafael David Tinoco (rafaeldtinoco)
 Status: In Progress

** Changed in: pacemaker (Ubuntu)
   Status: New => In Progress

** Changed in: pacemaker (Ubuntu)
   Importance: Undecided => Medium

** Changed in: pacemaker (Ubuntu)
 Assignee: (unassigned) => Rafael David Tinoco (rafaeldtinoco)

** Description changed:

+ When executing pacemaker i386 autopkgtests I realized that package
+ "resource-agents" wasn't available for i386. When discussing this with
+ @vorlon we came into the conclusion that some i386 only cluster packages
+ could be removed from the repository, towards the effort of having *only
+ the essential* packages available in i386 (to be run together with an
+ amd64 host).
  
- When executing pacemaker i386 autopkgtests I realized that package 
"resource-agents" wasn't available for i386. When discussing this with @vorlon 
we came into the conclusion that some i386 only cluster packages could be 
removed from the repository, towards the effort of having *only the essential* 
packages available in i386 (to be run together with an amd64 host).
+ IRC log:
  
+ """
   resource-agents i386 binary package
   the pacemaker binary is /not/ present on i386 in the release pocket
   that may have been an overly aggressive removal
   vorlon: are u keeping pacemaker because of dependencies ?
   yeah, I removed it when I shouldn't have
  (https://launchpad.net/ubuntu/focal/i386/pacemaker)
   rafaeldtinoco: pacemaker-dev is a build-dep of something else we 
need,
  see the referenced germinate output for full details
  
- 
- 
https://people.canonical.com/~ubuntu-archive/germinate-output/i386.focal/i386+build-depends
- 
+ https://people.canonical.com/~ubuntu-archive/germinate-
+ output/i386.focal/i386+build-depends
  
   pacemaker-dev is a build-dep of dlm
   and libesmtp-dev is a build-dep of pacemaker, not the other way 
around
   (and dlm is a build-dep of lvm2)
   ah gotcha
   dlm -> corosync -> pacemaker
   so, even though I removed the binary in error from the release 
pocket,
  the right answer is still for pacemaker/i386 binary to go away (leaving only 
the
  -dev and lib packages)
  
   do you want me to fix that up, or do you want to?
   to fix that we should do like we did to samba ?
   yeah
  
   looks like the binaries you'll need to drop are pacemaker,
  pacemaker-cli-utils, pacemaker-remote
   and I'll remove those from -proposed right now, so that those don't
  hold up migration
  
   but I'll hold off on adding the hint until they're dropped from the
  source
   deal, and next hint ill do with proper branch
+ """

** Also affects: pacemaker (Ubuntu Focal)
   Importance: Medium
 Assignee: Rafael David Tinoco (rafaeldtinoco)
   Status: In Progress

-- 
You received this bug notification because you are a member of Ubuntu
High Availability Team, which is subscribed to pacemaker in Ubuntu.
https://bugs.launchpad.net/bugs/1863437

Title:
  [focal] pacemaker i386 should

  1   2   3   4   >