Re: [Pacemaker] [Question] About quorum-policy=freeze and promote.

2014-05-09 Thread renayama19661014
Hi Andrew,

  Okay.
  I wish this problem is revised by the next release.
 
 crm_report?

I confirmed a problem again in PM1.2-rc1 and registered in Bugzilla.
 * http://bugs.clusterlabs.org/show_bug.cgi?id=5212

Towards Bugzilla, I attached the crm_report file.

Best Regards,
Hideo Yamauchi.

--- On Fri, 2014/5/9, Andrew Beekhof and...@beekhof.net wrote:

 
 On 9 May 2014, at 2:05 pm, renayama19661...@ybb.ne.jp wrote:
 
  Hi Andrew,
  
  Thank you for comment.
  
  Is it responsibility of the resource agent side to prevent a state of 
  these plural Master?
  
  No.
  
  In this scenario, no nodes have quorum and therefor no additional 
  instances should have been promoted.  Thats the definition of freeze :)
  Even if one partition DID have quorum, no instances should have been 
  promoted without fencing occurring first.
  
  Okay.
  I wish this problem is revised by the next release.
 
 crm_report?
 
  
  Many Thanks!
  Hideo Yamauchi.
  
  --- On Fri, 2014/5/9, Andrew Beekhof and...@beekhof.net wrote:
  
  
  On 8 May 2014, at 1:37 pm, renayama19661...@ybb.ne.jp wrote:
  
  Hi All,
  
  I composed Master/Slave resource of three nodes that set 
  quorum-policy=freeze.
  (I use Stateful in Master/Slave resource.)
  
  -
  Current DC: srv01 (3232238280) - partition with quorum
  Version: 1.1.11-830af67
  3 Nodes configured
  9 Resources configured
  
  
  Online: [ srv01 srv02 srv03 ]
  
  Resource Group: grpStonith1
       prmStonith1-1      (stonith:external/ssh): Started srv02 
  Resource Group: grpStonith2
       prmStonith2-1      (stonith:external/ssh): Started srv01 
  Resource Group: grpStonith3
       prmStonith3-1      (stonith:external/ssh): Started srv01 
  Master/Slave Set: msPostgresql [pgsql]
       Masters: [ srv01 ]
       Slaves: [ srv02 srv03 ]
  Clone Set: clnPingd [prmPingd]
       Started: [ srv01 srv02 srv03 ]
  -
  
  
  Master resource starts in all nodes when I interrupt the internal 
  communication of all nodes.
  
  -
  Node srv02 (3232238290): UNCLEAN (offline)
  Node srv03 (3232238300): UNCLEAN (offline)
  Online: [ srv01 ]
  
  Resource Group: grpStonith1
       prmStonith1-1      (stonith:external/ssh): Started srv02 
  Resource Group: grpStonith2
       prmStonith2-1      (stonith:external/ssh): Started srv01 
  Resource Group: grpStonith3
       prmStonith3-1      (stonith:external/ssh): Started srv01 
  Master/Slave Set: msPostgresql [pgsql]
       Masters: [ srv01 ]
       Slaves: [ srv02 srv03 ]
  Clone Set: clnPingd [prmPingd]
       Started: [ srv01 srv02 srv03 ]
  (snip)
  Node srv01 (3232238280): UNCLEAN (offline)
  Node srv03 (3232238300): UNCLEAN (offline)
  Online: [ srv02 ]
  
  Resource Group: grpStonith1
       prmStonith1-1      (stonith:external/ssh): Started srv02 
  Resource Group: grpStonith2
       prmStonith2-1      (stonith:external/ssh): Started srv01 
  Resource Group: grpStonith3
       prmStonith3-1      (stonith:external/ssh): Started srv01 
  Master/Slave Set: msPostgresql [pgsql]
       Masters: [ srv01 srv02 ]
       Slaves: [ srv03 ]
  Clone Set: clnPingd [prmPingd]
       Started: [ srv01 srv02 srv03 ]
  (snip)
  Node srv01 (3232238280): UNCLEAN (offline)
  Node srv02 (3232238290): UNCLEAN (offline)
  Online: [ srv03 ]
  
  Resource Group: grpStonith1
       prmStonith1-1      (stonith:external/ssh): Started srv02 
  Resource Group: grpStonith2
       prmStonith2-1      (stonith:external/ssh): Started srv01 
  Resource Group: grpStonith3
       prmStonith3-1      (stonith:external/ssh): Started srv01 
  Master/Slave Set: msPostgresql [pgsql]
       Masters: [ srv01 srv03 ]
       Slaves: [ srv02 ]
  Clone Set: clnPingd [prmPingd]
       Started: [ srv01 srv02 srv03 ]
  -
  
  I think even if the cluster loses Quorum, being promote the Master / 
  Slave resource that's specification of Pacemaker.
  
  Is it responsibility of the resource agent side to prevent a state of 
  these plural Master?
  
  No.
  
  In this scenario, no nodes have quorum and therefor no additional 
  instances should have been promoted.  Thats the definition of freeze :)
  Even if one partition DID have quorum, no instances should have been 
  promoted without fencing occurring first.
  
  * I think that drbd-RA has those functions.
  * But, there is no function in Stateful-RA.
  * As an example, I think that the mechanism such as drbd is necessary by 
  all means when I make a resource of Master/Slave newly.
  
  Will my understanding be wrong?
  
  Best Regards,
  Hideo Yamauchi.
  
  
  ___
  Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
  http://oss.clusterlabs.org/mailman/listinfo/pacemaker
  
  Project Home: http://www.clusterlabs.org
  Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
  Bugs: http://bugs.clusterlabs.org
  
  
 
 


Re: [Pacemaker] [Question] About quorum-policy=freeze and promote.

2014-05-08 Thread emmanuel segura
Why are you using ssh as stonith? i don't think the fencing is working
because your nodes are in unclean state


2014-05-08 5:37 GMT+02:00 renayama19661...@ybb.ne.jp:

 Hi All,

 I composed Master/Slave resource of three nodes that set
 quorum-policy=freeze.
 (I use Stateful in Master/Slave resource.)

 -
 Current DC: srv01 (3232238280) - partition with quorum
 Version: 1.1.11-830af67
 3 Nodes configured
 9 Resources configured


 Online: [ srv01 srv02 srv03 ]

  Resource Group: grpStonith1
  prmStonith1-1  (stonith:external/ssh): Started srv02
  Resource Group: grpStonith2
  prmStonith2-1  (stonith:external/ssh): Started srv01
  Resource Group: grpStonith3
  prmStonith3-1  (stonith:external/ssh): Started srv01
  Master/Slave Set: msPostgresql [pgsql]
  Masters: [ srv01 ]
  Slaves: [ srv02 srv03 ]
  Clone Set: clnPingd [prmPingd]
  Started: [ srv01 srv02 srv03 ]
 -


 Master resource starts in all nodes when I interrupt the internal
 communication of all nodes.

 -
 Node srv02 (3232238290): UNCLEAN (offline)
 Node srv03 (3232238300): UNCLEAN (offline)
 Online: [ srv01 ]

  Resource Group: grpStonith1
  prmStonith1-1  (stonith:external/ssh): Started srv02
  Resource Group: grpStonith2
  prmStonith2-1  (stonith:external/ssh): Started srv01
  Resource Group: grpStonith3
  prmStonith3-1  (stonith:external/ssh): Started srv01
  Master/Slave Set: msPostgresql [pgsql]
  Masters: [ srv01 ]
  Slaves: [ srv02 srv03 ]
  Clone Set: clnPingd [prmPingd]
  Started: [ srv01 srv02 srv03 ]
 (snip)
 Node srv01 (3232238280): UNCLEAN (offline)
 Node srv03 (3232238300): UNCLEAN (offline)
 Online: [ srv02 ]

  Resource Group: grpStonith1
  prmStonith1-1  (stonith:external/ssh): Started srv02
  Resource Group: grpStonith2
  prmStonith2-1  (stonith:external/ssh): Started srv01
  Resource Group: grpStonith3
  prmStonith3-1  (stonith:external/ssh): Started srv01
  Master/Slave Set: msPostgresql [pgsql]
  Masters: [ srv01 srv02 ]
  Slaves: [ srv03 ]
  Clone Set: clnPingd [prmPingd]
  Started: [ srv01 srv02 srv03 ]
 (snip)
 Node srv01 (3232238280): UNCLEAN (offline)
 Node srv02 (3232238290): UNCLEAN (offline)
 Online: [ srv03 ]

  Resource Group: grpStonith1
  prmStonith1-1  (stonith:external/ssh): Started srv02
  Resource Group: grpStonith2
  prmStonith2-1  (stonith:external/ssh): Started srv01
  Resource Group: grpStonith3
  prmStonith3-1  (stonith:external/ssh): Started srv01
  Master/Slave Set: msPostgresql [pgsql]
  Masters: [ srv01 srv03 ]
  Slaves: [ srv02 ]
  Clone Set: clnPingd [prmPingd]
  Started: [ srv01 srv02 srv03 ]
 -

 I think even if the cluster loses Quorum, being promote the Master /
 Slave resource that's specification of Pacemaker.

 Is it responsibility of the resource agent side to prevent a state of
 these plural Master?
  * I think that drbd-RA has those functions.
  * But, there is no function in Stateful-RA.
  * As an example, I think that the mechanism such as drbd is necessary by
 all means when I make a resource of Master/Slave newly.

 Will my understanding be wrong?

 Best Regards,
 Hideo Yamauchi.


 ___
 Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
 http://oss.clusterlabs.org/mailman/listinfo/pacemaker

 Project Home: http://www.clusterlabs.org
 Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
 Bugs: http://bugs.clusterlabs.org




-- 
esta es mi vida e me la vivo hasta que dios quiera
___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [Pacemaker] [Question] About quorum-policy=freeze and promote.

2014-05-08 Thread renayama19661014
Hi Emmanuel,

 Why are you using ssh as stonith? i don't think the fencing is working 
 because your nodes are in unclean state

No, STONITH is not carried out because all nodes lose quorum.
This is right movement of Pacemaker.

It is an example to use STONITH of ssh.

Best Regards,
Hideo Yamauchi.
--- On Thu, 2014/5/8, emmanuel segura emi2f...@gmail.com wrote:

 
 Why are you using ssh as stonith? i don't think the fencing is working 
 because your nodes are in unclean state
 
 
 
 
 2014-05-08 5:37 GMT+02:00  renayama19661...@ybb.ne.jp:
 Hi All,
 
 I composed Master/Slave resource of three nodes that set 
 quorum-policy=freeze.
 (I use Stateful in Master/Slave resource.)
 
 -
 Current DC: srv01 (3232238280) - partition with quorum
 Version: 1.1.11-830af67
 3 Nodes configured
 9 Resources configured
 
 
 Online: [ srv01 srv02 srv03 ]
 
  Resource Group: grpStonith1
      prmStonith1-1      (stonith:external/ssh): Started srv02
  Resource Group: grpStonith2
      prmStonith2-1      (stonith:external/ssh): Started srv01
  Resource Group: grpStonith3
      prmStonith3-1      (stonith:external/ssh): Started srv01
  Master/Slave Set: msPostgresql [pgsql]
      Masters: [ srv01 ]
      Slaves: [ srv02 srv03 ]
  Clone Set: clnPingd [prmPingd]
      Started: [ srv01 srv02 srv03 ]
 -
 
 
 Master resource starts in all nodes when I interrupt the internal 
 communication of all nodes.
 
 -
 Node srv02 (3232238290): UNCLEAN (offline)
 Node srv03 (3232238300): UNCLEAN (offline)
 Online: [ srv01 ]
 
  Resource Group: grpStonith1
      prmStonith1-1      (stonith:external/ssh): Started srv02
  Resource Group: grpStonith2
      prmStonith2-1      (stonith:external/ssh): Started srv01
  Resource Group: grpStonith3
      prmStonith3-1      (stonith:external/ssh): Started srv01
  Master/Slave Set: msPostgresql [pgsql]
      Masters: [ srv01 ]
      Slaves: [ srv02 srv03 ]
  Clone Set: clnPingd [prmPingd]
      Started: [ srv01 srv02 srv03 ]
 (snip)
 Node srv01 (3232238280): UNCLEAN (offline)
 Node srv03 (3232238300): UNCLEAN (offline)
 Online: [ srv02 ]
 
  Resource Group: grpStonith1
      prmStonith1-1      (stonith:external/ssh): Started srv02
  Resource Group: grpStonith2
      prmStonith2-1      (stonith:external/ssh): Started srv01
  Resource Group: grpStonith3
      prmStonith3-1      (stonith:external/ssh): Started srv01
  Master/Slave Set: msPostgresql [pgsql]
      Masters: [ srv01 srv02 ]
      Slaves: [ srv03 ]
  Clone Set: clnPingd [prmPingd]
      Started: [ srv01 srv02 srv03 ]
 (snip)
 Node srv01 (3232238280): UNCLEAN (offline)
 Node srv02 (3232238290): UNCLEAN (offline)
 Online: [ srv03 ]
 
  Resource Group: grpStonith1
      prmStonith1-1      (stonith:external/ssh): Started srv02
  Resource Group: grpStonith2
      prmStonith2-1      (stonith:external/ssh): Started srv01
  Resource Group: grpStonith3
      prmStonith3-1      (stonith:external/ssh): Started srv01
  Master/Slave Set: msPostgresql [pgsql]
      Masters: [ srv01 srv03 ]
      Slaves: [ srv02 ]
  Clone Set: clnPingd [prmPingd]
      Started: [ srv01 srv02 srv03 ]
 -
 
 I think even if the cluster loses Quorum, being promote the Master / Slave 
 resource that's specification of Pacemaker.
 
 Is it responsibility of the resource agent side to prevent a state of these 
 plural Master?
  * I think that drbd-RA has those functions.
  * But, there is no function in Stateful-RA.
  * As an example, I think that the mechanism such as drbd is necessary by all 
 means when I make a resource of Master/Slave newly.
 
 Will my understanding be wrong?
 
 Best Regards,
 Hideo Yamauchi.
 
 
 ___
 Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
 http://oss.clusterlabs.org/mailman/listinfo/pacemaker
 
 Project Home: http://www.clusterlabs.org
 Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
 Bugs: http://bugs.clusterlabs.org
 
 
 
 -- 
 esta es mi vida e me la vivo hasta que dios quiera

___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [Pacemaker] [Question] About quorum-policy=freeze and promote.

2014-05-08 Thread Andrew Beekhof

On 8 May 2014, at 1:37 pm, renayama19661...@ybb.ne.jp wrote:

 Hi All,
 
 I composed Master/Slave resource of three nodes that set 
 quorum-policy=freeze.
 (I use Stateful in Master/Slave resource.)
 
 -
 Current DC: srv01 (3232238280) - partition with quorum
 Version: 1.1.11-830af67
 3 Nodes configured
 9 Resources configured
 
 
 Online: [ srv01 srv02 srv03 ]
 
 Resource Group: grpStonith1
 prmStonith1-1  (stonith:external/ssh): Started srv02 
 Resource Group: grpStonith2
 prmStonith2-1  (stonith:external/ssh): Started srv01 
 Resource Group: grpStonith3
 prmStonith3-1  (stonith:external/ssh): Started srv01 
 Master/Slave Set: msPostgresql [pgsql]
 Masters: [ srv01 ]
 Slaves: [ srv02 srv03 ]
 Clone Set: clnPingd [prmPingd]
 Started: [ srv01 srv02 srv03 ]
 -
 
 
 Master resource starts in all nodes when I interrupt the internal 
 communication of all nodes.
 
 -
 Node srv02 (3232238290): UNCLEAN (offline)
 Node srv03 (3232238300): UNCLEAN (offline)
 Online: [ srv01 ]
 
 Resource Group: grpStonith1
 prmStonith1-1  (stonith:external/ssh): Started srv02 
 Resource Group: grpStonith2
 prmStonith2-1  (stonith:external/ssh): Started srv01 
 Resource Group: grpStonith3
 prmStonith3-1  (stonith:external/ssh): Started srv01 
 Master/Slave Set: msPostgresql [pgsql]
 Masters: [ srv01 ]
 Slaves: [ srv02 srv03 ]
 Clone Set: clnPingd [prmPingd]
 Started: [ srv01 srv02 srv03 ]
 (snip)
 Node srv01 (3232238280): UNCLEAN (offline)
 Node srv03 (3232238300): UNCLEAN (offline)
 Online: [ srv02 ]
 
 Resource Group: grpStonith1
 prmStonith1-1  (stonith:external/ssh): Started srv02 
 Resource Group: grpStonith2
 prmStonith2-1  (stonith:external/ssh): Started srv01 
 Resource Group: grpStonith3
 prmStonith3-1  (stonith:external/ssh): Started srv01 
 Master/Slave Set: msPostgresql [pgsql]
 Masters: [ srv01 srv02 ]
 Slaves: [ srv03 ]
 Clone Set: clnPingd [prmPingd]
 Started: [ srv01 srv02 srv03 ]
 (snip)
 Node srv01 (3232238280): UNCLEAN (offline)
 Node srv02 (3232238290): UNCLEAN (offline)
 Online: [ srv03 ]
 
 Resource Group: grpStonith1
 prmStonith1-1  (stonith:external/ssh): Started srv02 
 Resource Group: grpStonith2
 prmStonith2-1  (stonith:external/ssh): Started srv01 
 Resource Group: grpStonith3
 prmStonith3-1  (stonith:external/ssh): Started srv01 
 Master/Slave Set: msPostgresql [pgsql]
 Masters: [ srv01 srv03 ]
 Slaves: [ srv02 ]
 Clone Set: clnPingd [prmPingd]
 Started: [ srv01 srv02 srv03 ]
 -
 
 I think even if the cluster loses Quorum, being promote the Master / Slave 
 resource that's specification of Pacemaker.
 
 Is it responsibility of the resource agent side to prevent a state of these 
 plural Master?

No.

In this scenario, no nodes have quorum and therefor no additional instances 
should have been promoted.  Thats the definition of freeze :)
Even if one partition DID have quorum, no instances should have been promoted 
without fencing occurring first.

 * I think that drbd-RA has those functions.
 * But, there is no function in Stateful-RA.
 * As an example, I think that the mechanism such as drbd is necessary by all 
 means when I make a resource of Master/Slave newly.
 
 Will my understanding be wrong?
 
 Best Regards,
 Hideo Yamauchi.
 
 
 ___
 Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
 http://oss.clusterlabs.org/mailman/listinfo/pacemaker
 
 Project Home: http://www.clusterlabs.org
 Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
 Bugs: http://bugs.clusterlabs.org



signature.asc
Description: Message signed with OpenPGP using GPGMail
___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [Pacemaker] [Question] About quorum-policy=freeze and promote.

2014-05-08 Thread Andrew Beekhof

On 9 May 2014, at 2:05 pm, renayama19661...@ybb.ne.jp wrote:

 Hi Andrew,
 
 Thank you for comment.
 
 Is it responsibility of the resource agent side to prevent a state of these 
 plural Master?
 
 No.
 
 In this scenario, no nodes have quorum and therefor no additional instances 
 should have been promoted.  Thats the definition of freeze :)
 Even if one partition DID have quorum, no instances should have been 
 promoted without fencing occurring first.
 
 Okay.
 I wish this problem is revised by the next release.

crm_report?

 
 Many Thanks!
 Hideo Yamauchi.
 
 --- On Fri, 2014/5/9, Andrew Beekhof and...@beekhof.net wrote:
 
 
 On 8 May 2014, at 1:37 pm, renayama19661...@ybb.ne.jp wrote:
 
 Hi All,
 
 I composed Master/Slave resource of three nodes that set 
 quorum-policy=freeze.
 (I use Stateful in Master/Slave resource.)
 
 -
 Current DC: srv01 (3232238280) - partition with quorum
 Version: 1.1.11-830af67
 3 Nodes configured
 9 Resources configured
 
 
 Online: [ srv01 srv02 srv03 ]
 
 Resource Group: grpStonith1
  prmStonith1-1  (stonith:external/ssh): Started srv02 
 Resource Group: grpStonith2
  prmStonith2-1  (stonith:external/ssh): Started srv01 
 Resource Group: grpStonith3
  prmStonith3-1  (stonith:external/ssh): Started srv01 
 Master/Slave Set: msPostgresql [pgsql]
  Masters: [ srv01 ]
  Slaves: [ srv02 srv03 ]
 Clone Set: clnPingd [prmPingd]
  Started: [ srv01 srv02 srv03 ]
 -
 
 
 Master resource starts in all nodes when I interrupt the internal 
 communication of all nodes.
 
 -
 Node srv02 (3232238290): UNCLEAN (offline)
 Node srv03 (3232238300): UNCLEAN (offline)
 Online: [ srv01 ]
 
 Resource Group: grpStonith1
  prmStonith1-1  (stonith:external/ssh): Started srv02 
 Resource Group: grpStonith2
  prmStonith2-1  (stonith:external/ssh): Started srv01 
 Resource Group: grpStonith3
  prmStonith3-1  (stonith:external/ssh): Started srv01 
 Master/Slave Set: msPostgresql [pgsql]
  Masters: [ srv01 ]
  Slaves: [ srv02 srv03 ]
 Clone Set: clnPingd [prmPingd]
  Started: [ srv01 srv02 srv03 ]
 (snip)
 Node srv01 (3232238280): UNCLEAN (offline)
 Node srv03 (3232238300): UNCLEAN (offline)
 Online: [ srv02 ]
 
 Resource Group: grpStonith1
  prmStonith1-1  (stonith:external/ssh): Started srv02 
 Resource Group: grpStonith2
  prmStonith2-1  (stonith:external/ssh): Started srv01 
 Resource Group: grpStonith3
  prmStonith3-1  (stonith:external/ssh): Started srv01 
 Master/Slave Set: msPostgresql [pgsql]
  Masters: [ srv01 srv02 ]
  Slaves: [ srv03 ]
 Clone Set: clnPingd [prmPingd]
  Started: [ srv01 srv02 srv03 ]
 (snip)
 Node srv01 (3232238280): UNCLEAN (offline)
 Node srv02 (3232238290): UNCLEAN (offline)
 Online: [ srv03 ]
 
 Resource Group: grpStonith1
  prmStonith1-1  (stonith:external/ssh): Started srv02 
 Resource Group: grpStonith2
  prmStonith2-1  (stonith:external/ssh): Started srv01 
 Resource Group: grpStonith3
  prmStonith3-1  (stonith:external/ssh): Started srv01 
 Master/Slave Set: msPostgresql [pgsql]
  Masters: [ srv01 srv03 ]
  Slaves: [ srv02 ]
 Clone Set: clnPingd [prmPingd]
  Started: [ srv01 srv02 srv03 ]
 -
 
 I think even if the cluster loses Quorum, being promote the Master / 
 Slave resource that's specification of Pacemaker.
 
 Is it responsibility of the resource agent side to prevent a state of these 
 plural Master?
 
 No.
 
 In this scenario, no nodes have quorum and therefor no additional instances 
 should have been promoted.  Thats the definition of freeze :)
 Even if one partition DID have quorum, no instances should have been 
 promoted without fencing occurring first.
 
 * I think that drbd-RA has those functions.
 * But, there is no function in Stateful-RA.
 * As an example, I think that the mechanism such as drbd is necessary by 
 all means when I make a resource of Master/Slave newly.
 
 Will my understanding be wrong?
 
 Best Regards,
 Hideo Yamauchi.
 
 
 ___
 Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
 http://oss.clusterlabs.org/mailman/listinfo/pacemaker
 
 Project Home: http://www.clusterlabs.org
 Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
 Bugs: http://bugs.clusterlabs.org
 
 



signature.asc
Description: Message signed with OpenPGP using GPGMail
___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org