[Linux-HA] Antw: Re: location and orders : Question about a behavior ...

2011-08-04 Thread Ulrich Windl
 Dan Frincu df.clus...@gmail.com schrieb am 03.08.2011 um 13:28 in
Nachricht
CADQRkwiFCEUnq-i9Dtv6AbjQz4Z_e792=3is81zv1eqdrnj...@mail.gmail.com:
 Hi,
 
 On Wed, Aug 3, 2011 at 2:22 PM,  alain.mou...@bull.net wrote:
  Hi  Thanks
 
  I don't think the 1000 or 5000 value makes any difference,
 
 The values make little difference, it's about having a higher score atm.

Hi!

Isn't the stickyness effectively based on the failcount? We have one resource
that has a location constraint for one node with a weight of 50 and a
sticky ness of 10. The resource runs on a different node and shows no
tendency of moving back (not even after restarts).

Somehow the implementation of stickyness is not what one might expect. I'd
expect stickyness to be related to RUNNING resources. I see little sense in
keeping a resource on a node after restarting.

Why? Usually you want stickyness to prevent unexpected downtimes or negative
side effects caused by migration (e.g. all users loosing their connections).
But as a restart will have these side-effects anyway, I see little sense with
the current implementation.

Regards,
Ulrich

 
  so the rsc_options could make it work ?
 
 Yes, I believe so.
 
  But do you have also the order with a clone ?
 
 No.
 
  Because on other of my configurations, I have also
  property $id=cib-bootstrap-options \
 default-resource-stickiness=5000
  and the resource does not failback automatically ... so ...
  Could somebody explain ?
 
 Try the following:
 crm_verify -L 21 | grep stick
 
 And see what scores (weights) are given to resources. Based on these
 weights it might make more sense.
 
 HTH,
 Dan
 
  Thanks
  Alain
 
 
 
  De :Dan Frincu df.clus...@gmail.com
  A : General Linux-HA mailing list linux-ha@lists.linux-ha.org
  Date :  03/08/2011 13:00
  Objet : Re: [Linux-HA] location and orders : Question about a behavior
...
  Envoyé par :linux-ha-boun...@lists.linux-ha.org 
 
 
 
  Hi,
 
  On Tue, Aug 2, 2011 at 6:06 PM,  alain.mou...@bull.net wrote:
  Hi
 
  I have this simple configuration of locations and orders between
  resources
  group-1 , group-2 and clone-1
  (on a two nodes ha cluster with Pacemaker-1.1.2-7 /corosync-1.2.3-21) :
 
  location loc1-group-1   group-1 +100: node2
  location loc1-group-2   group-2 +100: node3
 
  order order-group-1   inf: group-1   clone-1
  order order-group-2   inf: group-2   clone-1
 
  property $id=cib-bootstrap-options \
 dc-version=1.1.2-f059ec7ced7a86f18e5490b67ebf4a0b963bccfe \
 cluster-infrastructure=openais \
 expected-quorum-votes=2 \
 stonith-enabled=true \
 no-quorum-policy=ignore \
 default-resource-stickiness=5000 \
 
  I use it as:
  rsc_defaults $id=rsc-options \
 resource-stickiness=1000
  Instead of:
  property $id=cib-bootstrap-options \
 default-resource-stickiness=5000
  And the behavior is the expected one, no failback.
 
  HTH,
  Dan
 
 
  (and no current cli- preferences)
 
  When I stop the node2, the group-1 is well migrated on node3
  But when node2 is up again, and that I start Pacemaker again on node2,
  the group-1 automatically comes back on node2 , and I wonder why ?
 
  I have other similar configuration with same location constraints and
  same
  default-resource-stickiness value, but without order with a clone
  resource,
  and the group does not come back automatically. But I don't understand
  why
  this order constraint would change this behavior ...
 
  Thanks for your help
  Alain Moullé
 
  ___
  Linux-HA mailing list
  Linux-HA@lists.linux-ha.org 
  http://lists.linux-ha.org/mailman/listinfo/linux-ha 
  See also: http://linux-ha.org/ReportingProblems 
 
 
 
 
  --
  Dan Frincu
  CCNA, RHCE
  ___
  Linux-HA mailing list
  Linux-HA@lists.linux-ha.org 
  http://lists.linux-ha.org/mailman/listinfo/linux-ha 
  See also: http://linux-ha.org/ReportingProblems 
  ___
  Linux-HA mailing list
  Linux-HA@lists.linux-ha.org 
  http://lists.linux-ha.org/mailman/listinfo/linux-ha 
  See also: http://linux-ha.org/ReportingProblems 
 
 
 
 
 -- 
 Dan Frincu
 CCNA, RHCE
 ___
 Linux-HA mailing list
 Linux-HA@lists.linux-ha.org 
 http://lists.linux-ha.org/mailman/listinfo/linux-ha 
 See also: http://linux-ha.org/ReportingProblems

 
 
___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Re: [Linux-HA] Antw: Re: location and orders : Question about a behavior ...

2011-08-04 Thread Andrew Beekhof
On Thu, Aug 4, 2011 at 4:28 PM, Ulrich Windl
ulrich.wi...@rz.uni-regensburg.de wrote:
 Dan Frincu df.clus...@gmail.com schrieb am 03.08.2011 um 13:28 in
 Nachricht
 CADQRkwiFCEUnq-i9Dtv6AbjQz4Z_e792=3is81zv1eqdrnj...@mail.gmail.com:
 Hi,

 On Wed, Aug 3, 2011 at 2:22 PM,  alain.mou...@bull.net wrote:
  Hi  Thanks
 
  I don't think the 1000 or 5000 value makes any difference,

 The values make little difference, it's about having a higher score atm.

 Hi!

 Isn't the stickyness effectively based on the failcount?

No.  A long long time ago there was failure-stickiness, but that was a
bad idea that we got rid of.

 We have one resource
 that has a location constraint for one node with a weight of 50 and a
 sticky ness of 10. The resource runs on a different node and shows no
 tendency of moving back (not even after restarts).

 Somehow the implementation of stickyness is not what one might expect. I'd
 expect stickyness to be related to RUNNING resources.

It is.

 I see little sense in
 keeping a resource on a node after restarting.

 Why? Usually you want stickyness to prevent unexpected downtimes or negative
 side effects caused by migration (e.g. all users loosing their connections).
 But as a restart will have these side-effects anyway, I see little sense with
 the current implementation.

 Regards,
 Ulrich


  so the rsc_options could make it work ?

 Yes, I believe so.

  But do you have also the order with a clone ?

 No.

  Because on other of my configurations, I have also
  property $id=cib-bootstrap-options \
         default-resource-stickiness=5000
  and the resource does not failback automatically ... so ...
  Could somebody explain ?

 Try the following:
 crm_verify -L 21 | grep stick

 And see what scores (weights) are given to resources. Based on these
 weights it might make more sense.

 HTH,
 Dan

  Thanks
  Alain
 
 
 
  De :    Dan Frincu df.clus...@gmail.com
  A :     General Linux-HA mailing list linux-ha@lists.linux-ha.org
  Date :  03/08/2011 13:00
  Objet : Re: [Linux-HA] location and orders : Question about a behavior
 ...
  Envoyé par :    linux-ha-boun...@lists.linux-ha.org
 
 
 
  Hi,
 
  On Tue, Aug 2, 2011 at 6:06 PM,  alain.mou...@bull.net wrote:
  Hi
 
  I have this simple configuration of locations and orders between
  resources
  group-1 , group-2 and clone-1
  (on a two nodes ha cluster with Pacemaker-1.1.2-7 /corosync-1.2.3-21) :
 
  location loc1-group-1   group-1 +100: node2
  location loc1-group-2   group-2 +100: node3
 
  order order-group-1   inf: group-1   clone-1
  order order-group-2   inf: group-2   clone-1
 
  property $id=cib-bootstrap-options \
         dc-version=1.1.2-f059ec7ced7a86f18e5490b67ebf4a0b963bccfe \
         cluster-infrastructure=openais \
         expected-quorum-votes=2 \
         stonith-enabled=true \
         no-quorum-policy=ignore \
         default-resource-stickiness=5000 \
 
  I use it as:
  rsc_defaults $id=rsc-options \
         resource-stickiness=1000
  Instead of:
  property $id=cib-bootstrap-options \
         default-resource-stickiness=5000
  And the behavior is the expected one, no failback.
 
  HTH,
  Dan
 
 
  (and no current cli- preferences)
 
  When I stop the node2, the group-1 is well migrated on node3
  But when node2 is up again, and that I start Pacemaker again on node2,
  the group-1 automatically comes back on node2 , and I wonder why ?
 
  I have other similar configuration with same location constraints and
  same
  default-resource-stickiness value, but without order with a clone
  resource,
  and the group does not come back automatically. But I don't understand
  why
  this order constraint would change this behavior ...
 
  Thanks for your help
  Alain Moullé
 
  ___
  Linux-HA mailing list
  Linux-HA@lists.linux-ha.org
  http://lists.linux-ha.org/mailman/listinfo/linux-ha
  See also: http://linux-ha.org/ReportingProblems
 
 
 
 
  --
  Dan Frincu
  CCNA, RHCE
  ___
  Linux-HA mailing list
  Linux-HA@lists.linux-ha.org
  http://lists.linux-ha.org/mailman/listinfo/linux-ha
  See also: http://linux-ha.org/ReportingProblems
  ___
  Linux-HA mailing list
  Linux-HA@lists.linux-ha.org
  http://lists.linux-ha.org/mailman/listinfo/linux-ha
  See also: http://linux-ha.org/ReportingProblems
 



 --
 Dan Frincu
 CCNA, RHCE
 ___
 Linux-HA mailing list
 Linux-HA@lists.linux-ha.org
 http://lists.linux-ha.org/mailman/listinfo/linux-ha
 See also: http://linux-ha.org/ReportingProblems



 ___
 Linux-HA mailing list
 Linux-HA@lists.linux-ha.org
 http://lists.linux-ha.org/mailman/listinfo/linux-ha
 See also: http://linux-ha.org/ReportingProblems
___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See 

Re: [Linux-HA] Antw: Re: location and orders : Question about a behavior ...

2011-08-04 Thread Maloja01
On 08/04/2011 08:28 AM, Ulrich Windl wrote:
 Hi!
 
 Isn't the stickyness effectively based on the failcount? We have one resource
 that has a location constraint for one node with a weight of 50 and a
 sticky ness of 10. The resource runs on a different node and shows no
 tendency of moving back (not even after restarts).

No stickiness has nothing to do with the failcount. The policy engine
could take both into account the stickiness (for RUNNING resources) and
the failcount for (RUNNING or non-running ressources).

If you ever had a on-start-failure of a resource on a node the failcount
is set to infinity which means, the resource could not be started at
this node.

If the policy engine needs to evaluate where to run a resource it uses
the location/antcolocation/cololaction constraints, failcounts,
stickiness and maybe some other scores to evaluate WHERE to run a resource.

So in my opinion the stiness does exactly what you are asking for.

Kind Regards
Fabian

 
 Somehow the implementation of stickyness is not what one might expect. I'd
 expect stickyness to be related to RUNNING resources. I see little sense in
 keeping a resource on a node after restarting.
 
 Why? Usually you want stickyness to prevent unexpected downtimes or negative
 side effects caused by migration (e.g. all users loosing their connections).
 But as a restart will have these side-effects anyway, I see little sense with
 the current implementation.
 
 Regards,
 Ulrich
___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


Re: [Linux-HA] Antw: Re: location and orders : Question about a behavior ...

2011-08-04 Thread alain . moulle
Hi,
and so ... about my initial question about automatic failback that was not 
expected for me with regards similar configuration 
on some other cluster of mine ? 
Thanks
Alain



De :Maloja01 maloj...@arcor.de
A : linux-ha@lists.linux-ha.org
Date :  04/08/2011 12:58
Objet : Re: [Linux-HA] Antw: Re: location and orders : Question about a 
behavior ...
Envoyé par :linux-ha-boun...@lists.linux-ha.org



On 08/04/2011 08:28 AM, Ulrich Windl wrote:
 Hi!
 
 Isn't the stickyness effectively based on the failcount? We have one 
resource
 that has a location constraint for one node with a weight of 50 and 
a
 sticky ness of 10. The resource runs on a different node and shows 
no
 tendency of moving back (not even after restarts).

No stickiness has nothing to do with the failcount. The policy engine
could take both into account the stickiness (for RUNNING resources) and
the failcount for (RUNNING or non-running ressources).

If you ever had a on-start-failure of a resource on a node the failcount
is set to infinity which means, the resource could not be started at
this node.

If the policy engine needs to evaluate where to run a resource it uses
the location/antcolocation/cololaction constraints, failcounts,
stickiness and maybe some other scores to evaluate WHERE to run a 
resource.

So in my opinion the stiness does exactly what you are asking for.

Kind Regards
Fabian

 
 Somehow the implementation of stickyness is not what one might expect. 
I'd
 expect stickyness to be related to RUNNING resources. I see little sense 
in
 keeping a resource on a node after restarting.
 
 Why? Usually you want stickyness to prevent unexpected downtimes or 
negative
 side effects caused by migration (e.g. all users loosing their 
connections).
 But as a restart will have these side-effects anyway, I see little sense 
with
 the current implementation.
 
 Regards,
 Ulrich
___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


[Linux-HA] DMC 0.9.7 - Pacemaker/Storage GUI

2011-08-04 Thread Rasto Levrinc
Hi,

This is the next DMC release 0.9.7. DMC is Pacemaker, Cluster Virtual Manager
and Storage/DRBD GUI written in Java.

The LVM plugins are now normal part of the GUI in the Storage view, where
you can create, delete physical volumes, volume groups, logical volumes,
resize them and create snapshots in the (whole) cluster. Goodbye command
line...

Some issues with KDE/Compiz were fixed or at least they were made much less
likely to appear. Most notably blank windows. Personally I haven't use KDE
since the 4.0 came out and just assumed that everything worked at least as
anything else on the KDE, but it didn't.

Some issues with applets were fixed as well, so go ahead and use your browser
for everything. :)

Screenshot:

http://oss.linbit.com/drbd-mc/img/drbd-mc-0.9.7.png


You can get DRBD:MC here:
http://oss.linbit.com/drbd-mc/


1. You don't need Java on the cluster servers only on your desktop.
2. Download the DMC-0.9.7.jar file.
3. Start it: java -jar DMC-0.9.7.jar, or click on the file.
4. It connects to the cluster via SSH.
5. If you use it without DRBD, you have to click Skip button on couple of
   places.
6. It is recommended to use it with SUN Java.

DRBD:MC is compatible with Heartbeat 2.1.3 to the Pacemaker 1.1.5 with
Corosync or Heartbeat and DRBD 8.

The most important changes:
* fix fonts in dialogs
* convert plugin code to normal dialogs
* add --debug level option
* add VG remove dialog
* add VG create to the dialogs
* add PV remove dialog
* show in the graph whether the block device is physical volume
* add PV create dialog
* add --out option to redirect the stdout
* fix distro detection so that it detects centos6
* fix unexpected jumping from Service view to VM view
* make default button work in dialogs again
* allow setting both bds to primary if allow-two-primaries is set
* don't allow to enter the same ids for the same resources
* don't create new dependent resource with existing id
* use checkboxes for colocation and order only constraints
* fix gray window problem with popups
* type to search feature for popups
* fix problem with not editable fields in an applet
* workarounds for gray windows
* fix null pointer exception, while removing a DRBD volume
* fix problem with menu nodes, while adding a clone

Rasto Levrinc

--

DI Rastislav Levrinc
Senior Software Engineer
___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


Re: [Linux-HA] Antw: Re: location and orders : Question about a behavior ...

2011-08-04 Thread Ulrich Windl
 Maloja01 maloj...@arcor.de schrieb am 04.08.2011 um 12:58 in Nachricht
4e3a7b5c.1030...@arcor.de:
 On 08/04/2011 08:28 AM, Ulrich Windl wrote:
  Hi!
  
  Isn't the stickyness effectively based on the failcount? We have one 
 resource
  that has a location constraint for one node with a weight of 50 and a
  sticky ness of 10. The resource runs on a different node and shows no
  tendency of moving back (not even after restarts).
 
 No stickiness has nothing to do with the failcount. The policy engine
 could take both into account the stickiness (for RUNNING resources) and
 the failcount for (RUNNING or non-running ressources).
 
 If you ever had a on-start-failure of a resource on a node the failcount
 is set to infinity which means, the resource could not be started at
 this node.

fabian,

I know that, and the errors were removed by crm_resource -C. Still the 
resource is happy where it is, and doesn't want to move away.

 
 If the policy engine needs to evaluate where to run a resource it uses
 the location/antcolocation/cololaction constraints, failcounts,
 stickiness and maybe some other scores to evaluate WHERE to run a resource.
 
 So in my opinion the stiness does exactly what you are asking for.

Unfortunately someone did a manual migrate yesterday, so I cannot show the 
scores that lead to the problem.

Regards,
Ulrich


___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


Re: [Linux-HA] Antw: Re: location and orders : Question about a behavior ...

2011-08-04 Thread Maloja01
Hi Ulrich,

I did not folow the complete thread, just jumped in - sorry. Is the
resource inside a resource group? In this case the stickiness is
multiplied. And sofor the stickiness could be greater than the location
role (score).

Regards
Fabian

On 08/04/2011 03:10 PM, Ulrich Windl wrote:
 Maloja01 maloj...@arcor.de schrieb am 04.08.2011 um 12:58 in Nachricht
 4e3a7b5c.1030...@arcor.de:
 On 08/04/2011 08:28 AM, Ulrich Windl wrote:
 Hi!

 Isn't the stickyness effectively based on the failcount? We have one
 resource
 that has a location constraint for one node with a weight of 50 and a
 sticky ness of 10. The resource runs on a different node and shows no
 tendency of moving back (not even after restarts).

 No stickiness has nothing to do with the failcount. The policy engine
 could take both into account the stickiness (for RUNNING resources) and
 the failcount for (RUNNING or non-running ressources).

 If you ever had a on-start-failure of a resource on a node the failcount
 is set to infinity which means, the resource could not be started at
 this node.

 fabian,

 I know that, and the errors were removed by crm_resource -C. Still the 
 resource is happy where it is, and doesn't want to move away.


 If the policy engine needs to evaluate where to run a resource it uses
 the location/antcolocation/cololaction constraints, failcounts,
 stickiness and maybe some other scores to evaluate WHERE to run a resource.

 So in my opinion the stiness does exactly what you are asking for.

 Unfortunately someone did a manual migrate yesterday, so I cannot show the 
 scores that lead to the problem.

 Regards,
 Ulrich


 ___
 Linux-HA mailing list
 Linux-HA@lists.linux-ha.org
 http://lists.linux-ha.org/mailman/listinfo/linux-ha
 See also: http://linux-ha.org/ReportingProblems

___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


Re: [Linux-HA] Heartbeat Restart is not same as Stop and Start

2011-08-04 Thread Rahul Kanna
Mike,

I checked the permission and those are fine.

If you can please check the restart script I have given below, it does not
touch the heartbeat lock file

*touch $LOCKDIR/$SUBSYS*

when the heartbeat is restared and I guess it is a problem. Is it not?

Btw, we have a product for some web application and as part of it we allow
Administrators to configure servers as redundant server and under lying we
use linux-ha to set up redundant servers.

Rahul
___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems