here is the config ....
<cib epoch="20" num_updates="0" admin_epoch="0" validate-with="pacemaker-1.2" cib-last-written="Wed Mar 9 00:56:57 2016" update-origin="server02" update-client="cibadmin" update-user="hacluster" crm_feature_set="3.0.8" have-quorum="1" dc-uuid="server01"> <configuration> <crm_config> <cluster_property_set id="cib-bootstrap-options"> <nvpair name="stonith-enabled" value="true" id="cib-bootstrap-options-stonith-enabled"/> <nvpair name="no-quorum-policy" value="ignore" id="cib-bootstrap-options-no-quorum-policy"/> <nvpair id="cib-bootstrap-options-dc-version" name="dc-version" value="1.1.11-3ca8c3b"/> <nvpair id="cib-bootstrap-options-cluster-infrastructure" name="cluster-infrastructure" value="classic openais (with plugin)"/> <nvpair id="cib-bootstrap-options-expected-quorum-votes" name="expected-quorum-votes" value="2"/> <nvpair name="stonith-action" value="reboot" id="cib-bootstrap-options-stonith-action"/> <nvpair name="stonith-timeout" value="150s" id="cib-bootstrap-options-stonith-timeout"/> </cluster_property_set> </crm_config> <nodes> <node id="server02" uname="server02"/> <node id="server01" uname="server01"/> </nodes> <resources> <primitive id="STONITH-server01" class="stonith" type="external/ipmi"> <operations> <op name="monitor" interval="0" timeout="60s" id="STONITH-server01-monitor-0"/> <op name="monitor" interval="300s" timeout="60s" on-fail="restart" id="STONITH-server01-monitor-300s"/> <op name="start" interval="0" timeout="60s" on-fail="restart" id="STONITH-server01-start-0"/> </operations> <instance_attributes id="STONITH-server01-instance_attributes"> <nvpair name="hostname" value="server01" id="STONITH-server01-instance_attributes-hostname"/> <nvpair name="ipaddr" value="server01-ipmi" id="STONITH-server01-instance_attributes-ipaddr"/> <nvpair name="userid" value="administrator" id="STONITH-server01-instance_attributes-userid"/> <nvpair name="passwd" value="To12" id="STONITH-server01-instance_attributes-passwd"/> <nvpair name="interface" value="lanplus" id="STONITH-server01-instance_attributes-interface"/> </instance_attributes> </primitive> <primitive id="STONITH-server02" class="stonith" type="external/ipmi"> <operations> <op name="monitor" interval="0" timeout="60s" id="STONITH-server02-monitor-0"/> <op name="monitor" interval="300s" timeout="60s" on-fail="restart" id="STONITH-server02-monitor-300s"/> <op name="start" interval="0" timeout="60s" on-fail="restart" id="STONITH-server02-start-0"/> </operations> <instance_attributes id="STONITH-server02-instance_attributes"> <nvpair name="hostname" value="server02" id="STONITH-server02-instance_attributes-hostname"/> <nvpair name="ipaddr" value="server02-ipmi" id="STONITH-server02-instance_attributes-ipaddr"/> <nvpair name="userid" value="administrator" id="STONITH-server02-instance_attributes-userid"/> <nvpair name="passwd" value="To12" id="STONITH-server02-instance_attributes-passwd"/> <nvpair name="interface" value="lanplus" id="STONITH-server02-instance_attributes-interface"/> </instance_attributes> </primitive> <primitive id="VIRTUAL-IP" class="ocf" provider="heartbeat" type="IPaddr2"> <instance_attributes id="VIRTUAL-IP-instance_attributes"> <nvpair name="ip" value="10.0.0.44" id="VIRTUAL-IP-instance_attributes-ip"/> </instance_attributes> <operations> <op name="monitor" timeout="20s" interval="10s" id="VIRTUAL-IP-monitor-10s"/> </operations> <meta_attributes id="VIRTUAL-IP-meta_attributes"> <nvpair name="is-managed" value="true" id="VIRTUAL-IP-meta_attributes-is-managed"/> <nvpair name="target-role" value="Started" id="VIRTUAL-IP-meta_attributes-target-role"/> </meta_attributes> </primitive> </resources> <constraints> <rsc_location id="LOC_STONITH_server01" rsc="STONITH-server01" score="INFINITY" node="server02"/> <rsc_location id="LOC_STONITH_server02" rsc="STONITH-server02" score="INFINITY" node="server01"/> </constraints> <rsc_defaults> <meta_attributes id="rsc-options"> <nvpair name="migration-threshold" value="5000" id="rsc-options-migration-threshold"/> <nvpair name="resource-stickiness" value="1000" id="rsc-options-resource-stickiness"/> </meta_attributes> </rsc_defaults> <op_defaults> <meta_attributes id="op-options"> <nvpair name="timeout" value="600" id="op-options-timeout"/> <nvpair name="record-pending" value="false" id="op-options-record-pending"/> </meta_attributes> </op_defaults> </configuration> </cib> On Wed, Mar 9, 2016 at 1:25 PM, emmanuel segura <emi2f...@gmail.com> wrote: > I think you should give the parameters to the stonith agent, anyway > show your config. > > 2016-03-09 5:29 GMT+01:00 vija ar <vja...@gmail.com>: > > I have configured SLEHA cluster on cisco ucs boxes with ipmi configured, > i > > have tested IPMI using impitool, however ipmitool to function neatly i > have > > to pass parameter -y i.e. <hex key> along with username and password, > > > > however to configure stonith there is no parameter in pacemaker to pass > > <hex key>? and due to which stonith is failing > > > > can you please let me know if there is any way to add it or is this a > bug? > > > > ******************* > > > > > > > > Mar 9 00:26:28 server02 stonith: external_status: 'ipmi status' failed > with > > rc 1 > > Mar 9 00:26:28 server02 stonith: external/ipmi device not accessible. > > Mar 9 00:26:28 server02 stonith-ng[99114]: notice: log_operation: > > Operation 'monitor' [99200] for device 'STONITH-server02' returned: -201 > > (Generic Pacemaker error) > > Mar 9 00:26:28 server02 stonith-ng[99114]: warning: log_operation: > > STONITH-server02:99200 [ Performing: stonith -t external/ipmi -S ] > > Mar 9 00:26:28 server02 stonith-ng[99114]: warning: log_operation: > > STONITH-server02:99200 [ logd is not runningfailed: 1 ] > > Mar 9 00:26:28 server02 crmd[99118]: error: process_lrm_event: LRM > > operation STONITH-server02_start_0 (call=13, status=4, cib-update=13, > > confirmed=true) Error > > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_cs_dispatch: > Update > > relayed from server01 > > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_trigger_update: > > Sending flush op to all hosts for: fail-count-STONITH-server02 (INFINITY) > > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_perform_update: > Sent > > update 35: fail-count-STONITH-server02=INFINITY > > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_cs_dispatch: > Update > > relayed from server01 > > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_trigger_update: > > Sending flush op to all hosts for: last-failure-STONITH-server02 > > (1457463388) > > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_perform_update: > Sent > > update 37: last-failure-STONITH-server02=1457463388 > > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_cs_dispatch: > Update > > relayed from server01 > > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_trigger_update: > > Sending flush op to all hosts for: fail-count-STONITH-server02 (INFINITY) > > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_perform_update: > Sent > > update 39: fail-count-STONITH-server02=INFINITY > > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_cs_dispatch: > Update > > relayed from server01 > > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_trigger_update: > > Sending flush op to all hosts for: last-failure-STONITH-server02 > > (1457463388) > > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_perform_update: > Sent > > update 41: last-failure-STONITH-server02=1457463388 > > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_cs_dispatch: > Update > > relayed from server01 > > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_trigger_update: > > Sending flush op to all hosts for: fail-count-STONITH-server02 (INFINITY) > > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_perform_update: > Sent > > update 43: fail-count-STONITH-server02=INFINITY > > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_cs_dispatch: > Update > > relayed from server01 > > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_trigger_update: > > Sending flush op to all hosts for: last-failure-STONITH-server02 > > (1457463388) > > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_perform_update: > Sent > > update 45: last-failure-STONITH-server02=1457463388 > > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_cs_dispatch: > Update > > relayed from server01 > > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_trigger_update: > > Sending flush op to all hosts for: fail-count-STONITH-server02 (INFINITY) > > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_perform_update: > Sent > > update 47: fail-count-STONITH-server02=INFINITY > > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_cs_dispatch: > Update > > relayed from server01 > > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_trigger_update: > > Sending flush op to all hosts for: last-failure-STONITH-server02 > > (1457463388) > > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_perform_update: > Sent > > update 49: last-failure-STONITH-server02=1457463388 > > Mar 9 00:26:28 server02 crmd[99118]: notice: process_lrm_event: LRM > > operation STONITH-server02_stop_0 (call=14, rc=0, cib-update=14, > > confirmed=true) ok > > Mar 9 00:26:28 server01 crmd[16809]: warning: status_from_rc: Action 9 > > (STONITH-server02_start_0) on server02 failed (target: 0 vs. rc: 1): > Error > > Mar 9 00:26:28 server01 crmd[16809]: warning: update_failcount: > Updating > > failcount for STONITH-server02 on server02 after failed start: rc=1 > > (update=INFINITY, time=1457463388) > > Mar 9 00:26:28 server01 crmd[16809]: warning: update_failcount: > Updating > > failcount for STONITH-server02 on server02 after failed start: rc=1 > > (update=INFINITY, time=1457463388) > > Mar 9 00:26:28 server01 crmd[16809]: warning: status_from_rc: Action 9 > > (STONITH-server02_start_0) on server02 failed (target: 0 vs. rc: 1): > Error > > Mar 9 00:26:28 server01 crmd[16809]: warning: update_failcount: > Updating > > failcount for STONITH-server02 on server02 after failed start: rc=1 > > (update=INFINITY, time=1457463388) > > Mar 9 00:26:28 server01 crmd[16809]: warning: update_failcount: > Updating > > failcount for STONITH-server02 on server02 after failed start: rc=1 > > (update=INFINITY, time=1457463388) > > Mar 9 00:26:28 server01 stonith: external_status: 'ipmi status' failed > with > > rc 1 > > Mar 9 00:26:28 server01 stonith: external/ipmi device not accessible. > > Mar 9 00:26:28 server01 stonith-ng[16805]: notice: log_operation: > > Operation 'monitor' [16891] for device 'STONITH-server01' returned: -201 > > (Generic Pacemaker error) > > Mar 9 00:26:28 server01 stonith-ng[16805]: warning: log_operation: > > STONITH-server01:16891 [ Performing: stonith -t external/ipmi -S ] > > Mar 9 00:26:28 server01 stonith-ng[16805]: warning: log_operation: > > STONITH-server01:16891 [ logd is not runningfailed: 1 ] > > Mar 9 00:26:28 server01 crmd[16809]: error: process_lrm_event: LRM > > operation STONITH-server01_start_0 (call=13, status=4, cib-update=49, > > confirmed=true) Error > > Mar 9 00:26:28 server01 crmd[16809]: warning: status_from_rc: Action 7 > > (STONITH-server01_start_0) on server01 failed (target: 0 vs. rc: 1): > Error > > Mar 9 00:26:28 server01 crmd[16809]: warning: update_failcount: > Updating > > failcount for STONITH-server01 on server01 after failed start: rc=1 > > (update=INFINITY, time=1457463388) > > Mar 9 00:26:28 server01 attrd[16807]: notice: attrd_trigger_update: > > Sending flush op to all hosts for: fail-count-STONITH-server01 (INFINITY) > > Mar 9 00:26:28 server01 crmd[16809]: warning: update_failcount: > Updating > > failcount for STONITH-server01 on server01 after failed start: rc=1 > > (update=INFINITY, time=1457463388) > > Mar 9 00:26:28 server01 crmd[16809]: warning: status_from_rc: Action 7 > > (STONITH-server01_start_0) on server01 failed (target: 0 vs. rc: 1): > Error > > Mar 9 00:26:28 server01 crmd[16809]: warning: update_failcount: > Updating > > failcount for STONITH-server01 on server01 after failed start: rc=1 > > (update=INFINITY, time=1457463388) > > Mar 9 00:26:28 server01 crmd[16809]: warning: update_failcount: > Updating > > failcount for STONITH-server01 on server01 after failed start: rc=1 > > (update=INFINITY, time=1457463388) > > Mar 9 00:26:28 server01 attrd[16807]: notice: attrd_perform_update: > Sent > > update 47: fail-count-STONITH-server01=INFINITY > > Mar 9 00:26:28 server01 crmd[16809]: notice: run_graph: Transition 3 > > (Complete=5, Pending=0, Fired=0, Skipped=2, Incomplete=0, > > Source=/var/lib/pacemaker/pengine/pe-input-70.bz2): Stopped > > Mar 9 00:26:28 server01 attrd[16807]: notice: attrd_trigger_update: > > Sending flush op to all hosts for: last-failure-STONITH-server01 > > (1457463388) > > Mar 9 00:26:28 server01 attrd[16807]: notice: attrd_perform_update: > Sent > > update 49: last-failure-STONITH-server01=1457463388 > > Mar 9 00:26:28 server01 attrd[16807]: notice: attrd_trigger_update: > > Sending flush op to all hosts for: fail-count-STONITH-server01 (INFINITY) > > Mar 9 00:26:28 server01 pengine[16808]: notice: unpack_config: On > loss of > > CCM Quorum: Ignore > > Mar 9 00:26:28 server01 pengine[16808]: warning: unpack_rsc_op_failure: > > Processing failed op start for STONITH-server02 on server01: unknown > error > > (1) > > Mar 9 00:26:28 server01 pengine[16808]: warning: unpack_rsc_op_failure: > > Processing failed op start for STONITH-server01 on server01: unknown > error > > (1) > > Mar 9 00:26:28 server01 pengine[16808]: warning: unpack_rsc_op_failure: > > Processing failed op start for STONITH-server01 on server01: unknown > error > > (1) > > Mar 9 00:26:28 server01 pengine[16808]: warning: unpack_rsc_op_failure: > > Processing failed op start for STONITH-server02 on server02: unknown > error > > (1) > > Mar 9 00:26:28 server01 pengine[16808]: warning: unpack_rsc_op_failure: > > Processing failed op start for STONITH-server02 on server02: unknown > error > > (1) > > Mar 9 00:26:28 server01 pengine[16808]: warning: unpack_rsc_op_failure: > > Processing failed op start for STONITH-server01 on server02: unknown > error > > (1) > > Mar 9 00:26:28 server01 pengine[16808]: warning: > common_apply_stickiness: > > Forcing STONITH-server02 away from server01 after 1000000 failures > (max=3) > > Mar 9 00:26:28 server01 pengine[16808]: warning: > common_apply_stickiness: > > Forcing STONITH-server01 away from server02 after 1000000 failures > (max=3) > > Mar 9 00:26:28 server01 pengine[16808]: warning: > common_apply_stickiness: > > Forcing STONITH-server02 away from server02 after 1000000 failures > (max=3) > > Mar 9 00:26:28 server01 pengine[16808]: notice: LogActions: Recover > > STONITH-server01 (Started server01) > > Mar 9 00:26:28 server01 pengine[16808]: notice: LogActions: Stop > > STONITH-server02 (server02) > > Mar 9 00:26:28 server01 pengine[16808]: notice: process_pe_message: > > Calculated Transition 4: /var/lib/pacemaker/pengine/pe-input-71.bz2 > > Mar 9 00:26:28 server01 attrd[16807]: notice: attrd_perform_update: > Sent > > update 51: fail-count-STONITH-server01=INFINITY > > Mar 9 00:26:28 server01 attrd[16807]: notice: attrd_trigger_update: > > Sending flush op to all hosts for: last-failure-STONITH-server01 > > (1457463388) > > Mar 9 00:26:28 server01 attrd[16807]: notice: attrd_perform_update: > Sent > > update 53: last-failure-STONITH-server01=1457463388 > > Mar 9 00:26:28 server01 attrd[16807]: notice: attrd_trigger_update: > > Sending flush op to all hosts for: fail-count-STONITH-server01 (INFINITY) > > Mar 9 00:26:28 server01 pengine[16808]: notice: unpack_config: On > loss of > > CCM Quorum: Ignore > > Mar 9 00:26:28 server01 pengine[16808]: warning: unpack_rsc_op_failure: > > Processing failed op start for STONITH-server02 on server01: unknown > error > > (1) > > Mar 9 00:26:28 server01 pengine[16808]: warning: unpack_rsc_op_failure: > > Processing failed op start for STONITH-server01 on server01: unknown > error > > (1) > > Mar 9 00:26:28 server01 pengine[16808]: warning: unpack_rsc_op_failure: > > Processing failed op start for STONITH-server01 on server01: unknown > error > > (1) > > Mar 9 00:26:28 server01 pengine[16808]: warning: unpack_rsc_op_failure: > > Processing failed op start for STONITH-server02 on server02: unknown > error > > (1) > > Mar 9 00:26:28 server01 pengine[16808]: warning: unpack_rsc_op_failure: > > Processing failed op start for STONITH-server02 on server02: unknown > error > > (1) > > Mar 9 00:26:28 server01 pengine[16808]: warning: unpack_rsc_op_failure: > > Processing failed op start for STONITH-server01 on server02: unknown > error > > (1) > > Mar 9 00:26:28 server01 pengine[16808]: warning: > common_apply_stickiness: > > Forcing STONITH-server01 away from server01 after 1000000 failures > (max=3) > > Mar 9 00:26:28 server01 pengine[16808]: warning: > common_apply_stickiness: > > Forcing STONITH-server02 away from server01 after 1000000 failures > (max=3) > > Mar 9 00:26:28 server01 pengine[16808]: warning: > common_apply_stickiness: > > Forcing STONITH-server01 away from server02 after 1000000 failures > (max=3) > > Mar 9 00:26:28 server01 pengine[16808]: warning: > common_apply_stickiness: > > Forcing STONITH-server02 away from server02 after 1000000 failures > (max=3) > > Mar 9 00:26:28 server01 pengine[16808]: notice: LogActions: Stop > > STONITH-server01 (server01) > > Mar 9 00:26:28 server01 pengine[16808]: notice: LogActions: Stop > > STONITH-server02 (server02) > > Mar 9 00:26:28 server01 pengine[16808]: notice: process_pe_message: > > Calculated Transition 5: /var/lib/pacemaker/pengine/pe-input-72.bz2 > > Mar 9 00:26:28 server01 attrd[16807]: notice: attrd_perform_update: > Sent > > update 55: fail-count-STONITH-server01=INFINITY > > Mar 9 00:26:28 server01 attrd[16807]: notice: attrd_trigger_update: > > Sending flush op to all hosts for: last-failure-STONITH-server01 > > (1457463388) > > Mar 9 00:26:28 server01 attrd[16807]: notice: attrd_perform_update: > Sent > > update 57: last-failure-STONITH-server01=1457463388 > > Mar 9 00:26:28 server01 pengine[16808]: notice: unpack_config: On > loss of > > CCM Quorum: Ignore > > Mar 9 00:26:28 server01 pengine[16808]: warning: unpack_rsc_op_failure: > > Processing failed op start for STONITH-server02 on server01: unknown > error > > (1) > > Mar 9 00:26:28 server01 pengine[16808]: warning: unpack_rsc_op_failure: > > Processing failed op start for STONITH-server01 on server01: unknown > error > > (1) > > Mar 9 00:26:28 server01 pengine[16808]: warning: unpack_rsc_op_failure: > > Processing failed op start for STONITH-server01 on server01: unknown > error > > (1) > > Mar 9 00:26:28 server01 pengine[16808]: warning: unpack_rsc_op_failure: > > Processing failed op start for STONITH-server02 on server02: unknown > error > > (1) > > Mar 9 00:26:28 server01 pengine[16808]: warning: unpack_rsc_op_failure: > > Processing failed op start for STONITH-server02 on server02: unknown > error > > (1) > > Mar 9 00:26:28 server01 pengine[16808]: warning: unpack_rsc_op_failure: > > Processing failed op start for STONITH-server01 on server02: unknown > error > > (1) > > Mar 9 00:26:28 server01 pengine[16808]: warning: > common_apply_stickiness: > > Forcing STONITH-server01 away from server01 after 1000000 failures > (max=3) > > Mar 9 00:26:28 server01 pengine[16808]: warning: > common_apply_stickiness: > > Forcing STONITH-server02 away from server01 after 1000000 failures > (max=3) > > Mar 9 00:26:28 server01 pengine[16808]: warning: > common_apply_stickiness: > > Forcing STONITH-server01 away from server02 after 1000000 failures > (max=3) > > Mar 9 00:26:28 server01 pengine[16808]: warning: > common_apply_stickiness: > > Forcing STONITH-server02 away from server02 after 1000000 failures > (max=3) > > Mar 9 00:26:28 server01 pengine[16808]: notice: LogActions: Stop > > STONITH-server01 (server01) > > Mar 9 00:26:28 server01 pengine[16808]: notice: LogActions: Stop > > STONITH-server02 (server02) > > Mar 9 00:26:28 server01 pengine[16808]: notice: process_pe_message: > > Calculated Transition 6: /var/lib/pacemaker/pengine/pe-input-73.bz2 > > Mar 9 00:26:28 server01 crmd[16809]: notice: do_te_invoke: Processing > > graph 6 (ref=pe_calc-dc-1457463388-32) derived from > > /var/lib/pacemaker/pengine/pe-input-73.bz2 > > Mar 9 00:26:28 server01 crmd[16809]: notice: te_rsc_command: > Initiating > > action 1: stop STONITH-server01_stop_0 on server01 (local) > > Mar 9 00:26:28 server01 crmd[16809]: notice: te_rsc_command: > Initiating > > action 2: stop STONITH-server02_stop_0 on server02 > > Mar 9 00:26:28 server01 crmd[16809]: notice: process_lrm_event: LRM > > operation STONITH-server01_stop_0 (call=14, rc=0, cib-update=55, > > confirmed=true) ok > > Mar 9 00:26:28 server01 crmd[16809]: notice: run_graph: Transition 6 > > (Complete=3, Pending=0, Fired=0, Skipped=0, Incomplete=0, > > Source=/var/lib/pacemaker/pengine/pe-input-73.bz2): Complete > > Mar 9 00:26:28 server01 crmd[16809]: notice: do_state_transition: > State > > transition S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS > > cause=C_FSA_INTERNAL origin=notify_crmd ] > > > > > > > > _______________________________________________ > > Users mailing list: Users@clusterlabs.org > > http://clusterlabs.org/mailman/listinfo/users > > > > Project Home: http://www.clusterlabs.org > > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > > Bugs: http://bugs.clusterlabs.org > > > > > > -- > .~. > /V\ > // \\ > /( )\ > ^`~'^ > > _______________________________________________ > Users mailing list: Users@clusterlabs.org > http://clusterlabs.org/mailman/listinfo/users > > Project Home: http://www.clusterlabs.org > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > Bugs: http://bugs.clusterlabs.org >
_______________________________________________ Users mailing list: Users@clusterlabs.org http://clusterlabs.org/mailman/listinfo/users Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org