Hello,

I'm trying to setup a cluster for shared storage with pacemaker, drbd, and
nfs.
I use two server and there is two network interface one each server, eth0
connected to the network, and eth1 direct link betwene the two node for drbd
replication,


                 Eth1: replication link
 --------------      ----------------
 | Server 1   |------| Server 2     |
 --------------      ----------------
      |                    |
     Eth0                 eth0
      |                    |  
         External Network
IPs
- server1 (storage01): eth0:10.1.1.70 eth1:192.168.0.2
- server2 (storage02): eth0:10.1.1.71 eth1:192.168.0.3

I can migrate the resource between the to node, but the problem is when i
make a hard power-off of server01 ( the server where is running the
resource) the second server (server02) doesn't make failover. (the resource
doesn't start) May you can take a look to my config file in the following

Thanks for your help

Thomas Schneider

Drbd.conf
------------------------------------------------------------------------
global {
        usage-count no;
}

common {
        syncer { rate 35M; }
}

resource nfs {
        protocol C;
        handlers {
                fence-peer "/usr/lib/drbd/crm-fence-peer.sh";
                after-resync-target "/usr/lib/drbd/crm-unfence-peer.sh";
        }
        startup {
                degr-wfc-timeout 120;    # 2 minutes.
        }
        disk {
                on-io-error   detach;
                fencing resource-only;
        }
        net {
                max-buffers     2048;
                max-epoch-size  2048;
                }
        syncer {
                rate 35M;
                al-extents 257;
        }
        on storage01.myriapulse.local {
                device     /dev/drbd0;
                disk       /dev/sda4;
                address    192.168.0.2:7788;
                meta-disk  /dev/sda3[0];
        }
        on storage02.myriapulse.local {
                device    /dev/drbd0;
                disk      /dev/sda4;
                address   192.168.0.3:7788;
                meta-disk /dev/sda3[0];
        }
}
----------------------------------------------------------------------------
----------

storage01:~# crm configure show
node storage01.myriapulse.local \
        attributes standby="off"
node storage02.myriapulse.local \
        attributes standby="off"
primitive drbd_nfs ocf:linbit:drbd \
        params drbd_resource="nfs" \
        op monitor interval="15s" \
        meta target-role="Started"
primitive fs_nfs ocf:heartbeat:Filesystem \
        params device="/dev/drbd/by-res/nfs" directory="/share"
fstype="ext3" \
        meta is-managed="true"
primitive ftp-server lsb:proftpd \
        op monitor interval="1min"
primitive ip_nfs ocf:heartbeat:IPaddr2 \
        params ip="10.1.1.69" nic="eth0"
primitive nfs-kernel-server lsb:nfs-kernel-server \
        op monitor interval="1min"
group nfs fs_nfs ip_nfs nfs-kernel-server \
        meta target-role="Started"
ms ms_drbd_nfs drbd_nfs \
        meta master-max="1" master-node-max="1" clone-max="2"
clone-node-max="1" notify="true"
location cli-standby-nfs nfs \
        rule $id="cli-standby-rule-nfs" -inf: #uname eq
storage02.myriapulse.local colocation ftp_on_nfs inf: ftp-server nfs
colocation nfs_on_drbd inf: nfs ms_drbd_nfs:Master order ftp_after_nfs inf:
nfs ftp-server order nfs_after_drbd inf: ms_drbd_nfs:promote nfs:start
property $id="cib-bootstrap-options" \
        dc-version="1.0.5-unknown" \
        cluster-infrastructure="openais" \
        expected-quorum-votes="2" \
        no-quorum-policy="ignore" \
        stonith-enabled="false"
storage01:~#
----------------------------------------------------------------------
Openais.conf

storage02:~# cat /etc/openais/openais.conf # Please read the openais.conf.5
manual page

totem {
        version: 2

        # How long before declaring a token lost (ms)
        token: 3000

        # How many token retransmits before forming a new configuration
        token_retransmits_before_loss_const: 10

        # How long to wait for join messages in the membership protocol (ms)
        join: 60

        # How long to wait for consensus to be achieved before starting a
new round of membership configuration (ms)
        consensus: 1500

        # Turn off the virtual synchrony filter
        vsftype: none

        # Number of messages that may be sent by one processor on receipt of
the token
        max_messages: 20

        # Limit generated nodeids to 31-bits (positive signed integers)
        clear_node_high_bit: yes

        # Disable encryption
        secauth: on

        # How many threads to use for encryption/decryption
        threads: 0

        # Optionally assign a fixed node id (integer)
        # nodeid: 1234

        # This specifies the mode of redundant ring, which may be none,
active, or passive.
        rrp_mode: passive

        interface {
                # The following values need to be set based on your
environment
                ringnumber: 0
                bindnetaddr: 10.1.1.0
                mcastaddr: 226.94.1.1
                mcastport: 5405
        }
        interface {
                # The following values need to be set based on your
environment
                ringnumber: 1
                bindnetaddr: 192.168.0.0
                mcastaddr: 226.94.2.1
                mcastport: 5405
        }
}

logging {
        to_stderr: yes
        debug: off
        timestamp: on
        to_file: /var/log/openais/openais.log
        to_syslog: yes
        syslog_facility: daemon
}

amf {
        mode: disabled
}

service {
        # Load the Pacemaker Cluster Resource Manager
        ver:       0
        name:      pacemaker
        use_mgmtd: yes
}

aisexec {
        user:   root
        group:  root
}

-------------------------------

Log

Oct 31 02:44:17 storage02 crm_attribute: [24179]: info: Invoked:crm_attribute 
-N storage02.myriapulse.local -n master-drbd_nfs:0 -Q -lreboot -v 10000 
Oct 31 02:44:32 storage02 crm_attribute: [24207]: info: Invoked:crm_attribute 
-N storage02.myriapulse.local -n master-drbd_nfs:0 -Q -lreboot -v 10000 
Oct 31 02:44:47 storage02 crm_attribute: [24234]: info: Invoked:crm_attribute 
-N storage02.myriapulse.local -n master-drbd_nfs:0 -Q -lreboot -v 10000 
Oct 31 02:45:01 storage02 /USR/SBIN/CRON[24237]: (root) CMD ([ 
-x/usr/lib/sysstat/sa1 ] && { [ -r "$DEFAULT" ] && . "$DEFAULT" ; [ "$ENABLED"= 
"true" ] && exec /usr/lib/sysstat/sa1 $SA1_OPTIONS 1 1 ; })
Oct 31 02:45:02 storage02 crm_attribute: [24264]: info: Invoked:crm_attribute 
-N storage02.myriapulse.local -n master-drbd_nfs:0 -Q -lreboot -v 10000 
Oct 31 02:45:17 storage02 crm_attribute: [24291]: info: Invoked:crm_attribute 
-N storage02.myriapulse.local -n master-drbd_nfs:0 -Q -lreboot -v 10000 
Oct 31 02:45:32 storage02 crm_attribute: [24318]: info: Invoked:crm_attribute 
-N storage02.myriapulse.local -n master-drbd_nfs:0 -Q -lreboot -v 10000 
Oct 31 02:45:47 storage02 crm_attribute: [24345]: info: Invoked:crm_attribute 
-N storage02.myriapulse.local -n master-drbd_nfs:0 -Q -lreboot -v 10000 
Oct 31 02:46:02 storage02 crm_attribute: [24372]: info: Invoked:crm_attribute 
-N storage02.myriapulse.local -n master-drbd_nfs:0 -Q -lreboot -v 10000 
Oct 31 02:46:17 storage02 crm_attribute: [24399]: info: Invoked:crm_attribute 
-N storage02.myriapulse.local -n master-drbd_nfs:0 -Q -lreboot -v 10000 
Oct 31 02:46:32 storage02 crm_attribute: [24426]: info: Invoked:crm_attribute 
-N storage02.myriapulse.local -n master-drbd_nfs:0 -Q -lreboot -v 10000 
Oct 31 02:46:47 storage02 crm_attribute: [24453]: info: Invoked:crm_attribute 
-N storage02.myriapulse.local -n master-drbd_nfs:0 -Q -lreboot -v 10000 
Oct 31 02:46:58 storage02 kernel: [419448.665572] e1000: eth1:e1000_watchdog: 
NIC Link is Down
Oct 31 02:47:01 storage02 openais[23815]: [TOTEM] The token was lost in 
theOPERATIONAL state.
Oct 31 02:47:01 storage02 openais[23815]: [TOTEM] Receive multicast socketrecv 
buffer size (262142 bytes).
Oct 31 02:47:01 storage02 openais[23815]: [TOTEM] Transmit multicast socketsend 
buffer size (262142 bytes).
Oct 31 02:47:01 storage02 openais[23815]: [TOTEM] Receive multicast socketrecv 
buffer size (262142 bytes).
Oct 31 02:47:01 storage02 openais[23815]: [TOTEM] Transmit multicast socketsend 
buffer size (262142 bytes).
Oct 31 02:47:01 storage02 openais[23815]: [TOTEM] entering GATHER state from2.
Oct 31 02:47:02 storage02 openais[23815]: [TOTEM] entering GATHER state from0.
Oct 31 02:47:02 storage02 openais[23815]: [TOTEM] Creating commit tokenbecause 
I am the rep.
Oct 31 02:47:02 storage02 openais[23815]: [TOTEM] Saving state aru b0 highseq 
received b0
Oct 31 02:47:02 storage02 openais[23815]: [TOTEM] Storing new sequence idfor 
ring 23c
Oct 31 02:47:02 storage02 openais[23815]: [TOTEM] entering COMMIT state.
Oct 31 02:47:02 storage02 openais[23815]: [TOTEM] position [0] member10.1.1.71:
Oct 31 02:47:02 storage02 openais[23815]: [TOTEM] previous ring seq 568 
rep10.1.1.70
Oct 31 02:47:02 storage02 openais[23815]: [TOTEM] aru b0 high delivered 
b0received flag 1
Oct 31 02:47:02 storage02 openais[23815]: [TOTEM] Did not need to originateany 
messages in recovery.
Oct 31 02:47:02 storage02 openais[23815]: [TOTEM] Sending initial ORF token572: 
quorum lost
Oct 31 02:47:02 storage02 cib: [23821]: info: ais_dispatch: Membership 
572:quorum still lost
Oct 31 02:47:02 storage02 crmd: [23825]: info: crm_update_peer: 
Nodestorage01.myriapulse.local: id=1174470922 state=lost (new) 
addr=r(0)ip(10.1.1.70) r(1) ip(192.168.0.2)  votes=1 born=544 
seen=568proc=00000000000000000000000000053312
Oct 31 02:47:02 storage02 crmd: [23825]: WARN: check_dead_member: Our DCnode 
(storage01.myriapulse.local) left the cluster
Oct 31 02:47:02 storage02 cib: [23821]: info: crm_update_peer: 
Nodestorage01.myriapulse.local: id=1174470922 state=lost (new) 
addr=r(0)ip(10.1.1.70) r(1) ip(192.168.0.2)  votes=1 born=544 
seen=568proc=00000000000000000000000000053312
Oct 31 02:47:02 storage02 crmd: [23825]: info: do_state_transition: 
Statetransition S_NOT_DC -> S_ELECTION [ input=I_ELECTION 
cause=C_FSA_INTERNALorigin=check_dead_member ]
Oct 31 02:47:02 storage02 crmd: [23825]: info: update_dc: Unset 
DCstorage01.myriapulse.local
Oct 31 02:47:02 storage02 crmd: [23825]: info: do_state_transition: 
Statetransition S_ELECTION -> S_INTEGRATION [ 
input=I_ELECTION_DCcause=C_FSA_INTERNAL origin=do_election_check ]
Oct 31 02:47:02 storage02 crmd: [23825]: info: do_te_control: Registering 
TEUUID: df06c9c7-f73a-4af0-962f-e96e6b3c8df6
Oct 31 02:47:02 storage02 crmd: [23825]: WARN:cib_client_add_notify_callback: 
Callback already present
Oct 31 02:47:02 storage02 crmd: [23825]: info: set_graph_functions: 
Settingcustom graph functions
Oct 31 02:47:02 storage02 crmd: [23825]: info: unpack_graph: Unpackedtransition 
-1: 0 actions in 0 synapses
Oct 31 02:47:02 storage02 crmd: [23825]: info: do_dc_takeover: Taking overDC 
status for this partition
Oct 31 02:47:02 storage02 cib: [23821]: info: cib_process_readwrite: We arenow 
in R/W mode
Oct 31 02:47:02 storage02 cib: [23821]: info: cib_process_request: 
Operationcomplete: op cib_master for section 'all' 
(origin=local/crmd/18,version=0.244.24): ok (rc=0)
Oct 31 02:47:02 storage02 cib: [23821]: info: cib_process_request: 
Operationcomplete: op cib_modify for section cib 
(origin=local/crmd/19,version=0.244.24): ok (rc=0)
Oct 31 02:47:02 storage02 cib: [23821]: info: cib_process_request: 
Operationcomplete: op cib_modify for section crm_config 
(origin=local/crmd/21,version=0.244.24): ok (rc=0)
Oct 31 02:47:02 storage02 openais[23815]: [CLM  ] CLM CONFIGURATION 
CHANGEoffers based on membership 572
Oct 31 02:47:02 storage02 openais[23815]: [CLM  ] New Configuration:complete: 
op cib_modify for section crm_config (origin=local/crmd/23,version=0.244.24): 
ok (rc=0)
Oct 31 02:47:02 storage02 crmd: [23825]: info: do_dc_join_offer_all: 
join-1:Waiting on 1 outstanding join acks
Oct 31 02:47:02 storage02 crmd: [23825]: info: ais_dispatch: Membership 
572:quorum still lost
Oct 31 02:47:02 storage02 openais[23815]: [CLM  ] #011r(0) ip(10.1.1.71)r(1) 
ip(192.168.0.3) 
Oct 31 02:47:02 storage02 openais[23815]: [CLM  ] Members Left:Checking for 
expired actions every 900000ms
Oct 31 02:47:02 storage02 crmd: [23825]: info: update_dc: Set DC 
tostorage02.myriapulse.local (3.0.1)
Oct 31 02:47:02 storage02 crmd: [23825]: info: ais_dispatch: Membership 
572:quorum still lost
Oct 31 02:47:02 storage02 openais[23815]: [CLM  ] #011r(0) ip(10.1.1.70)r(1) 
ip(192.168.0.2) 
Oct 31 02:47:02 storage02 openais[23815]: [CLM  ] Members Joined:Transitional 
membership event on ring 572: memb=1, new=0, lost=1
Oct 31 02:47:02 storage02 openais[23815]: [crm  ] info: pcmk_peer_update:memb: 
storage02.myriapulse.local 1191248138
Oct 31 02:47:02 storage02 openais[23815]: [crm  ] info: pcmk_peer_update:lost: 
storage01.myriapulse.local 1174470922
Oct 31 02:47:02 storage02 openais[23815]: [CLM  ] CLM CONFIGURATION CHANGE
Oct 31 02:47:02 storage02 openais[23815]: [CLM  ] #011r(0) ip(10.1.1.71)r(1) 
ip(192.168.0.3) 
Oct 31 02:47:02 storage02 openais[23815]: [CLM  ] Members Left:
Oct 31 02:47:02 storage02 openais[23815]: [crm  ] notice: 
pcmk_peer_update:Stable membership event on ring 572: memb=1, new=0, lost=0
Oct 31 02:47:02 storage02 openais[23815]: [crm  ] info: pcmk_peer_update:MEMB: 
storage02.myriapulse.local 1191248138
Oct 31 02:47:02 storage02 openais[23815]: [crm  ] 
info:ais_mark_unseen_peer_dead: Node storage01.myriapulse.local was not seen 
inthe previous transition
Oct 31 02:47:02 storage02 openais[23815]: [MAIN ] info: update_member: 
Node1174470922/storage01.myriapulse.local is now: lost
Oct 31 02:47:02 storage02 openais[23815]: [crm  ] 
info:send_member_notification: Sending membership update 572 to 2 children
Oct 31 02:47:02 storage02 openais[23815]: [SYNC ] This node is within 
theprimary component and will provide service.
Oct 31 02:47:02 storage02 openais[23815]: [TOTEM] entering OPERATIONALstate.
Oct 31 02:47:02 storage02 openais[23815]: [CLM  ] got nodejoin message10.1.1.71
Oct 31 02:47:02 storage02 cib: [23821]: info: cib_process_request: 
Operationcomplete: op cib_modify for section crm_config 
(origin=local/crmd/26,version=0.244.24): ok (rc=0)
Oct 31 02:47:02 storage02 crmd: [23825]: info: do_state_transition: 
Statetransition S_INTEGRATION -> S_FINALIZE_JOIN [ 
input=I_INTEGRATEDcause=C_FSA_INTERNAL origin=check_join_state ]
Oct 31 02:47:02 storage02 crmd: [23825]: info: do_state_transition: All 
1cluster nodes responded to the join offer.
Oct 31 02:47:02 storage02 crmd: [23825]: info: do_dc_join_finalize: 
join-1:Syncing the CIB from storage02.myriapulse.local to the rest of the 
cluster
Oct 31 02:47:02 storage02 crmd: [23825]: info: te_connect_stonith:Attempting 
connection to fencing daemon...
Oct 31 02:47:02 storage02 cib: [23821]: info: cib_process_request: 
Operationcomplete: op cib_modify for section crm_config 
(origin=local/crmd/29,version=0.244.24): ok (rc=0)
Oct 31 02:47:02 storage02 cib: [23821]: info: cib_process_request: 
Operationcomplete: op cib_sync for section 'all' 
(origin=local/crmd/30,version=0.244.24): ok (rc=0)
Oct 31 02:47:03 storage02 crm_attribute: [24480]: info: Invoked:crm_attribute 
-N storage02.myriapulse.local -n master-drbd_nfs:0 -Q -lreboot -v 10000 
Oct 31 02:47:03 storage02 crmd: [23825]: info: te_connect_stonith: 
ConnectedUpdating node state to member for storage02.myriapulse.local
Oct 31 02:47:03 storage02 cib: [23821]: info: cib_process_request: 
Operationcomplete: op cib_modify for section nodes 
(origin=local/crmd/31,version=0.244.24): ok (rc=0)
Oct 31 02:47:03 storage02 cib: [23821]: info: cib_process_request: 
Operationcomplete: op cib_delete for 
section//node_sta...@uname='storage02.myriapulse.local']/lrm 
(origin=local/crmd/32,version=0.244.25): ok (rc=0)
Oct 31 02:47:03 storage02 crmd: [23825]: info: erase_xpath_callback:Deletion of 
"//node_sta...@uname='storage02.myriapulse.local']/lrm": ok(rc=0)
Oct 31 02:47:03 storage02 crmd: [23825]: info: do_state_transition: 
Statetransition S_FINALIZE_JOIN -> S_POLICY_ENGINE [ 
input=I_FINALIZEDcause=C_FSA_INTERNAL origin=check_join_state ]
Oct 31 02:47:03 storage02 crmd: [23825]: info: do_state_transition: All 
1cluster nodes are eligible to run resources.
Oct 31 02:47:03 storage02 crmd: [23825]: info: do_dc_join_final: EnsuringDC, 
quorum and node attributes are up-to-date
Oct 31 02:47:03 storage02 crmd: [23825]: info: crm_update_quorum: 
Updatingquorum status to false (call=36)
Oct 31 02:47:03 storage02 crmd: [23825]: info: 
abort_transition_graph:do_te_invoke:190 - Triggered transition abort 
(complete=1) : Peer Cancelled
Oct 31 02:47:03 storage02 crmd: [23825]: info: do_pe_invoke: Query 
37:Requesting the current CIB: S_POLICY_ENGINE
Oct 31 02:47:03 storage02 attrd: [23823]: info: attrd_local_callback:Sending 
full refresh (origin=crmd)
Oct 31 02:47:03 storage02 attrd: [23823]: info: attrd_trigger_update:Sending 
flush op to all hosts for: probe_complete (true)
Oct 31 02:47:03 storage02 cib: [23821]: info: cib_process_request: 
Operationcomplete: op cib_modify for section nodes 
(origin=local/crmd/34,version=0.244.26): ok (rc=0)
Oct 31 02:47:03 storage02 crmd: [23825]: WARN: match_down_event: No matchfor 
shutdown action on storage01.myriapulse.local
Oct 31 02:47:03 storage02 crmd: [23825]: info: te_update_diff:Stonith/shutdown 
of storage01.myriapulse.local not matched
Oct 31 02:47:03 storage02 crmd: [23825]: info: 
abort_transition_graph:te_update_diff:191 - Triggered transition abort 
(complete=1, tag=node_state,id=storage01.myriapulse.local, magic=NA, 
cib=0.244.27) : Node failure
Oct 31 02:47:03 storage02 crmd: [23825]: info: do_pe_invoke: Query 
38:Requesting the current CIB: S_POLICY_ENGINE
Oct 31 02:47:03 storage02 cib: [23821]: info: log_data_element: cib:diff: -<cib 
have-quorum="1" dc-uuid="storage01.myriapulse.local" admin_epoch="0"epoch="244" 
num_updates="27" />
Oct 31 02:47:03 storage02 cib: [23821]: info: log_data_element: cib:diff: +<cib 
have-quorum="0" dc-uuid="storage02.myriapulse.local" admin_epoch="0"epoch="245" 
num_updates="1" />
Oct 31 02:47:03 storage02 cib: [23821]: info: cib_process_request: 
Operationcomplete: op cib_modify for section cib 
(origin=local/crmd/36,version=0.245.1): ok (rc=0)
Oct 31 02:47:03 storage02 crmd: [23825]: info: 
abort_transition_graph:need_abort:59 - Triggered transition abort (complete=1) 
: Non-status change
Oct 31 02:47:03 storage02 crmd: [23825]: info: need_abort: Aborting onchange to 
have-quorum
Oct 31 02:47:03 storage02 crmd: [23825]: info: do_pe_invoke: Query 
39:Requesting the current CIB: S_POLICY_ENGINE
Oct 31 02:47:03 storage02 crmd: [23825]: info: do_pe_invoke_callback:Invoking 
the PE: ref=pe_calc-dc-1256953623-10, seq=572, quorate=0
Oct 31 02:47:03 storage02 pengine: [23824]: notice: unpack_config: On lossof 
CCM Quorum: Ignore
Oct 31 02:47:03 storage02 pengine: [23824]: info: unpack_config: Nodescores: 
'red' = -INFINITY, 'yellow' = 0, 'green' = 0
Oct 31 02:47:03 storage02 pengine: [23824]: info: determine_online_status:Node 
storage02.myriapulse.local is online
Oct 31 02:47:03 storage02 pengine: [23824]: info: 
unpack_rsc_op:ftp-server_monitor_0 on storage02.myriapulse.local returned 0 
(ok) insteadof the expected value: 7 (not running)
Oct 31 02:47:03 storage02 pengine: [23824]: notice: unpack_rsc_op: 
Operationftp-server_monitor_0 found resource ftp-server active 
onstorage02.myriapulse.local
Oct 31 02:47:03 storage02 pengine: [23824]: notice: group_print: ResourceGroup: 
nfs
Oct 31 02:47:03 storage02 pengine: [23824]: notice: 
native_print:fs_nfs#011(ocf::heartbeat:Filesystem):#011Stopped 
Oct 31 02:47:03 storage02 pengine: [23824]: notice: 
native_print:ip_nfs#011(ocf::heartbeat:IPaddr2):#011Stopped 
Oct 31 02:47:03 storage02 pengine: [23824]: notice: 
native_print:nfs-kernel-server#011(lsb:nfs-kernel-server):#011Stopped 
Oct 31 02:47:03 storage02 pengine: [23824]: notice: clone_print:Master/Slave 
Set: ms_drbd_nfs
Oct 31 02:47:03 storage02 pengine: [23824]: notice: print_list: #011Slaves:[ 
storage02.myriapulse.local ]
Oct 31 02:47:03 storage02 attrd: [23823]: info: attrd_trigger_update:Sending 
flush op to all hosts for: master-drbd_nfs:0 (10000)
Oct 31 02:47:03 storage02 pengine: [23824]: notice: print_list: #011Stopped:[ 
drbd_nfs:1 ]
Oct 31 02:47:03 storage02 pengine: [23824]: notice: 
native_print:ftp-server#011(lsb:proftpd):#011Stopped 
Oct 31 02:47:03 storage02 pengine: [23824]: info: 
native_merge_weights:ms_drbd_nfs: Rolling back scores from fs_nfs
Oct 31 02:47:03 storage02 pengine: [23824]: WARN: native_color: 
Resourcedrbd_nfs:1 cannot run anywhere
Oct 31 02:47:03 storage02 pengine: [23824]: info: 
native_merge_weights:ms_drbd_nfs: Rolling back scores from fs_nfs
Oct 31 02:47:03 storage02 pengine: [23824]: info: master_color: 
Promotingdrbd_nfs:0 (Slave storage02.myriapulse.local)
Oct 31 02:47:03 storage02 pengine: [23824]: info: master_color: 
ms_drbd_nfs:Promoted 1 instances of a possible 1 to master
Oct 31 02:47:03 storage02 pengine: [23824]: info: native_merge_weights:fs_nfs: 
Rolling back scores from ip_nfs
Oct 31 02:47:03 storage02 pengine: [23824]: info: native_merge_weights:fs_nfs: 
Rolling back scores from ftp-server
Oct 31 02:47:03 storage02 pengine: [23824]: WARN: native_color: Resourcefs_nfs 
cannot run anywhere
Oct 31 02:47:03 storage02 pengine: [23824]: info: native_merge_weights:ip_nfs: 
Rolling back scores from nfs-kernel-server
Oct 31 02:47:03 storage02 pengine: [23824]: WARN: native_color: Resourceip_nfs 
cannot run anywhere
Oct 31 02:47:03 storage02 pengine: [23824]: WARN: native_color: 
Resourcenfs-kernel-server cannot run anywhere
Oct 31 02:47:03 storage02 pengine: [23824]: info: master_color: 
ms_drbd_nfs:Promoted 1 instances of a possible 1 to master
Oct 31 02:47:03 storage02 pengine: [23824]: WARN: native_color: 
Resourceftp-server cannot run anywhere
Oct 31 02:47:03 storage02 pengine: [23824]: info: RecurringOp: Cancellingaction 
drbd_nfs:0_monitor_15000 (Slave vs. Master)
Oct 31 02:47:03 storage02 pengine: [23824]: info: RecurringOp: Cancellingaction 
drbd_nfs:0_monitor_15000 (Slave vs. Master)
Oct 31 02:47:03 storage02 pengine: [23824]: notice: LogActions: Leaveresource 
fs_nfs#011(Stopped)
Oct 31 02:47:03 storage02 attrd: [23823]: info: attrd_trigger_update:Sending 
flush op to all hosts for: terminate (<null>)
Oct 31 02:47:03 storage02 pengine: [23824]: notice: LogActions: Leaveresource 
ip_nfs#011(Stopped)
Oct 31 02:47:03 storage02 pengine: [23824]: notice: LogActions: Leaveresource 
nfs-kernel-server#011(Stopped)
Oct 31 02:47:03 storage02 pengine: [23824]: notice: LogActions: 
Promotedrbd_nfs:0#011(Slave -> Master storage02.myriapulse.local)
Oct 31 02:47:03 storage02 pengine: [23824]: notice: LogActions: Leaveresource 
drbd_nfs:1#011(Stopped)
Oct 31 02:47:03 storage02 pengine: [23824]: notice: LogActions: Leaveresource 
ftp-server#011(Stopped)
Oct 31 02:47:03 storage02 crmd: [23825]: info: do_state_transition: 
Statetransition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ 
input=I_PE_SUCCESScause=C_IPC_MESSAGE origin=handle_response ]
Oct 31 02:47:03 storage02 cib: [24481]: info: write_cib_contents: 
Archivedprevious version as /var/lib/heartbeat/crm/cib-64.raw
Oct 31 02:47:03 storage02 attrd: [23823]: info: attrd_trigger_update:Sending 
flush op to all hosts for: master-drbd_nfs:1 (<null>)
Oct 31 02:47:03 storage02 crmd: [23825]: info: unpack_graph: Unpackedtransition 
0: 10 actions in 10 synapses
Oct 31 02:47:03 storage02 crmd: [23825]: info: do_te_invoke: Processinggraph 0 
(ref=pe_calc-dc-1256953623-10) derived from/var/lib/pengine/pe-warn-258.bz2
Oct 31 02:47:03 storage02 crmd: [23825]: info: te_rsc_command: Initiatingaction 
1: cancel drbd_nfs:0_monitor_15000 on storage02.myriapulse.local(local)
Oct 31 02:47:03 storage02 attrd: [23823]: info: attrd_trigger_update:Sending 
flush op to all hosts for: pingdnet1 (<null>)
Oct 31 02:47:03 storage02 attrd: [23823]: info: attrd_trigger_update:Sending 
flush op to all hosts for: shutdown (<null>)
Oct 31 02:47:03 storage02 crmd: [23825]: info: send_direct_ack: ACK'ingresource 
op drbd_nfs:0_monitor_15000 from1:0:0:df06c9c7-f73a-4af0-962f-e96e6b3c8df6: 
lrm_invoke-lrmd-1256953623-12
Oct 31 02:47:03 storage02 crmd: [23825]: info: process_te_message:Processing 
(N)ACK lrm_invoke-lrmd-1256953623-12 fromstorage02.myriapulse.local
Oct 31 02:47:03 storage02 crmd: [23825]: info: match_graph_event: 
Actiondrbd_nfs:0_monitor_15000 (1) confirmed on storage02.myriapulse.local 
(rc=0)
Oct 31 02:47:03 storage02 crmd: [23825]: info: te_pseudo_action: Pseudoaction 
27 fired and confirmed
Oct 31 02:47:03 storage02 crmd: [23825]: info: process_lrm_event: LRMoperation 
drbd_nfs:0_monitor_15000 (call=10, rc=-2, cib-update=0,confirmed=true) 
Cancelled unknown exec error
Oct 31 02:47:03 storage02 cib: [24481]: info: write_cib_contents: Wroteversion 
0.245.0 of the CIB to disk (digest:d1cd4b3ddc5831e6b702492a19e5311f)
Oct 31 02:47:03 storage02 pengine: [23824]: WARN: process_pe_message:Transition 
0: WARNINGs found during PE processing. PEngine Input stored 
in:/var/lib/pengine/pe-warn-258.bz2
Oct 31 02:47:03 storage02 crmd: [23825]: info: te_rsc_command: Initiatingaction 
54: notify drbd_nfs:0_pre_notify_promote_0 onstorage02.myriapulse.local (local)
Oct 31 02:47:03 storage02 pengine: [23824]: info: 
process_pe_message:Configuration WARNINGs found during PE processing.  Please 
run "crm_verify-L" to identify issues.
Oct 31 02:47:03 storage02 crmd: [23825]: info: do_lrm_rsc_op: 
Performingkey=54:0:0:df06c9c7-f73a-4af0-962f-e96e6b3c8df6 
op=drbd_nfs:0_notify_0 )
Oct 31 02:47:03 storage02 cib: [24481]: info: retrieveCib: Reading 
clusterconfiguration from: /var/lib/heartbeat/crm/cib.HQ5vBU 
(digest:/var/lib/heartbeat/crm/cib.Pn6wgG)
Oct 31 02:47:03 storage02 lrmd: [23822]: info: rsc:drbd_nfs:0:11: 
notifyoperation drbd_nfs:0_notify_0 (call=11, rc=0, cib-update=41, 
confirmed=true)complete ok
Oct 31 02:47:03 storage02 crmd: [23825]: info: match_graph_event: 
Actiondrbd_nfs:0_pre_notify_promote_0 (54) confirmed on 
storage02.myriapulse.local(rc=0)
Oct 31 02:47:03 storage02 crmd: [23825]: info: te_pseudo_action: Pseudoaction 
28 fired and confirmed
Oct 31 02:47:03 storage02 crmd: [23825]: info: te_pseudo_action: Pseudoaction 
25 fired and confirmed
Oct 31 02:47:03 storage02 crmd: [23825]: info: te_rsc_command: Initiatingaction 
12: promote drbd_nfs:0_promote_0 on storage02.myriapulse.local(local)
Oct 31 02:47:03 storage02 crmd: [23825]: info: do_lrm_rsc_op: 
Performingkey=12:0:0:df06c9c7-f73a-4af0-962f-e96e6b3c8df6 
op=drbd_nfs:0_promote_0 )
Oct 31 02:47:03 storage02 lrmd: [23822]: info: rsc:drbd_nfs:0:12: promotenot 
arrive in time.
Oct 31 02:47:04 storage02 kernel: [419455.436376] block drbd0: peer( Primary-> 
Unknown ) conn( Connected -> NetworkFailure ) pdsk( UpToDate -> DUnknown) 
Oct 31 02:47:04 storage02 kernel: [419455.436390] block drbd0: asenderterminated
Oct 31 02:47:04 storage02 kernel: [419455.436392] block drbd0: 
Terminatingasender thread
Oct 31 02:47:04 storage02 kernel: [419455.436319] block drbd0: short 
readexpecting header on sock: r=-512
Oct 31 02:47:04 storage02 kernel: [419455.504319] block drbd0: Connectionclosed
Oct 31 02:47:04 storage02 kernel: [419455.504319] block drbd0: 
conn(NetworkFailure -> Unconnected ) 
Oct 31 02:47:04 storage02 kernel: [419455.504319] block drbd0: 
receiverterminated
Oct 31 02:47:04 storage02 kernel: [419455.504319] block drbd0: 
Restartingreceiver thread
Oct 31 02:47:04 storage02 kernel: [419455.504319] block drbd0: 
receiver(re)started
Oct 31 02:47:04 storage02 kernel: [419455.536327] block drbd0: helpercommand: 
/sbin/drbdadm fence-peer minor-0
Oct 31 02:47:05 storage02 crm-fence-peer.sh[24533]: invoked for nfs
Oct 31 02:47:05 storage02 cibadmin: [24546]: info: Invoked: cibadmin -Q -t 1
Oct 31 02:47:06 storage02 cibadmin: [24558]: info: Invoked: cibadmin -Q -t 1
Oct 31 02:47:06 storage02 cibadmin: [24570]: info: Invoked: cibadmin -C 
-oconstraints -X <rsc_location 
rsc="ms_drbd_nfs"id="drbd-fence-by-handler-ms_drbd_nfs">#012  <rule 
role="Master"score="-INFINITY" 
id="drbd-fence-by-handler-rule-ms_drbd_nfs">#012<expression attribute="#uname" 
operation="ne"value="storage02.myriapulse.local"id="drbd-fence-by-handler-expr-ms_drbd_nfs"/>#012</rule>#012</rsc_location>
 
Oct 31 02:47:06 storage02 cib: [23821]: info: log_data_element: cib:diff: -<cib 
admin_epoch="0" epoch="245" num_updates="3" />
Oct 31 02:47:06 storage02 cib: [23821]: info: log_data_element: cib:diff: +<cib 
admin_epoch="0" epoch="246" num_updates="1" >
Oct 31 02:47:06 storage02 cib: [23821]: info: log_data_element: cib:diff: 
+<configuration >
Oct 31 02:47:06 storage02 cib: [23821]: info: log_data_element: cib:diff: 
+<constraints >
Oct 31 02:47:06 storage02 cib: [23821]: info: log_data_element: cib:diff: 
+<rsc_location rsc="ms_drbd_nfs" 
id="drbd-fence-by-handler-ms_drbd_nfs"__crm_diff_marker__="added:top" >
Oct 31 02:47:06 storage02 cib: [23821]: info: log_data_element: cib:diff: 
+<rule role="Master" 
score="-INFINITY"id="drbd-fence-by-handler-rule-ms_drbd_nfs" >
Oct 31 02:47:06 storage02 cib: [23821]: info: log_data_element: cib:diff: 
+<expression attribute="#uname" 
operation="ne"value="storage02.myriapulse.local"id="drbd-fence-by-handler-expr-ms_drbd_nfs"
 />
Oct 31 02:47:06 storage02 cib: [23821]: info: log_data_element: cib:diff: 
+</rule>
Oct 31 02:47:06 storage02 cib: [23821]: info: log_data_element: cib:diff: 
+</rsc_location>
Oct 31 02:47:06 storage02 cib: [23821]: info: log_data_element: cib:diff: 
+</constraints>
Oct 31 02:47:06 storage02 cib: [23821]: info: log_data_element: cib:diff: 
+</configuration>
Oct 31 02:47:06 storage02 cib: [23821]: info: log_data_element: cib:diff: 
+</cib>
Oct 31 02:47:06 storage02 crmd: [23825]: info: 
abort_transition_graph:need_abort:59 - Triggered transition abort (complete=0) 
: Non-status change
Oct 31 02:47:06 storage02 cib: [23821]: info: cib_process_request: 
Operationcomplete: op cib_create for section constraints 
(origin=local/cibadmin/2,version=0.246.1): ok (rc=0)
Oct 31 02:47:06 storage02 crmd: [23825]: info: update_abort_priority: 
Abortpriority upgraded from 0 to 1000000
Oct 31 02:47:06 storage02 crmd: [23825]: info: update_abort_priority: 
Abortaction done superceeded by restart
Oct 31 02:47:06 storage02 crmd: [23825]: info: need_abort: Aborting onchange to 
admin_epoch
Oct 31 02:47:06 storage02 kernel: [419458.059824] block drbd0: helpercommand: 
/sbin/drbdadm fence-peer minor-0 exit code 4 (0x400)
Oct 31 02:47:06 storage02 kernel: [419458.059824] block drbd0: fence-peerhelper 
returned 4 (peer was fenced)
Oct 31 02:47:06 storage02 lrmd: [23822]: info: RA 
output:(drbd_nfs:0:promote:stderr) 0: State change failed: (-7) Refusing to 
bePrimary while peer is not outdated
Oct 31 02:47:06 storage02 lrmd: [23822]: info: RA 
output:(drbd_nfs:0:promote:stderr) Command 'drbdsetup 0 primary' terminated 
withexit code 11
Oct 31 02:47:06 storage02 kernel: [419458.059824] block drbd0: conn(Unconnected 
-> WFConnection ) 
Oct 31 02:47:06 storage02 drbd[24504]: ERROR: nfs: Called drbdadm 
-c/etc/drbd.conf primary nfs
Oct 31 02:47:06 storage02 cib: [24571]: info: write_cib_contents: 
Archivedprevious version as /var/lib/heartbeat/crm/cib-65.raw
Oct 31 02:47:06 storage02 lrmd: [23822]: info: RA 
output:(drbd_nfs:0:promote:stderr) 2009/10/31_02:47:06 ERROR: nfs: Called 
drbdadm-c /etc/drbd.conf primary nfs
Oct 31 02:47:06 storage02 cib: [24571]: info: write_cib_contents: Wroteversion 
0.246.0 of the CIB to disk (digest:f4ae8df353f75d942fe9f71d498328ad)
Oct 31 02:47:06 storage02 cib: [24571]: info: retrieveCib: Reading 
clusterconfiguration from: /var/lib/heartbeat/crm/cib.WYi97O 
(digest:/var/lib/heartbeat/crm/cib.0KwMXG)
Oct 31 02:47:06 storage02 drbd[24504]: ERROR: nfs: Exit code 
11(drbd_nfs:0:promote:stderr) 2009/10/31_02:47:06 ERROR: nfs: Exit code 11
Oct 31 02:47:06 storage02 drbd[24504]: ERROR: nfs: Command output: 
(drbd_nfs:0:promote:stderr) 2009/10/31_02:47:06 ERROR: nfs: Command output: 
Oct 31 02:47:06 storage02 lrmd: [23822]: info: RA 
output:(drbd_nfs:0:promote:stdout) 
Oct 31 02:47:06 storage02 kernel: [419458.084631] block drbd0: helpercommand: 
/sbin/drbdadm fence-peer minor-0
Oct 31 02:47:07 storage02 crm-fence-peer.sh[24594]: invoked for nfs
Oct 31 02:47:07 storage02 kernel: [419459.283186] block drbd0: helpercommand: 
/sbin/drbdadm fence-peer minor-0 exit code 4 (0x400)
Oct 31 02:47:07 storage02 kernel: [419459.283186] block drbd0: fence-peerhelper 
returned 4 (peer was fenced)
Oct 31 02:47:07 storage02 kernel: [419459.283186] block drbd0: role(Secondary 
-> Primary ) pdsk( DUnknown -> Outdated ) 
Oct 31 02:47:07 storage02 kernel: [419459.343187] block drbd0: Creating 
newcurrent UUID
Oct 31 02:47:07 storage02 lrmd: [23822]: info: RA 
output:(drbd_nfs:0:promote:stdout) 
Oct 31 02:47:08 storage02 crmd: [23825]: info: process_lrm_event: LRMoperation 
drbd_nfs:0_promote_0 (call=12, rc=0, cib-update=42,confirmed=true) complete ok
Oct 31 02:47:08 storage02 crmd: [23825]: info: match_graph_event: 
Actiondrbd_nfs:0_promote_0 (12) confirmed on storage02.myriapulse.local (rc=0)
Oct 31 02:47:08 storage02 crmd: [23825]: info: te_pseudo_action: Pseudoaction 
26 fired and confirmed
Oct 31 02:47:08 storage02 crmd: [23825]: info: te_pseudo_action: Pseudoaction 
29 fired and confirmed
Oct 31 02:47:08 storage02 crmd: [23825]: info: te_rsc_command: Initiatingaction 
55: notify drbd_nfs:0_post_notify_promote_0 onstorage02.myriapulse.local (local)
Oct 31 02:47:08 storage02 crmd: [23825]: info: do_lrm_rsc_op: 
Performingkey=55:0:0:df06c9c7-f73a-4af0-962f-e96e6b3c8df6 
op=drbd_nfs:0_notify_0 )
Oct 31 02:47:08 storage02 lrmd: [23822]: info: rsc:drbd_nfs:0:13: 
notifycrm_attribute -N storage02.myriapulse.local -n master-drbd_nfs:0 -Q 
-lreboot -v 10000 
Oct 31 02:47:08 storage02 lrmd: [23822]: info: RA 
output:(drbd_nfs:0:notify:stdout) 
Oct 31 02:47:08 storage02 crmd: [23825]: info: process_lrm_event: LRMoperation 
drbd_nfs:0_notify_0 (call=13, rc=0, cib-update=43, confirmed=true)complete ok
Oct 31 02:47:08 storage02 crmd: [23825]: info: match_graph_event: 
Actiondrbd_nfs:0_post_notify_promote_0 (55) confirmed 
onstorage02.myriapulse.local (rc=0)
Oct 31 02:47:08 storage02 crmd: [23825]: info: te_pseudo_action: Pseudoaction 
30 fired and confirmed
Oct 31 02:47:08 storage02 crmd: [23825]: info: 
run_graph:====================================================
Oct 31 02:47:08 storage02 crmd: [23825]: notice: run_graph: Transition 
0(Complete=10, Pending=0, Fired=0, Skipped=0, 
Incomplete=0,Source=/var/lib/pengine/pe-warn-258.bz2): Complete
Oct 31 02:47:08 storage02 crmd: [23825]: info: te_graph_trigger: Transition0 is 
now complete
Oct 31 02:47:08 storage02 crmd: [23825]: info: do_state_transition: 
Statetransition S_TRANSITION_ENGINE -> S_POLICY_ENGINE [ 
input=I_PE_CALCcause=C_FSA_INTERNAL origin=notify_crmd ]
Oct 31 02:47:08 storage02 crmd: [23825]: info: do_state_transition: All 
1cluster nodes are eligible to run resources.
Oct 31 02:47:08 storage02 crmd: [23825]: info: do_pe_invoke: Query 
44:Requesting the current CIB: S_POLICY_ENGINE
Oct 31 02:47:08 storage02 crmd: [23825]: info: do_pe_invoke_callback:Invoking 
the PE: ref=pe_calc-dc-1256953628-16, seq=572, quorate=0
Oct 31 02:47:08 storage02 pengine: [23824]: notice: unpack_config: On lossof 
CCM Quorum: Ignore
Oct 31 02:47:08 storage02 pengine: [23824]: info: unpack_config: Nodescores: 
'red' = -INFINITY, 'yellow' = 0, 'green' = 0
Oct 31 02:47:08 storage02 pengine: [23824]: info: determine_online_status:Node 
storage02.myriapulse.local is online
Oct 31 02:47:08 storage02 pengine: [23824]: info: 
unpack_rsc_op:ftp-server_monitor_0 on storage02.myriapulse.local returned 0 
(ok) insteadof the expected value: 7 (not running)
Oct 31 02:47:08 storage02 pengine: [23824]: notice: unpack_rsc_op: 
Operationftp-server_monitor_0 found resource ftp-server active 
onstorage02.myriapulse.local
Oct 31 02:47:08 storage02 pengine: [23824]: notice: group_print: ResourceGroup: 
nfs
Oct 31 02:47:08 storage02 pengine: [23824]: notice: 
native_print:fs_nfs#011(ocf::heartbeat:Filesystem):#011Stopped 
Oct 31 02:47:08 storage02 pengine: [23824]: notice: 
native_print:ip_nfs#011(ocf::heartbeat:IPaddr2):#011Stopped 
Oct 31 02:47:08 storage02 pengine: [23824]: notice: 
native_print:nfs-kernel-server#011(lsb:nfs-kernel-server):#011Stopped 
Oct 31 02:47:08 storage02 pengine: [23824]: notice: clone_print:Master/Slave 
Set: ms_drbd_nfs
Oct 31 02:47:08 storage02 pengine: [23824]: notice: print_list: #011Masters:[ 
storage02.myriapulse.local ]
Oct 31 02:47:08 storage02 pengine: [23824]: notice: print_list: #011Stopped:[ 
drbd_nfs:1 ]
Oct 31 02:47:08 storage02 pengine: [23824]: notice: 
native_print:ftp-server#011(lsb:proftpd):#011Stopped 
Oct 31 02:47:08 storage02 pengine: [23824]: info: 
native_merge_weights:ms_drbd_nfs: Rolling back scores from fs_nfs
Oct 31 02:47:08 storage02 pengine: [23824]: WARN: native_color: 
Resourcedrbd_nfs:1 cannot run anywhere
Oct 31 02:47:08 storage02 pengine: [23824]: info: 
native_merge_weights:ms_drbd_nfs: Rolling back scores from fs_nfs
Oct 31 02:47:08 storage02 pengine: [23824]: info: master_color: 
Promotingdrbd_nfs:0 (Master storage02.myriapulse.local)
Oct 31 02:47:08 storage02 pengine: [23824]: info: master_color: 
ms_drbd_nfs:Promoted 1 instances of a possible 1 to master
Oct 31 02:47:08 storage02 pengine: [23824]: info: native_merge_weights:fs_nfs: 
Rolling back scores from ip_nfs
Oct 31 02:47:08 storage02 pengine: [23824]: info: native_merge_weights:fs_nfs: 
Rolling back scores from ftp-server
Oct 31 02:47:08 storage02 pengine: [23824]: WARN: native_color: Resourcefs_nfs 
cannot run anywhere
Oct 31 02:47:08 storage02 pengine: [23824]: info: native_merge_weights:ip_nfs: 
Rolling back scores from nfs-kernel-server
Oct 31 02:47:08 storage02 pengine: [23824]: WARN: native_color: Resourceip_nfs 
cannot run anywhere
Oct 31 02:47:08 storage02 pengine: [23824]: WARN: native_color: 
Resourcenfs-kernel-server cannot run anywhere
Oct 31 02:47:08 storage02 pengine: [23824]: info: master_color: 
ms_drbd_nfs:Promoted 1 instances of a possible 1 to master
Oct 31 02:47:08 storage02 pengine: [23824]: WARN: native_color: 
Resourceftp-server cannot run anywhere
Oct 31 02:47:08 storage02 pengine: [23824]: notice: LogActions: Leaveresource 
fs_nfs#011(Stopped)
Oct 31 02:47:08 storage02 pengine: [23824]: notice: LogActions: Leaveresource 
ip_nfs#011(Stopped)
Oct 31 02:47:08 storage02 pengine: [23824]: notice: LogActions: Leaveresource 
nfs-kernel-server#011(Stopped)
Oct 31 02:47:08 storage02 pengine: [23824]: notice: LogActions: Leaveresource 
drbd_nfs:0#011(Master storage02.myriapulse.local)
Oct 31 02:47:08 storage02 pengine: [23824]: notice: LogActions: Leaveresource 
drbd_nfs:1#011(Stopped)
Oct 31 02:47:08 storage02 pengine: [23824]: notice: LogActions: Leaveresource 
ftp-server#011(Stopped)
Oct 31 02:47:08 storage02 crmd: [23825]: info: do_state_transition: 
Statetransition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ 
input=I_PE_SUCCESScause=C_IPC_MESSAGE origin=handle_response ]
Oct 31 02:47:08 storage02 crmd: [23825]: WARN: destroy_action: Cancellingtimer 
for action 1 (src=44)
Oct 31 02:47:08 storage02 crmd: [23825]: info: unpack_graph: Unpackedtransition 
1: 0 actions in 0 synapses
Oct 31 02:47:08 storage02 crmd: [23825]: info: do_te_invoke: Processinggraph 1 
(ref=pe_calc-dc-1256953628-16) derived from/var/lib/pengine/pe-warn-259.bz2
Oct 31 02:47:08 storage02 crmd: [23825]: info: 
run_graph:====================================================
Oct 31 02:47:08 storage02 crmd: [23825]: notice: run_graph: Transition 
1(Complete=0, Pending=0, Fired=0, Skipped=0, 
Incomplete=0,Source=/var/lib/pengine/pe-warn-259.bz2): Complete
Oct 31 02:47:08 storage02 crmd: [23825]: info: te_graph_trigger: Transition1 is 
now complete
Oct 31 02:47:08 storage02 crmd: [23825]: info: notify_crmd: Transition 1status: 
done - <null>
Oct 31 02:47:08 storage02 crmd: [23825]: info: do_state_transition: 
Statetransition S_TRANSITION_ENGINE -> S_IDLE [ 
input=I_TE_SUCCESScause=C_FSA_INTERNAL origin=notify_crmd ]
Oct 31 02:47:08 storage02 crmd: [23825]: info: do_state_transition: 
StartingPEngine Recheck Timer
Oct 31 02:47:08 storage02 pengine: [23824]: WARN: process_pe_message:Transition 
1: WARNINGs found during PE processing. PEngine Input stored 
in:/var/lib/pengine/pe-warn-259.bz2
Oct 31 02:47:08 storage02 pengine: [23824]: info: process_pe_message: 
Configuration WARNINGs found during PE processing. Please run "crm_verify -L" 
to identify issues.

Thomas Schneider
Directeur des Opérations
Euskill SARL
Web: www.euskill.com
Mobile: +33 (0)6 19 26 47 76
Mail: [email protected]
5 rue de Phalsbourg
F-67000 Strasbourg



_______________________________________________
drbd-user mailing list
[email protected]
http://lists.linbit.com/mailman/listinfo/drbd-user

Reply via email to