Hi, See below.
> Message: 2 > Date: Tue, 8 Jun 2010 18:35:29 +0200 > From: Dejan Muhamedagic <[email protected]> > Subject: Re: [Linux-HA] nfsserver error > To: General Linux-HA mailing list <[email protected]> > Message-ID: <[email protected]> > Content-Type: text/plain; charset=us-ascii > Hi, > On Mon, Jun 07, 2010 at 05:15:13PM -0600, [email protected] wrote: > > Dejan, > > > > Thanks for the reply. See below : > > > > http://www.linux-ha.org/HaNFS > > > > > Message: 6 > > > Date: Mon, 7 Jun 2010 15:16:19 +0200 > > > From: Dejan Muhamedagic <[email protected]> > > > Subject: Re: [Linux-HA] nfsserver error > > > To: General Linux-HA mailing list <[email protected]> > > > Message-ID: <[email protected]> > > > Content-Type: text/plain; charset=us-ascii > > > > > Hi, > > > > > On Fri, Jun 04, 2010 at 02:49:20PM -0600, [email protected] wrote: > > > > I have a 2-node drbd/heartbeat cluster running on RHEL5.4(proliant > > > > dl380G6) I am attempting > > > > to add ha nfs(nfsserver)to and have run into an error. Specifically > > after > > > > adding the nfsserver resource > > > > and then running "crm resource cleanup nfsserver_nlsmtl" I have the > > > > resource running however I also have > > > > the below failed action : > > > > > > > > [r...@nlsmtl6 ~]# crm status > > > > ============ > > > > Last updated: Fri Jun 4 14:16:14 2010 > > > > Stack: Heartbeat > > > > Current DC: nlsmtl6 (16fd6af0-429e-402d-a5d8-a00a818f139a) - partition > > > > with quorum > > > > Version: 1.0.8-3225fc0d98c8fcd0f7b24f0134e89967136a9b00 > > > > 2 Nodes configured, unknown expected votes > > > > 3 Resources configured. > > > > ============ > > > > > > > > Online: [ nlsmtl5 nlsmtl6 ] > > > > > > > > Resource Group: grp_1 > > > > drbddisk_2 (heartbeat:drbddisk): Started nlsmtl5 > > > > Filesystem_3 (ocf::heartbeat:Filesystem): Started nlsmtl5 > > > > rc.primary_5 (lsb:rc.primary): Started nlsmtl5 > > > > IPaddr_nlsmtl (ocf::heartbeat:IPaddr): Started nlsmtl5 > > > > nfsserver_nlsmtl (ocf::heartbeat:nfsserver): Started nlsmtl5 > > > > CL_stonithset_node01 (stonith:external/riloe-iders): Started nlsmtl6 > > > > CL_stonithset_node02 (stonith:external/riloe-iders): Started nlsmtl5 > > > > > > > > Failed actions: > > > > nfsserver_nlsmtl_monitor_0 (node=nlsmtl6, call=12, rc=2, > > status=complete): > > > > invalid parameter > > > > > The only places where this error can occur is if nfs_ip or > > > nfs_shared_infodir are not set. Your configuration below looks > > > fine. Did you check the logs? There should be more log messages > > > from nfsserver. > > > > > Thanks, > > > > > Dejan > > > > To verify I zeroed my config and started from scratch. after loading my > > resources I get : > > > > [r...@nlsmtl5 config_nlsmtl]# crm_mon -1 > > ============ > > Last updated: Mon Jun 7 14:44:30 2010 > > Stack: Heartbeat > > Current DC: nlsmtl6 (6fa3ad00-4761-4e52-842c-36f002971200) - partition > > with quorum > > Version: 1.0.8-3225fc0d98c8fcd0f7b24f0134e89967136a9b00 > > 2 Nodes configured, unknown expected votes > > 3 Resources configured. > > ============ > > > > Online: [ nlsmtl5 nlsmtl6 ] > > > > Resource Group: grp_1 > > drbddisk_2 (heartbeat:drbddisk): Started nlsmtl5 > > Filesystem_3 (ocf::heartbeat:Filesystem): Started nlsmtl5 > > IPaddr_nlsmtl (ocf::heartbeat:IPaddr): Started nlsmtl5 > > rc.primary_5 (lsb:rc.primary): Started nlsmtl5 > > nfsserver_nlsmtl (ocf::heartbeat:nfsserver): Stopped > > CL_stonithset_node01 (stonith:external/riloe-iders): Started nlsmtl6 > > CL_stonithset_node02 (stonith:external/riloe-iders): Started nlsmtl5 > > > > Failed actions: > > nfsserver_nlsmtl_monitor_0 (node=nlsmtl5, call=6, rc=2, status=complete): > > invalid parameter > > nfsserver_nlsmtl_monitor_0 (node=nlsmtl6, call=6, rc=2, status=complete): > > invalid parameter > > [r...@nlsmtl5 config_nlsmtl]# > > > > Why do I have monitoring fail on both nodes - why is it running on nlsmtl6 > > at all ? > Don't see it running. But you still didn't get the right logs. > The problem is with a resource, so you need to lookup messages > either by lrmd or nfsserver. If you can't, then make hb_report > and attach that or post it somewhere. > Thanks, > Dejan Ok finally figured this out after a bunch of debug code ... On startup nfsserver monitor is invoked on BOTH nodes when all resources are STOPPED. If this just invoked a "/etc/init.d/nfs status" (rc=3) then all would be okay however the nfsserver resource has the following code : nfsserver_validate () { check_binary ${OCF_RESKEY_nfs_init_script} check_binary ${OCF_RESKEY_nfs_notify_cmd} if [ -z ${OCF_RESKEY_nfs_ip} ]; then exit $OCF_ERR_ARGS fi if [ -d $OCF_RESKEY_nfs_shared_infodir ]; then return $OCF_SUCCESS else exit $OCF_ERR_ARGS fi } if [ -n "$OCF_RESKEY_CRM_meta_clone" ]; then ocf_log err "THIS RA DO NOT SUPPORT CLONE MODE!" exit $OCF_ERR_CONFIGURED fi nfsserver_validate case $__OCF_ACTION in start) nfsserver_start ;; stop) nfsserver_stop ;; monitor) nfsserver_monitor ;; validate-all) nfsserver_validate ;; *) nfsserver_usage exit $OCF_ERR_UNIMPLEMENTED ;; esac The nfsserver_validate call checks the existence of $OCF_RESKEY_nfs_shared_infodir which fails because the drbd resource has not yet been started. This results with the crm_mon output above. Commenting out nfsserver_validate allows the resource to start. I can send further logs if needed. Still have a question of which services should be turned off(RHEL5.4) before nfsserver is invoked. I presume nfs and nfslock, what about rpcgssd, rpcidmapd ? Thanks > > Here is some of the log file from the DC : > > > > Jun 7 13:08:10 nlsmtl6 crmd: [17967]: info: te_rsc_command: Initiating > > action 10: monitor CL_stonithset_node02_monitor_0 on nlsmtl5 > > Jun 7 13:08:10 nlsmtl6 crmd: [17967]: info: te_rsc_command: Initiating > > action 18: monitor CL_stonithset_node02_monitor_0 on nlsmtl6 (local) > > Jun 7 13:08:10 nlsmtl6 logger: We are not PRIMARY ... > > Jun 7 13:08:10 nlsmtl6 lrmd: [17964]: notice: lrmd_rsc_new(): No > > lrm_rprovider field in message > > Jun 7 13:08:10 nlsmtl6 crmd: [17967]: info: do_lrm_rsc_op: Performing > > key=18:4:7:49711e59-c5cf-45a5-abb2-b0fe09acfce9 > > op=CL_stonithset_node02_monitor_0 ) > > Jun 7 13:08:10 nlsmtl6 lrmd: [17964]: info: rsc:CL_stonithset_node02:8: > > monitor > > Jun 7 13:08:10 nlsmtl6 crmd: [17967]: info: process_lrm_event: LRM > > operation drbddisk_2_monitor_0 (call=2, rc=7, cib-update=58, > > confirmed=true) not running > > Jun 7 13:08:10 nlsmtl6 crmd: [17967]: info: process_lrm_event: Result: > > stopped (Secondary) > > Jun 7 13:08:10 nlsmtl6 crmd: [17967]: info: process_lrm_event: LRM > > operation rc.primary_5_monitor_0 (call=5, rc=7, cib-update=59, > > confirmed=true) not running > > Jun 7 13:08:10 nlsmtl6 crmd: [17967]: info: process_lrm_event: LRM > > operation CL_stonithset_node02_monitor_0 (call=8, rc=7, cib-update=60, > > confirmed=true) not running > > Jun 7 13:08:10 nlsmtl6 crmd: [17967]: info: match_graph_event: Action > > drbddisk_2_monitor_0 (12) confirmed on nlsmtl6 (rc=0) > > Jun 7 13:08:10 nlsmtl6 crmd: [17967]: info: match_graph_event: Action > > rc.primary_5_monitor_0 (15) confirmed on nlsmtl6 (rc=0) > > Jun 7 13:08:10 nlsmtl6 crmd: [17967]: info: match_graph_event: Action > > CL_stonithset_node02_monitor_0 (18) confirmed on nlsmtl6 (rc=0) > > Jun 7 13:08:10 nlsmtl6 crmd: [17967]: info: process_lrm_event: LRM > > operation nfsserver_nlsmtl_monitor_0 (call=6, rc=2, cib-update=61, > > confirmed=true) invalid parameter > > Jun 7 13:08:10 nlsmtl6 crmd: [17967]: WARN: status_from_rc: Action 16 > > (nfsserver_nlsmtl_monitor_0) on nlsmtl6 failed (target: 7 vs. rc: 2): > > Error > > Jun 7 13:08:10 nlsmtl6 attrd: [17966]: info: attrd_ha_callback: flush > > message from nlsmtl6 > > Jun 7 13:08:10 nlsmtl6 attrd: [17966]: info: attrd_ha_callback: flush > > message from nlsmtl6 > > Jun 7 13:08:10 nlsmtl6 crmd: [17967]: info: abort_transition_graph: > > match_graph_event:272 - Triggered transition abort (complete=0, > > tag=lrm_rsc_op, id=nfsserver_nlsmtl_monitor_0, > > magic=0:2;16:4:7:49711e59-c5cf-45a5-abb2-b0fe09acfce9, cib=0.8.5) : Event > > failed > > Jun 7 13:08:10 nlsmtl6 crmd: [17967]: info: update_abort_priority: Abort > > priority upgraded from 0 to 1 > > Jun 7 13:08:10 nlsmtl6 crmd: [17967]: info: update_abort_priority: Abort > > action done superceeded by restart > > Jun 7 13:08:10 nlsmtl6 crmd: [17967]: info: match_graph_event: Action > > nfsserver_nlsmtl_monitor_0 (16) confirmed on nlsmtl6 (rc=4) > > Jun 7 13:08:10 nlsmtl6 crmd: [17967]: info: process_lrm_event: LRM > > operation IPaddr_nlsmtl_monitor_0 (call=4, rc=7, cib-update=62, > > confirmed=true) not running > > Jun 7 13:08:10 nlsmtl6 crmd: [17967]: info: process_lrm_event: LRM > > operation Filesystem_3_monitor_0 (call=3, rc=7, cib-update=63, > > confirmed=true) not running > > Jun 7 13:08:10 nlsmtl6 crmd: [17967]: info: match_graph_event: Action > > IPaddr_nlsmtl_monitor_0 (14) confirmed on nlsmtl6 (rc=0) > > Jun 7 13:08:10 nlsmtl6 crmd: [17967]: info: match_graph_event: Action > > Filesystem_3_monitor_0 (13) confirmed on nlsmtl6 (rc=0) > > Jun 7 13:08:11 nlsmtl6 lrmd: [17964]: info: rsc:CL_stonithset_node01:7: > > monitor > > Jun 7 13:08:11 nlsmtl6 crmd: [17967]: info: process_lrm_event: LRM > > operation CL_stonithset_node01_monitor_0 (call=7, rc=7, cib-update=64, > > confirmed=true) not running > > Jun 7 13:08:11 nlsmtl6 crmd: [17967]: info: match_graph_event: Action > > CL_stonithset_node01_monitor_0 (17) confirmed on nlsmtl6 (rc=0) > > Jun 7 13:08:11 nlsmtl6 crmd: [17967]: info: te_rsc_command: Initiating > > action 11: probe_complete probe_complete on nlsmtl6 (local) - no waiting > > Jun 7 13:08:11 nlsmtl6 crmd: [17967]: info: match_graph_event: Action > > rc.primary_5_monitor_0 (7) confirmed on nlsmtl5 (rc=0) > > Jun 7 13:08:11 nlsmtl6 crmd: [17967]: info: match_graph_event: Action > > drbddisk_2_monitor_0 (4) confirmed on nlsmtl5 (rc=0) > > Jun 7 13:08:11 nlsmtl6 crmd: [17967]: info: match_graph_event: Action > > IPaddr_nlsmtl_monitor_0 (6) confirmed on nlsmtl5 (rc=0) > > Jun 7 13:08:11 nlsmtl6 crmd: [17967]: info: match_graph_event: Action > > Filesystem_3_monitor_0 (5) confirmed on nlsmtl5 (rc=0) > > Jun 7 13:08:12 nlsmtl6 crmd: [17967]: info: match_graph_event: Action > > CL_stonithset_node02_monitor_0 (10) confirmed on nlsmtl5 (rc=0) > > Jun 7 13:08:12 nlsmtl6 crmd: [17967]: info: match_graph_event: Action > > CL_stonithset_node01_monitor_0 (9) confirmed on nlsmtl5 (rc=0) > > Jun 7 13:08:12 nlsmtl6 crmd: [17967]: WARN: status_from_rc: Action 8 > > (nfsserver_nlsmtl_monitor_0) on nlsmtl5 failed (target: 7 vs. rc: 2): > > Error > > Jun 7 13:08:13 nlsmtl6 crmd: [17967]: info: abort_transition_graph: > > match_graph_event:272 - Triggered transition abort (complete=0, > > tag=lrm_rsc_op, id=nfsserver_nlsmtl_monitor_0, > > magic=0:2;8:4:7:49711e59-c5cf-45a5-abb2-b0fe09acfce9, cib=0.8.15) : Event > > failed > > Jun 7 13:08:13 nlsmtl6 crmd: [17967]: info: match_graph_event: Action > > nfsserver_nlsmtl_monitor_0 (8) confirmed on nlsmtl5 (rc=4) > > Jun 7 13:08:13 nlsmtl6 crmd: [17967]: info: te_rsc_command: Initiating > > action 3: probe_complete probe_complete on nlsmtl5 - no waiting > > Jun 7 13:08:13 nlsmtl6 crmd: [17967]: info: run_graph: > > ==================================================== > > Jun 7 13:08:13 nlsmtl6 crmd: [17967]: notice: run_graph: Transition 4 > > (Complete=16, Pending=0, Fired=0, Skipped=25, Incomplete=0, > > Source=/var/lib/pengine/pe-input-4724.bz2): Stopped > > Jun 7 13:08:13 nlsmtl6 crmd: [17967]: info: te_graph_trigger: Transition > > 4 is now complete > > Jun 7 13:08:13 nlsmtl6 crmd: [17967]: info: do_state_transition: State > > transition S_TRANSITION_ENGINE -> S_POLICY_ENGINE [ input=I_PE_CALC > > cause=C_FSA_INTERNAL origin=notify_crmd ] > > Jun 7 13:08:13 nlsmtl6 crmd: [17967]: info: do_state_transition: All 2 > > cluster nodes are eligible to run resources. > > Jun 7 13:08:13 nlsmtl6 crmd: [17967]: info: do_pe_invoke: Query 65: > > Requesting the current CIB: S_POLICY_ENGINE > > Jun 7 13:08:13 nlsmtl6 crmd: [17967]: info: do_pe_invoke_callback: > > Invoking the PE: query=65, ref=pe_calc-dc-1275937693-40, seq=2, quorate=1 > > Jun 7 13:08:13 nlsmtl6 pengine: [19553]: notice: unpack_config: On loss > > of CCM Quorum: Ignore > > Jun 7 13:08:13 nlsmtl6 pengine: [19553]: info: unpack_config: Node > > scores: 'red' = -INFINITY, 'yellow' = 0, 'green' = 0 > > Jun 7 13:08:13 nlsmtl6 pengine: [19553]: info: determine_online_status: > > Node nlsmtl5 is online > > Jun 7 13:08:13 nlsmtl6 pengine: [19553]: ERROR: unpack_rsc_op: Hard error > > - nfsserver_nlsmtl_monitor_0 failed with rc=2: Preventing nfsserver_nlsmtl > > from re-starting on nlsmtl5 > > Jun 7 13:08:13 nlsmtl6 pengine: [19553]: info: determine_online_status: > > Node nlsmtl6 is online > > Jun 7 13:08:13 nlsmtl6 pengine: [19553]: ERROR: unpack_rsc_op: Hard error > > - nfsserver_nlsmtl_monitor_0 failed with rc=2: Preventing nfsserver_nlsmtl > > from re-starting on nlsmtl6 > > Jun 7 13:08:13 nlsmtl6 pengine: [19553]: notice: group_print: Resource > > Group: grp_1 > > Jun 7 13:08:13 nlsmtl6 pengine: [19553]: notice: native_print: drbddisk_2 > > (heartbeat:drbddisk): Stopped > > Jun 7 13:08:13 nlsmtl6 pengine: [19553]: notice: native_print: > > Filesystem_3 (ocf::heartbeat:Filesystem): Stopped > > Jun 7 13:08:13 nlsmtl6 pengine: [19553]: notice: native_print: > > IPaddr_nlsmtl (ocf::heartbeat:IPaddr): Stopped > > Jun 7 13:08:13 nlsmtl6 pengine: [19553]: notice: native_print: > > rc.primary_5 (lsb:rc.primary): Stopped > > Jun 7 13:08:13 nlsmtl6 pengine: [19553]: notice: native_print: > > nfsserver_nlsmtl (ocf::heartbeat:nfsserver): Stopped > > Jun 7 13:08:13 nlsmtl6 pengine: [19553]: notice: native_print: > > CL_stonithset_node01 (stonith:external/riloe-iders): Stopped > > Jun 7 13:08:13 nlsmtl6 pengine: [19553]: notice: native_print: > > CL_stonithset_node02 (stonith:external/riloe-iders): Stopped > > Jun 7 13:08:13 nlsmtl6 pengine: [19553]: info: native_merge_weights: > > drbddisk_2: Rolling back scores from Filesystem_3 > > Jun 7 13:08:13 nlsmtl6 pengine: [19553]: info: native_merge_weights: > > drbddisk_2: Rolling back scores from IPaddr_nlsmtl > > Jun 7 13:08:13 nlsmtl6 pengine: [19553]: info: native_merge_weights: > > drbddisk_2: Rolling back scores from rc.primary_5 > > Jun 7 13:08:13 nlsmtl6 pengine: [19553]: info: native_merge_weights: > > drbddisk_2: Rolling back scores from nfsserver_nlsmtl > > Jun 7 13:08:13 nlsmtl6 pengine: [19553]: info: native_merge_weights: > > Filesystem_3: Rolling back scores from IPaddr_nlsmtl > > Jun 7 13:08:13 nlsmtl6 pengine: [19553]: info: native_merge_weights: > > Filesystem_3: Rolling back scores from rc.primary_5 > > Jun 7 13:08:13 nlsmtl6 pengine: [19553]: info: native_merge_weights: > > Filesystem_3: Rolling back scores from nfsserver_nlsmtl > > Jun 7 13:08:13 nlsmtl6 pengine: [19553]: info: native_merge_weights: > > IPaddr_nlsmtl: Rolling back scores from rc.primary_5 > > Jun 7 13:08:13 nlsmtl6 pengine: [19553]: info: native_merge_weights: > > IPaddr_nlsmtl: Rolling back scores from nfsserver_nlsmtl > > Jun 7 13:08:13 nlsmtl6 pengine: [19553]: info: native_merge_weights: > > rc.primary_5: Rolling back scores from nfsserver_nlsmtl > > Jun 7 13:08:13 nlsmtl6 pengine: [19553]: info: native_color: Resource > > nfsserver_nlsmtl cannot run anywhere > > Jun 7 13:08:13 nlsmtl6 pengine: [19553]: notice: RecurringOp: Start > > recurring monitor (120s) for drbddisk_2 on nlsmtl5 > > Jun 7 13:08:13 nlsmtl6 pengine: [19553]: notice: RecurringOp: Start > > recurring monitor (120s) for Filesystem_3 on nlsmtl5 > > Jun 7 13:08:13 nlsmtl6 pengine: [19553]: notice: RecurringOp: Start > > recurring monitor (5s) for IPaddr_nlsmtl on nlsmtl5 > > Jun 7 13:08:13 nlsmtl6 pengine: [19553]: notice: RecurringOp: Start > > recurring monitor (120s) for rc.primary_5 on nlsmtl5 > > Jun 7 13:08:13 nlsmtl6 pengine: [19553]: notice: RecurringOp: Start > > recurring monitor (30s) for CL_stonithset_node01 on nlsmtl6 > > Jun 7 13:08:13 nlsmtl6 pengine: [19553]: notice: RecurringOp: Start > > recurring monitor (30s) for CL_stonithset_node02 on nlsmtl5 > > Jun 7 13:08:13 nlsmtl6 pengine: [19553]: notice: LogActions: Start > > drbddisk_2 (nlsmtl5) > > Jun 7 13:08:13 nlsmtl6 pengine: [19553]: notice: LogActions: Start > > Filesystem_3 (nlsmtl5) > > Jun 7 13:08:13 nlsmtl6 pengine: [19553]: notice: LogActions: Start > > IPaddr_nlsmtl (nlsmtl5) > > Jun 7 13:08:13 nlsmtl6 pengine: [19553]: notice: LogActions: Start > > rc.primary_5 (nlsmtl5) > > Jun 7 13:08:13 nlsmtl6 pengine: [19553]: notice: LogActions: Leave > > resource nfsserver_nlsmtl (Stopped) > > Jun 7 13:08:13 nlsmtl6 pengine: [19553]: notice: LogActions: Start > > CL_stonithset_node01 (nlsmtl6) > > Jun 7 13:08:13 nlsmtl6 pengine: [19553]: notice: LogActions: Start > > CL_stonithset_node02 (nlsmtl5) > > Jun 7 13:08:13 nlsmtl6 crmd: [17967]: info: do_state_transition: State > > transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS > > cause=C_IPC_MESSAGE origin=handle_response > > > > At this point if I try : > > crm(live)resource# cleanup nfsserver_nlsmtl > > Cleaning up nfsserver_nlsmtl on nlsmtl5 > > Cleaning up nfsserver_nlsmtl on nlsmtl6 > > crm(live)resource# > > > > - we get : > > > > [r...@nlsmtl5 config_nlsmtl]# crm_mon -1 > > ============ > > Last updated: Mon Jun 7 15:20:40 2010 > > Stack: Heartbeat > > Current DC: nlsmtl6 (6fa3ad00-4761-4e52-842c-36f002971200) - partition > > with quorum > > Version: 1.0.8-3225fc0d98c8fcd0f7b24f0134e89967136a9b00 > > 2 Nodes configured, unknown expected votes > > 3 Resources configured. > > ============ > > > > Online: [ nlsmtl5 nlsmtl6 ] > > > > Resource Group: grp_1 > > drbddisk_2 (heartbeat:drbddisk): Started nlsmtl5 > > Filesystem_3 (ocf::heartbeat:Filesystem): Started nlsmtl5 > > IPaddr_nlsmtl (ocf::heartbeat:IPaddr): Started nlsmtl5 > > rc.primary_5 (lsb:rc.primary): Started nlsmtl5 > > nfsserver_nlsmtl (ocf::heartbeat:nfsserver): Started nlsmtl5 > > CL_stonithset_node01 (stonith:external/riloe-iders): Started nlsmtl6 > > CL_stonithset_node02 (stonith:external/riloe-iders): Started nlsmtl5 > > > > Failed actions: > > nfsserver_nlsmtl_monitor_0 (node=nlsmtl6, call=11, rc=2, status=complete): > > invalid parameter > > [r...@nlsmtl5 config_nlsmtl]# > > > > The resource is now running however because it has failed on nlsmtl6 I > > cannot failover to > > nlsmtl6. How to I cleat the failed action ? > > > > Regardless I'm still not sure wht my correct setup is before I attempt to > > add nfsserver. > > Which services should be running on nlsmtl5/6. nfslock ? rpcidmapd ? I > > found the below > > url however as shown this is not supported. > > > > http://www.linux-ha.org/HaNFS > > > > Thanks > > > > [r...@nlsmtl6 ~]# > > > > > > > > At this point I cannot migrate to nlsmtl6. I have not been able to > > find > > > > much > > > > documentation on nfsserver. I've found > > > > http://linux-ha.org/doc/re-ra-nfsserver.html > > > > but nothing else on setup. > > > > > > > > My config is : > > > > > > > > r...@nlsmtl5 init.d]# crm configure show > > > > node $id="16fd6af0-429e-402d-a5d8-a00a818f139a" nlsmtl6 \ > > > > attributes standby="off" > > > > node $id="2f6b429e-74c3-482e-bf20-5a6b0c94cd46" nlsmtl5 \ > > > > attributes standby="off" > > > > primitive CL_stonithset_node01 stonith:external/riloe-iders \ > > > > op monitor interval="30s" timeout="20s" on-fail="ignore" \ > > > > op start interval="0" timeout="60s" on-fail="restart" \ > > > > params hostlist="nlsmtl5" ilo_hostname="nlsmtl5-ilo" > > ilo_user="Heartbeat" > > > > ilo_password="xxx" ilo_can_reset="0" ilo_protocol="2.0" > > > > ilo_powerdown_method="button" > > > > primitive CL_stonithset_node02 stonith:external/riloe-iders \ > > > > op monitor interval="30s" timeout="20s" on-fail="ignore" \ > > > > op start interval="0" timeout="60s" on-fail="restart" \ > > > > params hostlist="nlsmtl6" ilo_hostname="nlsmtl6-ilo" > > ilo_user="Heartbeat" > > > > ilo_password="xxx" ilo_can_reset="0" ilo_protocol="2.0" > > > > ilo_powerdown_method="button" > > > > primitive Filesystem_3 ocf:heartbeat:Filesystem \ > > > > op monitor interval="120s" timeout="60s" \ > > > > params device="/dev/drbd0" directory="/drbd" fstype="ext3" > > > > options="defaults" > > > > primitive IPaddr_nlsmtl ocf:heartbeat:IPaddr \ > > > > op monitor interval="5s" timeout="5s" \ > > > > params ip="165.115.204.222" > > > > primitive drbddisk_2 heartbeat:drbddisk \ > > > > op monitor interval="120s" timeout="60s" \ > > > > params 1="r0" > > > > primitive nfsserver_nlsmtl ocf:heartbeat:nfsserver \ > > > > op monitor interval="30s" timeout="60s" \ > > > > params nfs_init_script="/etc/init.d/nfs" > > nfs_notify_cmd="/sbin/rpc.statd" > > > > nfs_shared_infodir="/drbd/nfs" nfs_ip="165.115.204.222" > > > > > > primitive rc.primary_5 lsb:rc.primary \ > > > > op monitor interval="120s" timeout="60s" > > > > group grp_1 drbddisk_2 Filesystem_3 rc.primary_5 IPaddr_nlsmtl > > > > nfsserver_nlsmtl > > > > location node-1-dont-run CL_stonithset_node01 -inf: nlsmtl5 > > > > location node-2-dont-run CL_stonithset_node02 -inf: nlsmtl6 > > > > location rsc_location_group_1 grp_1 100: nlsmtl5 > > > > property $id="cib-bootstrap-options" \ > > > > dc-version="1.0.8-3225fc0d98c8fcd0f7b24f0134e89967136a9b00" \ > > > > cluster-infrastructure="Heartbeat" \ > > > > no-quorum-policy="ignore" \ > > > > last-lrm-refresh="1275682003" > > > > > > > > My packages are : > > > > drbd-pacemaker-8.3.7-1 > > > > heartbeat-3.0.2-2.el5 > > > > pacemaker-1.0.8-2.el5 > > > > pacemaker-libs-1.0.8-2.el5 > > > > cluster-glue-1.0.3-1.el5 > > > > cluster-glue-libs-1.0.3-1.el5 > > > > corosynclib-1.2.0-1.el5 > > > > corosync-1.2.0-1.el5 > > > > > > > > I looked at /usr/lib/ocf/resource.d/heartbeat/nfsserver and > > > > nfsserver_monitor which is pretty simple. > > > > nfsserver_monitor () > > > > { > > > > fn=`/bin/mktemp` > > > > ${OCF_RESKEY_nfs_init_script} status > $fn 2>&1 > > > > rc=$? > > > > ocf_log debug `cat $fn` > > > > rm -f $fn > > > > > > > > #Adapte LSB status code to OCF return code > > > > if [ $rc -eq 0 ]; then > > > > return $OCF_SUCCESS > > > > elif [ $rc -eq 3 ]; then > > > > return $OCF_NOT_RUNNING > > > > else > > > > return $OCF_ERR_GENERIC > > > > fi > > > > } > > > > > > > > In my case I presume it would call "/etc/init.d/nfs status" which on > > : > > > > nlsmtl5 returns : > > > > [r...@nlsmtl5 init.d]# /etc/init.d/nfs status > > > > rpc.mountd (pid 24994) is running... > > > > nfsd (pid 24991 24990 24989 24988 24987 24980 24979 24969) is > > running... > > > > rpc.rquotad (pid 24963) is running... > > > > [r...@nlsmtl5 init.d]# > > > > - return code is 0 > > > > > > > > nlsmtl6 returns : > > > > [r...@nlsmtl6 ~]# /etc/init.d/nfs status > > > > rpc.mountd is stopped > > > > nfsd is stopped > > > > rpc.rquotad is stopped > > > > [r...@nlsmtl6 ~]# > > > > - return code is 3 > > > > > > > > Why am I getting a rc=2 and how can I debug ? Am I missing something > > on > > > > setup ? Is > > > > this the best way to run nfs ? > > > > > > > > Thanks > > > > > > > > > > _______________________________________________ > > Linux-HA mailing list > > [email protected] > > http://lists.linux-ha.org/mailman/listinfo/linux-ha > > See also: http://linux-ha.org/ReportingProblems _______________________________________________ Linux-HA mailing list [email protected] http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems
