I have a 2-node drbd/heartbeat cluster running on RHEL5.4(proliant 
dl380G6) I am attempting
to add ha nfs(nfsserver)to and have run into an error. Specifically after 
adding the nfsserver resource
and then running "crm resource cleanup nfsserver_nlsmtl" I have the 
resource running however I also have
the below failed action :

[r...@nlsmtl6 ~]# crm status
============
Last updated: Fri Jun  4 14:16:14 2010
Stack: Heartbeat
Current DC: nlsmtl6 (16fd6af0-429e-402d-a5d8-a00a818f139a) - partition 
with quorum
Version: 1.0.8-3225fc0d98c8fcd0f7b24f0134e89967136a9b00
2 Nodes configured, unknown expected votes
3 Resources configured.
============

Online: [ nlsmtl5 nlsmtl6 ]

Resource Group: grp_1
drbddisk_2 (heartbeat:drbddisk):   Started nlsmtl5
Filesystem_3       (ocf::heartbeat:Filesystem):    Started nlsmtl5
rc.primary_5       (lsb:rc.primary):       Started nlsmtl5
IPaddr_nlsmtl      (ocf::heartbeat:IPaddr):        Started nlsmtl5
nfsserver_nlsmtl   (ocf::heartbeat:nfsserver):     Started nlsmtl5
CL_stonithset_node01   (stonith:external/riloe-iders): Started nlsmtl6
CL_stonithset_node02   (stonith:external/riloe-iders): Started nlsmtl5

Failed actions:
nfsserver_nlsmtl_monitor_0 (node=nlsmtl6, call=12, rc=2, status=complete): 
invalid parameter
[r...@nlsmtl6 ~]#

At this point I cannot migrate to nlsmtl6.  I have not been able to find 
much 
documentation on nfsserver. I've found 
http://linux-ha.org/doc/re-ra-nfsserver.html
but nothing else on setup.

My config is :

r...@nlsmtl5 init.d]# crm configure show
node $id="16fd6af0-429e-402d-a5d8-a00a818f139a" nlsmtl6 \
attributes standby="off"
node $id="2f6b429e-74c3-482e-bf20-5a6b0c94cd46" nlsmtl5 \
attributes standby="off"
primitive CL_stonithset_node01 stonith:external/riloe-iders \
op monitor interval="30s" timeout="20s" on-fail="ignore" \
op start interval="0" timeout="60s" on-fail="restart" \
params hostlist="nlsmtl5" ilo_hostname="nlsmtl5-ilo" ilo_user="Heartbeat" 
ilo_password="xxx" ilo_can_reset="0" ilo_protocol="2.0" 
ilo_powerdown_method="button"
primitive CL_stonithset_node02 stonith:external/riloe-iders \
op monitor interval="30s" timeout="20s" on-fail="ignore" \
op start interval="0" timeout="60s" on-fail="restart" \
params hostlist="nlsmtl6" ilo_hostname="nlsmtl6-ilo" ilo_user="Heartbeat" 
ilo_password="xxx" ilo_can_reset="0" ilo_protocol="2.0" 
ilo_powerdown_method="button"
primitive Filesystem_3 ocf:heartbeat:Filesystem \
op monitor interval="120s" timeout="60s" \
params device="/dev/drbd0" directory="/drbd" fstype="ext3" 
options="defaults"
primitive IPaddr_nlsmtl ocf:heartbeat:IPaddr \
op monitor interval="5s" timeout="5s" \
params ip="165.115.204.222"
primitive drbddisk_2 heartbeat:drbddisk \
op monitor interval="120s" timeout="60s" \
params 1="r0"
primitive nfsserver_nlsmtl ocf:heartbeat:nfsserver \
op monitor interval="30s" timeout="60s" \
params nfs_init_script="/etc/init.d/nfs" nfs_notify_cmd="/sbin/rpc.statd" 
nfs_shared_infodir="/drbd/nfs" nfs_ip="165.115.204.222"
primitive rc.primary_5 lsb:rc.primary \
op monitor interval="120s" timeout="60s"
group grp_1 drbddisk_2 Filesystem_3 rc.primary_5 IPaddr_nlsmtl 
nfsserver_nlsmtl
location node-1-dont-run CL_stonithset_node01 -inf: nlsmtl5
location node-2-dont-run CL_stonithset_node02 -inf: nlsmtl6
location rsc_location_group_1 grp_1 100: nlsmtl5
property $id="cib-bootstrap-options" \
dc-version="1.0.8-3225fc0d98c8fcd0f7b24f0134e89967136a9b00" \
cluster-infrastructure="Heartbeat" \
no-quorum-policy="ignore" \
last-lrm-refresh="1275682003" 

My packages are :
drbd-pacemaker-8.3.7-1
heartbeat-3.0.2-2.el5
pacemaker-1.0.8-2.el5
pacemaker-libs-1.0.8-2.el5
cluster-glue-1.0.3-1.el5
cluster-glue-libs-1.0.3-1.el5
corosynclib-1.2.0-1.el5
corosync-1.2.0-1.el5

I looked at  /usr/lib/ocf/resource.d/heartbeat/nfsserver and 
nfsserver_monitor which is pretty simple.
nfsserver_monitor ()
{
fn=`/bin/mktemp`
${OCF_RESKEY_nfs_init_script} status > $fn 2>&1
rc=$?
ocf_log debug `cat $fn`
rm -f $fn

#Adapte LSB status code to OCF return code
if [ $rc -eq 0 ]; then
return $OCF_SUCCESS
elif [ $rc -eq 3 ]; then
return $OCF_NOT_RUNNING
else
return $OCF_ERR_GENERIC
fi
}

In my case I presume it would call  "/etc/init.d/nfs status"  which on :
nlsmtl5 returns :
[r...@nlsmtl5 init.d]# /etc/init.d/nfs status
rpc.mountd (pid 24994) is running...
nfsd (pid 24991 24990 24989 24988 24987 24980 24979 24969) is running...
rpc.rquotad (pid 24963) is running...
[r...@nlsmtl5 init.d]#
- return code is 0

nlsmtl6 returns :
[r...@nlsmtl6 ~]# /etc/init.d/nfs status
rpc.mountd is stopped
nfsd is stopped
rpc.rquotad is stopped
[r...@nlsmtl6 ~]#
- return code is 3

Why am I getting a rc=2 and how can I debug ? Am I missing something on 
setup ? Is 
this the best way to run nfs ?

Thanks


_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to