>>> Matthew Spah <[email protected]> schrieb am 08.01.2013 um 00:18 in Nachricht <caoqsdi4urf7bjofsa-4aau+dzfyugldkx27ysqcjzw2fhn8...@mail.gmail.com>: > Everything looks okay... the strings being exported weren't being resolved. > I'm going to update the exportfs ocf now from github. > > If it makes any difference.. I can get the cluster up for a few seconds > then the p_exportfs_nfs_monitor action starts to time out, and it all goes > down hill from there.
Nameserver (i.e. address to name resolution) timeout, I guess... Maybe play with ocf-tester, testing the exportfs RA. Regards, Ulrich > > On Mon, Jan 7, 2013 at 9:17 AM, William Seligman < > [email protected]> wrote: > > > On 1/4/13 7:10 PM, Matthew Spah wrote: > > > Hey everyone, > > > > > > I've just recently built up a pacemaker cluster and have begun testing > > it. > > > Everything has been going great until after Christmas break.. I fired up > > > the cluster to find this going on. > > > > > > ============ > > > Last updated: Fri Jan 4 16:06:41 2013 > > > Last change: Fri Jan 4 16:02:13 2013 via crmd on emserver1 > > > Stack: openais > > > Current DC: emserver1 - partition with quorum > > > Version: 1.1.6-9971ebba4494012a93c03b40a2c58ec0eb60f50c > > > 2 Nodes configured, 2 expected votes > > > 9 Resources configured. > > > ============ > > > > > > Online: [ emserver1 emserver2 ] > > > > > > Master/Slave Set: ms_drbd_nfs [p_drbd_nfs] > > > Masters: [ emserver2 ] > > > Slaves: [ emserver1 ] > > > Clone Set: cl_lsb_nfsserver [p_lsb_nfsserver] > > > Started: [ emserver1 emserver2 ] > > > Resource Group: g_nfs > > > p_fs_nfs (ocf::heartbeat:Filesystem): Started emserver2 > > > p_exportfs_nfs (ocf::heartbeat:exportfs): Started emserver2 > > > (unmanaged) FAILED > > > p_ip_nfs (ocf::heartbeat:IPaddr2): Stopped > > > Clone Set: cl_exportfs_root [p_exportfs_root] > > > Started: [ emserver2 ] > > > Stopped: [ p_exportfs_root:1 ] > > > > > > Failed actions: > > > p_exportfs_root:0_start_0 (node=emserver1, call=10, rc=-2, > > status=Timed > > > Out): unknown exec error > > > p_exportfs_root:1_monitor_30000 (node=emserver2, call=11, rc=7, > > > status=complete): not running > > > p_exportfs_nfs_stop_0 (node=emserver2, call=39, rc=-2, status=Timed > > > Out): unknown exec error > > > > > > > > > I've been reading through documentation to figure out what is going on. > > If > > > you guys could point me in the right direction that would be a huge > > help. :) > > > > > > Here is my configuration... > > > node emserver1 > > > node emserver2 > > > primitive p_drbd_nfs ocf:linbit:drbd \ > > > params drbd_resource="r0" \ > > > op monitor interval="15" role="Master" \ > > > op monitor interval="30" role="Slave" > > > primitive p_exportfs_nfs ocf:heartbeat:exportfs \ > > > params fsid="1" directory="/srv/nfs" options="rw,crossmnt" > > > clientspec="10.1.10.0/255.255.255.0" \ > > > op monitor interval="30s" > > > primitive p_exportfs_root ocf:heartbeat:exportfs \ > > > params fsid="0" directory="/srv" options="rw,crossmnt" > > clientspec=" > > > 10.1.10.0/255.255.255.0" \ > > > op monitor interval="30s" > > > primitive p_fs_nfs ocf:heartbeat:Filesystem \ > > > params device="/dev/drbd1" directory="/srv/nfs" fstype="ext3" \ > > > op monitor interval="10s" > > > primitive p_ip_nfs ocf:heartbeat:IPaddr2 \ > > > params ip="10.1.10.10" cidr_netmask="24" iflabel="NFSV_IP" \ > > > op monitor interval="30s" > > > primitive p_lsb_nfsserver lsb:nfs-kernel-server \ > > > op monitor interval="30s" > > > group g_nfs p_fs_nfs p_exportfs_nfs p_ip_nfs > > > ms ms_drbd_nfs p_drbd_nfs \ > > > meta master-max="1" master-node-max="1" clone-max="2" > > > clone-node-max="1" notify="true" > > > clone cl_exportfs_root p_exportfs_root > > > clone cl_lsb_nfsserver p_lsb_nfsserver > > > colocation c_nfs_on_drbd inf: g_nfs ms_drbd_nfs:Master > > > colocation c_nfs_on_root inf: g_nfs cl_exportfs_root > > > order o_drbd_before_nfs inf: ms_drbd_nfs:promote g_nfs:start > > > order o_root_before_nfs inf: cl_exportfs_root g_nfs:start > > > property $id="cib-bootstrap-options" \ > > > dc-version="1.1.6-9971ebba4494012a93c03b40a2c58ec0eb60f50c" \ > > > cluster-infrastructure="openais" \ > > > expected-quorum-votes="2" \ > > > stonith-enabled="false" \ > > > no-quorum-policy="ignore" \ > > > maintenance-mode="false" \ > > > last-lrm-refresh="1357344133" > > > rsc_defaults $id="rsc-options" \ > > > resource-stickiness="200" > > > > I've had problems like this with the exportfs resource. Here are some > > things to > > check: > > > > - You didn't list the software versions. In particular, look at the > > version of > > your resource-agents package. There have been some recent changes to the > > ocf:heartbeat:exportfs script that improve the pattern-matching in its > > monitor > > action. > > > > - The ocf:heartbeat:exportfs monitor works by comparing the clientspec > > parameter > > with the output of the exportfs command. Check when you export to > > 10.1.10.0 that > > the output of exportfs returns exactly that string, instead of a resolved > > name. > > > > It may help to give a concrete example: I exported a partition via > > ocf:heartbeat:exportfs to clientspec=mail.nevis.columbia.edu. The monitor > > action > > always failed, until I realized that mail.nevis.columbia.edu was an alias > > for > > franklin.nevis.columbia.edu; that was the name that appeared in the > > output of > > /usr/sbin/exportfs. > > > > Hope this helps. > > -- > > William Seligman | Phone: (914) 591-2823 > > Nevis Labs, Columbia Univ | > > PO Box 137 | > > Irvington NY 10533 USA | http://www.nevis.columbia.edu/~seligman/ > > > > > > _______________________________________________ > > Linux-HA mailing list > > [email protected] > > http://lists.linux-ha.org/mailman/listinfo/linux-ha > > See also: http://linux-ha.org/ReportingProblems > > > > _______________________________________________ Linux-HA mailing list [email protected] http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems
