On Thu, Jul 19, 2007 at 05:12:56PM +0200, Sebastian Reitenbach wrote:
> Hi list,
> 
> I am going to build an active active NFS server, where one exports a public 
> directory, and the other the home directories. In case one fails, both 
> should be exported by the remaining one server. 
> I have a shared storage on a SAN, connected to both servers, I use the 
> Filesystem ocf script to mount/umount the partitions (ext3, ocfs2 doesn't 
> have ACL's, and I do not get GFS2 to work). Therefore I cannot run a nfs 
> server clone, because I cannot umount the partition when the nfs server 
> still lives on it, and the shared IP is wandering. I only see the LSB Script 
> available for managing the nfsserver, but with the LSB script, only one NFS 
> server can be started or stopped.
> 
> 
> So I have to configure two NFS resources using the LSB script, so that both 
> can life on different servers. But now when I manually tell on nfs resource 
> to move to another server, then both nfs resources will not be available for 
> a short time. I also saw some problems when a dead node comes back into the 
> cluster, also both nfs server resouces were not available for a short time.
> 
> An other option would be to create a OCF script (I haven't found one) to 
> manage the nfsserver. In the manual page of rpc.mountd I have seen that it 
> is possible to specify a exports file and the port automatically. But I 
> don't know what kind of other problems I might get, or whether it will be 
> possible to run two nfsservers in parallel.
> 
> anybody has an idea?

You can start here: http://linux-ha.org/HaNFS

> 
> kind regards
> Sebastian
> 
> _______________________________________________
> Linux-HA mailing list
> [email protected]
> http://lists.linux-ha.org/mailman/listinfo/linux-ha
> See also: http://linux-ha.org/ReportingProblems

-- 
Dejan
_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to