On 06/06/2012 08:16 AM, Torsten Rosenberger wrote:
> Am 06.06.2012 05:50, schrieb Yount, William D:
>> I understand what heartbeat does in the general sense. Actually configuring 
>> it correctly and making it work the way it is supposed to is the problem.
>>
>> I have read the official DRBD/Heartbeat documentation 
>> (http://www.linbit.com/fileadmin/tech-guides/ha-nfs.pdf). That covers a LVM 
>> situation that isn't applicable to me. I use LVM but just have one logical 
>> volume so no need to group them.
>>
>> I have been able to cobble together a set of steps based off of the official 
>> documentation and other guides. Different documentation takes different 
>> approaches and they often contain contradictory information.
>>
>> I have two servers with two 2tb hard drives each. I am using software RAID 
>> with logical volumes. I have one 50gb LV for the OS, one 30gb LV for swap 
>> and one 1.7tb volume for Storage. All I want is to mirror that 1.7tb LV 
>> across servers and then have pacemaker/heartbeat switch over the second 
>> server. 
>>
>> I am not sure if I need to define nfs-kernel-server, LVM, exportFS and drbd0 
>> as services. I am using the LCMC application to monitor the configuration. 
>>
>> Using the steps that I attached, if the primary server goes down, the 
>> secondary does nothing. It doesn't mount /dev/drbd0 to /Storage and it 
>> doesn't start accepting traffic on 10.89.99.30. 
>>
>>
>>
>>
>>
>>
>> -----Original Message-----
>> From: Marcel Kraan [mailto:[email protected]] 
>> Sent: Tuesday, June 05, 2012 5:19 PM
>> To: Yount, William D
>> Cc: Felix Frank; [email protected]
>> Subject: Re: [DRBD-user] Fault Tolerant NFS
>>
>> This is what heartbeat does.
>> It mount the drbd disk  and start all the programs that are given in the 
>> haresources the virtual ip will be on the second server up and running.
>> so basically your 1servers becomes the second.
>> when the 1st server come up again he will take it over again.
>>
>> i can shutdown the first or second server without going down.. (maybe 5 or 
>> 10 seconds for switching)
>>
>> works great...
>>
>> On 5 jun. 2012, at 23:59, Yount, William D wrote:
>>
>>> I am looking for a fault tolerant solution. By this, I mean I want there to 
>>> be an automatic switch over if one of the two storage servers goes down 
>>> with no human intervention. 
>>>
>>> Initially, I followed this guide: 
>>> https://help.ubuntu.com/community/HighlyAvailableNFS
>>> That works fine, but there are several steps that require human 
>>> intervention in case of a server failure:
>>>     Promote secondary server to primary
>>>     Mount drbd partition to export path
>>>     Restart nfs-kernel-server (if necessary)
>>>
>>> I was trying to get dual primaries setup, thinking that if one goes out the 
>>> other will take over automatically. There just seems to be so many moving 
>>> pieces that don't always work they way they are supposed to. I have been 
>>> reading all the material I can get my hands on but a lot of it seems 
>>> contradictory or only applicable on certain OS versions with certain 
>>> versions of OCFS2, DRBD and Pacemaker. 
>>>
>>> It doesn't matter to me if it is master/slave or dual primaries. I am just 
>>> trying to find something that actually works.
>>>
>>>
>>>
>>> -----Original Message-----
>>> From: Felix Frank [mailto:[email protected]]
>>> Sent: Tuesday, June 05, 2012 2:42 AM
>>> To: Yount, William D
>>> Cc: [email protected]
>>> Subject: Re: [DRBD-user] Fault Tolerant NFS
>>>
>>> On 06/05/2012 07:41 AM, Yount, William D wrote:
>>>> Does anyone have a good resource for setting up a fault tolerant NFS 
>>>> cluster using DRBD? I am currently using DRBD, Pacemaker, Corosync 
>>>> and
>>>> OCFS2 on Ubuntu 12.04.
>>> Those are all right, but I don't really see how OCFS2 is required.
>>> Dual-primary? Not needed for HA NFS.
>>>
>>> But it should still work.
>>>
>>>> High availability doesn't meet my needs. I have spent quite a while 
>>>> reading and trying out every combination of settings, but nothing 
>>>> seems to work properly.
>>> What are the exact limitations you're facing? Stale mounts after failover?
>>> _______________________________________________
>>> drbd-user mailing list
>>> [email protected]
>>> http://lists.linbit.com/mailman/listinfo/drbd-user
>>
>>
>> _______________________________________________
>> drbd-user mailing list
>> [email protected]
>> http://lists.linbit.com/mailman/listinfo/drbd-user
> Hello
> 
> check the 'No Quorum Policy' in the pacemaker CRM config default is stop, i
> changed it to suicied

I may be wrong here but setting the policy to "suicide" in a two node
cluster means if there is a split then *both* nodes will commit suicide,
no? So what you really want is to set the policy to "ignore"?

Regards,
  Dennis

_______________________________________________
drbd-user mailing list
[email protected]
http://lists.linbit.com/mailman/listinfo/drbd-user

Reply via email to