On Tue Jan 31 03:47:11 MST 2012 Lars Ellenberg wrote:

> On Mon, Jan 30, 2012 at 05:42:34PM -0500, William Seligman wrote:
>> I'm trying to follow the directions for setting up a dual-primary DRBD setup
>> with CMAN and Pacemaker. I'm stuck at an annoying spot: Pacemaker won't 
>> promote
>> the DRBD resources to primary at either node.
>> 
>> 
>> Here's the result of crm_mon:
>> 
>> Last updated: Mon Jan 30 17:07:03 2012
>> Stack: cman
>> Current DC: hypatia-tb - partition with quorum
>> Version: 1.1.5-5.el6-01e86afaaa6d4a8c4836f68df80ababd6ca3902f
>> 2 Nodes configured, unknown expected votes
>> 2 Resources configured.
>> ============
>> 
>> Online: [ orestes-tb hypatia-tb ]
>> 
>>  Master/Slave Set: AdminClone [AdminDrbd]
>>      Slaves: [ hypatia-tb orestes-tb ]
> 
>> crm configure show:
>> 
>> node hypatia-tb
>> node orestes-tb
>> primitive AdminDrbd ocf:linbit:drbd \
>>      params drbd_resource="admin" \
>>      op monitor interval="60s" role="Master" \
> 
> You are missing an additional monitor op for role=Slave
> make sure it has a different interval than the one for role=Master.
> 
> e.g.
>       op monitor interval="59s" role="Slave" \
> 
>>      op stop interval="0" timeout="320" \
>>      op start interval="0" timeout="240"

I put that in, but it didn't change my basic problem: Neither instance of
AdminDrbd is promoted on either node.

>> primitive Clvmd lsb:clvmd
>> ms AdminClone AdminDrbd \
>>      meta master-max="2" master-node-max="1" clone-max="2" clone-node-max="1"
>> notify="true"
>> clone ClvmdClone Clvmd
>> colocation ClvmdWithAdmin inf: ClvmdClone AdminClone:Master
>> order AdminBeforeClvmd inf: AdminClone:promote ClvmdClone:start
>> property $id="cib-bootstrap-options" \
>>      dc-version="1.1.5-5.el6-01e86afaaa6d4a8c4836f68df80ababd6ca3902f" \
>>      cluster-infrastructure="cman" \
>>      stonith-enabled="false"
> 
> Also remember that, for dual-primary DRBD,
> working and tested fencing both on cluster level (stonith) and on drbd
> level (fence-peer) is mandatory.
> Unless you don't care for data integrity.

I'll get to that. I'm just starting out on this configuration. I don't want to
put in STONITH just yet, otherwise I'll have to do recovery after every typo.
I'll put in STONITH and test it when I get to installing the KVM resources. But
until I solve this problem, I can't get to that stage.

>> DRBD looks OK:
>> 
>> # cat /proc/drbd
>> version: 8.4.0 (api:1/proto:86-100)
>> GIT-hash: 28753f559ab51b549d16bcf487fe625d5919c49c build by gardner@, 
>> 2012-01-25
>> 19:10:28
>>  0: cs:Connected ro:Secondary/Secondary ds:UpToDate/UpToDate C r-----
>>     ns:0 nr:0 dw:0 dr:0 al:0 bm:0 lo:0 pe:0 ua:0 ap:0 ep:1 wo:b oos:0

Any clues as to what I can look at to track the source of the problem?

-- 
Bill Seligman             | Phone: (914) 591-2823
Nevis Labs, Columbia Univ | mailto://[email protected]
PO Box 137                |
Irvington NY 10533 USA    | http://www.nevis.columbia.edu/~seligman/

Attachment: smime.p7s
Description: S/MIME Cryptographic Signature

_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to