Hi Jason

any chance you started drbd at boot or the drbd device was active at the
time you started the cluster resource? If so, read the introduction of
the howto again and correct your setup.

Jason Fitzpatrick wrote:
> Hi Dominik
> 
> I have upgraded to HB 2.9xx and have been following the instructions that
> you provided (thanks for those) and have added a resource as follows
> 
> crm
> configure
> primitive Storage1 ocf:heartbeat:drbd \
> params drbd_resource=Storage1 \
> op monitor role=Master interval=59s timeout=30s \
> op monitor role=Slave interval=60s timeout=30s
> ms DRBD_Storage Storage1 \
> meta clone-max=2 notify=true globally-unique=false target-role=stopped
> commit
> exit
> 
> no errors are reported and the resource is visable from within the hb_gui
> 
> when I try to bring the resource online with
> 
> crm resource start DRBD_Storage
> 
> I see the resource attempt to come online and then fail, it seems to be
> starting the services, changing the status of the devices to attached (from
> detached) but not setting any device to master
> 
> the following is from the ha-log
> 
> crmd[8020]: 2009/02/10_17:22:32 info: do_lrm_rsc_op: Performing
> key=7:166:0:b57f7f7c-4e2d-4134-9c14-b1a2b7db11a7 op=Storage1:1_start_0 )
> lrmd[8016]: 2009/02/10_17:22:32 info: rsc:Storage1:1: start
> lrmd[8016]: 2009/02/10_17:22:32 info: RA output: (Storage1:1:start:stdout)
> /dev/drbd0: Failure: (124) Device is attached to a disk (use detach first)
> Command
>  'drbdsetup /dev/drbd0 disk /dev/sdb /dev/sdb internal --set-defaults
> --create-device --on-io-error=pass_on' terminated with exit code 10

This looks like "drbdadm up" is failing because the device is already
attached to the lower level storage device.

Regards
Dominik

> drbd[22270]:    2009/02/10_17:22:32 ERROR: Storage1 start: not in Secondary
> mode after start.
> crmd[8020]: 2009/02/10_17:22:32 info: process_lrm_event: LRM operation
> Storage1:1_start_0 (call=189, rc=1, cib-update=380, confirmed=true) complete
> unknown e
> rror
> .
> 
> I have checked the DRBD device Storage1 and it is in secondary mode after
> the start, and should I choose I can make it primary on either node
> 
> Thanks
> 
> Jason
> 
> 2009/2/10 Jason Fitzpatrick <jayfitzpatr...@gmail.com>
> 
>> Thanks,
>>
>> This was the latest version in the Fedora Repos, I will upgrade and see
>> what happens
>>
>> Jason
>>
>> 2009/2/10 Dominik Klein <d...@in-telegence.net>
>>
>> Jason Fitzpatrick wrote:
>>>>> Hi All
>>>>>
>>>>> I am having a hell of a time trying to get heartbeat to fail over my
>>> DRBD
>>>>> harddisk and am hoping for some help.
>>>>>
>>>>> I have a 2 node cluster, heartbeat is working as I am able to fail over
>>> IP
>>>>> Addresses and services successfully, but when I try to fail over my
>>> DRBD
>>>>> resource from secondary to primary I am hitting a brick wall, I can
>>> fail
>>>>> over the DRBD resource manually so I know that it does work at some
>>> level
>>>>> DRBD version 8.3 Heartbeat version heartbeat-2.1.3-1.fc9.i386
>>> Please upgrade. Thats too old for reliable master/slave behaviour.
>>> Preferrably upgrade to pacemaker and ais or heartbeat 2.99. Read
>>> http://www.clusterlabs.org/wiki/Install for install notes.
>>>
>>>>> and using
>>>>> heartbeat-gui to configure
>>> Don't use the gui to configure complex (ie clone or master/slave)
>>> resources.
>>>
>>> Once you upgraded to the latest pacemaker, please refer to
>>> http://www.clusterlabs.org/wiki/DRBD_HowTo_1.0 for drbd's cluster
>>> configuration.
>>>
>>> Regards
>>> Dominik
>>>
>>>>> DRBD Resource is called Storage1, the 2 nodes are connected via 2
>>> x-over
>>>>> cables (1 heartbeat 1 Replication)
>>>>>
>>>>> I have stripped down my config to the bare bones and tried every option
>>>>> that I can think off but know that I am missing something simple,
>>>>>
>>>>> I have attached my cib.xml but have removed domain names from the
>>> systems
>>>>> for privacy reasons
>>>>>
>>>>> Thanks in advance
>>>>>
>>>>> Jason
>>>>>
>>>>>  <cib admin_epoch="0" have_quorum="true" ignore_dtd="false"
>>>>> cib_feature_revision="2.0" num_peers="2" generated="true"
>>>>> ccm_transition="22" dc_uuid="9d8abc28-4fa3-408a-a695-fb36b0d67a48"
>>>>> epoch="733" num_updates="1" cib-last-written="Mon Feb  9 18:31:19
>>> 2009">
>>>>>    <configuration>
>>>>>      <crm_config>
>>>>>        <cluster_property_set id="cib-bootstrap-options">
>>>>>          <attributes>
>>>>>            <nvpair id="cib-bootstrap-options-dc-version"
>>> name="dc-version"
>>>>> value="2.1.3-node: 552305612591183b1628baa5bc6e903e0f1e26a3"/>
>>>>>            <nvpair name="last-lrm-refresh"
>>>>> id="cib-bootstrap-options-last-lrm-refresh" value="1234204278"/>
>>>>>          </attributes>
>>>>>        </cluster_property_set>
>>>>>      </crm_config>
>>>>>      <nodes>
>>>>>        <node id="df707752-d5fb-405a-8ca7-049e25a227b7"
>>> uname="lpissan1001"
>>>>> type="normal">
>>>>>          <instance_attributes
>>>>> id="nodes-df707752-d5fb-405a-8ca7-049e25a227b7">
>>>>>            <attributes>
>>>>>              <nvpair id="standby-df707752-d5fb-405a-8ca7-049e25a227b7"
>>>>> name="standby" value="off"/>
>>>>>            </attributes>
>>>>>          </instance_attributes>
>>>>>        </node>
>>>>>        <node id="9d8abc28-4fa3-408a-a695-fb36b0d67a48"
>>> uname="lpissan1002"
>>>>> type="normal">
>>>>>          <instance_attributes
>>>>> id="nodes-9d8abc28-4fa3-408a-a695-fb36b0d67a48">
>>>>>            <attributes>
>>>>>              <nvpair id="standby-9d8abc28-4fa3-408a-a695-fb36b0d67a48"
>>>>> name="standby" value="off"/>
>>>>>            </attributes>
>>>>>          </instance_attributes>
>>>>>        </node>
>>>>>      </nodes>
>>>>>      <resources>
>>>>>        <master_slave id="Storage1">
>>>>>          <meta_attributes id="Storage1_meta_attrs">
>>>>>            <attributes>
>>>>>              <nvpair id="Storage1_metaattr_target_role"
>>> name="target_role"
>>>>> value="started"/>
>>>>>              <nvpair id="Storage1_metaattr_clone_max" name="clone_max"
>>>>> value="2"/>
>>>>>              <nvpair id="Storage1_metaattr_clone_node_max"
>>>>> name="clone_node_max" value="1"/>
>>>>>              <nvpair id="Storage1_metaattr_master_max"
>>> name="master_max"
>>>>> value="1"/>
>>>>>              <nvpair id="Storage1_metaattr_master_node_max"
>>>>> name="master_node_max" value="1"/>
>>>>>              <nvpair id="Storage1_metaattr_notify" name="notify"
>>>>> value="true"/>
>>>>>              <nvpair id="Storage1_metaattr_globally_unique"
>>>>> name="globally_unique" value="false"/>
>>>>>            </attributes>
>>>>>          </meta_attributes>
>>>>>          <primitive id="Storage1" class="ocf" type="drbd"
>>>>> provider="heartbeat">
>>>>>            <instance_attributes id="Storage1_instance_attrs">
>>>>>              <attributes>
>>>>>                <nvpair id="273a1bb2-4867-42dd-a9e5-7cebbf48ef3b"
>>>>> name="drbd_resource" value="Storage1"/>
>>>>>              </attributes>
>>>>>            </instance_attributes>
>>>>>            <operations>
>>>>>              <op id="9ddc0ce9-4090-4546-a7d5-787fe47de872"
>>> name="monitor"
>>>>> description="master" interval="29" timeout="10" start_delay="1m"
>>>>> role="Master"/>
>>>>>              <op id="56a7508f-fa42-46f8-9924-3b284cdb97f0"
>>> name="monitor"
>>>>> description="slave" interval="29" timeout="10" start_delay="1m"
>>>>> role="Slave"/>
>>>>>            </operations>
>>>>>          </primitive>
>>>>>        </master_slave>
>>>>>      </resources>
>>>>>      <constraints/>
>>>>>    </configuration>
>>>>>  </cib>
>>>>>
>>>>>
>>>> _______________________________________________
>>>> Linux-HA mailing list
>>>> Linux-HA@lists.linux-ha.org
>>>> http://lists.linux-ha.org/mailman/listinfo/linux-ha
>>>> See also: http://linux-ha.org/ReportingProblems
>>>>
>>> _______________________________________________
>>> Linux-HA mailing list
>>> Linux-HA@lists.linux-ha.org
>>> http://lists.linux-ha.org/mailman/listinfo/linux-ha
>>> See also: http://linux-ha.org/ReportingProblems
>>>
>>
> _______________________________________________
> Linux-HA mailing list
> Linux-HA@lists.linux-ha.org
> http://lists.linux-ha.org/mailman/listinfo/linux-ha
> See also: http://linux-ha.org/ReportingProblems
> 

_______________________________________________
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to