On Mon, Mar 31, 2008 at 2:25 PM, Carlos Alexandre de Souza da Silva
<[EMAIL PROTECTED]> wrote:
> Greetings,
>
>
>
> Something is very wrong with my drbd. Whenever my master machine goes
> down and it comes back, I get something like this when the service is
> started and the node is promoted by heartbeat.
Try to configure the after-splitbrain policies according to your needs
.... default is 'disconnect', "drbdsetup repdata show" reveals your
complete config including defaults. Have a look at the example config
coming with drbd, it is very well documented.
Regards,
Andreas
>
>
>
> [EMAIL PROTECTED] ~]# cat /proc/drbd
>
> version: 8.0.11 (api:86/proto:86)
>
> GIT-hash: b3fe2bdfd3b9f7c2f923186883eb9e2a0d3a5b1b build by
> [EMAIL PROTECTED], 2008-03-09 10:26:55
>
> 0: cs:WFConnection st:Primary/Unknown ds:UpToDate/DUnknown C r---
>
> ns:0 nr:0 dw:0 dr:0 al:0 bm:0 lo:0 pe:0 ua:0 ap:0
>
> resync: used:0/31 hits:0 misses:0 starving:0 dirty:0 changed:0
>
> act_log: used:0/127 hits:0 misses:0 starving:0 dirty:0 changed:0
>
> [EMAIL PROTECTED] ~]#
>
>
>
>
>
> And when this happens, the nodes does not sync anymore, looks like one
> can't see the other one. The problem gets fixed if I recreated the meta
> disks but if I have to do this then I dont have a reliable fail over
> solution.
>
>
>
> Does anyone have any advice to give me about this ??
>
>
>
> My drbd.conf and cib.xml are posted below.
>
>
>
> global { usage-count no; }
>
> resource repdata {
>
> protocol C;
>
> startup { wfc-timeout 0; degr-wfc-timeout 120; }
>
> disk { on-io-error detach; } # or panic, ...
>
> net { cram-hmac-alg "sha1"; shared-secret "C4c4b4sh1"; } # don't
> forget to choose a secret for auth !
>
> syncer { rate 10M; }
>
> on pgmaster.blumar.com.br {
>
> device /dev/drbd0;
>
> disk /dev/sda6;
>
> address 30.30.30.30:7788;
>
> meta-disk /dev/sda7[0];
>
> }
>
> on pgslave.blumar.com.br {
>
> device /dev/drbd0;
>
> disk /dev/hda3;
>
> address 30.30.30.29:7788;
>
> meta-disk /dev/hda4[0];
>
> }
>
> }
>
>
>
>
>
> <cib generated="true" admin_epoch="0" have_quorum="true"
> ignore_dtd="false" num_peers="2" cib_feature_revision="2.0"
> ccm_transition="5" dc_uuid="56ec2257-b0e1-4395-8ca2-ff2f96151b55"
> epoch="100" num_updates="1" cib-last-written="Fri Mar 28 17:33:41 2008">
>
> <configuration>
>
> <crm_config>
>
> <cluster_property_set id="cib-bootstrap-options">
>
> <attributes>
>
> <nvpair id="cib-bootstrap-options-dc-version"
> name="dc-version" value="2.1.3-node:
> 552305612591183b1628baa5bc6e903e0f1e26a3"/>
>
> <nvpair name="last-lrm-refresh"
> id="cib-bootstrap-options-last-lrm-refresh" value="1206725922"/>
>
> </attributes>
>
> </cluster_property_set>
>
> </crm_config>
>
> <nodes>
>
> <node id="56ec2257-b0e1-4395-8ca2-ff2f96151b55"
> uname="pgslave.blumar.com.br" type="normal"/>
>
> <node id="fb01ec99-1ece-40ed-a737-f3816e69368c"
> uname="pgmaster.blumar.com.br" type="normal"/>
>
> </nodes>
>
> <resources>
>
> <master_slave id="array_drbd">
>
> <meta_attributes id="array_drbd_meta_attrs">
>
> <attributes>
>
> <nvpair name="target_role"
> id="array_drbd_metaattr_target_role" value="stopped"/>
>
> <nvpair id="array_drbd_metaattr_clone_max" name="clone_max"
> value="2"/>
>
> <nvpair id="array_drbd_metaattr_clone_node_max"
> name="clone_node_max" value="1"/>
>
> <nvpair id="array_drbd_metaattr_master_max"
> name="master_max" value="1"/>
>
> <nvpair id="array_drbd_metaattr_master_node_max"
> name="master_node_max" value="1"/>
>
> <nvpair id="array_drbd_metaattr_notify" name="notify"
> value="true"/>
>
> <nvpair id="array_drbd_metaattr_globally_unique"
> name="globally_unique" value="false"/>
>
> </attributes>
>
> </meta_attributes>
>
> <primitive id="resource_drbd" class="ocf" type="drbd"
> provider="heartbeat">
>
> <instance_attributes id="resource_drbd_instance_attrs">
>
> <attributes>
>
> <nvpair id="19b32757-d764-410b-8892-376598313f94"
> name="drbd_resource" value="repdata"/>
>
> </attributes>
>
> </instance_attributes>
>
> </primitive>
>
> </master_slave>
>
> <primitive class="ocf" type="Filesystem" provider="heartbeat"
> id="resource_fs">
>
> <meta_attributes id="resource_fs_meta_attrs">
>
> <attributes>
>
> <nvpair name="target_role"
> id="resource_fs_metaattr_target_role" value="stopped"/>
>
> </attributes>
>
> </meta_attributes>
>
> <instance_attributes id="resource_fs_instance_attrs">
>
> <attributes>
>
> <nvpair id="edf7c582-3d98-4605-bbca-a176f0b8ec0d"
> name="device" value="/dev/drbd0"/>
>
> <nvpair id="a71f6fa0-2be3-485a-8467-cfb92e54bf08"
> name="directory" value="/repdata"/>
>
> <nvpair id="73af8c10-3244-44d9-b883-43c4c5e87685"
> name="fstype" value="ext3"/>
>
> </attributes>
>
> </instance_attributes>
>
> </primitive>
>
> <primitive class="ocf" type="IPaddr2" provider="heartbeat"
> id="ip_postgres">
>
> <meta_attributes id="ip_postgres_meta_attrs">
>
> <attributes>
>
> <nvpair name="target_role"
> id="ip_postgres_metaattr_target_role" value="stopped"/>
>
> </attributes>
>
> </meta_attributes>
>
> <instance_attributes id="ip_postgres_instance_attrs">
>
> <attributes>
>
> <nvpair id="e1e0d015-aafa-49fb-aab1-877ecb54b4b7" name="ip"
> value="10.3.3.1"/>
>
> </attributes>
>
> </instance_attributes>
>
> </primitive>
>
> <primitive class="ocf" type="pgsql" provider="heartbeat"
> id="resource_postgres">
>
> <meta_attributes id="resource_postgres_meta_attrs">
>
> <attributes>
>
> <nvpair name="target_role"
> id="resource_postgres_metaattr_target_role" value="stopped"/>
>
> </attributes>
>
> </meta_attributes>
>
> <instance_attributes id="resource_postgres_instance_attrs">
>
> <attributes>
>
> <nvpair id="9b48a9af-cea7-4758-b392-072e8eb68634"
> name="pgdata" value="/repdata/pg"/>
>
> </attributes>
>
> </instance_attributes>
>
> </primitive>
>
> </resources>
>
> <constraints>
>
> <rsc_order id="drbd_antes_fs" from="resource_fs" action="start"
> to="array_drbd" to_action="promote"/>
>
> <rsc_colocation id="fs_com_drbd" to="array_drbd" to_role="master"
> from="resource_fs" score="INFINITY"/>
>
> <rsc_colocation id="ip_com_fs" from="ip_postgres"
> to="resource_fs" score="INFINITY"/>
>
> <rsc_order id="ip_antes_pg" from="ip_postgres" type="before"
> to="resource_postgres"/>
>
> <rsc_location id="drbd_pgmaster" rsc="array_drbd">
>
> <rule id="prefered_drbd_pgmaster" score="100">
>
> <expression attribute="#uname"
> id="68a231b9-57e0-4bcb-840c-bce74e429a4c" operation="eq"
> value="pgmaster.blumar.com.br"/>
>
> </rule>
>
> </rsc_location>
>
> <rsc_colocation id="postgres_com_ip" from="resource_postgres"
> to="ip_postgres" score="INFINITY"/>
>
> </constraints>
>
> </configuration>
>
> </cib>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
> Carlos Alexandre
> Analista de Suporte
>
> Direct Phone: + 55 21 2142-9383
> www.blumar.com.br <http://www.blumar.com.br>
>
>
>
>
>
>
>
>
>
>
>
>
>
> Av. Borges de Medeiros, 633 - Sala 405 a 408 - Leblon - Rio de janeiro -
> Brasil - 22430-041
> www.blumar.com.br <http://www.blumar.com.br> phone: +55 21 2142 9300
> fax: +55 21 2511 3739
>
>
>
>
>
>
> _______________________________________________
> Linux-HA mailing list
> [email protected]
> http://lists.linux-ha.org/mailman/listinfo/linux-ha
> See also: http://linux-ha.org/ReportingProblems
>
_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems