Thanks a lot, now it is working as I think it shall do.
-----Mensagem original----- De: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] Em nome de Andreas Kurz Enviada em: segunda-feira, 31 de março de 2008 10:42 Para: General Linux-HA mailing list Assunto: Re: [Linux-HA] HEARTBEAT + DRBD - Not working as intended On Mon, Mar 31, 2008 at 2:25 PM, Carlos Alexandre de Souza da Silva <[EMAIL PROTECTED]> wrote: > Greetings, > > > > Something is very wrong with my drbd. Whenever my master machine goes > down and it comes back, I get something like this when the service is > started and the node is promoted by heartbeat. Try to configure the after-splitbrain policies according to your needs .... default is 'disconnect', "drbdsetup repdata show" reveals your complete config including defaults. Have a look at the example config coming with drbd, it is very well documented. Regards, Andreas > > > > [EMAIL PROTECTED] ~]# cat /proc/drbd > > version: 8.0.11 (api:86/proto:86) > > GIT-hash: b3fe2bdfd3b9f7c2f923186883eb9e2a0d3a5b1b build by > [EMAIL PROTECTED], 2008-03-09 10:26:55 > > 0: cs:WFConnection st:Primary/Unknown ds:UpToDate/DUnknown C r--- > > ns:0 nr:0 dw:0 dr:0 al:0 bm:0 lo:0 pe:0 ua:0 ap:0 > > resync: used:0/31 hits:0 misses:0 starving:0 dirty:0 changed:0 > > act_log: used:0/127 hits:0 misses:0 starving:0 dirty:0 changed:0 > > [EMAIL PROTECTED] ~]# > > > > > > And when this happens, the nodes does not sync anymore, looks like one > can't see the other one. The problem gets fixed if I recreated the meta > disks but if I have to do this then I dont have a reliable fail over > solution. > > > > Does anyone have any advice to give me about this ?? > > > > My drbd.conf and cib.xml are posted below. > > > > global { usage-count no; } > > resource repdata { > > protocol C; > > startup { wfc-timeout 0; degr-wfc-timeout 120; } > > disk { on-io-error detach; } # or panic, ... > > net { cram-hmac-alg "sha1"; shared-secret "C4c4b4sh1"; } # don't > forget to choose a secret for auth ! > > syncer { rate 10M; } > > on pgmaster.blumar.com.br { > > device /dev/drbd0; > > disk /dev/sda6; > > address 30.30.30.30:7788; > > meta-disk /dev/sda7[0]; > > } > > on pgslave.blumar.com.br { > > device /dev/drbd0; > > disk /dev/hda3; > > address 30.30.30.29:7788; > > meta-disk /dev/hda4[0]; > > } > > } > > > > > > <cib generated="true" admin_epoch="0" have_quorum="true" > ignore_dtd="false" num_peers="2" cib_feature_revision="2.0" > ccm_transition="5" dc_uuid="56ec2257-b0e1-4395-8ca2-ff2f96151b55" > epoch="100" num_updates="1" cib-last-written="Fri Mar 28 17:33:41 2008"> > > <configuration> > > <crm_config> > > <cluster_property_set id="cib-bootstrap-options"> > > <attributes> > > <nvpair id="cib-bootstrap-options-dc-version" > name="dc-version" value="2.1.3-node: > 552305612591183b1628baa5bc6e903e0f1e26a3"/> > > <nvpair name="last-lrm-refresh" > id="cib-bootstrap-options-last-lrm-refresh" value="1206725922"/> > > </attributes> > > </cluster_property_set> > > </crm_config> > > <nodes> > > <node id="56ec2257-b0e1-4395-8ca2-ff2f96151b55" > uname="pgslave.blumar.com.br" type="normal"/> > > <node id="fb01ec99-1ece-40ed-a737-f3816e69368c" > uname="pgmaster.blumar.com.br" type="normal"/> > > </nodes> > > <resources> > > <master_slave id="array_drbd"> > > <meta_attributes id="array_drbd_meta_attrs"> > > <attributes> > > <nvpair name="target_role" > id="array_drbd_metaattr_target_role" value="stopped"/> > > <nvpair id="array_drbd_metaattr_clone_max" name="clone_max" > value="2"/> > > <nvpair id="array_drbd_metaattr_clone_node_max" > name="clone_node_max" value="1"/> > > <nvpair id="array_drbd_metaattr_master_max" > name="master_max" value="1"/> > > <nvpair id="array_drbd_metaattr_master_node_max" > name="master_node_max" value="1"/> > > <nvpair id="array_drbd_metaattr_notify" name="notify" > value="true"/> > > <nvpair id="array_drbd_metaattr_globally_unique" > name="globally_unique" value="false"/> > > </attributes> > > </meta_attributes> > > <primitive id="resource_drbd" class="ocf" type="drbd" > provider="heartbeat"> > > <instance_attributes id="resource_drbd_instance_attrs"> > > <attributes> > > <nvpair id="19b32757-d764-410b-8892-376598313f94" > name="drbd_resource" value="repdata"/> > > </attributes> > > </instance_attributes> > > </primitive> > > </master_slave> > > <primitive class="ocf" type="Filesystem" provider="heartbeat" > id="resource_fs"> > > <meta_attributes id="resource_fs_meta_attrs"> > > <attributes> > > <nvpair name="target_role" > id="resource_fs_metaattr_target_role" value="stopped"/> > > </attributes> > > </meta_attributes> > > <instance_attributes id="resource_fs_instance_attrs"> > > <attributes> > > <nvpair id="edf7c582-3d98-4605-bbca-a176f0b8ec0d" > name="device" value="/dev/drbd0"/> > > <nvpair id="a71f6fa0-2be3-485a-8467-cfb92e54bf08" > name="directory" value="/repdata"/> > > <nvpair id="73af8c10-3244-44d9-b883-43c4c5e87685" > name="fstype" value="ext3"/> > > </attributes> > > </instance_attributes> > > </primitive> > > <primitive class="ocf" type="IPaddr2" provider="heartbeat" > id="ip_postgres"> > > <meta_attributes id="ip_postgres_meta_attrs"> > > <attributes> > > <nvpair name="target_role" > id="ip_postgres_metaattr_target_role" value="stopped"/> > > </attributes> > > </meta_attributes> > > <instance_attributes id="ip_postgres_instance_attrs"> > > <attributes> > > <nvpair id="e1e0d015-aafa-49fb-aab1-877ecb54b4b7" name="ip" > value="10.3.3.1"/> > > </attributes> > > </instance_attributes> > > </primitive> > > <primitive class="ocf" type="pgsql" provider="heartbeat" > id="resource_postgres"> > > <meta_attributes id="resource_postgres_meta_attrs"> > > <attributes> > > <nvpair name="target_role" > id="resource_postgres_metaattr_target_role" value="stopped"/> > > </attributes> > > </meta_attributes> > > <instance_attributes id="resource_postgres_instance_attrs"> > > <attributes> > > <nvpair id="9b48a9af-cea7-4758-b392-072e8eb68634" > name="pgdata" value="/repdata/pg"/> > > </attributes> > > </instance_attributes> > > </primitive> > > </resources> > > <constraints> > > <rsc_order id="drbd_antes_fs" from="resource_fs" action="start" > to="array_drbd" to_action="promote"/> > > <rsc_colocation id="fs_com_drbd" to="array_drbd" to_role="master" > from="resource_fs" score="INFINITY"/> > > <rsc_colocation id="ip_com_fs" from="ip_postgres" > to="resource_fs" score="INFINITY"/> > > <rsc_order id="ip_antes_pg" from="ip_postgres" type="before" > to="resource_postgres"/> > > <rsc_location id="drbd_pgmaster" rsc="array_drbd"> > > <rule id="prefered_drbd_pgmaster" score="100"> > > <expression attribute="#uname" > id="68a231b9-57e0-4bcb-840c-bce74e429a4c" operation="eq" > value="pgmaster.blumar.com.br"/> > > </rule> > > </rsc_location> > > <rsc_colocation id="postgres_com_ip" from="resource_postgres" > to="ip_postgres" score="INFINITY"/> > > </constraints> > > </configuration> > > </cib> > > > > > > > > > > > > > > > > > > > > Carlos Alexandre > Analista de Suporte > > Direct Phone: + 55 21 2142-9383 > www.blumar.com.br <http://www.blumar.com.br> > > > > > > > > > > > > > > Av. Borges de Medeiros, 633 - Sala 405 a 408 - Leblon - Rio de janeiro - > Brasil - 22430-041 > www.blumar.com.br <http://www.blumar.com.br> phone: +55 21 2142 9300 > fax: +55 21 2511 3739 > > > > > > > _______________________________________________ > Linux-HA mailing list > [email protected] > http://lists.linux-ha.org/mailman/listinfo/linux-ha > See also: http://linux-ha.org/ReportingProblems > _______________________________________________ Linux-HA mailing list [email protected] http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems _______________________________________________ Linux-HA mailing list [email protected] http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems
