Thanks a lot, now it is working as I think it shall do.


-----Mensagem original-----
De: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] Em nome de Andreas Kurz
Enviada em: segunda-feira, 31 de março de 2008 10:42
Para: General Linux-HA mailing list
Assunto: Re: [Linux-HA] HEARTBEAT + DRBD - Not working as intended

On Mon, Mar 31, 2008 at 2:25 PM, Carlos Alexandre de Souza da Silva
<[EMAIL PROTECTED]> wrote:
> Greetings,
>
>
>
>  Something is very wrong with my drbd. Whenever my master machine goes
>  down and it comes back, I get something like this when the service is
>  started and the node is promoted by heartbeat.

Try to configure the after-splitbrain policies according to your needs
.... default is 'disconnect', "drbdsetup repdata show" reveals your
complete config including defaults. Have a look at the example config
coming with drbd, it is very well documented.

Regards,
Andreas


>
>
>
>  [EMAIL PROTECTED] ~]# cat /proc/drbd
>
>  version: 8.0.11 (api:86/proto:86)
>
>  GIT-hash: b3fe2bdfd3b9f7c2f923186883eb9e2a0d3a5b1b build by
>  [EMAIL PROTECTED], 2008-03-09 10:26:55
>
>   0: cs:WFConnection st:Primary/Unknown ds:UpToDate/DUnknown C r---
>
>     ns:0 nr:0 dw:0 dr:0 al:0 bm:0 lo:0 pe:0 ua:0 ap:0
>
>         resync: used:0/31 hits:0 misses:0 starving:0 dirty:0 changed:0
>
>         act_log: used:0/127 hits:0 misses:0 starving:0 dirty:0 changed:0
>
>  [EMAIL PROTECTED] ~]#
>
>
>
>
>
>  And when this happens, the nodes does not sync anymore, looks like one
>  can't see the other one. The problem gets fixed if I recreated the meta
>  disks but if I have to do this then I dont have a reliable fail over
>  solution.
>
>
>
>  Does anyone have any advice to give me about this ??
>
>
>
>  My drbd.conf and cib.xml are posted below.
>
>
>
>  global { usage-count no; }
>
>  resource repdata {
>
>   protocol C;
>
>   startup { wfc-timeout 0; degr-wfc-timeout     120; }
>
>   disk { on-io-error detach; } # or panic, ...
>
>   net {  cram-hmac-alg "sha1"; shared-secret "C4c4b4sh1"; } # don't
>  forget to choose a secret for auth !
>
>   syncer { rate 10M; }
>
>   on pgmaster.blumar.com.br {
>
>     device /dev/drbd0;
>
>     disk /dev/sda6;
>
>     address 30.30.30.30:7788;
>
>     meta-disk /dev/sda7[0];
>
>   }
>
>   on pgslave.blumar.com.br {
>
>     device /dev/drbd0;
>
>     disk /dev/hda3;
>
>     address 30.30.30.29:7788;
>
>     meta-disk /dev/hda4[0];
>
>   }
>
>  }
>
>
>
>
>
>  <cib generated="true" admin_epoch="0" have_quorum="true"
>  ignore_dtd="false" num_peers="2" cib_feature_revision="2.0"
>  ccm_transition="5" dc_uuid="56ec2257-b0e1-4395-8ca2-ff2f96151b55"
>  epoch="100" num_updates="1" cib-last-written="Fri Mar 28 17:33:41 2008">
>
>    <configuration>
>
>      <crm_config>
>
>        <cluster_property_set id="cib-bootstrap-options">
>
>          <attributes>
>
>            <nvpair id="cib-bootstrap-options-dc-version"
>  name="dc-version" value="2.1.3-node:
>  552305612591183b1628baa5bc6e903e0f1e26a3"/>
>
>            <nvpair name="last-lrm-refresh"
>  id="cib-bootstrap-options-last-lrm-refresh" value="1206725922"/>
>
>          </attributes>
>
>        </cluster_property_set>
>
>      </crm_config>
>
>      <nodes>
>
>        <node id="56ec2257-b0e1-4395-8ca2-ff2f96151b55"
>  uname="pgslave.blumar.com.br" type="normal"/>
>
>        <node id="fb01ec99-1ece-40ed-a737-f3816e69368c"
>  uname="pgmaster.blumar.com.br" type="normal"/>
>
>      </nodes>
>
>      <resources>
>
>        <master_slave id="array_drbd">
>
>          <meta_attributes id="array_drbd_meta_attrs">
>
>            <attributes>
>
>              <nvpair name="target_role"
>  id="array_drbd_metaattr_target_role" value="stopped"/>
>
>              <nvpair id="array_drbd_metaattr_clone_max" name="clone_max"
>  value="2"/>
>
>              <nvpair id="array_drbd_metaattr_clone_node_max"
>  name="clone_node_max" value="1"/>
>
>              <nvpair id="array_drbd_metaattr_master_max"
>  name="master_max" value="1"/>
>
>              <nvpair id="array_drbd_metaattr_master_node_max"
>  name="master_node_max" value="1"/>
>
>              <nvpair id="array_drbd_metaattr_notify" name="notify"
>  value="true"/>
>
>              <nvpair id="array_drbd_metaattr_globally_unique"
>  name="globally_unique" value="false"/>
>
>            </attributes>
>
>          </meta_attributes>
>
>          <primitive id="resource_drbd" class="ocf" type="drbd"
>  provider="heartbeat">
>
>            <instance_attributes id="resource_drbd_instance_attrs">
>
>              <attributes>
>
>                <nvpair id="19b32757-d764-410b-8892-376598313f94"
>  name="drbd_resource" value="repdata"/>
>
>              </attributes>
>
>            </instance_attributes>
>
>          </primitive>
>
>        </master_slave>
>
>        <primitive class="ocf" type="Filesystem" provider="heartbeat"
>  id="resource_fs">
>
>          <meta_attributes id="resource_fs_meta_attrs">
>
>            <attributes>
>
>              <nvpair name="target_role"
>  id="resource_fs_metaattr_target_role" value="stopped"/>
>
>            </attributes>
>
>          </meta_attributes>
>
>          <instance_attributes id="resource_fs_instance_attrs">
>
>            <attributes>
>
>              <nvpair id="edf7c582-3d98-4605-bbca-a176f0b8ec0d"
>  name="device" value="/dev/drbd0"/>
>
>              <nvpair id="a71f6fa0-2be3-485a-8467-cfb92e54bf08"
>  name="directory" value="/repdata"/>
>
>              <nvpair id="73af8c10-3244-44d9-b883-43c4c5e87685"
>  name="fstype" value="ext3"/>
>
>            </attributes>
>
>          </instance_attributes>
>
>        </primitive>
>
>        <primitive class="ocf" type="IPaddr2" provider="heartbeat"
>  id="ip_postgres">
>
>          <meta_attributes id="ip_postgres_meta_attrs">
>
>            <attributes>
>
>              <nvpair name="target_role"
>  id="ip_postgres_metaattr_target_role" value="stopped"/>
>
>            </attributes>
>
>          </meta_attributes>
>
>          <instance_attributes id="ip_postgres_instance_attrs">
>
>            <attributes>
>
>              <nvpair id="e1e0d015-aafa-49fb-aab1-877ecb54b4b7" name="ip"
>  value="10.3.3.1"/>
>
>            </attributes>
>
>          </instance_attributes>
>
>        </primitive>
>
>        <primitive class="ocf" type="pgsql" provider="heartbeat"
>  id="resource_postgres">
>
>          <meta_attributes id="resource_postgres_meta_attrs">
>
>            <attributes>
>
>              <nvpair name="target_role"
>  id="resource_postgres_metaattr_target_role" value="stopped"/>
>
>            </attributes>
>
>          </meta_attributes>
>
>          <instance_attributes id="resource_postgres_instance_attrs">
>
>            <attributes>
>
>              <nvpair id="9b48a9af-cea7-4758-b392-072e8eb68634"
>  name="pgdata" value="/repdata/pg"/>
>
>            </attributes>
>
>          </instance_attributes>
>
>        </primitive>
>
>      </resources>
>
>      <constraints>
>
>        <rsc_order id="drbd_antes_fs" from="resource_fs" action="start"
>  to="array_drbd" to_action="promote"/>
>
>        <rsc_colocation id="fs_com_drbd" to="array_drbd" to_role="master"
>  from="resource_fs" score="INFINITY"/>
>
>        <rsc_colocation id="ip_com_fs" from="ip_postgres"
>  to="resource_fs" score="INFINITY"/>
>
>        <rsc_order id="ip_antes_pg" from="ip_postgres" type="before"
>  to="resource_postgres"/>
>
>        <rsc_location id="drbd_pgmaster" rsc="array_drbd">
>
>          <rule id="prefered_drbd_pgmaster" score="100">
>
>            <expression attribute="#uname"
>  id="68a231b9-57e0-4bcb-840c-bce74e429a4c" operation="eq"
>  value="pgmaster.blumar.com.br"/>
>
>          </rule>
>
>        </rsc_location>
>
>        <rsc_colocation id="postgres_com_ip" from="resource_postgres"
>  to="ip_postgres" score="INFINITY"/>
>
>      </constraints>
>
>    </configuration>
>
>   </cib>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>  Carlos Alexandre
>  Analista de Suporte
>
>  Direct Phone: + 55 21 2142-9383
>  www.blumar.com.br <http://www.blumar.com.br>
>
>
>
>
>
>
>
>
>
>
>
>
>
>  Av. Borges de Medeiros, 633 - Sala 405 a 408 - Leblon - Rio de janeiro -
>  Brasil - 22430-041
>  www.blumar.com.br <http://www.blumar.com.br>  phone: +55 21 2142 9300
>  fax: +55 21 2511 3739
>
>
>
>
>
>
> _______________________________________________
>  Linux-HA mailing list
>  [email protected]
>  http://lists.linux-ha.org/mailman/listinfo/linux-ha
>  See also: http://linux-ha.org/ReportingProblems
>
_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems
_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to