On Fri, May 2, 2008 at 2:56 PM, Travis Sidelinger <[EMAIL PROTECTED]> wrote:
> Hello everyone,
>
>  This is my first time setting up heartbeat, and I'm running into a few
> problems.
>
>  My objective is to have a 2.x cluster of 4+ servers with the following
> resources: different apache instances, ip addresses, application servers,
> databases, NFS, and ISCSI.
>
>  Currently, I have a version 1.x cluster of 3 servers up and running with
> apache and IP address resources.

highly dangerous!

>
>  Documentation on setting up a 1.x verses a 2.x configuration has been a
> little confusing.
>
>  I've setup a 4th server to help develop a 2.x configuration.  The follow
> messages keep showing up in the logs.
>
>  ---------------------------------
>  May  2 08:56:05 ocdcweb037 heartbeat: [27911]: WARN: string2msg_ll: node
> [ocdcweb034] failed authentication
>  May  2 08:56:07 ocdcweb037 heartbeat: [27911]: WARN: string2msg_ll: node
> [ocdcweb034] failed authentication
>  May  2 08:56:09 ocdcweb037 heartbeat: [27911]: WARN: string2msg_ll: node
> [ocdcweb034] failed authentication
>  May  2 08:56:09 ocdcweb037 heartbeat: [27911]: WARN: string2msg_ll: node
> [ocdcweb034] failed authentication
>  May  2 08:56:11 ocdcweb037 heartbeat: [27911]: WARN: string2msg_ll: node
> [ocdcweb034] failed authentication
>  May  2 08:56:13 ocdcweb037 heartbeat: [27911]: WARN: string2msg_ll: node
> [ocdcweb034] failed authentication
>  May  2 08:56:15 ocdcweb037 heartbeat: [27911]: WARN: string2msg_ll: node
> [ocdcweb034] failed authentication
>
>  ---------------------------------
>
>  How can I fix this?

i'm guessing you either forgot to copy the authkeys file to the new
node or not all the nodes are listed in ha.cf

>  My ha.cf file looks like this:
>
>  -------------------------------------
>  autojoin none           # autojoin other nodes
>  crm on                  # Use version 1.x or version 2.x style
>  logfacility local0      # syslog facility
>  use_logd no             # another logging service
>  debug 0                 # levels 0-255
>  #bcast eth0             # node(s) to send heartbeats on
>  ucast eth0 ocdcweb037   # nodes to send heartbeats to
>  keepalive 2             # time between heartbeats
>  warntime 10             # time before a late warning shows in the logs
>  deadtime 30             # time before the node is pronounced dead
>  initdead 120            # deadtime after a reboot, gives time for the
> network to come up
>  udpport 694             # udp port for heartbeat broadcast
>  node ocdcweb037         # cluster node
>  auto_failback on        # enables favorite member node
>
>  #apiauth  mgmtd   uid=hacluster
>  respawn   root    /usr/lib64/heartbeat/mgmtd -v
>  #respawn  root    /sbin/evmsd
>  #apiauth  evms    uid=hacluster,root
>
>  # 2.x settings
>  #apiauth stonithd        uid=root
>  #apiauth crmd            uid=hacluster
>  #apiauth cib             uid=hacluster
>  #respawn hacluster       ccm
>  #respawn hacluster       cib
>  #respawn root            stonithd
>  #respawn root            lrmd
>  #respawn hacluster       crmd
>  -------------------------------------
>
>  No resources have been configured yet for the 2.x configuration.
>
>  Thanks
>
>  -Travis Sidelinger
>  _______________________________________________
>  Linux-HA mailing list
>  [email protected]
>  http://lists.linux-ha.org/mailman/listinfo/linux-ha
>  See also: http://linux-ha.org/ReportingProblems
>
_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to