resource configuration?

On Tue, Nov 3, 2009 at 9:52 PM, Wilfredo Ocasio
<[email protected]> wrote:
> Simple 2 node active / passive configuration; Oracle DB resource only starts 
> on one node. 2nd node keeps failing to start, but Oracle DB will start 
> manually.
> Any help is appreciated .
>
> Ha.cf :
> use_logd on
> auto_failback off
> autojoin other
> traditional_compression false
> # debug 3
> node lnxhat1
> node lnxhat2
> deadtime 10
> keepalive 2
> udpport 694
> ucast hsi0 172.27.100.204
> ucast hsi0 172.27.100.205
> crm on
> ping 172.22.4.1 172.27.220.10 172.27.220.12
> respawn root /usr/lib64/heartbeat/pingd -m 2000 -d 5s -a my_ping_set
>
>
> crm_verify[9452]: 2009/11/03_14:47:26 ERROR: unpack_rsc_op: Remapping 
> resource_oracle_start_0 (rc=1) on lnxhat1 to an ERROR
>
> # ptest -LVVVVVV
> ptest[9519]: 2009/11/03_14:48:00 info: main: =#=#=#=#= Getting XML =#=#=#=#=
> ptest[9519]: 2009/11/03_14:48:00 info: main: Reading XML from: live cluster
> ptest[9519]: 2009/11/03_14:48:00 notice: main: Required feature set: 2.0
> ptest[9519]: 2009/11/03_14:48:00 debug: cluster_option: Using default value 
> 'stop' for cluster option 'no-quorum-policy'
> ptest[9519]: 2009/11/03_14:48:00 debug: cluster_option: Using default value 
> 'true' for cluster option 'symmetric-cluster'
> ptest[9519]: 2009/11/03_14:48:00 debug: cluster_option: Using default value 
> 'reboot' for cluster option 'stonith-action'
> ptest[9519]: 2009/11/03_14:48:00 debug: cluster_option: Using default value 
> '0' for cluster option 'default-resource-stickiness'
> ptest[9519]: 2009/11/03_14:48:00 debug: cluster_option: Using default value 
> '0' for cluster option 'default-resource-failure-stickiness'
> ptest[9519]: 2009/11/03_14:48:00 debug: cluster_option: Using default value 
> 'true' for cluster option 'is-managed-default'
> ptest[9519]: 2009/11/03_14:48:00 debug: cluster_option: Using default value 
> '60s' for cluster option 'cluster-delay'
> ptest[9519]: 2009/11/03_14:48:00 debug: cluster_option: Using default value 
> '30' for cluster option 'batch-limit'
> ptest[9519]: 2009/11/03_14:48:00 debug: cluster_option: Using default value 
> '20s' for cluster option 'default-action-timeout'
> ptest[9519]: 2009/11/03_14:48:00 debug: cluster_option: Using default value 
> 'true' for cluster option 'stop-orphan-resources'
> ptest[9519]: 2009/11/03_14:48:00 debug: cluster_option: Using default value 
> 'true' for cluster option 'stop-orphan-actions'
> ptest[9519]: 2009/11/03_14:48:00 debug: cluster_option: Using default value 
> 'false' for cluster option 'remove-after-stop'
> ptest[9519]: 2009/11/03_14:48:00 debug: cluster_option: Using default value 
> '-1' for cluster option 'pe-error-series-max'
> ptest[9519]: 2009/11/03_14:48:00 debug: cluster_option: Using default value 
> '-1' for cluster option 'pe-warn-series-max'
> ptest[9519]: 2009/11/03_14:48:00 debug: cluster_option: Using default value 
> '-1' for cluster option 'pe-input-series-max'
> ptest[9519]: 2009/11/03_14:48:00 debug: cluster_option: Using default value 
> 'true' for cluster option 'startup-fencing'
> ptest[9519]: 2009/11/03_14:48:00 debug: cluster_option: Using default value 
> 'true' for cluster option 'start-failure-is-fatal'
> ptest[9519]: 2009/11/03_14:48:00 debug: unpack_config: Default action 
> timeout: 20s
> ptest[9519]: 2009/11/03_14:48:00 debug: unpack_config: Default stickiness: 0
> ptest[9519]: 2009/11/03_14:48:00 debug: unpack_config: Default failure 
> stickiness: 0
> ptest[9519]: 2009/11/03_14:48:00 debug: unpack_config: STONITH of failed 
> nodes is enabled
> ptest[9519]: 2009/11/03_14:48:00 debug: unpack_config: Cluster is symmetric - 
> resources can run anywhere by default
> ptest[9519]: 2009/11/03_14:48:00 debug: unpack_config: On loss of CCM Quorum: 
> Stop ALL resources
> ptest[9519]: 2009/11/03_14:48:00 info: determine_online_status: Node lnxhat1 
> is online
> ptest[9519]: 2009/11/03_14:48:00 debug: common_apply_stickiness: 
> fail-count-resource_oracle: INFINITY
> ptest[9519]: 2009/11/03_14:48:00 WARN: unpack_rsc_op: Processing failed op 
> resource_listener_monitor_0 on lnxhat1: Timed Out
> ptest[9519]: 2009/11/03_14:48:00 ERROR: unpack_rsc_op: Remapping 
> resource_oracle_start_0 (rc=1) on lnxhat1 to an ERROR
> ptest[9519]: 2009/11/03_14:48:00 WARN: unpack_rsc_op: Processing failed op 
> resource_oracle_start_0 on lnxhat1: Error
> ptest[9519]: 2009/11/03_14:48:00 WARN: unpack_rsc_op: Compatability handling 
> for failed op resource_oracle_start_0 on lnxhat1
> ptest[9519]: 2009/11/03_14:48:00 info: determine_online_status: Node lnxhat2 
> is online
> ptest[9519]: 2009/11/03_14:48:00 WARN: unpack_rsc_op: Processing failed op 
> resource_listener_monitor_0 on lnxhat2: Timed Out
> ptest[9519]: 2009/11/03_14:48:00 notice: group_print: Resource Group: rg_A
> ptest[9519]: 2009/11/03_14:48:00 notice: native_print:     resource_IP  
> (ocf::heartbeat:IPaddr):        Started lnxhat2
> ptest[9519]: 2009/11/03_14:48:00 notice: native_print:     resource_oracle    
>   (ocf::heartbeat:oracle):        Started lnxhat2
> ptest[9519]: 2009/11/03_14:48:00 notice: native_print:     resource_listener  
>   (ocf::heartbeat:oralsnr):       Started lnxhat2
> ptest[9519]: 2009/11/03_14:48:00 notice: clone_print: Clone Set: 1
> ptest[9519]: 2009/11/03_14:48:00 notice: native_print:     resource_stonith:0 
>   (stonith:ssh):  Stopped
> ptest[9519]: 2009/11/03_14:48:00 notice: native_print:     resource_stonith:1 
>   (stonith:ssh):  Stopped
> ptest[9519]: 2009/11/03_14:48:00 notice: clone_print: Clone Set: ocfs2cloneset
> ptest[9519]: 2009/11/03_14:48:00 notice: native_print:     ocfs2clone:0 
> (ocf::heartbeat:Filesystem):    Started lnxhat1
> ptest[9519]: 2009/11/03_14:48:00 notice: native_print:     ocfs2clone:1 
> (ocf::heartbeat:Filesystem):    Started lnxhat2
> ptest[9519]: 2009/11/03_14:48:00 debug: native_assign_node: Assigning lnxhat2 
> to resource_IP
> ptest[9519]: 2009/11/03_14:48:00 debug: native_assign_node: Assigning lnxhat2 
> to resource_oracle
> ptest[9519]: 2009/11/03_14:48:00 debug: native_assign_node: Assigning lnxhat2 
> to resource_listener
> ptest[9519]: 2009/11/03_14:48:00 debug: native_assign_node: All nodes for 
> resource resource_stonith:0 are unavailable, unclean or shutting down
> ptest[9519]: 2009/11/03_14:48:00 WARN: native_color: Resource 
> resource_stonith:0 cannot run anywhere
> ptest[9519]: 2009/11/03_14:48:00 debug: native_assign_node: All nodes for 
> resource resource_stonith:1 are unavailable, unclean or shutting down
> ptest[9519]: 2009/11/03_14:48:00 WARN: native_color: Resource 
> resource_stonith:1 cannot run anywhere
> ptest[9519]: 2009/11/03_14:48:00 debug: clone_color: Allocated 0 1 instances 
> of a possible 2
> ptest[9519]: 2009/11/03_14:48:00 debug: native_assign_node: Assigning lnxhat1 
> to ocfs2clone:0
> ptest[9519]: 2009/11/03_14:48:00 debug: native_assign_node: Assigning lnxhat2 
> to ocfs2clone:1
> ptest[9519]: 2009/11/03_14:48:00 debug: clone_color: Allocated 2 
> ocfs2cloneset instances of a possible 2
> ptest[9519]: 2009/11/03_14:48:00 notice: NoRoleChange: Leave resource 
> resource_IP       (lnxhat2)
> ptest[9519]: 2009/11/03_14:48:00 notice: NoRoleChange: Leave resource 
> resource_oracle   (lnxhat2)
> ptest[9519]: 2009/11/03_14:48:00 notice: NoRoleChange: Leave resource 
> resource_listener (lnxhat2)
> ptest[9519]: 2009/11/03_14:48:00 debug: child_starting_constraints: 1 has no 
> active children
> ptest[9519]: 2009/11/03_14:48:00 debug: child_stopping_constraints: 1 has no 
> active children
> ptest[9519]: 2009/11/03_14:48:00 notice: NoRoleChange: Leave resource 
> ocfs2clone:0      (lnxhat1)
> ptest[9519]: 2009/11/03_14:48:00 notice: NoRoleChange: Leave resource 
> ocfs2clone:1      (lnxhat2)
> ***556*** [r...@lnxhat1:/var/lib/heartbeat/crm] *** uid=0 ***
>  #
>
> ________________________________
> The information contained in this e-mail message is intended only for the 
> personal and confidential use of the recipient(s) named above. This message 
> may be an attorney-client communication and/or work product and as such is 
> privileged and confidential. If the reader of this message is not the 
> intended recipient or an agent responsible for delivering it to the intended 
> recipient, you are hereby notified that you have received this document in 
> error and that any review, dissemination, distribution, or copying of this 
> message is strictly prohibited. If you have received this communication in 
> error, please notify us immediately by e-mail, and delete the original 
> message.
> _______________________________________________
> Linux-HA mailing list
> [email protected]
> http://lists.linux-ha.org/mailman/listinfo/linux-ha
> See also: http://linux-ha.org/ReportingProblems
>
_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to