three clone resources 
1 - dlm-clone
2 - o2cb-clone
3 - ocf2-clone

the 'ocfs2-clone' resource does not start, please help me

# crm configure show
node node1
node node2 \
        attributes standby="off"
primitive dlm ocf:pacemaker:controld \
        operations $id="dlm-operations" \
        op monitor interval="10" timeout="20" start-delay="0"
primitive o2cb ocf:ocfs2:o2cb \
        operations $id="o2cb-operations" \
        op monitor interval="10" timeout="20"
primitive ocfs2 ocf:heartbeat:Filesystem \
        operations $id="ocfs2-operations" \
        op monitor interval="20" timeout="40" \
        op start interval="0" timeout="60" \
        op stop interval="0" timeout="60" \
        params
device="/dev/disk/by-id/scsi-149455400000000000000000001000000a20600000f000000" 
directory="/sapmnt" fstype="ocfs2"
clone dlm-clone dlm \
        meta clone-max="2" target-role="started"
clone o2cb-clone o2cb \
        meta clone-max="2" target-role="started"
clone ocfs2-clone ocfs2 \
        meta clone-max="2" target-role="started"
colocation o2cb-with-dlm inf: o2cb-clone dlm-clone
colocation ocfs2-with-o2cb inf: ocfs2-clone o2cb-clone
order o2cb-after-dlm : dlm-clone o2cb-clone
order ocfs2-after-o2cb : o2cb-clone ocfs2-clone
property $id="cib-bootstrap-options" \
        dc-version="1.1.2-2e096a41a5f9e184a1c1537c82c6da1093698eb5" \
        cluster-infrastructure="openais" \
        expected-quorum-votes="2" \
        last-lrm-refresh="1299240986" \
        no-quorum-policy="ignore" \
        stonith-enabled="false"

# crm_verify -LV
crm_verify[6127]: 2011/03/04_20:19:37 WARN: unpack_rsc_op: Processing
failed op ocfs2:1_start_0 on node2: unknown error (1)
crm_verify[6127]: 2011/03/04_20:19:37 WARN: unpack_rsc_op: Processing
failed op ocfs2:0_start_0 on node1: unknown error (1)

from /var/log/messages
Mar  4 20:08:27 node1 crmd: [4325]: info: te_rsc_command: Initiating
action 25: start ocfs2:1_start_0 on node2
Mar  4 20:08:27 node1 crmd: [4325]: info: process_lrm_event: LRM
operation o2cb:0_monitor_10000 (call=8, rc=0, cib-update=39,
confirmed=false) ok
Mar  4 20:08:27 node1 crmd: [4325]: info: match_graph_event: Action
o2cb:0_monitor_10000 (16) confirmed on node1 (rc=0)
Mar  4 20:08:28 node1 Filesystem[4495]: INFO: Running start
for /dev/disk/by-id/scsi-149455400000000000000000001000000a20600000f000000 on 
/sapmnt
Mar  4 20:08:28 node1 crmd: [4325]: info: match_graph_event: Action
o2cb:1_monitor_10000 (18) confirmed on node2 (rc=0)
Mar  4 20:08:28 node1 lrmd: [4322]: info: RA output:
(ocfs2:0:start:stderr) mount.ocfs2
Mar  4 20:08:28 node1 lrmd: [4322]: info: RA output:
(ocfs2:0:start:stderr) : 
Mar  4 20:08:28 node1 lrmd: [4322]: info: RA output:
(ocfs2:0:start:stderr) Cluster stack specified does not match the one
currently running
Mar  4 20:08:28 node1 lrmd: [4322]: info: RA output:
(ocfs2:0:start:stderr)  
Mar  4 20:08:28 node1 lrmd: [4322]: info: RA output:
(ocfs2:0:start:stderr) while trying to join the group
Mar  4 20:08:28 node1 lrmd: [4322]: info: RA output:
(ocfs2:0:start:stderr)  
Mar  4 20:08:28 node1 Filesystem[4495]: ERROR: Couldn't mount
filesystem 
/dev/disk/by-id/scsi-149455400000000000000000001000000a20600000f000000 on 
/sapmnt
Mar  4 20:08:28 node1 crmd: [4325]: info: process_lrm_event: LRM
operation ocfs2:0_start_0 (call=9, rc=1, cib-update=40, confirmed=true)
unknown error
Mar  4 20:08:28 node1 crmd: [4325]: WARN: status_from_rc: Action 23
(ocfs2:0_start_0) on node1 failed (target: 0 vs. rc: 1): Error
Mar  4 20:08:28 node1 crmd: [4325]: WARN: update_failcount: Updating
failcount for ocfs2:0 on node1 after failed start: rc=1
(update=INFINITY, time=1299251308)
Mar  4 20:08:28 node1 crmd: [4325]: info: abort_transition_graph:
match_graph_event:276 - Triggered transition abort (complete=0,
tag=lrm_rsc_op, id=ocfs2:0_start_0,
magic=0:1;23:3:0:9f874490-01c6-4282-89cd-95ac1bb1f412, cib=0.441.18) :
Event failed
Mar  4 20:08:28 node1 crmd: [4325]: info: update_abort_priority: Abort
priority upgraded from 0 to 1
Mar  4 20:08:28 node1 crmd: [4325]: info: update_abort_priority: Abort
action done superceeded by restart
Mar  4 20:08:28 node1 crmd: [4325]: info: match_graph_event: Action
ocfs2:0_start_0 (23) confirmed on node1 (rc=4)
Mar  4 20:08:28 node1 attrd: [4323]: info: find_hash_entry: Creating
hash entry for fail-count-ocfs2:0
Mar  4 20:08:28 node1 attrd: [4323]: info: attrd_trigger_update: Sending
flush op to all hosts for: fail-count-ocfs2:0 (INFINITY)
Mar  4 20:08:28 node1 attrd: [4323]: info: attrd_perform_update: Sent
update 17: fail-count-ocfs2:0=INFINITY
Mar  4 20:08:28 node1 crmd: [4325]: info: abort_transition_graph:
te_update_diff:146 - Triggered transition abort (complete=0,
tag=transient_attributes, id=node1, magic=NA, cib=0.441.19) : Transient
attribute: update
Mar  4 20:08:28 node1 crmd: [4325]: info: update_abort_priority: Abort
priority upgraded from 1 to 1000000
Mar  4 20:08:28 node1 attrd: [4323]: info: find_hash_entry: Creating
hash entry for last-failure-ocfs2:0
Mar  4 20:08:28 node1 crmd: [4325]: info: update_abort_priority: 'Event
failed' abort superceeded
Mar  4 20:08:28 node1 attrd: [4323]: info: attrd_trigger_update: Sending
flush op to all hosts for: last-failure-ocfs2:0 (1299251308)
Mar  4 20:08:28 node1 attrd: [4323]: info: attrd_perform_update: Sent
update 20: last-failure-ocfs2:0=1299251308
Mar  4 20:08:28 node1 crmd: [4325]: info: abort_transition_graph:
te_update_diff:146 - Triggered transition abort (complete=0,
tag=transient_attributes, id=node1, magic=NA, cib=0.441.20) : Transient
attribute: update
Mar  4 20:08:28 node1 crmd: [4325]: WARN: status_from_rc: Action 25
(ocfs2:1_start_0) on node2 failed (target: 0 vs. rc: 1): Error
Mar  4 20:08:28 node1 crmd: [4325]: WARN: update_failcount: Updating
failcount for ocfs2:1 on node2 after failed start: rc=1
(update=INFINITY, time=1299251308)
Mar  4 20:08:28 node1 crmd: [4325]: info: abort_transition_graph:
match_graph_event:276 - Triggered transition abort (complete=0,
tag=lrm_rsc_op, id=ocfs2:1_start_0,
magic=0:1;25:3:0:9f874490-01c6-4282-89cd-95ac1bb1f412, cib=0.441.21) :
Event failed
Mar  4 20:08:28 node1 crmd: [4325]: info: match_graph_event: Action
ocfs2:1_start_0 (25) confirmed on node2 (rc=4)
Mar  4 20:08:28 node1 crmd: [4325]: info: te_pseudo_action: Pseudo
action 28 fired and confirmed
Mar  4 20:08:25 node1 o2cb[4424]: INFO: Stack glue driver not loaded
Mar  4 20:08:25 node1 o2cb[4424]: INFO: Starting o2cb:0
Mar  4 20:08:25 node1 kernel: [  333.664427] ocfs2: Registered cluster
interface user
Mar  4 20:08:25 node1 kernel: [  333.710269] OCFS2 Node Manager 1.5.0
Mar  4 20:08:25 node1 kernel: [  333.770793] OCFS2 1.5.0
Mar  4 20:08:25 node1 ocfs2_controld.pcmk: Core dumps
enabled: /var/lib/openais
Mar  4 20:08:25 node1 corosync[4312]:   [pcmk  ] info: pcmk_notify:
Enabling node notifications for child 4469 (0x80d46f0)
Mar  4 20:08:25 node1 ocfs2_controld[4469]: Cluster connection
established.  Local node id: 1
Mar  4 20:08:25 node1 ocfs2_controld[4469]: Added Pacemaker as client 1
with fd 7
Mar  4 20:08:25 node1 ocfs2_controld[4469]: Initializing CKPT service
(try 1)
Mar  4 20:08:25 node1 ocfs2_controld[4469]: Connected to CKPT service
with handle 0x327b23c600000000
Mar  4 20:08:25 node1 ocfs2_controld[4469]: Opening checkpoint
"ocfs2:controld:00000001" (try 1)
Mar  4 20:08:25 node1 ocfs2_controld[4469]: Opened checkpoint
"ocfs2:controld:00000001" with handle 0x6633487300000000
Mar  4 20:08:25 node1 ocfs2_controld[4469]: Writing to section
"daemon_max_protocol" on checkpoint "ocfs2:controld:00000001" (try 1)
Mar  4 20:08:25 node1 ocfs2_controld[4469]: Creating section
"daemon_max_protocol" on checkpoint "ocfs2:controld:00000001" (try 1)
Mar  4 20:08:25 node1 ocfs2_controld[4469]: Created section
"daemon_max_protocol" on checkpoint "ocfs2:controld:00000001"
Mar  4 20:08:25 node1 ocfs2_controld[4469]: Writing to section
"ocfs2_max_protocol" on checkpoint "ocfs2:controld:00000001" (try 1)
Mar  4 20:08:25 node1 ocfs2_controld[4469]: Creating section
"ocfs2_max_protocol" on checkpoint "ocfs2:controld:00000001" (try 1)
Mar  4 20:08:25 node1 ocfs2_controld[4469]: Created section
"ocfs2_max_protocol" on checkpoint "ocfs2:controld:00000001"
Mar  4 20:08:25 node1 ocfs2_controld[4469]: Starting join for group
"ocfs2:controld"
Mar  4 20:08:25 node1 ocfs2_controld[4469]: cpg_join succeeded
Mar  4 20:08:25 node1 ocfs2_controld[4469]: setup done
Mar  4 20:08:25 node1 ocfs2_controld[4469]: confchg called
Mar  4 20:08:25 node1 ocfs2_controld[4469]: ocfs2_controld (group
"ocfs2:controld") confchg: members 1, left 0, joined 1
Mar  4 20:08:25 node1 ocfs2_controld[4469]: CPG is live, we are the
first daemon
Mar  4 20:08:25 node1 ocfs2_controld[4469]: Opening checkpoint
"ocfs2:controld" (try 1)
Mar  4 20:08:25 node1 ocfs2_controld[4469]: Opened checkpoint
"ocfs2:controld" with handle 0x2ae8944a00000001
Mar  4 20:08:25 node1 ocfs2_controld[4469]: Writing to section
"daemon_protocol" on checkpoint "ocfs2:controld" (try 1)
Mar  4 20:08:25 node1 ocfs2_controld[4469]: Creating section
"daemon_protocol" on checkpoint "ocfs2:controld" (try 1)
Mar  4 20:08:25 node1 ocfs2_controld[4469]: Created section
"daemon_protocol" on checkpoint "ocfs2:controld"
Mar  4 20:08:25 node1 ocfs2_controld[4469]: Writing to section
"ocfs2_protocol" on checkpoint "ocfs2:controld" (try 1)
Mar  4 20:08:25 node1 ocfs2_controld[4469]: Creating section
"ocfs2_protocol" on checkpoint "ocfs2:controld" (try 1)
Mar  4 20:08:25 node1 ocfs2_controld[4469]: Created section
"ocfs2_protocol" on checkpoint "ocfs2:controld"
Mar  4 20:08:25 node1 ocfs2_controld[4469]: Daemon protocol is 1.0
Mar  4 20:08:25 node1 ocfs2_controld[4469]: fs protocol is 1.0
Mar  4 20:08:25 node1 ocfs2_controld[4469]: Connecting to dlm_controld
Mar  4 20:08:25 node1 ocfs2_controld[4469]: Opening control device
Mar  4 20:08:25 node1 ocfs2_controld[4469]: Starting to listen for
mounters
Mar  4 20:08:25 node1 ocfs2_controld[4469]: new listening connection 4
Mar  4 20:08:25 node1 cluster-dlm[4411]: process_listener: client
connection 5 fd 14
Mar  4 20:08:25 node1 ocfs2_controld[4469]: confchg called
Mar  4 20:08:25 node1 ocfs2_controld[4469]: ocfs2_controld (group
"ocfs2:controld") confchg: members 2, left 0, joined 1
Mar  4 20:08:27 node1 crmd: [4325]: info: process_lrm_event: LRM
operation o2cb:0_start_0 (call=7, rc=0, cib-update=38, confirmed=true)
ok
Mar  4 20:08:27 node1 crmd: [4325]: info: match_graph_event: Action
o2cb:0_start_0 (15) confirmed on node1 (rc=0)
Mar  4 20:08:27 node1 crmd: [4325]: info: te_rsc_command: Initiating
action 16: monitor o2cb:0_monitor_10000 on node1 (local)
Mar  4 20:08:27 node1 crmd: [4325]: info: do_lrm_rsc_op: Performing
key=16:3:0:9f874490-01c6-4282-89cd-95ac1bb1f412
op=o2cb:0_monitor_10000 )
Mar  4 20:08:27 node1 crmd: [4325]: info: match_graph_event: Action
o2cb:1_start_0 (17) confirmed on node2 (rc=0)
Mar  4 20:08:27 node1 crmd: [4325]: info: te_rsc_command: Initiating
action 18: monitor o2cb:1_monitor_10000 on node2
Mar  4 20:08:27 node1 crmd: [4325]: info: te_pseudo_action: Pseudo
action 20 fired and confirmed
Mar  4 20:08:27 node1 crmd: [4325]: info: te_pseudo_action: Pseudo
action 27 fired and confirmed
Mar  4 20:08:27 node1 crmd: [4325]: info: te_rsc_command: Initiating
action 23: start ocfs2:0_start_0 on node1 (local)
Mar  4 20:08:27 node1 crmd: [4325]: info: do_lrm_rsc_op: Performing
key=23:3:0:9f874490-01c6-4282-89cd-95ac1bb1f412 op=ocfs2:0_start_0 )
Mar  4 20:08:27 node1 lrmd: [4322]: info: rsc:ocfs2:0:9: start
Mar  4 20:08:27 node1 crmd: [4325]: info: te_rsc_command: Initiating
action 25: start ocfs2:1_start_0 on node2
Mar  4 20:08:27 node1 crmd: [4325]: info: process_lrm_event: LRM
operation o2cb:0_monitor_10000 (call=8, rc=0, cib-update=39,
confirmed=false) ok
Mar  4 20:08:27 node1 crmd: [4325]: info: match_graph_event: Action
o2cb:0_monitor_10000 (16) confirmed on node1 (rc=0)
Mar  4 20:08:28 node1 Filesystem[4495]: INFO: Running start
for /dev/disk/by-id/scsi-149455400000000000000000001000000a20600000f000000 on 
/sapmnt
Mar  4 20:08:28 node1 crmd: [4325]: info: match_graph_event: Action
o2cb:1_monitor_10000 (18) confirmed on node2 (rc=0)
Mar  4 20:08:28 node1 lrmd: [4322]: info: RA output:
(ocfs2:0:start:stderr) mount.ocfs2
Mar  4 20:08:28 node1 lrmd: [4322]: info: RA output:
(ocfs2:0:start:stderr) : 
Mar  4 20:08:28 node1 lrmd: [4322]: info: RA output:
(ocfs2:0:start:stderr) Cluster stack specified does not match the one
currently running
Mar  4 20:08:28 node1 lrmd: [4322]: info: RA output:
(ocfs2:0:start:stderr)  
Mar  4 20:08:28 node1 lrmd: [4322]: info: RA output:
(ocfs2:0:start:stderr) while trying to join the group
Mar  4 20:08:28 node1 lrmd: [4322]: info: RA output:
(ocfs2:0:start:stderr)  
Mar  4 20:08:28 node1 Filesystem[4495]: ERROR: Couldn't mount
filesystem 
/dev/disk/by-id/scsi-149455400000000000000000001000000a20600000f000000 on 
/sapmnt
Mar  4 20:08:28 node1 crmd: [4325]: info: process_lrm_event: LRM
operation ocfs2:0_start_0 (call=9, rc=1, cib-update=40, confirmed=true)
unknown error
Mar  4 20:08:28 node1 crmd: [4325]: WARN: status_from_rc: Action 23
(ocfs2:0_start_0) on node1 failed (target: 0 vs. rc: 1): Error
Mar  4 20:08:28 node1 crmd: [4325]: WARN: update_failcount: Updating
failcount for ocfs2:0 on node1 after failed start: rc=1
(update=INFINITY, time=1299251308)
Mar  4 20:08:28 node1 crmd: [4325]: info: abort_transition_graph:
match_graph_event:276 - Triggered transition abort (complete=0,
tag=lrm_rsc_op, id=ocfs2:0_start_0,
magic=0:1;23:3:0:9f874490-01c6-4282-89cd-95ac1bb1f412, cib=0.441.18) :
Event failed
Mar  4 20:08:28 node1 crmd: [4325]: info: update_abort_priority: Abort
priority upgraded from 0 to 1
Mar  4 20:08:28 node1 crmd: [4325]: info: update_abort_priority: Abort
action done superceeded by restart
Mar  4 20:08:28 node1 crmd: [4325]: info: match_graph_event: Action
ocfs2:0_start_0 (23) confirmed on node1 (rc=4)
Mar  4 20:08:28 node1 attrd: [4323]: info: find_hash_entry: Creating
hash entry for fail-count-ocfs2:0
Mar  4 20:08:28 node1 attrd: [4323]: info: attrd_trigger_update: Sending
flush op to all hosts for: fail-count-ocfs2:0 (INFINITY)
Mar  4 20:08:28 node1 attrd: [4323]: info: attrd_perform_update: Sent
update 17: fail-count-ocfs2:0=INFINITY
Mar  4 20:08:28 node1 crmd: [4325]: info: abort_transition_graph:
te_update_diff:146 - Triggered transition abort (complete=0,
tag=transient_attributes, id=node1, magic=NA, cib=0.441.19) : Transient
attribute: update
Mar  4 20:08:28 node1 crmd: [4325]: info: update_abort_priority: Abort
priority upgraded from 1 to 1000000
Mar  4 20:08:28 node1 attrd: [4323]: info: find_hash_entry: Creating
hash entry for last-failure-ocfs2:0
Mar  4 20:08:28 node1 crmd: [4325]: info: update_abort_priority: 'Event
failed' abort superceeded
Mar  4 20:08:28 node1 attrd: [4323]: info: attrd_trigger_update: Sending
flush op to all hosts for: last-failure-ocfs2:0 (1299251308)
Mar  4 20:08:28 node1 attrd: [4323]: info: attrd_perform_update: Sent
update 20: last-failure-ocfs2:0=1299251308
Mar  4 20:08:28 node1 crmd: [4325]: info: abort_transition_graph:
te_update_diff:146 - Triggered transition abort (complete=0,
tag=transient_attributes, id=node1, magic=NA, cib=0.441.20) : Transient
attribute: update
Mar  4 20:08:28 node1 crmd: [4325]: WARN: status_from_rc: Action 25
(ocfs2:1_start_0) on node2 failed (target: 0 vs. rc: 1): Error
Mar  4 20:08:28 node1 crmd: [4325]: WARN: update_failcount: Updating
failcount for ocfs2:1 on node2 after failed start: rc=1
(update=INFINITY, time=1299251308)
Mar  4 20:08:28 node1 crmd: [4325]: info: abort_transition_graph:
match_graph_event:276 - Triggered transition abort (complete=0,
tag=lrm_rsc_op, id=ocfs2:1_start_0,
magic=0:1;25:3:0:9f874490-01c6-4282-89cd-95ac1bb1f412, cib=0.441.21) :
Event failed
Mar  4 20:08:28 node1 crmd: [4325]: info: match_graph_event: Action
ocfs2:1_start_0 (25) confirmed on node2 (rc=4)
Mar  4 20:08:28 node1 crmd: [4325]: info: te_pseudo_action: Pseudo
action 28 fired and confirmed
Mar  4 20:08:28 node1 pengine: [4324]: WARN: unpack_rsc_op: Processing
failed op ocfs2:1_start_0 on node2: unknown error (1)
Mar  4 20:08:28 node1 pengine: [4324]: info: determine_online_status:
Node node1 is online
Mar  4 20:08:28 node1 pengine: [4324]: WARN: unpack_rsc_op: Processing
failed op ocfs2:0_start_0 on node1: unknown error (1)
Mar  4 20:08:28 node1 pengine: [4324]: notice: clone_print:  Clone Set:
dlm-clone
Mar  4 20:08:28 node1 pengine: [4324]: notice: short_print:
Started: [ node1 node2 ]
Mar  4 20:08:28 node1 pengine: [4324]: notice: clone_print:  Clone Set:
o2cb-clone
Mar  4 20:08:28 node1 pengine: [4324]: notice: short_print:
Started: [ node1 node2 ]
Mar  4 20:08:28 node1 pengine: [4324]: notice: clone_print:  Clone Set:
ocfs2-clone
Mar  4 20:08:28 node1 pengine: [4324]: notice: native_print:
ocfs2:0 (ocf::heartbeat:Filesystem):    Started node1 FAILED
Mar  4 20:08:28 node1 pengine: [4324]: notice: native_print:
ocfs2:1 (ocf::heartbeat:Filesystem):    Started node2 FAILED
Mar  4 20:08:28 node1 pengine: [4324]: info: get_failcount: ocfs2-clone
has failed INFINITY times on node1
Mar  4 20:08:28 node1 pengine: [4324]: WARN: common_apply_stickiness:
Forcing ocfs2-clone away from node1 after 1000000 failures (max=1000000)
Mar  4 20:08:28 node1 pengine: [4324]: info: get_failcount: ocfs2-clone
has failed INFINITY times on node1
Mar  4 20:08:28 node1 pengine: [4324]: WARN: common_apply_stickiness:
Forcing ocfs2-clone away from node1 after 1000000 failures (max=1000000)
Mar  4 20:08:28 node1 pengine: [4324]: info: get_failcount: ocfs2-clone
has failed INFINITY times on node2
Mar  4 20:08:28 node1 pengine: [4324]: WARN: common_apply_stickiness:
Forcing ocfs2-clone away from node2 after 1000000 failures (max=1000000)
Mar  4 20:08:28 node1 pengine: [4324]: info: get_failcount: ocfs2-clone
has failed INFINITY times on node2
Mar  4 20:08:28 node1 pengine: [4324]: WARN: common_apply_stickiness:
Forcing ocfs2-clone away from node2 after 1000000 failures (max=1000000)
Mar  4 20:08:28 node1 pengine: [4324]: info: native_merge_weights:
dlm-clone: Rolling back scores from ocfs2-clone
Mar  4 20:08:28 node1 pengine: [4324]: info: native_merge_weights:
o2cb-clone: Rolling back scores from ocfs2-clone
Mar  4 20:08:28 node1 pengine: [4324]: info: native_color: Resource
ocfs2:0 cannot run anywhere
Mar  4 20:08:28 node1 pengine: [4324]: info: native_color: Resource
ocfs2:1 cannot run anywhere
Mar  4 20:08:28 node1 pengine: [4324]: notice: LogActions: Leave
resource dlm:0  (Started node1)
Mar  4 20:08:28 node1 pengine: [4324]: notice: LogActions: Leave
resource dlm:1  (Started node2)
Mar  4 20:08:28 node1 pengine: [4324]: notice: LogActions: Leave
resource o2cb:0 (Started node1)
Mar  4 20:08:28 node1 pengine: [4324]: notice: LogActions: Leave
resource o2cb:1 (Started node2)
Mar  4 20:08:28 node1 pengine: [4324]: notice: LogActions: Stop resource
ocfs2:0 (node1)
Mar  4 20:08:28 node1 pengine: [4324]: notice: LogActions: Stop resource
ocfs2:1 (node2)

--
Regards,
Muhammad Sharfuddin | NDS Technologies Pvt Ltd | cell: +92-333-2144823 |
UAN: +92-21-111-111-142 ext: 113

The London Stock Exchange moves to SUSE Linux
http://www.computerworlduk.com/news/open-source/3260727/london-stock-exchange-in-historic-linux-go-live/
http://www.zdnet.com/blog/open-source/the-london-stock-exchange-moves-to-novell-linux/8285

Your Linux is Ready
http://www.novell.com/linux 

_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to