Send Linux-ha-cvs mailing list submissions to
[email protected]
To subscribe or unsubscribe via the World Wide Web, visit
http://lists.community.tummy.com/mailman/listinfo/linux-ha-cvs
or, via email, send a message with subject or body 'help' to
[EMAIL PROTECTED]
You can reach the person managing the list at
[EMAIL PROTECTED]
When replying, please edit your Subject line so it is more specific
than "Re: Contents of Linux-ha-cvs digest..."
Today's Topics:
1. Linux-HA CVS: cts by andrew from
([email protected])
2. Linux-HA CVS: cts by andrew from
([email protected])
3. Linux-HA CVS: cts by andrew from
([email protected])
----------------------------------------------------------------------
Message: 1
Date: Wed, 19 Apr 2006 08:51:42 -0600 (MDT)
From: [email protected]
Subject: [Linux-ha-cvs] Linux-HA CVS: cts by andrew from
To: [EMAIL PROTECTED]
Message-ID: <[EMAIL PROTECTED]>
linux-ha CVS committal
Author : andrew
Host :
Project : linux-ha
Module : cts
Dir : linux-ha/cts
Modified Files:
CTSlab.py.in
Log Message:
typo
===================================================================
RCS file: /home/cvs/linux-ha/linux-ha/cts/CTSlab.py.in,v
retrieving revision 1.60
retrieving revision 1.61
diff -u -3 -r1.60 -r1.61
--- CTSlab.py.in 19 Apr 2006 13:48:50 -0000 1.60
+++ CTSlab.py.in 19 Apr 2006 14:51:42 -0000 1.61
@@ -783,10 +783,10 @@
traceback.print_exc(50, logmethod)
tests.summarize()
- if test.Stats["failure"] > 0:
+ if tests.Stats["failure"] > 0:
sys.exit(test.Stats["failure"])
- elif test.Stats["success"] != NumIter:
+ elif tests.Stats["success"] != NumIter:
cm.Env.log("No failure count but success != requested iterations")
sys.exit(1)
------------------------------
Message: 2
Date: Wed, 19 Apr 2006 08:52:32 -0600 (MDT)
From: [email protected]
Subject: [Linux-ha-cvs] Linux-HA CVS: cts by andrew from
To: [EMAIL PROTECTED]
Message-ID: <[EMAIL PROTECTED]>
linux-ha CVS committal
Author : andrew
Host :
Project : linux-ha
Module : cts
Dir : linux-ha/cts
Modified Files:
CTSlab.py.in
Log Message:
tabs->spaces
===================================================================
RCS file: /home/cvs/linux-ha/linux-ha/cts/CTSlab.py.in,v
retrieving revision 1.61
retrieving revision 1.62
diff -u -3 -r1.61 -r1.62
--- CTSlab.py.in 19 Apr 2006 14:51:42 -0000 1.61
+++ CTSlab.py.in 19 Apr 2006 14:52:32 -0000 1.62
@@ -653,9 +653,9 @@
node_list = config.Parameters["node"]
if DoBSC:
- NumIter = 2
- Version = 2
- LimitNodes = 1
+ NumIter = 2
+ Version = 2
+ LimitNodes = 1
if LogFacility == None:
if config.Parameters.has_key("logfacility"):
@@ -672,7 +672,7 @@
# alt_list = []
# for node in node_list:
-# alt_list.append(string.lower(node))
+# alt_list.append(string.lower(node))
# node_list = alt_list
Environment = CtsLab(node_list)
@@ -696,7 +696,7 @@
Environment["logfacility"] = LogFacility
if config.Parameters.has_key("use_logd"):
- Environment["use_logd"] = 1
+ Environment["use_logd"] = 1
if Version == 2:
from CM_LinuxHAv2 import LinuxHAv2
@@ -708,15 +708,15 @@
Environment["reset"] = Stonith(sttype="external/ssh",
parm=string.join(node_list, " "))
if DoBSC:
- Environment["DoBSC"] = 1
- Environment["ClobberCIB"] = 1
- Environment["CIBResource"] = 0
+ Environment["DoBSC"] = 1
+ Environment["ClobberCIB"] = 1
+ Environment["CIBResource"] = 0
Environment["logger"] = (FileLog(Environment), StdErrLog(Environment))
- scenario = Scenario([ BasicSanityCheck(Environment) ])
+ scenario = Scenario([ BasicSanityCheck(Environment) ])
else:
- scenario = Scenario(
- [ InitClusterManager(Environment), PacketLoss(Environment)])
-
+ scenario = Scenario(
+ [ InitClusterManager(Environment), PacketLoss(Environment)])
+
Environment.SupplyDefaults()
# Your basic start up the world type of test scenario...
@@ -728,7 +728,7 @@
# Create the Cluster Manager object
cm = Environment['CMclass'](Environment)
if TruncateLog:
- cm.log("Truncating %s" % LogFile)
+ cm.log("Truncating %s" % LogFile)
lf = open(LogFile, "w");
if lf != None:
lf.truncate(0)
@@ -748,16 +748,16 @@
out=cm.rsh.readaline(node, "@sbindir@/crm_uuid")
if not out:
cm.log(" * %s: __undefined_uuid__" % node)
- else:
- out = out[:-1]
- cm.log(" * %s: %s" % (node, out))
-
+ else:
+ out = out[:-1]
+ cm.log(" * %s: %s" % (node, out))
+
Audits = AuditList(cm)
Tests = []
if Environment["DoBSC"]:
test = BSC_AddResource(cm)
- Tests.append(test)
+ Tests.append(test)
elif TestCase != None:
for test in TestList(cm):
if test.name == TestCase:
@@ -772,7 +772,7 @@
for test in Tests :
cm.log(str(test.name));
sys.exit(0)
-
+
tests = RandomTests(scenario, cm, Tests, Audits)
Environment.RandomTests = tests
try :
@@ -780,13 +780,13 @@
except :
cm.Env.log("Exception by %s" % sys.exc_info()[0])
for logmethod in Environment["logger"]:
- traceback.print_exc(50, logmethod)
+ traceback.print_exc(50, logmethod)
tests.summarize()
if tests.Stats["failure"] > 0:
- sys.exit(test.Stats["failure"])
+ sys.exit(test.Stats["failure"])
elif tests.Stats["success"] != NumIter:
cm.Env.log("No failure count but success != requested iterations")
- sys.exit(1)
-
+ sys.exit(1)
+
------------------------------
Message: 3
Date: Wed, 19 Apr 2006 08:53:46 -0600 (MDT)
From: [email protected]
Subject: [Linux-ha-cvs] Linux-HA CVS: cts by andrew from
To: [EMAIL PROTECTED]
Message-ID: <[EMAIL PROTECTED]>
linux-ha CVS committal
Author : andrew
Host :
Project : linux-ha
Module : cts
Dir : linux-ha/cts
Modified Files:
CM_LinuxHAv2.py.in
Log Message:
tabs->spaces
some extra logs i'm interested in
===================================================================
RCS file: /home/cvs/linux-ha/linux-ha/cts/CM_LinuxHAv2.py.in,v
retrieving revision 1.146
retrieving revision 1.147
diff -u -3 -r1.146 -r1.147
--- CM_LinuxHAv2.py.in 19 Apr 2006 12:16:00 -0000 1.146
+++ CM_LinuxHAv2.py.in 19 Apr 2006 14:53:46 -0000 1.147
@@ -51,10 +51,10 @@
def __init__(self, Environment, randseed=None):
HeartbeatCM.__init__(self, Environment, randseed=randseed)
- self.cib_installed = 0
+ self.cib_installed = 0
self.config = None
self.cluster_monitor = 0
- self.use_short_names = 1
+ self.use_short_names = 1
self.update({
"Name" : "linux-ha-v2",
"DeadTime" : 300,
@@ -66,7 +66,7 @@
"StatusCmd" : "@sbindir@/crmadmin -S %s 2>/dev/null",
"EpocheCmd" : "@sbindir@/ccm_tool -e",
"QuorumCmd" : "@sbindir@/ccm_tool -q",
- "CibQuery" : "@sbindir@/cibadmin -Ql",
+ "CibQuery" : "@sbindir@/cibadmin -Ql",
"ParitionCmd" : "@sbindir@/ccm_tool -p",
"IsRscRunning" : "@libdir@/heartbeat/lrmadmin -E %s monitor 0 0
EVERYTIME 2>/dev/null|grep return",
"ExecuteRscOp" : "@libdir@/heartbeat/lrmadmin -E %s %s 0 0
EVERYTIME 2>/dev/null",
@@ -85,11 +85,11 @@
# This wont work if we have multiple partitions
# Use: "Pat:They_started" : "%s crmd:.*State transition.*->
S_NOT_DC",
"Pat:They_started" : "Updating node state to member for %s",
- "Pat:We_started" : "crmd.*%s: State transition.*-> S_IDLE",
- "Pat:We_stopped" : "heartbeat.*%s.*Heartbeat shutdown complete",
+ "Pat:We_started" : "crmd.*%s: State transition.*-> S_IDLE",
+ "Pat:We_stopped" : "heartbeat.*%s.*Heartbeat shutdown complete",
"Pat:Logd_stopped" : "%s logd:.*Exiting write process",
"Pat:They_stopped" : "%s crmd:.*LOST:.* %s ",
- "Pat:All_stopped" : "heartbeat.*%s.*Heartbeat shutdown complete",
+ "Pat:All_stopped" : "heartbeat.*%s.*Heartbeat shutdown complete",
"Pat:They_dead" : "node %s.*: is dead",
"Pat:TransitionComplete" : "Transition status: Complete: complete",
@@ -114,14 +114,20 @@
r"Confirm not received from",
r"Welcome reply not received from",
r"Attempting to schedule .* after a stop",
- r"Resource .* was active at shutdown",
- r"duplicate entries for call_id",
+ r"Resource .* was active at shutdown",
+ r"duplicate entries for call_id",
+ r"Search terminated:",
+ r"No need to invoke the TE",
+ r":global_timer_callback",
+ r"Updating failcount for ",
+ r"Faking parameter digest creation",
+ r"Parameters to .* action changed:",
),
})
del self["Standby"]
- if self.Env["DoBSC"]:
- del self["Pat:They_stopped"]
- del self["Pat:Logd_stopped"]
+ if self.Env["DoBSC"]:
+ del self["Pat:They_stopped"]
+ del self["Pat:Logd_stopped"]
self.check_transitions = 0
self.check_elections = 0
@@ -142,21 +148,21 @@
def install_config(self, node):
if not self.CIBsync.has_key(node) and self.Env["ClobberCIB"] == 1:
self.CIBsync[node] = 1
- self.rsh.remote_py(node, "os", "system", "rm -f
@HA_VARLIBDIR@/heartbeat/crm/cib.xml")
- self.rsh.remote_py(node, "os", "system", "rm -f
@HA_VARLIBDIR@/heartbeat/crm/cib.xml.sig")
+ self.rsh.remote_py(node, "os", "system", "rm -f
@HA_VARLIBDIR@/heartbeat/crm/cib.xml")
+ self.rsh.remote_py(node, "os", "system", "rm -f
@HA_VARLIBDIR@/heartbeat/crm/cib.xml.sig")
- # Only install the CIB on the first node, all the other ones will
pick it up from there
- if self.cib_installed == 1:
- return None
+ # Only install the CIB on the first node, all the other ones will
pick it up from there
+ if self.cib_installed == 1:
+ return None
- self.cib_installed = 1
+ self.cib_installed = 1
if self.Env["CIBfilename"] == None:
self.debug("Installing Generated CIB on node %s" %(node))
- warnings.filterwarnings("ignore")
- cib_file=os.tmpnam()
- warnings.resetwarnings()
+ warnings.filterwarnings("ignore")
+ cib_file=os.tmpnam()
+ warnings.resetwarnings()
os.system("rm -f "+cib_file)
- self.debug("Creating new CIB for " + node + " in: " + cib_file)
+ self.debug("Creating new CIB for " + node + " in: " + cib_file)
os.system("echo \'" + self.default_cts_cib + "\' > " +
cib_file)
if 0!=self.rsh.echo_cp(None, cib_file, node,
"@HA_VARLIBDIR@/heartbeat/crm/cib.xml"):
raise ValueError("Can not create CIB on %s "%node)
@@ -166,8 +172,8 @@
self.debug("Installing CIB (%s) on node %s"
%(self.Env["CIBfilename"], node))
if 0!=self.rsh.cp(self.Env["CIBfilename"], "root@" +
(self["CIBfile"]%node)):
raise ValueError("Can not scp file to %s "%node)
-
- self.rsh.remote_py(node, "os", "system", "chown @HA_CCMUSER@
@HA_VARLIBDIR@/heartbeat/crm/cib.xml")
+
+ self.rsh.remote_py(node, "os", "system", "chown @HA_CCMUSER@
@HA_VARLIBDIR@/heartbeat/crm/cib.xml")
def prepare(self):
'''Finish the Initialization process. Prepare to test...'''
@@ -313,28 +319,28 @@
return None
def Configuration(self):
- if self.config:
- return self.config.getElementsByTagName('configuration')[0]
+ if self.config:
+ return self.config.getElementsByTagName('configuration')[0]
- warnings.filterwarnings("ignore")
- cib_file=os.tmpnam()
- warnings.resetwarnings()
-
- os.system("rm -f "+cib_file)
+ warnings.filterwarnings("ignore")
+ cib_file=os.tmpnam()
+ warnings.resetwarnings()
+
+ os.system("rm -f "+cib_file)
if self.Env["ClobberCIB"] == 1:
if self.Env["CIBfilename"] == None:
- self.debug("Creating new CIB in: " + cib_file)
+ self.debug("Creating new CIB in: " + cib_file)
os.system("echo \'"+ self.default_cts_cib +"\' > "+ cib_file)
else:
os.system("cp "+self.Env["CIBfilename"]+" "+cib_file)
- else:
- if 0 != self.rsh.echo_cp(
- self.Env["nodes"][0], "@HA_VARLIBDIR@/heartbeat/crm/cib.xml",
None, cib_file):
+ else:
+ if 0 != self.rsh.echo_cp(
+ self.Env["nodes"][0], "@HA_VARLIBDIR@/heartbeat/crm/cib.xml",
None, cib_file):
raise ValueError("Can not copy file to %s, maybe permission
denied"%cib_file)
- self.config = parse(cib_file)
- os.remove(cib_file)
+ self.config = parse(cib_file)
+ os.remove(cib_file)
return self.config.getElementsByTagName('configuration')[0]
@@ -354,10 +360,10 @@
for grp in groups:
for rsc in rscs:
if rsc in grp.childNodes:
- if self.use_short_names:
- resource = HAResource(self,rsc)
- else:
- resource = HAResource(self,rsc,grp.getAttribute('id'))
+ if self.use_short_names:
+ resource = HAResource(self,rsc)
+ else:
+ resource = HAResource(self,rsc,grp.getAttribute('id'))
ResourceList.append(resource)
for inc in incs:
@@ -375,10 +381,10 @@
rsc.inc_no = i
rsc.inc_name = inc_name
rsc.inc_max = max
- if self.use_short_names:
- rsc.rid = rsc.rid + ":%d"%i
- else:
- rsc.rid = inc_name+":"+rsc.rid + ":%d"%i
+ if self.use_short_names:
+ rsc.rid = rsc.rid + ":%d"%i
+ else:
+ rsc.rid = inc_name+":"+rsc.rid + ":%d"%i
rsc.Instance = rsc.rid
ResourceList.append(rsc)
return ResourceList
@@ -394,10 +400,10 @@
GroupList.append(group)
for rsc in rscs:
if rsc in grp.childNodes:
- if self.use_short_names:
- resource = HAResource(self,rsc)
- else:
- resource = HAResource(self,rsc,grp.getAttribute('id'))
+ if self.use_short_names:
+ resource = HAResource(self,rsc)
+ else:
+ resource = HAResource(self,rsc,grp.getAttribute('id'))
group.append(resource)
return GroupList
@@ -462,11 +468,11 @@
if self.ShouldBeStatus[node] == self["up"]:
quorum = self.rsh.readaline(node, self["QuorumCmd"])
if string.find(quorum, "1") != -1:
- return 1
+ return 1
elif string.find(quorum, "0") != -1:
- return 0
- else:
- self.log("WARN: Unexpected quorum test result from "+ node
+":"+ quorum)
+ return 0
+ else:
+ self.log("WARN: Unexpected quorum test result from "+ node
+":"+ quorum)
return 0
def Components(self):
@@ -497,7 +503,7 @@
out=self.rsh.readaline(node, self["StandbyQueryCmd"]%node)
if not out:
return "off"
- out = out[:-1]
+ out = out[:-1]
self.debug("Standby result: "+out)
return out
@@ -509,7 +515,7 @@
return True
current_status = self.StandbyStatus(node)
- cmd = self["StandbyCmd"] % (node, status)
+ cmd = self["StandbyCmd"] % (node, status)
ret = self.rsh(node, cmd)
return True
------------------------------
_______________________________________________
Linux-ha-cvs mailing list
[email protected]
http://lists.community.tummy.com/mailman/listinfo/linux-ha-cvs
End of Linux-ha-cvs Digest, Vol 29, Issue 102
*********************************************