I have recently taken a job working in the Performance and Capacity Mgmt
area of the company, but used to do z/Linux support (and AIX support). Our
setup is odd, they moved z/Linux support to reporting to the AIX area about
two years ago. I hadn't had much time prior to that to really learn the
guts of Velocity. Now that I've moved areas, I'm hoping I can get some
help on where else to look in Velocity to find some indicators of what
might be the problem with a particular applications on z/Linux. Some
background...there is 1 production guest (wasp50) running all our internal
WebSphere applications (3 appb servers on the guest and 5G memory). The
test workload is split between 2 z/Linux guests with 2 app servers on each.
What is happening is that our Claims area is running a batch process
through z/Linux twice a day and as you can see from the numbers below, it
really cranks the system. So far, I can dig to the point that it's a java
process that's using up most the time, but are there other screens in
particular that can help me point the developers in the right direction as
to what they can do to make this process run better (although the good news
is it won't run during our prime shift once it's in production)? I have
access to the browser based screens for Velocity, but could also talk to
the zVM admin to have him look at stuff through the mainframe screens. At
this point, the AIX area is pushing to move everything off of z/Linux and
to AIX because it will 'run better' there and they've done a good job of
convincing management. Any thoughts would be appreciated.
ESAMAIN
<---Users----> Transact. <Processor> Cap- <--Storage (MB)->
<-Paging--> <-----I/O-----> <MiniDisk> Spool Communications
<-avg number-> per Avg. Utilization ture Fixed Active Stor
<pages/sec> <-DASD--> Other <-Cache--> Page <-per second->
Time On Actv In Q Sec. Time CPUs Total Virt. Ratio User Resid. Load
XStore DASD Rate Resp Rate Rate %Hit Rate IUCV VMCF
-------- ---- ---- ---- ---- ---- ---- ----- ----- ----- ----- ------ ----
------ ---- ---- ---- ----- ----- ---- ----- ------ ------
08:27:00 24 20 7.0 3.4 0.17 2 153.3 150.2 100 67 12063 0.4
1 0 36 1 0 13 100 0 198 0
08:26:00 24 21 8.0 4.0 0.13 2 149.6 146.4 100 67 12064 0.4
0 0 35 1 0 1 100 0 197 0
08:25:00 24 20 10.0 3.6 0.14 2 150.0 146.7 100 67 12063 0.4
0 0 34 1 0 1 100 0 194 0
08:24:00 24 20 8.0 4.0 0.13 2 153.5 150.2 100 67 12063 0.4
0 0 34 1 0 1 95.1 0 196 0
08:23:00 24 20 10.0 3.4 0.16 2 164.0 161.0 100 67 12063 0.4
0 0 23 2 0 1 100 0 194 0
08:22:00 24 20 9.0 4.0 0.13 2 153.6 150.7 100 67 12063 0.4
0 0 26 2 0 1 95.7 0 196 0
08:21:00 24 20 8.0 3.7 0.14 2 144.7 141.6 100 67 12063 0.4
0 0 26 2 0 1 100 0 194 0
08:20:00 24 20 9.0 4.1 0.10 2 138.5 135.1 100 67 12063 0.4
0 0 23 2 0 1 97.7 0 196 0
08:19:00 24 20 9.0 3.7 0.11 2 182.1 179.2 100 67 12063 0.4
0 0 26 2 0 1 100 0 194 0
08:18:00 24 21 8.0 3.9 0.13 2 155.1 151.9 100 67 12064 0.4
0 0 24 2 0 7 95.5 0 196 0
08:17:00 24 20 10.0 4.1 0.10 2 153.7 150.6 100 67 12063 0.4
0 0 24 2 0 6 97.3 0 194 0
ESAUMENU:ESATUSRS
<------CPU time-------> <----Main Storage (pages)----->
<---------Paging (pages)----------> <Spooling(pages)> Qed Resid Frame
Address
UserID <----(seconds)----> T:V <Resident> Lock <---WSSize---->
<-------Allocated-----> <---I/O---> <---I/O---> Pg+ at List
Spaces
Time /Class Total Virt Rat Total Actv -ed Total Actv Avg
Total ExStg Pref NPref Read Write Alloc Read Write Spl Reset Reord Avg
Max
-------- -------- ---------- -------- --- ----- ---- ---- ----- ---- ----
----- ----- ----- ----- ----- ----- ----- ----- ----- --- ----- ----- ---
---
08:30:00 System: 94.000 92.329 1.0 3088K 3M 824 3169K 3M 132K
158K 75085 0 83323 0 0 2624 0 0 0 0 0 0
0
08:30:00 WAST51 56.188 55.793 1.0 711K 711K 24 717K 717K 717K
4771 0 0 4771 0 0 0 0 0 0 0 0 0
0
08:30:00 WASP50 33.188 32.760 1.0 1273K 1M 25 1311K 1M 1M
36251 1295 0 34956 0 0 0 0 0 0 0 0 0
0
08:30:00 WAST50 2.031 1.959 1.0 692K 692K 24 717K 717K 717K
23824 16747 0 7077 0 0 0 0 0 0 0 0 0
0
08:30:00 LNPRODR2 1.917 1.362 1.4 106K 106K 176 105K 105K 105K
15251 4971 0 10280 0 0 0 0 0 0 0 0 0
0
08:30:00 LNOUS 0.155 0.151 1.0 119K 119K 13 131K 131K 131K
11214 8217 0 2997 0 0 0 0 0 0 0 0 0
0
08:30:00 DTCVSW1 0.148 0.000 3K 84 84 34 50 50 50
2502 1 0 2501 0 0 1 0 0 0 0 0 0
0
08:30:00 LNS10AP 0.095 0.093 1.0 107K 107K 13 107K 107K 107K
23617 20452 0 3165 0 0 0 0 0 0 0 0 0
0
08:30:00 LINUX1 0.090 0.087 1.0 75490 75K 13 75456 75K 75K
26449 22441 0 4008 0 0 0 0 0 0 0 0 0
0
08:30:00 ESATCP 0.060 0.045 1.4 513 513 1 512 512 512
647 17 0 630 0 0 2 0 0 0 0 0 0 0
08:30:00 ESAWRITE 0.018 0.017 1.1 639 639 1 638 638 638
467 22 0 445 0 0 945 0 0 0 0 0 0 0
ESAXMENU:ESALNXP
08:26:00 WAST51 snmpd 2352 1 2351 0.2 0.1 0.0 0 0 -10
0.1 0.1 0.0 0 0 6483 2496 2 0 0 0
08:26:00 WAST51 qpmon 4840 4811 4810 0.1 0.0 0.1 0 0 0
0.1 0.0 0.1 0 0 20K 2440 0 0 0 0
08:26:00 WAST51 smbd 6730 2528 2528 0.3 0.1 0.2 0 0 0
0.2 0.1 0.1 0 0 14K 4232 0 0 0 0
08:26:00 WAST51 java 28704 1 28704 0.2 0.0 0.2 0 0 0
0.1 0.0 0.1 0 0 463K 232K 0 0 0 0
08:26:00 WAST51 java 28847 1 28847 91.7 3.4 88.3 0 0 0
55.0 2.0 53.0 0 0 1M 1M 19 0 0 0
08:26:00 WAST51 java 29117 1 29117 0.5 0.1 0.4 0 0 0
0.3 0.1 0.3 0
Thanks!
Joell Chockley
System Capacity/Performance Specialist
Blue Cross Blue Shield of KS
1133 Topeka Blvd
Topeka, KS 66629-0001
Work (785)291-7837
CONFIDENTIALITY NOTICE: This email message and any attachments are for the sole
use of the intended recipient(s) and may contain proprietary, confidential,
trade secret or privileged information. Any unauthorized review use,
disclosure or distribution is prohibited and may be a violation of law. If you
are not the intended recipient or a person responsible for delivering this
message to an intended recipient, please contact the sender by reply email and
destroy all copies of the original message.
----------------------------------------------------------------------
For LINUX-390 subscribe / signoff / archive access instructions,
send email to [email protected] with the message: INFO LINUX-390 or visit
http://www.marist.edu/htbin/wlvindex?LINUX-390