Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-02-18 Thread Walter Medenbach
Two years back we had an issue where the throughput of a database utility
dropped considerably (job had to be cancelled because it was too slow)
because it decided to not use more memory to do it's processing. This was
found to be  because the page datasets were more than 50% utilized. The
notes below are from my analysis of the problem at the time.

Some software will believe that the system is storage constrained if the
page datasets are at or near 50% full.   The software affected   use the
SYSEVENT STGTEST macro to determine how much storage it can use without
affecting system paging performance. The returned values are used to set
maximum sizes for such things as memory work areas. If these are small then
the programs will be inefficient and use significantly more cpu and do more
IOs. This should be considered when determining page dataset size.

SYSEVENT STGTST behaviour from Authorized assembler Services reference
manual:

After SRM returns, each word contains a storage amount that represents a
specific number of frames. Before you choose a number to use as the basis
for decision, be aware of how your decision affects the performance of the
system. The meaning of the returned values is:
Use of the first number will affect system performance very little, if at
all.
Use of the second number might affect system performance to some degree.
Use of the third number might substantially affect system performance.

If you base decisions on the value in the second or third word, SRM may
have to take processor storage away from other programs and replace it with
auxiliary storage.
If you are running your system in workload management goal mode, the value
returned in the third word will always be the same as the value returned in
the second word.

APAR OA20116 has a note throws some light on how the second word returned
from SYSEVENT STGTST is calculated:

Problem conclusion
The logic in IRAEVREQ was changed in a way, that the maximum of
the return value 1 and return value 2 is used as return value 2
and return value 3.

Additionally the logic in IRAEVREQ ensures that the value
returned in words 2 and 3 do not drive the system into an
auxiliary storage shortage.  The value returned in words 2 and 3
will only fill the AUX subsystem up to 50%.

Examples:
If the AUX subsystem is filled by 25% and the return value 1
contains 1000 frames.  Then the return value 2 and 3 are set to
1000 frames + 25% of the AUX slots.

Walter Medenbach

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-02-14 Thread Mark Jacobs

On 02/13/12 20:03, Jim Mulder wrote:

snip


  Just to follow up on this, APAR OA38742 has been
opened.  This problem was introduced in z/OS 1.13.

Jim Mulder   z/OS System Test   IBM Corp.  Poughkeepsie,  NY

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN

   


I tried to look at the apar in IBMLINK and received this error;

An error has occurred:

   * An error occurred accessing the database: RPA0 401.

Is it a security/integrity apar?

--
Mark Jacobs
Time Customer Service
Tampa, FL


Learn from yesterday, live for today, hope for tomorrow.
The important thing is to not stop questioning.

- Albert Einstein

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-02-14 Thread Barbara Nitz
   Just to follow up on this, APAR OA38742 has been
 opened.  This problem was introduced in z/OS 1.13.

I tried to look at the apar in IBMLINK and received this error;

An error has occurred:

* An error occurred accessing the database: RPA0 401.

Is it a security/integrity apar?

I get the same error. If it is an integrity apar, then *all* ASM apars dealing 
with abend066 or ilrcpbld errors are integrity apars (as well as information 
apars). My guess is that the servicelink entry into the apar database is 
currently broken.

I called support directly for the ASM/RSM shenanigans that we currently have 
(at z/OS 1.12) to see the apar text that SIS denied me and the ptf number.

Barbara Nitz

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-02-14 Thread Mary Anne Matyaz
I get the same error on OA38542 and OA37992, so I think it's just SIS itself. I 
opened a problem. 

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-02-14 Thread Jan Vanbrabant
No problem with me (in Belgium).



*OA38742: DIFFERENCE in ASMIORQR / ASMIORQC COUNTS become LARGE ENOUGH to
AFFECT FRAME STEAL PROCESSING and ASMIORQR INCREMENTED TOO HIGH*

During AUX paging, I/O is scheduled to drive input and/or
output I/O to auxiliary page packs.  When an I/O error occurs and ASM I/O
completion exit gets control in ILRCMP, the abnormal termination entry
point ABNMTERM: gets control for the I/O error.

When this happens, the ASM I/O is redriven, causing the ASVT
ASM 'received' counts to be incremented twice, thus never allowing the
'completed' counts (ASMIORQC) to equal the 'received' counts (AMIORQR).  If
this continues for too long, the difference between the received-completed
counts grows large enough to begin impacting RSM frame steal processing.


Apar has just been openend for a few days, Barbara.


   - Submitted date  2012-02-09
   - Last modified date 2012-02-10



Hence no PTF yet;  no aparfix either.

Jan

   -


On Tue, Feb 14, 2012 at 1:29 PM, Barbara Nitz nitz-...@gmx.net wrote:

Just to follow up on this, APAR OA38742 has been
  opened.  This problem was introduced in z/OS 1.13.
 
 I tried to look at the apar in IBMLINK and received this error;
 
 An error has occurred:
 
 * An error occurred accessing the database: RPA0 401.
 
 Is it a security/integrity apar?

 I get the same error. If it is an integrity apar, then *all* ASM apars
 dealing with abend066 or ilrcpbld errors are integrity apars (as well as
 information apars). My guess is that the servicelink entry into the apar
 database is currently broken.

 I called support directly for the ASM/RSM shenanigans that we currently
 have (at z/OS 1.12) to see the apar text that SIS denied me and the ptf
 number.

 Barbara Nitz

 --
 For IBM-MAIN subscribe / signoff / archive access instructions,
 send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-02-14 Thread Mary Anne Matyaz
SIS seems to be working again. 

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-02-13 Thread Jim Mulder
IBM Mainframe Discussion List IBM-MAIN@bama.ua.edu wrote on 02/03/2012 
12:19:04 PM:

 BTW if you have not already seen it, please look at our PMR 
81276.227.000 
 for an agonizing saga of RSM/ASM misadventures. 

 JO.Skip Robinson
 SCE Infrastructure Technology Services

 Just to follow up on this, APAR OA38742 has been 
opened.  This problem was introduced in z/OS 1.13.

Jim Mulder   z/OS System Test   IBM Corp.  Poughkeepsie,  NY

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-02-05 Thread Martin Packer
Paging (if you'll pardon the pun :-)  ) Don Deese and wondering what 
checks he has in CPExpert in this area. Might provide inspiration to z/OS 
Development - if Barbara's comment had spurred them to revisit this area.

Cheers, Martin

Martin Packer,
Mainframe Performance Consultant, zChampion
Worldwide Banking Center of Excellence, IBM

+44-7802-245-584

email: martin_pac...@uk.ibm.com

Twitter / Facebook IDs: MartinPacker
Blog: 
https://www.ibm.com/developerworks/mydeveloperworks/blogs/MartinPacker

IBM Mainframe Discussion List IBM-MAIN@bama.ua.edu wrote on 04/02/2012 
16:03:22:

 From:
 
 Barbara Nitz nitz-...@gmx.net
 
 To:
 
 IBM-MAIN@bama.ua.edu, 
 
 Date:
 
 04/02/2012 16:35
 
 Subject:
 
 Re: Very Lage Page Datasets (was ASM and HiperPAV)
 
 Sent by:
 
 IBM Mainframe Discussion List IBM-MAIN@bama.ua.edu
 
 Charles,
 
 I believe you're referring to the ASM_LOCAL_SLOT_USAGE check.
 Yes, I do.
 
  If so, this check runs at a 30 minute interval and checks the slot
 usage for each in-use local paging data set regardless of whether it
 was defined 'statically' or added via PAGE ADD. So... it does 
 eventually recognize a changed configuration. Some of the ASM checks
 are set to re-run when a PAGE ADD/DELETE is issued, but 
 ASM_LOCAL_SLOT_USAGE is currently not one of them.  Open to 
suggestions
 
 As far as I am concerned, there is 'intelligence' missing in that 
 check (my usual complaint with health checks).
 
 In our installation, we only have one device geometry, behind the 
 same controller, and all page data sets are same size. So I expect 
 the health check to recognize that adding a page ds to the overall 
 config will relieve the 'performance impact' of 30% overall usage. 
 Unfortunately, ASM does NOT prefer the new page ds, it is one of 
 many, and as I indicated before, has never filled to the same slot 
 usage as the others. Only IPL has ever evened out slot usage (and I 
 am not talking about 1% difference, more like 20% difference on the 
 same geometry and size! I think Skip mentioned this, too.) 
 
 So the old, already-at-30%-usage locals and the new one are treated 
 equal, so slot usage on the old ones increases over the 30% mark, 
 and the check would still trip every 30 minutes because of that (if 
 I hadn't set it to only check once per day until the next IPL). 
 After it had alerted us to the condition once, it becomes basically 
 useless for the life of the IPL.
 
 In my opinion, this health check should have the intelligence to 
 recognize something about the configuration (same controller, same 
 size data set) and change the checking parms accordingly, especially
 when a page data set was added. The check should obviously (as Kees 
 indicated) behave differently if different percentage slot usage is 
 due to different geometry and/or different size page ds. 
 
 Now that I have established a daily graphic that shows slot 
 utilization (as shown by Barry via MXG), I might get rid if that 
 check completely and check my graphic instead. :-)
 
 Barbara
 
 --
 For IBM-MAIN subscribe / signoff / archive access instructions,
 send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN
 






Unless stated otherwise above:
IBM United Kingdom Limited - Registered in England and Wales with number 
741598. 
Registered office: PO Box 41, North Harbour, Portsmouth, Hampshire PO6 3AU






--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-02-04 Thread Barbara Nitz
Charles,

I believe you're referring to the ASM_LOCAL_SLOT_USAGE check.
Yes, I do.

 If so, this check runs at a 30 minute interval and checks the slot usage for 
 each in-use local paging data set regardless of whether it was defined 
 'statically' or added via PAGE ADD. So... it does eventually recognize a 
 changed configuration. Some of the ASM checks are set to re-run when a PAGE 
 ADD/DELETE is issued, but ASM_LOCAL_SLOT_USAGE is currently not one of them.  
 Open to suggestions

As far as I am concerned, there is 'intelligence' missing in that check (my 
usual complaint with health checks).

In our installation, we only have one device geometry, behind the same 
controller, and all page data sets are same size. So I expect the health check 
to recognize that adding a page ds to the overall config will relieve the 
'performance impact' of 30% overall usage. 
Unfortunately, ASM does NOT prefer the new page ds, it is one of many, and as I 
indicated before, has never filled to the same slot usage as the others. Only 
IPL has ever evened out slot usage (and I am not talking about 1% difference, 
more like 20% difference on the same geometry and size! I think Skip mentioned 
this, too.) 

So the old, already-at-30%-usage locals and the new one are treated equal, so 
slot usage on the old ones increases over the 30% mark, and the check would 
still trip every 30 minutes because of that (if I hadn't set it to only check 
once per day until the next IPL). After it had alerted us to the condition 
once, it becomes basically useless for the life of the IPL.

In my opinion, this health check should have the intelligence to recognize 
something about the configuration (same controller, same size data set) and 
change the checking parms accordingly, especially when a page data set was 
added. The check should obviously (as Kees indicated) behave differently if 
different percentage slot usage is due to different geometry and/or different 
size page ds. 

Now that I have established a daily graphic that shows slot utilization (as 
shown by Barry via MXG), I might get rid if that check completely and check my 
graphic instead. :-)

Barbara

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-02-03 Thread Vernooij, CP - SPLXM
Are you sure that this occurs at 50% Aux. storage utilization? IRA201E
defines 'Critical storage shortage' at 85%.
Kees.

Skip Robinson jo.skip.robin...@sce.com wrote in message
news:of7017747b.2f20bada-on88257999.0017af50-88257999.0018d...@sce.com
...
 IEE711I SYSTEM DUMP NOT TAKEN. A CRITICAL AUXILIARY STORAGE SHORTAGE 
 EXISTS 
 
 From the message manual:
 
 If the reason field shows A CRITICAL AUXILIARY STORAGE SHORTAGE
EXISTS, 
 first you need to ensure that enough DASD resource is available for 
 captured dumps to be written out. Then, consider adding additional 
 auxiliary storage (paging) resources, because SVC dumps will not be 
 allowed again until the auxiliary storage utilization drops below 35%.
See 
 the system programmer response for message IRA201E for additional 
 information about auxiliary storage utilization concerns.
 
 According to conversations at SHARE, the threshold for this condition
is 
 too conservative. AFAIK there is no APAR open to fix it. It's
especially 
 frustrating if you want to take a dump to find out who's using up AUX.
;-(
 
 .
 .
 JO.Skip Robinson
 SCE Infrastructure Technology Services
 Electric Dragon Team Paddler 
 SHARE MVS Program Co-Manager
 626-302-7535 Office
 323-715-0595 Mobile
 jo.skip.robin...@sce.com
 
 
 
 From:   Mark Zelden m...@mzelden.com
 To: IBM-MAIN@bama.ua.edu
 Date:   02/02/2012 07:35 PM
 Subject:Re: Very Lage Page Datasets (was ASM and HiperPAV)
 Sent by:IBM Mainframe Discussion List IBM-MAIN@bama.ua.edu
 
 
 
 On Thu, 2 Feb 2012 16:50:44 -0800, Skip Robinson 
 jo.skip.robin...@sce.com wrote:
 
 There is a much lower limit to worry about than the one that prevents
new
 works from starting. At around 50%, SVC dump will fail with 'ASM
 shortage'. This barrier has been discussed recently at SHARE. IBM
agrees
 that SVCDUMP's ASM calculation as implemented is too strict, but it
still
 carries the day. With no SVC dumps possible, many would consider a
system
 hobbled.
 
 
 Interesting.  Is there an APAR that has more detail as to when this 
 happens? 
 Not too long ago I know we were getting warning messages about an
 LPAR that had hit 50% and I'm pretty sure I saw an email from a 
 coworker that they manually took an SVCDUMP prior to adding some
 additional page volumes. 
 
 Regards,
 
 Mark
 
 
 
 --
 For IBM-MAIN subscribe / signoff / archive access instructions,
 send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN

For information, services and offers, please visit our web site: 
http://www.klm.com. This e-mail and any attachment may contain confidential and 
privileged material intended for the addressee only. If you are not the 
addressee, you are notified that no part of the e-mail or any attachment may be 
disclosed, copied or distributed, and that any other action related to this 
e-mail or attachment is strictly prohibited, and may be unlawful. If you have 
received this e-mail by error, please notify the sender immediately by return 
e-mail, and delete this message. 

Koninklijke Luchtvaart Maatschappij NV (KLM), its subsidiaries and/or its 
employees shall not be liable for the incorrect or incomplete transmission of 
this e-mail or any attachments, nor responsible for any delay in receipt. 
Koninklijke Luchtvaart Maatschappij N.V. (also known as KLM Royal Dutch 
Airlines) is registered in Amstelveen, The Netherlands, with registered number 
33014286



--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-02-03 Thread Barbara Nitz
Are you sure that this occurs at 50% Aux. storage utilization? IRA201E
defines 'Critical storage shortage' at 85%.

See cd command, parm AUXMGMT:
SDUMP,AUXMGMT=ON or OFF
Specifies when SDUMP data captures should stop.
ON: No new dumps are allowed when auxiliary storage usage reaches 50%. New 
dumps are allowed again only after the auxiliary storage usage drops below 35%. 
Current SDUMP data capture stops when auxiliary storage usage exceeds 68%, 
generating a partial dump.

Don't assume that sdump uses the same lingo as ASM :-)

Barbara

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-02-03 Thread Ron Hawkins
Jim,

Thanks. That's exactly the sort of update I needed.

Ron

 -Original Message-
 From: IBM Mainframe Discussion List [mailto:IBM-MAIN@bama.ua.edu] On
Behalf Of
 Jim Mulder
 Sent: Thursday, February 02, 2012 10:14 AM
 To: IBM-MAIN@bama.ua.edu
 Subject: Re: [IBM-MAIN] Very Lage Page Datasets (was ASM and HiperPAV)
 
  Another element of paging that has not been referenced is the ability
  to handle all of the swap set size in parallel. If the swap set size
  is 120 pages then the old practice was to have at least four LOCALS so
  each
 thirty
  page block of pages could be swapped-in in parallel. While swapping,
 like
  paging, is not as prevalent as it once was I'm wondering if the swap
  set size is still one of the principal guidelines for the number of
  locals
 that
  should be defined.
 
   Starting with z/OS 1.8, physical swapping is no longer done at all.
 Block paging has not been done for quite a while either.  There can be
some
 trimming done for address spaces when they get logically swapped, and
before
 global LRU is done. So those pages might get written to contiguous slots
to
 help the throughput of the output I/O.  But with no swapping and block
paging,
 they will come back in via individual page faults, with no relation to the
 order in which they were written, and probably as separate I/O operations.
 
   I am not trying to say that is a good thing, just saying how it works
now,
 so that you don't spend time trying to design your paging configuration to
 accommodate former behavior of the operating system.
 
 Jim Mulder   z/OS System Test   IBM Corp.  Poughkeepsie,  NY
 
 --
 For IBM-MAIN subscribe / signoff / archive access instructions, send email
to
 lists...@bama.ua.edu with the message: INFO IBM-MAIN

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-02-03 Thread Bob Shannon
 At around 50%, SVC dump will fail with 'ASM shortage'.

I think at some point, maybe 1.13, AUXMGMT=YES became the default which 
prevented taking dumps at a lower aux storage percentage. You can return to the 
previous default threshold by issuing this command:

CD SET,SDUMP,AUXMGMT=OFF

Reference OA30726

Bob Shannon
Rocket Software

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-02-03 Thread Charles Mari
Barbara,   
  
In regards to:

 'I like the ASM health check that tells us that usage is 30% or more. (In 
 fact, I send an automated exception email every time this happens.) I hate 
 that ASM does not recognize that a  new page data set was added. That health 
 check stupidly doesn't recognize a changed config and still spits out the 
 warning.'   
   
I believe you're referring to the ASM_LOCAL_SLOT_USAGE check. If so, this check 
runs at a 30 minute interval and checks the slot usage for each in-use local 
paging data set regardless of whether it was defined 'statically' or added via 
PAGE ADD. So... it does eventually recognize a changed configuration. Some of 
the ASM checks are set to re-run when a PAGE ADD/DELETE is issued, but 
ASM_LOCAL_SLOT_USAGE is currently not one of them.  Open to suggestions 
  
  
Regards,  
Charles Mari - IBM RSM/ASM Development 

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-02-03 Thread Mark Zelden
On Thu, 2 Feb 2012 20:31:12 -0800, Skip Robinson jo.skip.robin...@sce.com 
wrote:


According to conversations at SHARE, the threshold for this condition is
too conservative. AFAIK there is no APAR open to fix it. It's especially
frustrating if you want to take a dump to find out who's using up AUX. ;-(


There's always RMF III.   I've also used this REXX exec posted to IBM-MAIN
by David Barnard-Brown on 09/13/2001:

http://groups.google.com/group/bit.listserv.ibm-main/msg/33117230ea79eb5a?hl=en

--
Mark Zelden - Zelden Consulting Services - z/OS, OS/390 and MVS   
mailto:m...@mzelden.com
Mark's MVS Utilities: http://www.mzelden.com/mvsutil.html 
Systems Programming expert at http://expertanswercenter.techtarget.com/

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-02-03 Thread Mark Zelden
On Thu, 2 Feb 2012 23:37:06 -0600, Barbara Nitz nitz-...@gmx.net wrote:

Mark, this came into z/OS with either 1.11 or 1.12. Some of my pagedels were an
 attempt to test this behaviour. I could not test it.

I saw it on this list of changed messages on z/OS 1.11.   


Skip, try ipcs active, and then the undocumented command (IP) ILRSLOTC.
 This used to be in the hidden IPCS panel that I have forgotten how to 
 get at (there were SHARE presentations about it).

Option 2.6i


But I've been stuck on 1.11 since 2010 do to moving my client to 2 new
data centers in 2011 and I've never seen that message and I know we
have had systems over 50% aux usage and were able to take dumps.  

I believe what's being said here, but on face value critical is after
message IRA201E (85%).  Or perhaps IRA200E (70%).What seems
wrong is that message IEE711I says it won't allow dumps again
until you get under 35%.  

Regards,

Mark
--
Mark Zelden - Zelden Consulting Services - z/OS, OS/390 and MVS   
mailto:m...@mzelden.com
Mark's MVS Utilities: http://www.mzelden.com/mvsutil.html 
Systems Programming expert at http://expertanswercenter.techtarget.com/

--
Mark Zelden - Zelden Consulting Services - z/OS, OS/390 and MVS   
mailto:m...@mzelden.com
Mark's MVS Utilities: http://www.mzelden.com/mvsutil.html 
Systems Programming expert at http://expertanswercenter.techtarget.com/

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-02-03 Thread Mary Anne Matyaz
Barbara wrote: Skip, try ipcs active, and then the undocumented command (IP) 
ILRSLOTC. This used to be in the hidden IPCS panel that I have forgotten how to 
get at


2.6i is the hidden IPCS panel. 

MA

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-02-03 Thread Skip Robinson
Charles,

My suggestion would be to add PA/PD to the set of conditions that redrive 
the ASM_LOCAL_SLOT_USAGE check. I know that I've occasionally added space 
in an attempt relieve a shortage and been dismayed that my effort *at the 
time* appeared not to have helped. 

BTW if you have not already seen it, please look at our PMR 81276.227.000 
for an agonizing saga of RSM/ASM misadventures. 

.
.
JO.Skip Robinson
SCE Infrastructure Technology Services
Electric Dragon Team Paddler 
SHARE MVS Program Co-Manager
626-302-7535 Office
323-715-0595 Mobile
jo.skip.robin...@sce.com



From:   Charles Mari cm...@us.ibm.com
To: IBM-MAIN@bama.ua.edu
Date:   02/03/2012 05:46 AM
Subject:Re: Very Lage Page Datasets (was ASM and HiperPAV)
Sent by:IBM Mainframe Discussion List IBM-MAIN@bama.ua.edu



Barbara, 
 
In regards to:

 'I like the ASM health check that tells us that usage is 30% or more. 
(In fact, I send an automated exception email every time this happens.) I 
hate that ASM does not recognize that a  new page data set was added. That 
health check stupidly doesn't recognize a changed config and still spits 
out the warning.' 
 
I believe you're referring to the ASM_LOCAL_SLOT_USAGE check. If so, this 
check runs at a 30 minute interval and checks the slot usage for each 
in-use local paging data set regardless of whether it was defined 
'statically' or added via PAGE ADD. So... it does eventually recognize a 
changed configuration. Some of the ASM checks are set to re-run when a 
PAGE ADD/DELETE is issued, but ASM_LOCAL_SLOT_USAGE is currently not one 
of them.  Open to suggestions 
 
Regards, 
Charles Mari - IBM RSM/ASM Development 


--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-02-03 Thread Vernooij, CP - SPLXM
I think the health check's way of working, to check each page dataset
for its 30%, in stead of the entire page dataset configuration, is not
such a bad one. AFAIK and some earlier posts in this thread suggested it
too: ASM has one or more groups of page datasets, depending on
devicetype, channel configuration etc. If you have a well balanced page
configuration, there will be 1 group. It selects one group and
circulates through it to select a page dataset. This means that if some
are 30% full, others are 30%, all have an equal chance of being
selected. So with this mechanisme in mind, it is worth while signalling
than 1 page dataset is 30%, because this can cause AMS inefficiency in
spite of the filling of the others.

(Unless I am mistaken or outdated on ASM's algorithmes).

Kees.


Charles Mari cm...@us.ibm.com wrote in message
news:9695857008538097.wa.cmarius.ibm@bama.ua.edu...
 Barbara,   
   
 In regards to:
 
  'I like the ASM health check that tells us that usage is 30% or
more. (In fact, I send an automated exception email every time this
happens.) I hate that ASM does not recognize that a  new page data set
was added. That health check stupidly doesn't recognize a changed config
and still spits out the warning.'   

 I believe you're referring to the ASM_LOCAL_SLOT_USAGE check. If so,
this check runs at a 30 minute interval and checks the slot usage for
each in-use local paging data set regardless of whether it was defined
'statically' or added via PAGE ADD. So... it does eventually recognize a
changed configuration. Some of the ASM checks are set to re-run when a
PAGE ADD/DELETE is issued, but ASM_LOCAL_SLOT_USAGE is currently not one
of them.  Open to suggestions   
   
 Regards,  
 Charles Mari - IBM RSM/ASM Development 
 
 --
 For IBM-MAIN subscribe / signoff / archive access instructions,
 send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN

For information, services and offers, please visit our web site: 
http://www.klm.com. This e-mail and any attachment may contain confidential and 
privileged material intended for the addressee only. If you are not the 
addressee, you are notified that no part of the e-mail or any attachment may be 
disclosed, copied or distributed, and that any other action related to this 
e-mail or attachment is strictly prohibited, and may be unlawful. If you have 
received this e-mail by error, please notify the sender immediately by return 
e-mail, and delete this message. 

Koninklijke Luchtvaart Maatschappij NV (KLM), its subsidiaries and/or its 
employees shall not be liable for the incorrect or incomplete transmission of 
this e-mail or any attachments, nor responsible for any delay in receipt. 
Koninklijke Luchtvaart Maatschappij N.V. (also known as KLM Royal Dutch 
Airlines) is registered in Amstelveen, The Netherlands, with registered number 
33014286



--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-02-02 Thread Mark Zelden
On Wed, 1 Feb 2012 23:12:48 -0600, Barbara Nitz nitz-...@gmx.net wrote:

So if I have 5 3390-27 locals and they are all equally used at 50%, the
algorithms (CPU usage, not I/O) are going to pick one of them, then do the
page outs.  That paging will find contiguous slots and should be efficient.

BTW, this is just an example, we still try to keep our 3390-27 local usage
at 30% just like we always did with smaller local page datasets in
the past.

I wonder what if any studies on this have been done in the lab.
It would be nice if an IBM performance expert  like Kathy Walsh
could weigh in.

I had the 'honour' of deleting and adding several local page data sets on 
several lpars. They were a mixture of 1.10 and 1.12, I think. What I did 
observe (and that clashed with what I thought I knew about ASM) is the 
following:

1) Adding one or more locals, I expected them to first fill up to about the 
same percentage as the ones that were already in use (same size page ds, much 
faster -new- controller). No such luck. It looked to me like *all* of them 
were filling up (percentage wise) in about the same manner. Meaning that the 
'old' locals had about 27%, the new ones right after add 0%. A day later the 
old ones had 35%, the new ones 8%. About the same behaviour when adding locals 
of the same size on the same controller - we only have one DASD controller per 
sysplex, and having two was the time when we migrated from one to the other.


Sounds like they all had contiguous slots and were used evenly based on that.  

2) A pagedel finishes MUCH faster than it ever did. It looked like ASM is 
actively shifting slots away from the to-be-deleted page data set. A pagedel 
finishes in under a minute. This used to be a really slow process because 
nothing was actively done.

3) In one case, I had two locals and pagedel'd one of them. Usage of the 
remaining one went up to 46%, pagedel was extremely fast. I then added a new 
local (on a new, different, much faster controller). Usage of that one stayed 
at 0% for a long time, which also surprised me.

*WARNING*If you are going to be replacing page datasets, better to use the 
REPLACE
option of PAGEDEL as opposed to DELETE then doing a PAGEADD.  Not doing so
can lead to ESQA shortages.  See the MVS commands manual for more detail. 

snip



6) I bemoan IBMs failure to give us a good means of figuring out who is using 
HVSHARE or HVCOMMON storage and how much storage-above-the-bar is actually 
*used*, i.e. backed. As far as I know, there still isn't any tracking done for 
HVCOMMON storage, no means of reporting about it. No way to know who is 
excessively using common storage above the bar. Same for HVSHARE. Unless 
you're named Jim Mulder and know where to look in a dump, us lowly customers 
cannot even check that in a dump. Am I mistaken in the reporting capabilities? 
Has that been fixed by now? Or is it another means of IBM trying to sell  
software service contracts to get that done only by IBM? Not to mention the 
frustration until you find someone who can actually *do* it.


Not sure how much info you are looking for, but there is RMF III and my RXSTOR64
exec on my web site / CBT file 434.   If you want to see what the sample output
looks like, here is a screen shot - http://www.mzelden.com/rxstor64.html

Regards,

Mark
--
Mark Zelden - Zelden Consulting Services - z/OS, OS/390 and MVS   
mailto:m...@mzelden.com
Mark's MVS Utilities: http://www.mzelden.com/mvsutil.html 
Systems Programming expert at http://expertanswercenter.techtarget.com/

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-02-02 Thread Jim Mulder
 Another element of paging that has not been referenced is the ability to
 handle all of the swap set size in parallel. If the swap set size is 120
 pages then the old practice was to have at least four LOCALS so each 
thirty
 page block of pages could be swapped-in in parallel. While swapping, 
like
 paging, is not as prevalent as it once was I'm wondering if the swap set
 size is still one of the principal guidelines for the number of locals 
that
 should be defined.

  Starting with z/OS 1.8, physical swapping is no longer done at all.
Block paging has not been done for quite a while either.  There can
be some trimming done for address spaces when they get logically
swapped, and before global LRU is done. So those pages
might get written to contiguous slots to help the throughput of the 
output I/O.  But with no swapping and block paging, they will come back
in via individual page faults, with no relation to the order in which
they were written, and probably as separate I/O operations.

  I am not trying to say that is a good thing, just saying how it 
works now, so that you don't spend time trying to design your paging
configuration to accommodate former behavior of the operating system. 

Jim Mulder   z/OS System Test   IBM Corp.  Poughkeepsie,  NY

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-02-02 Thread Jim Mulder
 I wonder what if any studies on this have been done in the lab.
 It would be nice if an IBM performance expert  like Kathy Walsh
 could weigh in.

  The last performance studies I remember for paging were 
around the time of MVS XA/SP2.2.0.  Very little has been done
in the area of paging performance since then except for the 
PAV stuff to allow two concurrent operations to a page data set. 

 
 I had the 'honour' of deleting and adding several local page data 
 sets on several lpars. They were a mixture of 1.10 and 1.12, I 
 think. What I did observe (and that clashed with what I thought I 
 knew about ASM) is the following:
 
 1) Adding one or more locals, I expected them to first fill up to 
 about the same percentage as the ones that were already in use (same
 size page ds, much faster -new- controller). No such luck. It looked
 to me like *all* of them were filling up (percentage wise) in about 
 the same manner. Meaning that the 'old' locals had about 27%, the 
 new ones right after add 0%. A day later the old ones had 35%, the 
 new ones 8%. About the same behaviour when adding locals of the same
 size on the same controller - we only have one DASD controller per 
 sysplex, and having two was the time when we migrated from one to the 
other.

  The page data set selection algorithm considers service time for
the devices, but not the full percentage.  One could argue that 
the full percentage should be considered, since it affects the 
likihood of finding contiguous slots, and the CPU time to find
available slots, but that is not how it currently works. 
 
 2) A pagedel finishes MUCH faster than it ever did. It looked like 
 ASM is actively shifting slots away from the to-be-deleted page data
 set. A pagedel finishes in under a minute. This used to be a really 
 slow process because nothing was actively done.

  Pagedel has always done active removal.  There were some 
problems with doing active removal of VIO in the original SP3.1.0
implementation, but that was fixed in SP3.1.3.
 
 6) I bemoan IBMs failure to give us a good means of figuring out who
 is using HVSHARE or HVCOMMON storage and how much storage-above-the-
 bar is actually *used*, i.e. backed. As far as I know, there still 
 isn't any tracking done for HVCOMMON storage, no means of reporting 
 about it. No way to know who is excessively using common storage 
 above the bar. Same for HVSHARE. Unless you're named Jim Mulder and 
 know where to look in a dump, us lowly customers cannot even check 
 that in a dump. Am I mistaken in the reporting capabilities? Has 
 that been fixed by now? Or is it another means of IBM trying to sell
  software service contracts to get that done only by IBM? Not to
 mention the frustration until you find someone who can actually *do* it.

 IPCS has  RSMDATA HVCOMMON.  That at least tells you the owner of
the memory objects. 

  For HVCOMMON which is obtained via the IARST64 macro,
the IARST64 macro says:

There  is  diagnostic support for 64 bit cell pools, created
by  IARST64,  in IPCS via the CBFORMAT command.  In order to
locate  the  cell  pool  of interest, you need to follow the
pointers from HP1, to HP2, to the CPHD.  For common storage,
the  HP1  is  located in the ECVT.  CBF ECVT will format the
ECVT, then do a FIND on HP1.  Extract the address of the HP1
from the ECVT and 
CBF addrhp1 STR(HP1) 
will  format  the HP1.   Each entry in the HP1 represents an
attribute  set  (storage  key, storage type (pageable, DREF,
FIXED),  and  Fetch  Protection (ON or OFF)) The output from
this  command  will  contain  CBF commands for any connected
HP2s.Select  the  CBF  command of interest and run it to
format  the HP2.   The HP2 consists of pointers to cell pool
headers  for  different sizes.   Choose the size of interest
and select the command that will look like this: 
CBF addrcphd STR(IAXCPHD). 
This will format the cell pool header.  To see details about
all  of  the  cells  in  the  pool,  use  the EXIT option as
follows: 
CBF addrcphd STR(IAXCPHD) EXIT. 
For  private  storage, the HP1 is anchored in the STCB.  The
quickest  way to locate the HP1 is to run the SUMMARY FORMAT
command  for  the address space of interest.  Locate the TCB
that  owns  the  storage of interest and then scroll down to
the  formatted STCB.   The HP1 field contains the address of
the HP1.  From here, the processing is the same as described
for common storage above. 
You can also use the EXIT option as follows: 
CBF addrhp1 STR(HP1) EXIT 
to  produce a report that summarizes the storage usage under
that HP1. 

 
Jim Mulder   z/OS System Test   IBM Corp.  Poughkeepsie,  NY

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-02-02 Thread Vernooij, CP - SPLXM
Jim Mulder d10j...@us.ibm.com wrote in message
news:of15dad2d4.112617be-on85257998.0064c1d8-85257998.0066c...@us.ibm.c
om...
  I wonder what if any studies on this have been done in the lab.
  It would be nice if an IBM performance expert  like Kathy Walsh
  could weigh in.
 
   The last performance studies I remember for paging were 
 around the time of MVS XA/SP2.2.0.  Very little has been done
 in the area of paging performance since then except for the 
 PAV stuff to allow two concurrent operations to a page data set. 
 
  

This almost 'requires' a new study, especially since those ancient
studies require/recommend us to leave 70+% of a 3390-27 (17GB per page
dataset) free for some algorithme that required it decades ago for
devices from decades ago.

I still am under the impression that leaving 70% of a 3390-3 free will
equal leaving them same 2GB free on a 3390-27. Well ok, to avoid the 70%
full messages, I am willing to leave 7.5GB free, but I would like to see
studies that proves it benificial to leave the other 10GB free.

Thanks Jim, for letting the bull out of the barn... Someone has to get
it in again.

Kees.

For information, services and offers, please visit our web site: 
http://www.klm.com. This e-mail and any attachment may contain confidential and 
privileged material intended for the addressee only. If you are not the 
addressee, you are notified that no part of the e-mail or any attachment may be 
disclosed, copied or distributed, and that any other action related to this 
e-mail or attachment is strictly prohibited, and may be unlawful. If you have 
received this e-mail by error, please notify the sender immediately by return 
e-mail, and delete this message. 

Koninklijke Luchtvaart Maatschappij NV (KLM), its subsidiaries and/or its 
employees shall not be liable for the incorrect or incomplete transmission of 
this e-mail or any attachments, nor responsible for any delay in receipt. 
Koninklijke Luchtvaart Maatschappij N.V. (also known as KLM Royal Dutch 
Airlines) is registered in Amstelveen, The Netherlands, with registered number 
33014286



--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-02-02 Thread Mark Zelden
On Thu, 2 Feb 2012 13:42:45 -0500, Jim Mulder d10j...@us.ibm.com wrote:

 I wonder what if any studies on this have been done in the lab.
 It would be nice if an IBM performance expert  like Kathy Walsh
 could weigh in.


Thanks for jumping in Jim!   I hoped you or Peter would eventually
to clear up some of this FUD. 


  The last performance studies I remember for paging were
around the time of MVS XA/SP2.2.0.  Very little has been done
in the area of paging performance since then except for the
PAV stuff to allow two concurrent operations to a page data set.



Even though zero demand paging is best and what many shops
strive / configure for these days, I think it's time for a new study
based on modern architecture and what the OS now does. 
 

  The page data set selection algorithm considers service time for
the devices, but not the full percentage.  One could argue that
the full percentage should be considered, since it affects the
likihood of finding contiguous slots, and the CPU time to find
available slots, but that is not how it currently works.


I'm going to take a giant leap of faith here and say there is zero to
negligible performance impact in having a farm of 3390-27 local
page datasets running at 50% full compared to having them
at 30% (or less).   Maybe 69% is fine too.  No higher - we certainly
don't want to hit that 70% MCCASMT1 threshold!  :-)


Regards,

Mark
--
Mark Zelden - Zelden Consulting Services - z/OS, OS/390 and MVS   
mailto:m...@mzelden.com
Mark's MVS Utilities: http://www.mzelden.com/mvsutil.html 
Systems Programming expert at http://expertanswercenter.techtarget.com/

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-02-02 Thread Martin Packer
I liked Barry's plot the curve approach to this i.e. Relating occupation 
to performance.

Cheers, Martin

Martin Packer,
Mainframe Performance Consultant, zChampion
Worldwide Banking Center of Excellence, IBM

+44-7802-245-584

email: martin_pac...@uk.ibm.com

Twitter / Facebook IDs: MartinPacker
Blog: 
https://www.ibm.com/developerworks/mydeveloperworks/blogs/MartinPacker





Unless stated otherwise above:
IBM United Kingdom Limited - Registered in England and Wales with number 
741598. 
Registered office: PO Box 41, North Harbour, Portsmouth, Hampshire PO6 3AU






--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-02-02 Thread Mark Zelden
On Thu, 2 Feb 2012 21:48:17 +, Martin Packer martin_pac...@uk.ibm.com 
wrote:

I liked Barry's plot the curve approach to this i.e. Relating occupation
to performance.


What is the SSCH count in VMAC74?  SIO74CNT ? 

With a static environment for the most part (except after IPL when all the
subsystems are coming active), and enough real storage to (usually) have a
demand paging rate of zero, and page data sets =30% full, how will I
find where the knee is located?

A lab environment is needed.   I guess I could do this in a sandbox
LPAR, but I'm not *that* interested.   

Mark
--
Mark Zelden - Zelden Consulting Services - z/OS, OS/390 and MVS   
mailto:m...@mzelden.com
Mark's MVS Utilities: http://www.mzelden.com/mvsutil.html 
Systems Programming expert at http://expertanswercenter.techtarget.com/

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-02-02 Thread Skip Robinson
There is a much lower limit to worry about than the one that prevents new 
works from starting. At around 50%, SVC dump will fail with 'ASM 
shortage'. This barrier has been discussed recently at SHARE. IBM agrees 
that SVCDUMP's ASM calculation as implemented is too strict, but it still 
carries the day. With no SVC dumps possible, many would consider a system 
hobbled. 

.
.
JO.Skip Robinson
SCE Infrastructure Technology Services
Electric Dragon Team Paddler 
SHARE MVS Program Co-Manager
626-302-7535 Office
323-715-0595 Mobile
jo.skip.robin...@sce.com



From:   Mark Zelden m...@mzelden.com
To: IBM-MAIN@bama.ua.edu
Date:   02/02/2012 01:02 PM
Subject:Re: Very Lage Page Datasets (was ASM and HiperPAV)
Sent by:IBM Mainframe Discussion List IBM-MAIN@bama.ua.edu



On Thu, 2 Feb 2012 13:42:45 -0500, Jim Mulder d10j...@us.ibm.com wrote:

 I wonder what if any studies on this have been done in the lab.
 It would be nice if an IBM performance expert  like Kathy Walsh
 could weigh in.


Thanks for jumping in Jim!   I hoped you or Peter would eventually
to clear up some of this FUD. 


  The last performance studies I remember for paging were
around the time of MVS XA/SP2.2.0.  Very little has been done
in the area of paging performance since then except for the
PAV stuff to allow two concurrent operations to a page data set.



Even though zero demand paging is best and what many shops
strive / configure for these days, I think it's time for a new study
based on modern architecture and what the OS now does. 
 

  The page data set selection algorithm considers service time for
the devices, but not the full percentage.  One could argue that
the full percentage should be considered, since it affects the
likihood of finding contiguous slots, and the CPU time to find
available slots, but that is not how it currently works.


I'm going to take a giant leap of faith here and say there is zero to
negligible performance impact in having a farm of 3390-27 local
page datasets running at 50% full compared to having them
at 30% (or less).   Maybe 69% is fine too.  No higher - we certainly
don't want to hit that 70% MCCASMT1 threshold!  :-)


Regards,

Mark



--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-02-02 Thread Mark Zelden
On Thu, 2 Feb 2012 16:50:44 -0800, Skip Robinson jo.skip.robin...@sce.com 
wrote:

There is a much lower limit to worry about than the one that prevents new
works from starting. At around 50%, SVC dump will fail with 'ASM
shortage'. This barrier has been discussed recently at SHARE. IBM agrees
that SVCDUMP's ASM calculation as implemented is too strict, but it still
carries the day. With no SVC dumps possible, many would consider a system
hobbled.


Interesting.  Is there an APAR that has more detail as to when this happens? 
Not too long ago I know we were getting warning messages about an
LPAR that had hit 50% and I'm pretty sure I saw an email from a 
coworker that they manually took an SVCDUMP prior to adding some
additional page volumes. 

Regards,

Mark
--
Mark Zelden - Zelden Consulting Services - z/OS, OS/390 and MVS   
mailto:m...@mzelden.com
Mark's MVS Utilities: http://www.mzelden.com/mvsutil.html 
Systems Programming expert at http://expertanswercenter.techtarget.com/

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-02-02 Thread Skip Robinson
IEE711I SYSTEM DUMP NOT TAKEN. A CRITICAL AUXILIARY STORAGE SHORTAGE 
EXISTS 

From the message manual:

If the reason field shows A CRITICAL AUXILIARY STORAGE SHORTAGE EXISTS, 
first you need to ensure that enough DASD resource is available for 
captured dumps to be written out. Then, consider adding additional 
auxiliary storage (paging) resources, because SVC dumps will not be 
allowed again until the auxiliary storage utilization drops below 35%. See 
the system programmer response for message IRA201E for additional 
information about auxiliary storage utilization concerns.

According to conversations at SHARE, the threshold for this condition is 
too conservative. AFAIK there is no APAR open to fix it. It's especially 
frustrating if you want to take a dump to find out who's using up AUX. ;-(

.
.
JO.Skip Robinson
SCE Infrastructure Technology Services
Electric Dragon Team Paddler 
SHARE MVS Program Co-Manager
626-302-7535 Office
323-715-0595 Mobile
jo.skip.robin...@sce.com



From:   Mark Zelden m...@mzelden.com
To: IBM-MAIN@bama.ua.edu
Date:   02/02/2012 07:35 PM
Subject:Re: Very Lage Page Datasets (was ASM and HiperPAV)
Sent by:IBM Mainframe Discussion List IBM-MAIN@bama.ua.edu



On Thu, 2 Feb 2012 16:50:44 -0800, Skip Robinson 
jo.skip.robin...@sce.com wrote:

There is a much lower limit to worry about than the one that prevents new
works from starting. At around 50%, SVC dump will fail with 'ASM
shortage'. This barrier has been discussed recently at SHARE. IBM agrees
that SVCDUMP's ASM calculation as implemented is too strict, but it still
carries the day. With no SVC dumps possible, many would consider a system
hobbled.


Interesting.  Is there an APAR that has more detail as to when this 
happens? 
Not too long ago I know we were getting warning messages about an
LPAR that had hit 50% and I'm pretty sure I saw an email from a 
coworker that they manually took an SVCDUMP prior to adding some
additional page volumes. 

Regards,

Mark



--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-02-02 Thread Barbara Nitz
If the reason field shows A CRITICAL AUXILIARY STORAGE SHORTAGE EXISTS,
first you need to ensure that enough DASD resource is available for
captured dumps to be written out. Then, consider adding additional
auxiliary storage (paging) resources, because SVC dumps will not be
allowed again until the auxiliary storage utilization drops below 35%. See
the system programmer response for message IRA201E for additional
information about auxiliary storage utilization concerns.

According to conversations at SHARE, the threshold for this condition is
too conservative. AFAIK there is no APAR open to fix it. It's especially
frustrating if you want to take a dump to find out who's using up AUX. ;-(

Mark, this came into z/OS with either 1.11 or 1.12. Some of my pagedels were an 
attempt to test this behaviour. I could not test it.

Skip, try ipcs active, and then the undocumented command (IP) ILRSLOTC. This 
used to be in the hidden IPCS panel that I have forgotten how to get at (there 
were SHARE presentations about it). This command works on active storage. No 
need for a dump in an AUX shortage to determine who holds the slots:
VERSION 02/18/2010  
ASID=00A4 JOB=DSW1DBM1 SLOTCOUNT=99EB VIO COUNT=
ASID=00A3 JOB=DSNXDBM1 SLOTCOUNT=8B0F VIO COUNT=
ASID=002E JOB=ZFS  SLOTCOUNT=2F26 VIO COUNT=

Jim, we talked about the need for an HVCOMMON storage tracking tool before. In 
a pinch, someone with a lot of knowledge of IPCS could probably find out who 
did what. I seem to remember you agreeing that the procedure you summarized 
neatly in your post :-) is not something one wants to do manually. Has anyone 
submitted a requirement for this on my behalf? I am still willing to have the 
requirement submitted, and I think that some regulars on ibmmain would agree 
that it were a good thing! Some might even concur with such a requirement.

*WARNING*If you are going to be replacing page datasets, better to
 use the REPLACE option of PAGEDEL as opposed to DELETE then doing
 a PAGEADD.  Not doing so can lead to ESQA shortages.
Thanks for the warning. Unfortunately, my spaceadmin needed to reuse the name 
of the page data set, so there was no way I could use REPLACE. I'll keep that 
in mind for our pageds redesign, though.

  Starting with z/OS 1.8, physical swapping is no longer done at all.
Block paging has not been done for quite a while either.  There can
be some trimming done for address spaces when they get logically
swapped, and before global LRU is done. So those pages
might get written to contiguous slots to help the throughput of the
output I/O.  But with no swapping and block paging, they will come back
in via individual page faults, with no relation to the order in which
they were written, and probably as separate I/O operations.
That explains why (MXG) fields BLKPAGE  BLKSAUIN PGBKAUIN are still filled. 
I'll ignore this for my colourful pictures, then!

  Pagedel has always done active removal.  There were some
problems with doing active removal of VIO in the original SP3.1.0
implementation, but that was fixed in SP3.1.3.
This is interesting, given that I wasn't around for SP3. I just remember that 
it used to take hours to empty a page data set, SP4 right up until z/OS 1.4 (I 
think might have been the last time I tried).

Even though zero demand paging is best and what many shops
strive / configure for these days, I think it's time for a new study
based on modern architecture and what the OS now does.
I agree. Unfortunately, we have lots of demand paging (we are a small 
installation, after all), and all the newfangled applications are just 
storage-hungry in the typical clicker way of what's a few Terabyte more. 
Maybe then ASM would better use available space and better thresholds. I guess 
I'll need to bite the bullit and try to find the knee.

Barbara

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-02-02 Thread Skip Robinson
Wow. Very interesting. PFA was just complaining about super hog address 
space DA20DBM1. Here's what ILRSLOTC shows as the top piggy:

VERSION 02/18/2010 
ASID=0079 JOB=DA20DBM1 SLOTCOUNT=00029E21 VIO COUNT=

I love it when a picture comes together. 

.
.
JO.Skip Robinson
SCE Infrastructure Technology Services
Electric Dragon Team Paddler 
SHARE MVS Program Co-Manager
626-302-7535 Office
323-715-0595 Mobile
jo.skip.robin...@sce.com



From:   Barbara Nitz nitz-...@gmx.net
To: IBM-MAIN@bama.ua.edu
Date:   02/02/2012 09:47 PM
Subject:Re: Very Lage Page Datasets (was ASM and HiperPAV)
Sent by:IBM Mainframe Discussion List IBM-MAIN@bama.ua.edu



If the reason field shows A CRITICAL AUXILIARY STORAGE SHORTAGE EXISTS,
first you need to ensure that enough DASD resource is available for
captured dumps to be written out. Then, consider adding additional
auxiliary storage (paging) resources, because SVC dumps will not be
allowed again until the auxiliary storage utilization drops below 35%. 
See
the system programmer response for message IRA201E for additional
information about auxiliary storage utilization concerns.

According to conversations at SHARE, the threshold for this condition is
too conservative. AFAIK there is no APAR open to fix it. It's especially
frustrating if you want to take a dump to find out who's using up AUX. 
;-(

Mark, this came into z/OS with either 1.11 or 1.12. Some of my pagedels 
were an attempt to test this behaviour. I could not test it.

Skip, try ipcs active, and then the undocumented command (IP) ILRSLOTC. 
This used to be in the hidden IPCS panel that I have forgotten how to get 
at (there were SHARE presentations about it). This command works on active 
storage. No need for a dump in an AUX shortage to determine who holds the 
slots:
VERSION 02/18/2010 
ASID=00A4 JOB=DSW1DBM1 SLOTCOUNT=99EB VIO COUNT=
ASID=00A3 JOB=DSNXDBM1 SLOTCOUNT=8B0F VIO COUNT=
ASID=002E JOB=ZFS  SLOTCOUNT=2F26 VIO COUNT=

Jim, we talked about the need for an HVCOMMON storage tracking tool 
before. In a pinch, someone with a lot of knowledge of IPCS could probably 
find out who did what. I seem to remember you agreeing that the procedure 
you summarized neatly in your post :-) is not something one wants to do 
manually. Has anyone submitted a requirement for this on my behalf? I am 
still willing to have the requirement submitted, and I think that some 
regulars on ibmmain would agree that it were a good thing! Some might even 
concur with such a requirement.

*WARNING*If you are going to be replacing page datasets, better to
 use the REPLACE option of PAGEDEL as opposed to DELETE then doing
 a PAGEADD.  Not doing so can lead to ESQA shortages.
Thanks for the warning. Unfortunately, my spaceadmin needed to reuse the 
name of the page data set, so there was no way I could use REPLACE. I'll 
keep that in mind for our pageds redesign, though.

  Starting with z/OS 1.8, physical swapping is no longer done at all.
Block paging has not been done for quite a while either.  There can
be some trimming done for address spaces when they get logically
swapped, and before global LRU is done. So those pages
might get written to contiguous slots to help the throughput of the
output I/O.  But with no swapping and block paging, they will come back
in via individual page faults, with no relation to the order in which
they were written, and probably as separate I/O operations.
That explains why (MXG) fields BLKPAGE  BLKSAUIN PGBKAUIN are still 
filled. I'll ignore this for my colourful pictures, then!

  Pagedel has always done active removal.  There were some
problems with doing active removal of VIO in the original SP3.1.0
implementation, but that was fixed in SP3.1.3.
This is interesting, given that I wasn't around for SP3. I just remember 
that it used to take hours to empty a page data set, SP4 right up until 
z/OS 1.4 (I think might have been the last time I tried).

Even though zero demand paging is best and what many shops
strive / configure for these days, I think it's time for a new study
based on modern architecture and what the OS now does.
I agree. Unfortunately, we have lots of demand paging (we are a small 
installation, after all), and all the newfangled applications are just 
storage-hungry in the typical clicker way of what's a few Terabyte more. 
Maybe then ASM would better use available space and better thresholds. I 
guess I'll need to bite the bullit and try to find the knee.

Barbara


--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-02-02 Thread Barbara Nitz
Wow. Very interesting. PFA was just complaining about super hog address
space DA20DBM1. Here's what ILRSLOTC shows as the top piggy:

VERSION 02/18/2010
ASID=0079 JOB=DA20DBM1 SLOTCOUNT=00029E21 VIO COUNT=

I love it when a picture comes together.

Yeah, and the DB admins will tell you that this is normal. Happens here all the 
time. On the pretext of 'someone else is the hog and we are just the innocent 
bystanders. We *need* this'.

Guess why I complain about the code that just gets a HUGE chunk of storage, 
touches every page (causing it to get backed) and then to become a slot in some 
page dataset, probably never to get touched again. :-(
And I haven't really found a way to find those that hog storage above the bar 
and never really use it. 

Barbara

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-02-02 Thread Mike Schwab
How about a new keyword DUMP=dsn,dsn similar to LOCAL=dsn,dsn for a
Local paging dsn used only for dumps?

On Thu, Feb 2, 2012 at 10:31 PM, Skip Robinson jo.skip.robin...@sce.com wrote:
 IEE711I SYSTEM DUMP NOT TAKEN. A CRITICAL AUXILIARY STORAGE SHORTAGE
 EXISTS

 From the message manual:

 If the reason field shows A CRITICAL AUXILIARY STORAGE SHORTAGE EXISTS,
 first you need to ensure that enough DASD resource is available for
 captured dumps to be written out. Then, consider adding additional
 auxiliary storage (paging) resources, because SVC dumps will not be
 allowed again until the auxiliary storage utilization drops below 35%. See
 the system programmer response for message IRA201E for additional
 information about auxiliary storage utilization concerns.

 According to conversations at SHARE, the threshold for this condition is
 too conservative. AFAIK there is no APAR open to fix it. It's especially
 frustrating if you want to take a dump to find out who's using up AUX. ;-(
-- 
Mike A Schwab, Springfield IL USA
Where do Forest Rangers go to get away from it all?

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-02-01 Thread Mark Zelden
On Tue, 31 Jan 2012 22:49:53 -0600, Barbara Nitz nitz-...@gmx.net wrote:

Writing to contiguous slots and over allocation is mentioned, but unless I
missed it the old ROT (and health check) of not having more than 30%
of the slots allocated is not specifically addressed.   Certainly with 4K
pages (for the most part) and 3390-27 (or bigger) that 30% ROT doesn't
apply anymore?50% of a mod-27 is still a helava lot of free slots.

I think it still applies. My understanding has always been that the 30% usage 
(after which paging effectiveness drastically drops) applies to the algorithm 
used on the in-storage control blocks to pick the next free slot in a page 
data set. Unless that algorithm was redesigned, 30% of 44.9GB per page dataset 
is what you should not exceed (just as the health check says) in AUX usage. 
Redesign of that is IMHO unlikely, just as using more than 2 IOs on a page 
data set simultaneously would require (an unlikely) redesign.


That sounds right as far as the algorithm, but I thought the paging 
effectiveness 
was related to likelihood of  not being able to find contiguous slots for group
page outs after the 30% usage (based on old technology).  
  
So if I have 5 3390-27 locals and they are all equally used at 50%, the
algorithms (CPU usage, not I/O) are going to pick one of them, then do the
page outs.  That paging will find contiguous slots and should be efficient. 

BTW, this is just an example, we still try to keep our 3390-27 local usage
at 30% just like we always did with smaller local page datasets in 
the past.  

I wonder what if any studies on this have been done in the lab. 
It would be nice if an IBM performance expert  like Kathy Walsh
could weigh in.  

Regards,

Mark
--
Mark Zelden - Zelden Consulting Services - z/OS, OS/390 and MVS   
mailto:m...@mzelden.com
Mark's MVS Utilities: http://www.mzelden.com/mvsutil.html 
Systems Programming expert at http://expertanswercenter.techtarget.com/

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-02-01 Thread Martin Packer
I'm sure I don't count as like Kathy* :-) but...

Contiguous Slot Allocation Algorithm is still in place so I also tout the 
30% but...

1) I say this is not a falling off a cliff thing but hazard it to be 
more gradual than that - so a dynamic.
2) I also suggest people aim for free paging space generally 1.5x the 
LPAR's memory.

Item 2 is a ROT I made up myself. :-) It's motivated by the need to have 
something to dump into - and it leans in the same direction as the Cont 
Slot Alloc Algorithm. I'm not sure the 1.5x number is right...

... I consider both the 30% and my 1.5x as STARTING POINTS. And I 
emphasise good paging subsystem design and adequate memory provision - 
even now.

So I'm really glad we're having this conversation.

Martin

Martin Packer,
Mainframe Performance Consultant, zChampion
Worldwide Banking Center of Excellence, IBM

+44-7802-245-584

email: martin_pac...@uk.ibm.com

Twitter / Facebook IDs: MartinPacker
Blog: 
https://www.ibm.com/developerworks/mydeveloperworks/blogs/MartinPacker

* Kathy's the real expert to defer to



From:
Mark Zelden m...@mzelden.com
To:
IBM-MAIN@bama.ua.edu, 
Date:
01/02/2012 15:48
Subject:
Re: Very Lage Page Datasets (was ASM and HiperPAV)
Sent by:
IBM Mainframe Discussion List IBM-MAIN@bama.ua.edu



On Tue, 31 Jan 2012 22:49:53 -0600, Barbara Nitz nitz-...@gmx.net wrote:

Writing to contiguous slots and over allocation is mentioned, but unless 
I
missed it the old ROT (and health check) of not having more than 30%
of the slots allocated is not specifically addressed.   Certainly with 
4K
pages (for the most part) and 3390-27 (or bigger) that 30% ROT doesn't
apply anymore?50% of a mod-27 is still a helava lot of free slots.

I think it still applies. My understanding has always been that the 30% 
usage (after which paging effectiveness drastically drops) applies to the 
algorithm used on the in-storage control blocks to pick the next free slot 
in a page data set. Unless that algorithm was redesigned, 30% of 44.9GB 
per page dataset is what you should not exceed (just as the health check 
says) in AUX usage. Redesign of that is IMHO unlikely, just as using more 
than 2 IOs on a page data set simultaneously would require (an unlikely) 
redesign.


That sounds right as far as the algorithm, but I thought the paging 
effectiveness 
was related to likelihood of  not being able to find contiguous slots for 
group
page outs after the 30% usage (based on old technology). 
 
So if I have 5 3390-27 locals and they are all equally used at 50%, the
algorithms (CPU usage, not I/O) are going to pick one of them, then do the
page outs.  That paging will find contiguous slots and should be 
efficient. 

BTW, this is just an example, we still try to keep our 3390-27 local usage
at 30% just like we always did with smaller local page datasets in 
the past. 

I wonder what if any studies on this have been done in the lab. 
It would be nice if an IBM performance expert  like Kathy Walsh
could weigh in. 

Regards,

Mark
--
Mark Zelden - Zelden Consulting Services - z/OS, OS/390 and MVS 
mailto:m...@mzelden.com 
Mark's MVS Utilities: http://www.mzelden.com/mvsutil.html 
Systems Programming expert at http://expertanswercenter.techtarget.com/

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN








Unless stated otherwise above:
IBM United Kingdom Limited - Registered in England and Wales with number 
741598. 
Registered office: PO Box 41, North Harbour, Portsmouth, Hampshire PO6 3AU






--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-02-01 Thread Barry Merrill
MXG added variable SLOTUTIL back in '97 with this change text:

Change 15.064  Variable SLOTUTIL is added to dataset TYPE71 to measure
VMAC71 the percentage of local page dataset slots that are in
Apr 28, 1997   use.  Find the maximum value of SLOTUTIL during the month
   to make sure you have enough page dataset slots defined.
   SLOTUTIL should always be less than 25% (because the
   ASM's contiguous slot allocation algorithm can move 30
   pages in one SSCH only when there are 30 contiguous free
   slots, and at utilizations above 25%, ASM begins to not
   find 30 slots, so it tries to find 15, then 8... which
   causes lots of extra SSCHs to your page volumes at the
   very worst possible time - those few times when paging
   becomes a performance problem!).  I have preached this
   concept, but had not provided the variable (and the value
   I used in class turns out to need to be changed):
 SLOTUTIL=(SLOTLOMN-SLOTUNMN)/SLOTLOMN;
   compares the minimum number of defined local slots with
   the minimum number of unused local slots to calculate the
   maximum utilization of slots during the interval.

That note was based on a CMG or SHARE presentation I had attended
years earlier, when the contiguous slot allocation algorithm was first
being used, and the presentation was a smooth curve, output from a
model, rather than actual measurements, so there was no real knee of
the curve, but somewhere in the 25-30% range was noted as the beginning 
of possible pain.

Since one of the consequences of breaking the contiguous slot allocation
is an increase in the number of SSCHs to the paging volumes, and since
you all have dedicated devices, you should be able to plot the SSCH count
to your local page datasets from RMF 74 records versus the SLOTUTIL from
the 71 to see where your site's knee is located.

Barry Merrill 

-Original Message-
From: IBM Mainframe Discussion List [mailto:IBM-MAIN@bama.ua.edu] On Behalf
Of Barbara Nitz
Sent: Tuesday, January 31, 2012 10:50 PM
To: IBM-MAIN@bama.ua.edu
Subject: Re: Very Lage Page Datasets (was ASM and HiperPAV)

Writing to contiguous slots and over allocation is mentioned, but 
unless I missed it the old ROT (and health check) of not having more than
30%
of the slots allocated is not specifically addressed.   Certainly with 4K
pages (for the most part) and 3390-27 (or bigger) that 30% ROT doesn't
apply anymore?50% of a mod-27 is still a helava lot of free slots.

I think it still applies. My understanding has always been that the 30%
usage (after which paging effectiveness drastically drops) applies to the
algorithm used on the in-storage control blocks to pick the next free slot
in a page data set. Unless that algorithm was redesigned, 30% of 44.9GB per
page dataset is what you should not exceed (just as the health check says)
in AUX usage. Redesign of that is IMHO unlikely, just as using more than 2
IOs on a page data set simultaneously would require (an unlikely) redesign.

Sometimes the need for the appearance of an autonomic, self-healing 
system becomes more important than the need for an autonomic,
self-healing system.
;-) But, you're also saying close to 50% of health checks are useful, 
so that's good thing.
I consider about 30 of the 180-200 checks (1.12) useful. Otherwise I'll stay
out of *this* discussion. 

Barbara

--
For IBM-MAIN subscribe / signoff / archive access instructions, send email
to lists...@bama.ua.edu with the message: INFO IBM-MAIN

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-02-01 Thread Barbara Nitz
So if I have 5 3390-27 locals and they are all equally used at 50%, the
algorithms (CPU usage, not I/O) are going to pick one of them, then do the
page outs.  That paging will find contiguous slots and should be efficient.

BTW, this is just an example, we still try to keep our 3390-27 local usage
at 30% just like we always did with smaller local page datasets in
the past.

I wonder what if any studies on this have been done in the lab.
It would be nice if an IBM performance expert  like Kathy Walsh
could weigh in.

I had the 'honour' of deleting and adding several local page data sets on 
several lpars. They were a mixture of 1.10 and 1.12, I think. What I did 
observe (and that clashed with what I thought I knew about ASM) is the 
following:

1) Adding one or more locals, I expected them to first fill up to about the 
same percentage as the ones that were already in use (same size page ds, much 
faster -new- controller). No such luck. It looked to me like *all* of them were 
filling up (percentage wise) in about the same manner. Meaning that the 'old' 
locals had about 27%, the new ones right after add 0%. A day later the old ones 
had 35%, the new ones 8%. About the same behaviour when adding locals of the 
same size on the same controller - we only have one DASD controller per 
sysplex, and having two was the time when we migrated from one to the other.

2) A pagedel finishes MUCH faster than it ever did. It looked like ASM is 
actively shifting slots away from the to-be-deleted page data set. A pagedel 
finishes in under a minute. This used to be a really slow process because 
nothing was actively done.

3) In one case, I had two locals and pagedel'd one of them. Usage of the 
remaining one went up to 46%, pagedel was extremely fast. I then added a new 
local (on a new, different, much faster controller). Usage of that one stayed 
at 0% for a long time, which also surprised me.

4) I like the ASM health check that tells us that usage is 30% or more. (In 
fact, I send an automated exception email every time this happens.) I hate that 
ASM does not recognize that a  new page data set was added. That health check 
stupidly doesn't recognize a changed config and still spits out the warning. 
ASM also doesn't do anything active to level out slot usage on a new local. 
Usage only levels out after the next IPL.

5) I wonder if the behaviour I witnessed is due to applications (written by 
clickers with no z/OS clue) taking *a lot* - in the GB range - of above- 
the-bar storage, getting that backed by 'initializing' and then never use it, 
causing all those backed frames to become slots for the live of the IPL. Yes, I 
am talking primarily about the stuff that has feet in OMVS, where typically 
clickers write the code.

6) I bemoan IBMs failure to give us a good means of figuring out who is using 
HVSHARE or HVCOMMON storage and how much storage-above-the-bar is actually 
*used*, i.e. backed. As far as I know, there still isn't any tracking done for 
HVCOMMON storage, no means of reporting about it. No way to know who is 
excessively using common storage above the bar. Same for HVSHARE. Unless you're 
named Jim Mulder and know where to look in a dump, us lowly customers cannot 
even check that in a dump. Am I mistaken in the reporting capabilities? Has 
that been fixed by now? Or is it another means of IBM trying to sell  
software service contracts to get that done only by IBM? Not to mention the 
frustration until you find someone who can actually *do* it.

Barry, thank you very much for pointing out the MXG SLOTUTIL thing. I am now 
off to reading in the TYPE71 records and doing nice coloured pictures!

Barbara

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-02-01 Thread Ron Hawkins
All,

Another element of paging that has not been referenced is the ability to
handle all of the swap set size in parallel. If the swap set size is 120
pages then the old practice was to have at least four LOCALS so each thirty
page block of pages could be swapped-in in parallel. While swapping, like
paging, is not as prevalent as it once was I'm wondering if the swap set
size is still one of the principal guidelines for the number of locals that
should be defined.

For HDS VSP customers there is a new facility called MF-HDP that allows for
some very wide striping of volumes across the available spindles. If you are
using or plan to use MF-HDP then LOCALS would be very good candidates for
HDP pool volumes. You can allocate a 3390-3, 9, 27, 54 or A as a virtual
volume within the pool, but initially the space you use will be a 672 track
block that contains the volume label, VTOC, Index and VVDS. Then when you
define and format you LOCAL you will only use space equal to the size of the
page dataset rounded up to 672 tracks.

So if you want to allocate a 3390-54 for your locals, but only make them
5000 CYLS in size you should go for it, because the 60020 CYLS of empty
space won't actually use any space in the HDP pool. If you handled this
concept on Iceberg and the RVA then you're well on your way to wrapping your
head around this with MF-HDP?

The other advantage of this is the wide striping I mentioned. Each 30 page
set of contiguous slots will be within the same page, but the page is
striped across the underlying parity group disks. There won't be much
advantage for block page-in for each set of thirty pages, but you don't have
to worry about hand placing all your locals across the parity groups. The
wide striping will uniformly distribute all the page datasets across all the
underlying parity groups and disks. If you have 128 parity groups of R6
6D+2P then the read miss and destage activity of your locals, no matter how
many, will be uniformly spread across 1024 disk drives. 

Ignoring UCB constraints, it kind of makes minimizing the number of locals
an academic exercise. If you think you need eight locals then allocate
sixteen that are half the size on 3390-A. You will still only use the same
amount of disk space.

Ron

 -Original Message-
 From: IBM Mainframe Discussion List [mailto:IBM-MAIN@bama.ua.edu] On
Behalf Of
 Barbara Nitz
 Sent: Wednesday, February 01, 2012 9:13 PM
 To: IBM-MAIN@bama.ua.edu
 Subject: Re: [IBM-MAIN] Very Lage Page Datasets (was ASM and HiperPAV)
 
 So if I have 5 3390-27 locals and they are all equally used at 50%, the
 algorithms (CPU usage, not I/O) are going to pick one of them, then do
 the page outs.  That paging will find contiguous slots and should be
 efficient.
 
 BTW, this is just an example, we still try to keep our 3390-27 local
 usage at 30% just like we always did with smaller local page datasets
 in the past.
 

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-01-31 Thread Mark Zelden
On Mon, 30 Jan 2012 22:35:02 +, Martin Packer martin_pac...@uk.ibm.com 
wrote:

 Because
of the virtualisation within modern disk controllers robustness favours
more, smaller.


What 6-9 3390-27 page volumes isn't robust enough?  :-)

And regardless of PAV and where the physical location is on the emulated
DASD, if you put 5 smaller ones on  _multiple_ mod-27s, isn't there more of
a chance some of them end up on the same physical disk(s). 

One thing this thread didn't cover are some of the performance
recommendations in the init and tuning guide related to large DASD.  
 
Recommendation #1, does say to only have one page data set per device. 
and in the z/OS 1.11 manual I have open it even has the updated bars
on the left hand side.  

Writing to contiguous slots and over allocation is mentioned, but unless I 
missed it the old ROT (and health check) of not having more than 30% 
of the slots allocated is not specifically addressed.   Certainly with 4K
pages (for the most part) and 3390-27 (or bigger) that 30% ROT doesn't
apply anymore?50% of a mod-27 is still a helava lot of free slots.  

Mark
--
Mark Zelden - Zelden Consulting Services - z/OS, OS/390 and MVS   
mailto:m...@mzelden.com
Mark's MVS Utilities: http://www.mzelden.com/mvsutil.html 
Systems Programming expert at http://expertanswercenter.techtarget.com/

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-01-31 Thread Vernooij, CP - SPLXM
Mark Zelden m...@mzelden.com wrote in message
news:9474272146773836.wa.markmzelden@bama.ua.edu...
 On Mon, 30 Jan 2012 22:35:02 +, Martin Packer
martin_pac...@uk.ibm.com wrote:
 
  Because
 of the virtualisation within modern disk controllers robustness
favours
 more, smaller.
 
 
 What 6-9 3390-27 page volumes isn't robust enough?  :-)
 
 And regardless of PAV and where the physical location is on the
emulated
 DASD, if you put 5 smaller ones on  _multiple_ mod-27s, isn't there
more of
 a chance some of them end up on the same physical disk(s). 
 
 One thing this thread didn't cover are some of the performance
 recommendations in the init and tuning guide related to large DASD.  
  
 Recommendation #1, does say to only have one page data set per device.

 and in the z/OS 1.11 manual I have open it even has the updated bars
 on the left hand side.  
 
 Writing to contiguous slots and over allocation is mentioned, but
unless I 
 missed it the old ROT (and health check) of not having more than 30%

 of the slots allocated is not specifically addressed.   Certainly with
4K
 pages (for the most part) and 3390-27 (or bigger) that 30% ROT doesn't
 apply anymore?50% of a mod-27 is still a helava lot of free slots.

 
 Mark
 --

I discussed the last issue many years ago with Greg Dyck, who then
(afaik) was the expert on paging, with regards to 30% free slots on
3390-03 and -09. He stated that simulations still recommended 30% usage,
in spite of the helave lot of free slots on the -09.

In that same discussion I learned that ASM reserves 2 exposures per
pagedataset, so it was no problem putting more than 1 pagedataset on a
volume. ASM just ensured 2n-1 PAVs statically assigned to that volume.
This seems contradicting with Rec #1 you mention.

About health checks: I consider at least 50% of them being writen on a
lost Friday afternoon, because of some target to just deliver health
checks.

Kees.

For information, services and offers, please visit our web site: 
http://www.klm.com. This e-mail and any attachment may contain confidential and 
privileged material intended for the addressee only. If you are not the 
addressee, you are notified that no part of the e-mail or any attachment may be 
disclosed, copied or distributed, and that any other action related to this 
e-mail or attachment is strictly prohibited, and may be unlawful. If you have 
received this e-mail by error, please notify the sender immediately by return 
e-mail, and delete this message. 

Koninklijke Luchtvaart Maatschappij NV (KLM), its subsidiaries and/or its 
employees shall not be liable for the incorrect or incomplete transmission of 
this e-mail or any attachments, nor responsible for any delay in receipt. 
Koninklijke Luchtvaart Maatschappij N.V. (also known as KLM Royal Dutch 
Airlines) is registered in Amstelveen, The Netherlands, with registered number 
33014286



--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-01-31 Thread Mark Zelden
On Tue, 31 Jan 2012 08:48:30 -0600, Mark Zelden m...@mzelden.com wrote:

On Mon, 30 Jan 2012 22:35:02 +, Martin Packer martin_pac...@uk.ibm.com 
wrote:

 Because
of the virtualisation within modern disk controllers robustness favours
more, smaller.


What 6-9 3390-27 page volumes isn't robust enough?  :-)

And regardless of PAV and where the physical location is on the emulated
DASD, if you put 5 smaller ones on  _multiple_ mod-27s, isn't there more of
a chance some of them end up on the same physical disk(s).

One thing this thread didn't cover are some of the performance
recommendations in the init and tuning guide related to large DASD.

Recommendation #1, does say to only have one page data set per device.
and in the z/OS 1.11 manual I have open it even has the updated bars
on the left hand side.

Writing to contiguous slots and over allocation is mentioned, but unless I
missed it the old ROT (and health check) of not having more than 30%
of the slots allocated is not specifically addressed.   Certainly with 4K
pages (for the most part) and 3390-27 (or bigger) that 30% ROT doesn't
apply anymore?50% of a mod-27 is still a helava lot of free slots.


Remove (for the most part) above - I momentarily forgot that 1M pages
are not pageable.  At least not now... 

--
Mark Zelden - Zelden Consulting Services - z/OS, OS/390 and MVS   
mailto:m...@mzelden.com
Mark's MVS Utilities: http://www.mzelden.com/mvsutil.html 
Systems Programming expert at http://expertanswercenter.techtarget.com/

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-01-31 Thread Edward Jaffe

On 1/31/2012 7:07 AM, Vernooij, CP - SPLXM wrote:

About health checks: I consider at least 50% of them being writen on a
lost Friday afternoon, because of some target to just deliver health
checks.


Sometimes the need for the appearance of an autonomic, self-healing system 
becomes more important than the need for an autonomic, self-healing system.  
;-) But, you're also saying close to 50% of health checks are useful, so that's  
good thing.


--
Edward E Jaffe
Phoenix Software International, Inc
831 Parkview Drive North
El Segundo, CA 90245
310-338-0400 x318
edja...@phoenixsoftware.com
http://www.phoenixsoftware.com/

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN


Re: Very Lage Page Datasets (was ASM and HiperPAV)

2012-01-31 Thread Barbara Nitz
Writing to contiguous slots and over allocation is mentioned, but unless I
missed it the old ROT (and health check) of not having more than 30%
of the slots allocated is not specifically addressed.   Certainly with 4K
pages (for the most part) and 3390-27 (or bigger) that 30% ROT doesn't
apply anymore?50% of a mod-27 is still a helava lot of free slots.

I think it still applies. My understanding has always been that the 30% usage 
(after which paging effectiveness drastically drops) applies to the algorithm 
used on the in-storage control blocks to pick the next free slot in a page data 
set. Unless that algorithm was redesigned, 30% of 44.9GB per page dataset is 
what you should not exceed (just as the health check says) in AUX usage. 
Redesign of that is IMHO unlikely, just as using more than 2 IOs on a page data 
set simultaneously would require (an unlikely) redesign.

Sometimes the need for the appearance of an autonomic, self-healing system
becomes more important than the need for an autonomic, self-healing system.
;-) But, you're also saying close to 50% of health checks are useful, so that's
good thing.
I consider about 30 of the 180-200 checks (1.12) useful. Otherwise I'll stay 
out of *this* discussion. 

Barbara

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: INFO IBM-MAIN