Actually, giving the question some serious consideration, at a reasonably good cache hit ratio the impact of an infinite number of jobs randomly reading the same dataset will not be significantly different to an infinite number of jobs accessing many datasets and volumes on the same shared channels.
At high cache hit rates elongation usually comes about due to path busy, whether that be Pend for channel busy on ESCON, frame interleaving on FICON, or the effects of microprocessor utilisation at either end of the channel. For random access with poor cache hit ratios, or sequential read, the RAID scheme will have a much greater bearing on throughput, as will the scheme employed for pre-fetch in the case of sequential read. For Sequential read, good old RAID-1 is probably at the bottom of the pile in this case as it only employs two spindles, and the pre-fetch scheme used by the only MF RAID-1 vendor simply flip-flops between the spindles instead of pre-fetching from both concurrently. RAID-10 schemes that can pre-fetch from all the disks in parallel would be at the top, with RAID-5/6 not far behind and standard RAID-10 a bit behind RAID-5/6. For random it would pretty much depend on how many spindles can be used concurrently as one volume. RAID-10, RAID-5 and RAID-6 using eight drives would all have the same performance for a single dataset. > > The SWAG ROT is that there is no free lunch, and each additional job > will degrade performance to some degree. The first few may not be > measurable. The degradation may follow a classic 'knee curve' (rise > slowly to a point then abruptly get much worse). > Hiperbatch is a great tool, but last time I looked it did not support EFDS :( Pre-loading is not the only way to use Hiperbatch, as it has a great 'catch algorithm' that keeps track of the leading and following jobs allowing to get bursts of speed from the Hiperbatch buffer. The best way I have found to Hiperbatch without pre loading is to kick off all the jobs that will be using hiperbatch at the same time and give them the same service class. You want them to synchronise around the same area of the file so one job does the reading, and the rest are just behind reading from Hiperbatch. If some jobs are accessing too many different parts of the file then Hiperbatch does not work that well. Finally, if the datasets is 4-5GB and you want to have a ton of jobs read it as fast as possible, then on HDS storage consider putting the file in FlashAccess if you have HDS Arrays. This is functionally the same as Solid State Disk (remember those), and it won't matter what sort of disk and parity scheme you are using. Ron ---------------------------------------------------------------------- For IBM-MAIN subscribe / signoff / archive access instructions, send email to [EMAIL PROTECTED] with the message: GET IBM-MAIN INFO Search the archives at http://bama.ua.edu/archives/ibm-main.html

