Pawel,

I have been out of the sorting business for a while, but the target you
can go for is what I call 2T, where T is the time to read in the data.

The 2T comes from the fact that you have to read it in and then write it
out.

Writing to sort work files should not be the problem, hopefully the sort
is writing to several of them in parallel during the sort phase and
reading from them in parallel during the merge phase.

EXCP's are an accountant's measurement, not a tuner's measurement.  Look
at your SMF data and see what the channel connect times are for SORTIN
and SORTOUT.  If the elapsed time is significantly greater than those
two added together then you have to start looking for the bottleneck.

My first reaction to your 17M/second was a question.  Is that data
sorted divided by total time?  I assume it is.  If so then you are
getting about 34M/second on input and output, and that is not a bad
number.

Tape and DASD inputs and output will give you different numbers.  There
tends to be more interference with tape files.  Fewer channels, no PAV.

Back when DASD was real 3390, 4M/second was the target, so you are doing
pretty good.  I couch that with the fact that I have not run throughput
tests on the various DASD subsystems available today.  34M may be great
or it may be just average.  I don't know.

The best thing you can do is give each sort enough memory.  You will
need to ask Frank Yeager what the target amounts of memory are for a 10G
sort.  My guess is about 17M to 24M.  A 100G sort might need 50M to 75M.
A 1G sort might need 5M to 7M.

What you are attempting to do is make it so there is no need to do an
intermediate merge, or in the case of blockset sorting, that the final
merge remains efficient.

I would not attempt to get a 10G sort to run as a turnaround or incore
sort.  It will put a large strain on RSM/ASM and not buy you that much.

I hope Frank weighs in on this.

Chris Blaicher
BMC Software, Inc.
These are my own personal comments and do not reflect the views of my
employer.  Please do your own evaluation of any statements or
suggestions made.

-----Original Message-----
From: IBM Mainframe Discussion List [mailto:[EMAIL PROTECTED] On
Behalf Of Pawel Leszczynski
Sent: Sunday, March 09, 2008 4:06 PM
To: [email protected]
Subject: how fast can I sort on mainframe (using DFSORT)?

Hello everybody,

I realize subject is VEEEERY broad and my question VEEEERY general,
but...

Recently in our shop we are reviewing our whole batch processing.
Most of the time of EOD processing is consumed by sort of many big
sequential 
files. (One such file has approximately order of 10GB, 10mln records)
I listed few tens of such batch jobs (the longest-lasting ones)
and computed mean sorting rate.
It appeared to be about 1GB/min ~ 17MB/sec

I suppose it's very poor result(???).

Can you tell me how much I can improve this?

These batch jobs are little CPU-consuming (~10% of one CPU),
I suppose that major concern is to:

-limit EXCPS (1)
-increase throughput rate from DASD to central storage (2)

I realize that sorting whole file in central storage (hiperspace
sorting) would 
eliminate need to use work files and EXCPS to them.
How much central storage is needed to handle in-storage sorting for
let's say 10 GB file???

How much can I  improve (2)?

Can you tell me what is mean sorting rate in your installation?
Before starting I would just like to know if I can achieve substantial 
improvement.

TIA,
Pawel Leszczynski
PKO BP SA

----------------------------------------------------------------------
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to [EMAIL PROTECTED] with the message: GET IBM-MAIN INFO
Search the archives at http://bama.ua.edu/archives/ibm-main.html

Reply via email to