Ted,

I'm wondering if your primary rule of not compressing a file unless it will
exceed its architectural limit may have blocked the opportunity for you to
come across cases where compression is not a waste of time.

Synchronous remote copy is one area where compression of datasets created or
updated in the batch critical path can buy back the impact of write
elongation, and sometimes leave you with improved elapsed times. For example
where TC, SRDF or PPRC introduce a 100% write response time impact in a
father-to-son update, but the input and output files are compressed to 40%
of their original size. That's a net run time improvement of 20% due to IO
reduction.

Another example are datasets written or updated once, and then read by many
programs. Even a reduction in size of only 40% results in a corresponding
reduction in IO activity multiplied by the number of times the file is read.
Using LZW compression means that there is only a small increase in CPU Time
to decompress the file multiple times, usually around 15-20% of the CPU cost
of compressing the file. The more you read the dataset, the more benefit you
obtain.

These are two examples where I have used DFSMS compression to make
significant improvements to batch run time. The only hiccup was when IBM
moved the compression assist instructions to firmware on the G3 CMOS
processor (it may have been G4). They were back in the hardware by G6 and
all the advantages of asymmetric CPU cost for compress/decompress returned. 

Huffman may not be an appropriate compression algorithm for this because the
implementations I have experience with (e.g. DFSMShsm, DFSMSdss) have an
equal cost for compress and decompress. The symmetrical CPU cost heavily
impacts the value of a compressed input file.

Ron  

> -----Original Message-----
> From: IBM Mainframe Discussion List [mailto:[email protected]] On
Behalf Of
> Ted MacNEIL
> Sent: Tuesday, July 06, 2010 1:41 PM
> To: [email protected]
> Subject: Re: [IBM-MAIN] VSAM Max Lrecl?
> 
> >I'm a "results oriented" guy; formal proofs, elegant or otherwise, don't
> "float my boat" anywhere near as well as concrete results that I can
> quantify.
> 
> I tend to agree with you; without the time to do an empirical study on
> compression what do you have?
> 
> I've never had a compressed file come out larger than the original input.
> 
> I've heard of others that have, but have never seen it.
> 
> And, I wrote a Huffman programme in 1977, as part of the same course that
> introduced me to the theory that we're discussing.
> 
> And, in the early days of BBS, COMPUSERVE, AOL, and the INTERNET,
compression
> was important due to bandwidth concerns.
> 
> Just like it was when DASD was expensive.
> On disk, I only use compression when a file is going to exceed its
> architectural limit.
> 
> Using it to save space (or bandwidth) is a waste of time, productivity,
and
> CPU.
> 
> 
> -
> I'm a SuperHero with neither powers, nor motivation!
> Kimota!
> 
> ----------------------------------------------------------------------
> For IBM-MAIN subscribe / signoff / archive access instructions,
> send email to [email protected] with the message: GET IBM-MAIN INFO
> Search the archives at http://bama.ua.edu/archives/ibm-main.html

----------------------------------------------------------------------
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to [email protected] with the message: GET IBM-MAIN INFO
Search the archives at http://bama.ua.edu/archives/ibm-main.html

Reply via email to