In a message to All <03 Nov 94 13:02> [EMAIL PROTECTED] wrote:

 Ia> On 03 Nov 94 04:35:00 +0000, Johnathan Taylor said:
..
 >> A gzip'd tar must first be decompressed in its entirety.

 Ia> Your answer is quite correct, but it
 Ia> is to a different question...  It
 Ia> is true that if a gzip file gets
 Ia> corrupted then that's it.  However I

I didn't understand the question as it didn't apear relevent to what I 
said/meant in the text you quoted prior to it, so I guessed at what you wanted 
and got it wrong, obviously;-)

 Ia> think specially catering for corrupted
 Ia> files is not the main aim of a
 Ia> compression/archive system.

It's not, it's just a valuable advantage:-)

 Ia> Anyway, the explanation that I
 Ia> actually wanted was of the following
 Ia> sentence:

 >jet> Plus I'd want
 >jet> to be able combine assosiated files
 >jet> into a single archive not a seperate
 >jet> lose-able bunch of seperatly compressed files!

 Ia> So why is a single compressed file any
 Ia> more lose-able than any other kind of
 Ia> file?

What I meant by a bunch of lose-able sperate files is when you have 1000 
readme files for a 1000 different programs or whats.new etc. Keeping all files 
related to a particular program in one archive with random access to 
individual members allows easy access to locate a single programs assosiated 
files without having to resort to sticking each bunch of files in their own 
directory or resorting to obtuse 32bit HEX digit names to uniquly ID 
individual compressed files requiring some form of data-base to re-create 
human readable named access to individual files...
Does that explain what I meant? If not don't worry about it!

 Ia> How is this different from LZ77?

AFAIK LZ77 is a distinct 2-stage process and requires the addition of the 
encoding tree prior to starting of the actual encoded bit-stream, Whilst LZHUF 
does it on the fly so to speak, this also means that LZHUFF can actually even 
compress small files of less than 100bytes but LZ77 cannot!

Plus dictionary size for dictionary size LZHUF generally produces better 
compression than LZ77 it's only problem is that the encode alogerithm is a bit 
more processor intensive which is probably why LZHUF -lh5- only uses an 8k 
dictionary whilst LZ77 has HAD to grow to 32k(ZIP) or 64k(RAR) to beat it on 
99% of file types.
 >> These ARE ALL Documented in various places on
 >> the nets, so those with direct access
 >> to all those net tools should be able
 >> to locate them themselves!

 Ia> There are quite a lot of things on the
 Ia> net actually, so you will have to
 Ia> narrow down the search a little more than that.

Heck I duno what the facility of the online tools are! Isn't it called archie 
or somthing... ask it to look for LHA or LZHUF that's one reason I capitalise 
those words.....

 >>  >> The thing about gzip is that it uses
 >>  >> very little RAM to decompress (apart
 >>  >> from having to store each block of the
 >>  >> file

 >> I don't know WHO wrote the above
 >> paragraph imc maybe... but GET REAL! gzip and
 >> PKUNZIP2.04g etc are required to keep a
 >> running 32K ring-buffer OR rely on
 >> flawless random file access of the
 >> output stream in order to get at the 32k
 >> sliding dictionary!

 Ia> Which is implied by "storing each
 Ia> block of the file" (if the blocks are 32K.

Didn't you know that it uses a 32k sliding window?

 Ia> Note: I did _not_ mean disk blocks.

So? does this mean that it uses any less ram just because you chose not to 
name that 'block' size? 32k is half the available address space! then you need 
to add the program code size, main tables size input and output file handling 
interface into Sam-Dos.... I know decompression *CAN* be done with a z80 but 
why bother when the Sam cannot compress the archives and probably never will 
be able to do DeflatX compression! The only use I'd see for such a util is to 
inflate tar.z file from a unix box... If that was the requirement then I'd do 
what I do now and inflate them on the unix box and LHA re-compress them and 
use PMEXT.COM under prodos or port LHA to sam-dos which should be much easier 
than gzip!

 Ia> Do you remember how this started?  By
 Ia> me saying I might write one.

I do remeber but have you started yet? Well assuming you were serious may I 
suggest that you don't *AIM* for gzip as its only a partial product! Which is 
one reason why I am so against the waste of effort in writing it for the SAM. 

The full product is a ZIP and UNZIP! Actually IF you did that then yould only 
need to make the UNZIP portion handle deflated files as the ZIP portion could 
still achieve reasonable ratios using the older implode alogerithm which is 
the same as deflate but uses a 4 or 8k sliding dictionary and IS supported by 
the latest PKUNZIP and the latest portable UNZIP50p? infozip project.. And is 
more likely to fit reasonably into the sam scheme of things:-)
And I'm sure Si'll be very glad to hear of a ZIP/UNZIP for sam native mode if 
you can write it:-)

 >> As if deflate was the BEST lossless
 >> compression method... Ever heard of RAR?

 Ia> No I haven't.

Well there's loads of stuff you seem to be missing info on.. If you've got 
live net access I suggest you go digging through some of those internet sites 
and have a read about what the real alternatives are for yourself, after all 
if you don't believe me go get the tech docs and read for your self:)

 >> Oh btw LZHuff can compress some LZW
 >> encoded stuff a bit further!

 Ia> Perhaps, but you are better off
 Ia> uncompressing the LZW first before trying
 Ia> another compression method.

Of course, but it proves that LZHuff is more robust when compressing awkward 
data than LZW. LZ77 gets around the stuation of uncompressable stuff by 
resorting to temporarily giving up and simply storing the awkward 32k section, 
LZHUFF generally speaking doesn't come across this sort of situation often 
enough to warrent such a drastic change in format as was done when PKZIP204 
was released.

A reason I feel LHARC format is better for the SAM than the ZIP format is that 
although both forms can hold all the sam-specific file attribute info, the 
LHARC format can also be implemented on the humble 48k speccy with disks and 
could contain speccy file info too! It just seems more logical to use a format 
that's free from license restrictions and easy to implement as a non-rambound 
utility file serialization utility that works without needing vast amount of 
ram to function in.

Oh btw even the old -lh1- LZHUF method is amazingly more effective than the 
old 16-bit unix.Z compress method at compression and resource requirements.
When I receive tar.Z archives on my SAM I always zcat 'em and re-compress them 
straight away with PMARC (the z80 LZHUFF derivative). That reduces the overal 
size of the compressed tar by at least another 3rd smaller than the 
unix-compressed tar size. Until I can sort out a clean floppy to detar it and 
archive the tar members properly so they're useable but still much smaller 
than the original tar.Z archive!

Would you imc mind if we let this thread taper off please as I've got other 
things I need to be doing rather than aurguing the toss about somthing as 
trivial as this to save you wasted effort when I want to get on with other 
more pressing stuff and you've got to re-aquaint yourself with your recently 
obtained SAM;-)
Cheers
Johnathan.

___ Olms 1.60 [Evaluation]
+-------------------------------------------------------------------+
| Standard disclaimer:  The views of this user are strictly his own |
|  ===> Gated @ Centronics BBS [centron.com] +44-1473-273246] <===  |
+-------------------------------------------------------------------+

Reply via email to