Re: [Hampshire] [hants] compression

Top Page

Reply to this message
Author: Alan Bell
Date:  
To: isaacclose, Hampshire LUG Discussion List
Subject: Re: [Hampshire] [hants] compression
there may be some academic projects to compress particular types of data
in particularly clever ways, but you have listed all the main general
purpose lossless tools. If possible you might use lossy compression,
which could mean recoding image content at lower quality settings, but
this really applies to photo type images, video and audio.
Most of the command line compressors have a default setting which is a
reasonable tradeoff between speed and compression. Have a look at the
man pages to see if there are flags you can flip to scrunch things up a
bit tighter.
Finally, the reason things are .tar.gz is that they are slapped together
first with tar, then the whole archive is compressed with gzip. This is
more efficient than .gz.tar because the compressor can use similarities
between different files. For example if you had a directory of fairy
tales and they all started "Once upon a time" the compressor could store
that string once for the whole collection and just insert it at the
start of each on decompression.


Isaac Close wrote:
> hello all,
>
> I'm (trying) to find the 'best' compression software/algorithm about.
> Best as-in best compression ratio most, cpu time and memory footprint are not a problem.
>
> I'm already well aware of 7zip, bzip2, rzip and ofcourse gzip, but looking on google is not putting me beyond those.
>
> tar very much,
>
> Isaac.
>
>
>
>
>
>
>