On 2013-02-24 10:12, dd wrote:
> but, wait a second, there is another way to look at it: if in all this
> compression (and then decompression for use) one byte gets corrupted
> then suddenly ALL 100 of those 1 MB files are corrupted…if each of
That is true of any compression method.
For example, if you do a tar archive backup, which is then compressed (a
…tar.gz or .tgz), a single error in the compression and all is lost.
Other methods compress each file individually, then archive the lot.
This way, an error only destroys one file.
There are more sophisticated methods of backup or archival. I have used,
in MsDos times, proprietary methods that combined compression with
forward error recovery, which is a step ahead of simple error detection.
I have not seen any Linux backup software using this, unless we use our
own combination with par or par2, not even in the oss repo.
(search for ‘forward error recovery’. It is used by the NASA for their
llllong distance communications, for example)
> not only that, but just like the old (last century) DoubleSpace
> application [which “doubled” the apparent size of your disk via
> compression] there is a speed penalty to pay because every time you
> retrieve anything it has to be uncompressed (by the CPUs)…
It is negligible compared to the read speed of mechanical devices such
as hard disks
Even at the times doublespace was designed and used, you actually gained
speed on most hardware, because hard disks were very slow compared to
recent hardware; having to write or read half the sectors made the
overall speed faster. Curious, eh?
Of course, for that you use fast algorithms with low rate of
compression. If the intention is backup or archival, you usually
compress more, speed is less of an issue.
Interestingly, NTFS does have directory compression, a successor of
doublespace. You can tag individual files or entire trees for
compression. Ext2/3/4 has flags to mark files as compressed, but the
code to achieve this was never implemented. Shame on Linux, IMNSHO.
btrfs does have compression if you enable it, I heard. Of course, if you
use replication in that filesystem space increases…
> now, you have not explained your situation and your need, but i’d
> say that the price of 1 MB (or 10,000 MB) of storage has fallen so much
> in the last 10 years that if you are working with the ‘average’ home’s
> storage needs you might be better off in the long run to leave this new
> look at an OLD trick alone and just buy a few more gigabytes of local
> drives…on the other hand, if you are a google data-farm
> administrator there might be millions of dollars to save–but a word to
> the wise, before jumping in on this i’d wanna:
Compression is nice. My desktop has about 2.5TB of storage. To do a
backup I need at least that external space, maybe duplicated for
alternating backups. A 3TB Seagate HD costs between 100 and 300 euros at
Alternate. Plus box, tax, P&P.
I don’t call that “cheap”.
However, I know nothing about this particular product the OP asks about.
I’m curious, though.
–
Cheers / Saludos,
Carlos E. R.
(from 12.1 x86_64 “Asparagus” at Telcontar)