Compressing packages with bzip2 instead gzip?
Paul Sladen
ubuntu at paul.sladen.org
Tue Jan 17 12:14:36 GMT 2006
On Mon, 16 Jan 2006, Phillip Susi wrote:
> Which is why we want to use 7zip, since it is twice as fast at
> decompressing as bzip2, and gets better compression.
I don't think you can generalise like. One of the (many) algorithms that
7zip selects between is bzip2, the other being bascially zlib with a huge
(out of spec) window size.
The reason that 7zip does better on jump-infested i386 code is that it runs
a pre-processor over the binary to convert relative jumps into absolute
values (and the absolute values are more likely to repeat).
I think there is a case for adding support for a post-processing stage to
dpkg/cloop (like PNG does with predictors). Secondly, if bzip2 speed is a
problem, I suspect there is room for optimisation (refactoring, _not_
turning on --omg-faster) and using bzip2 compression with a smaller
windowsize (<900kB) if the difference is less than 1% in size.
(For perspective, one seek() on the CD image is going to cost about the same
as the difference in execution speed of zlib/bzip2).
Delta-debs on Mirrors.
For mirror distribution, my conclusion from 12months ago was that zsync
(pre-computed, client-size executed rsync) against the data.tar regenerated
by dpkg-repack is the way to go.
Since this would go into APT, dpkg stays clean and
I like the idea in somebody else's email of only providing zsyncs against
files published on CD or the last 10days worth.
I think a *top* priority should be delta-diffs on the Package lists, but the
dapper-does-it-48-times-a-day churn on these is massive.
-Paul
--
This country is covered in white fluffy snow. Helsinki, FI
More information about the ubuntu-devel
mailing list