Re: LZMA and the Install Sets?

2006-04-18 Thread Antti Harri

On Mon, 17 Apr 2006, Nick Holland wrote:


Oh?  Where are your numbers?


Here's a benchmark about gzip vs lzma vs bzip2:

http://tukaani.org/lzma/benchmarks

Lasse Collin is also working on a more sane
tool for the algorithm but I don't have enough
information about that so if someone is interested
he/she should contact Lasse directly.

--
Antti Harri



Re: LZMA and the Install Sets?

2006-04-18 Thread Frank Denis

Le Mon, Apr 17, 2006 at 11:55:34PM -0400, Nick Holland ecrivait :

~/comptest $ time lzma e comp39d.tar comp39d.tar.lz
7m5.59s real 6m54.79s user 0m0.59s system
(maximum RAM used: around 80M, I think)

Comments:
rzip and lzma turned in some good numbers (REALLY good numbers),


 The lzma package is useless.
 
 Although the algorithm is also LZMA, every time I tried, p7zip was better.

2x faster and always with greater compression (both at max compression level).

448032  comp39.tar
155904  comp39.tgz
42912   comp39.tar.lzma
34272   comp39.tar.7z

 lzma : 403.22s user 0.45s system 87% cpu 7:43.96 total
 7za  : 211.07s user 2.48s system 89% cpu 3:58.25 total
 
 http://p7zip.sourceforge.net/
 
 Sure, LZMA compression needs memory, though.


--
Frank Denis - frank [at] nailbox.fr
Young Nails / Akzentz nail tech



Re: LZMA and the Install Sets?

2006-04-17 Thread Ted Unangst
On 4/15/06, [EMAIL PROTECTED] [EMAIL PROTECTED] wrote:
 Long time ago I read about LZMA at undeadly.org.
 I just wanted to ask if somebody realy works on a BSD-Implementation or not.
 Because LZMA-Compressions are realy awesome and for the decompression it
 needs less time and less memory then bzip2 it could be even an alternative
 for slow Hardware (VAX?).

you obviously missed the part about the preposterously slow compression times.



Re: LZMA and the Install Sets?

2006-04-17 Thread sebastian . rother
 you obviously missed the part about the preposterously slow compression
times.

Well..
You missed my point. I know you don`t crosscompile but you can compress
all Archives on a stronger Computer. So a VAX isn`t the best hardware to
start with LZMA compressing. Yes..

Something else: You compress install Sets for CDs mostly just once but
they get decompressed many times. :)
And if somebody wanna change these install Sets he could use another Box
for the compression. But I don4t think there`s anybody out there who just
owns a 486 and a VAX.

You would save a lot Bandwith and also storage space on the CDs.
E.g. the src gets compressed very well and the decompression isn`t that
much slower.


That`s why I asked.. and because I read about it on undeadly.org (-:

Sebastian



Re: LZMA and the Install Sets?

2006-04-17 Thread Marco Peereboom
And did you see any code?

On Mon, Apr 17, 2006 at 08:16:06PM +0200, [EMAIL PROTECTED] wrote:
  you obviously missed the part about the preposterously slow compression
 times.
 
 Well..
 You missed my point. I know you don`t crosscompile but you can compress
 all Archives on a stronger Computer. So a VAX isn`t the best hardware to
 start with LZMA compressing. Yes..
 
 Something else: You compress install Sets for CDs mostly just once but
 they get decompressed many times. :)
 And if somebody wanna change these install Sets he could use another Box
 for the compression. But I don4t think there`s anybody out there who just
 owns a 486 and a VAX.
 
 You would save a lot Bandwith and also storage space on the CDs.
 E.g. the src gets compressed very well and the decompression isn`t that
 much slower.
 
 
 That`s why I asked.. and because I read about it on undeadly.org (-:
 
 Sebastian



Re: LZMA and the Install Sets?

2006-04-17 Thread Nick Holland

curiosity got the best of me here.  Found I had a machine with several
major compression programs on it, and while this tired argument isn't
worth this, I found it the idea of a comparison interesting, so I
thought I'd share...

[EMAIL PROTECTED] wrote:

you obviously missed the part about the preposterously slow compression

times.

Well..
You missed my point. I know you don`t crosscompile but you can compress
all Archives on a stronger Computer. So a VAX isn`t the best hardware to
start with LZMA compressing. Yes..


ick.  no.


Something else: You compress install Sets for CDs mostly just once but
they get decompressed many times. :)


depends.  I build sets often on slow machines.  More often than I install
'em.  In OpenBSD, I don't get much of a vote, but I get more than you. :)


And if somebody wanna change these install Sets he could use another Box
for the compression. But I don4t think there`s anybody out there who just
owns a 486 and a VAX.


As a NATIVE BUILDING OS, we assume you have only one machine.  OpenBSD
does not need a second machine to build, and hopefully will not.  When
you need a second machine, you might as well be cross building.  Heck,
you ARE cross-building.


You would save a lot Bandwith and also storage space on the CDs.


Bandwidth is cheap.
If you can cut the time to download from 30 minutes to 3 minutes, you
have accomplished something.  Changing it from 30 minutes to 15 minutes,
I don't think that changes a thing.  You will still get up and go do
something else (or you need a job).

Far less painful to all than changing compression would probably be
DVD media.  If you are going to inconvience some people, heck, at least
give them a REAL benefit, not 20% more.


E.g. the src gets compressed very well and the decompression isn`t that
much slower.


Oh?  Where are your numbers?

Being you have brought this topic up in the past, I'm going to show you
some real numbers.

Took comp39.tgz for i386 and ran these tests on a slow amd64 system (yeah,
choice of an i386 file for running the test on an amd64 is kinda strange,
but it really doesn't matter much).  Unzipped it, and copied it several
times, ran these compressions and watched RAM usage using top:

~/comptest $ time bzip2 comp39b.tar
1m7.03s real 0m55.97s user 0m0.43s system
(maximum RAM used: around 8M)

~/comptest $ time rzip comp39c.tar
0m36.42s real 0m26.62s user 0m1.14s system
(maximum RAM used: around 118M)

~/comptest $ time lzma e comp39d.tar comp39d.tar.lz
7m5.59s real 6m54.79s user 0m0.59s system
(maximum RAM used: around 80M, I think)

~/comptest $ time gzip -9 comp39e.tar
1m16.87s real 1m15.28s user 0m0.42s system
(maximum RAM used: around 700k)

Results:
~/comptest $ ls -l comp*
-rw-r--r--  1 njholland  njholland   75288260 Apr 17 19:20 comp39.tgz
-rw-r--r--  1 njholland  njholland  218347520 Apr 17 19:25 comp39a.tar
-rw-r--r--  1 njholland  njholland   50369737 Apr 17 19:26 comp39b.tar.bz2
-rw-r--r--  1 njholland  njholland   25860279 Apr 17 19:29 comp39c.tar.rz
-rw-r--r--  1 njholland  njholland   20849017 Apr 17 19:38 comp39d.tar.lz
-rw-r--r--  1 njholland  njholland   75288272 Apr 17 19:45 comp39e.tar.gz

Comments:
rzip and lzma turned in some good numbers (REALLY good numbers),
but the RAM required was absolutely absurd (at least, for this
application).  Your average mac68k or 486 would be in swap hell for
a week, and many good firewall systems would take many times as long
to load.  I'm rather amazed that rzip was such a screamer in
compression speed for this app, I've never seen it outrun anything
before.  I repeated the test, btw, consistent.

bzip2 seems to be an answer in search of a problem.  It does things a
little better for a big price.

gzip is the only one with a RAM footprint that is acceptable for a
multi-platform OS.  Really, the amount of change in our lives that
switching to a better compressor would make is not worth the
difference here.

HOWEVER, I did spot one interesting thing...first time I ran the gzip
test, I got a funny answer:

~/comptest $ time gzip comp39e.tar
0m29.39s real 0m26.70s user 0m0.44s system

[EMAIL PROTECTED]
~/comptest $ ls -l comp39   
   ~/comptest $ ls -l comp39*
-rw-r--r--  1 njholland  njholland   75288260 Apr 17 19:20 comp39.tgz
  ...
-rw-r--r--  1 njholland  njholland   75990958 Apr 17 19:26 comp39e.tar.gz

Since the size didn't line up, I figured that was probably due the
wrong compression factor specified -- apparently, the file sets are
built with a -9.  Note that in my humble opinion, the -9 on gzip that
tar and/or the build process uses is Just Not Worth It -- for a tiny
improvement in file size, you wait well over twice as long.  You think
that doesn't matter?  Re-pack a mac68k build some day.

ok, one last nail:
~ $ ls -l `which rzip` `which lzma` `which bzip2` `which gzip`
-r-xr-xr-x  6 root  bin 29248 Feb