Philip Rowlands <[EMAIL PROTECTED]> wrote:

> On Thu, 18 Jan 2007, Jim Meyering wrote:
>
>> I've done some more timings, but with two more sizes of input.
>> Here's the summary, comparing straight sort with sort --comp=gzip:
>>
>>  2.7GB:   6.6% speed-up
>>  10.0GB: 17.8% speed-up
>
> It would be interesting to see the individual stats returned by wait4(2)
> from the child, to separate CPU seconds spent in sort itself, and in the
> compression/decompression forks.
>
> I think allowing an environment variable to define the compressor is a
> good idea, so long as there's a corresponding --nocompress override
> available from the command line.
>
>>  $ seq 9999999 > k
>>  $ cat k k k k k k k k k > j
>>  $ cat j j j j > sort-in
>>  $ wc -c sort-in
>>  2839999968 sort-in
>
> I had to use "seq -f %.0f" to get this filesize.

Odd.
Here's what those generate for me:

  $ seq 9999999 > k
  $ wc -c < k
  78888888

  $ tail -1 k
  9999999

The remaining "cat" commands merely write 36 copies of that data to sort-in:

  $ (wc -c < k|tr -d '\n'; echo '* 36')|bc
  2839999968

What happens differently for you?


_______________________________________________
Bug-coreutils mailing list
Bug-coreutils@gnu.org
http://lists.gnu.org/mailman/listinfo/bug-coreutils

Reply via email to