Arun's attempt to send this bounced, so I'll try...
-- Owen
From: Arun C Murthy <[EMAIL PROTECTED]>
Date: Mon, 23 Oct 2006 23:41:30 +0530
Subject: Re: Error in reading block compressed sequence file
To: hadoop-dev@lucene.apache.org
On Sun, Oct 22, 2006 at 04:31:23PM -0700, Runping Qi wrote:
I got an error when I tried to read a blocked compressed sequence
file I
just created. Anybody can offer any clues?
The file was created through a SequenceFile.Writer object:
The writer was created as follows:
Configuration conf = new Configuration();
FileSystem fs = FileSystem.getNamed
("local", conf);
SequenceFile.Writer theWriter =
SequenceFile.createWriter(fs, conf, new Path(outP), Text.class,
ULTRecordJT.class,
CompressionType.BLOCK, new GzipCodec());
^^^^^^^^^^^^^^^^
You are using the GzipCodec which (for now, should change with
HADOOP-538) won't work with SequenceFiles. Can you replace that
with org.apache.hadoop.io.compress.DefaultCodec and give it a whirl?
The gzip basically is the zlib+headers, while the DefaultCodec is
zlib only... so compression should be identical but for the headers.
thanks,
Arun