I suppose we could do with a simple identity mapping/identity reducing
example/tool that can easily be reutilized for purposes such as these.
Could you file a JIRA on this?
The -text is like -cat but has codec and some file format detection.
Hopefully it should work for your case.
On Fri, Aug 5,
Instead of hd fs -put hundreds of files of X megs, I want to do it once on a
gzipped (or zipped) archive, one file, much smaller total megs. Then I want to
decompress the archive on HDFS? I can't figure out what hd fs type command
would do such a thing.
Thanks.
Keith,
The 'hadoop fs -text' tool does decompress a file given to it if
needed/able, but what you could also do is run a distributed mapreduce
job that converts from compressed to decompressed, that'd be much
faster.
On Fri, Aug 5, 2011 at 4:58 AM, Keith Wiley kwi...@keithwiley.com wrote: