It does work for me. I have to BOTH ship the extra jar using -file AND
include in classpath on local system (via setting HADOOP_CLASSPATH).
I'm not sure what "nothing happened" means. BTW, I'm using the 0.16.2
release.
On Friday 04 April 2008 10:19:54 am Francesco Tamberi wrote:
> I already tr
On 4/4/08 10:18 AM, "Francesco Tamberi" <[EMAIL PROTECTED]> wrote:
> Thank for your fast reply!
>
> Ted Dunning ha scritto:
>> Take a looks at the way that the text input format moves to the next line
>> after a split point.
>>
>>
> I'm not sure to understand.. is my way correct or are you
I already tried that... nothing happened...
Thank you,
-- Francesco
Ted Dunning ha scritto:
I saw that, but I don't know if it will put a jar into the classpath at the
other end.
On 4/4/08 9:56 AM, "Yuri Pradkin" <[EMAIL PROTECTED]> wrote:
There is a -file option to streaming that
-file
Thank for your fast reply!
Ted Dunning ha scritto:
Take a looks at the way that the text input format moves to the next line
after a split point.
I'm not sure to understand.. is my way correct or are you suggesting
another one?
There are a couple of possible problems with your input format
I saw that, but I don't know if it will put a jar into the classpath at the
other end.
On 4/4/08 9:56 AM, "Yuri Pradkin" <[EMAIL PROTECTED]> wrote:
> There is a -file option to streaming that
> -file File/dir to be shipped in the Job jar file
>
> On Friday 04 April 2008 09:24:59 am Te
There is a -file option to streaming that
-file File/dir to be shipped in the Job jar file
On Friday 04 April 2008 09:24:59 am Ted Dunning wrote:
> At one point, it
> was necessary to unpack the streaming.jar file and put your own classes and
> jars into that. Last time I looked
Take a looks at the way that the text input format moves to the next line
after a split point.
There are a couple of possible problems with your input format not found
problem.
First, is your input in a package? If so, you need to provide a complete
name for the class.
Secondly, you have to gi
Hi All,
I have a streaming tool chain written in c++/python that performs some
operations on really big text files (gigabytes order); the chain reads files
and writes its result to standard output.
The chain needs to read well structured files and so I need to control how
hadoop splits files: i