Hey Andy,
You're correct that 0.20.203 doesn't have append. Your best bet is to
build a version of the append branch or switch to
CDH3u0.
-Joey
On Tue, Jun 7, 2011 at 6:31 PM, Zhong, Sheng wrote:
> Thanks! The issue has been resolved by removing some bad blks...
>
> But St.Ack,
>
> We do want a
I assumed before reading the split API that it is the actual split, my bad.
Thanks alot Harsh, it's working great!
Mark
any comments?
From: Shantian Purkad
To: "common-user@hadoop.apache.org"
Sent: Tuesday, June 7, 2011 3:53 PM
Subject: Linear scalability question
Hi,
I have a question on the linear scalability of Hadoop.
We have a situation where we have to do reduce side jo
Hi,
I am trying to rewrite and improve some old code using Map side join
such as TupleWritable, KeyValueTextInputFormat, etc. The reference
materials I have are based on old API (0.19.x). Since Hadoop is
updating rapidly, I am wondering is there any new functions / API /
framework about Map
Or not. I misread your error, that shouldn't be it, sorry.
On Wed, Jun 8, 2011 at 10:21 PM, Harsh J wrote:
> Madhu,
>
> Are you multi-threading in your Reducer code by any chance? MTOF may
> not be thread-safe in the release you're using. Using MultipleOutputs
> is recommended right now, if this
Madhu,
Are you multi-threading in your Reducer code by any chance? MTOF may
not be thread-safe in the release you're using. Using MultipleOutputs
is recommended right now, if this is the cause/case.
On Wed, Jun 8, 2011 at 7:58 PM, Madhu Ramanna wrote:
> Hello,
>
> We're using CDH3b3 0.20.2 hadoo
Mark,
The InputSplit is something of a meta class you ought to use to get
path, offset and length information from. Your RecordReader
implementation in the InputFormat would ideally be wrapping two
instantiated RecordReaders made from the same InputSplit meta
information. The InputSplit object doe
I have a question though for Harsh case... I wrote my custom inputFormat
which will create an array of recordReaders and give them to the MapRunner.
Will that mean multiple copies of the inputSplit are all in memory? or will
there be one copy pointed by all of them .. as if they were pointers ?
T
Hi
I am experiencing a lot of tasks failures while running any Hadoop
application.
In particular, I get the following warnings:
Error initializing attempt_201106081500_0018_r_00_0:
java.io.IOException: Could not obtain block: blk_-7386162385184325734_1214
file=/home/hadoop/data/mapred/system/jo
Thanks for the replies, but input doesn't have 'clone' I don't know why ...
so I'll have to write my custom inputFormat ... I was hoping for an easier
way though.
Thank you,
Mark
On Wed, Jun 8, 2011 at 1:58 AM, Harsh J wrote:
> Or if that does not work for any reason (haven't tried it really),
Hello,
We're using CDH3b3 0.20.2 hadoop. In our map reduce jobs we've extended
MultipleTextOutputFormat to override checkOutputSpecs() and
generateFileNameForKeyValue() returning
relative path based on key. I don't have multiple jobs running with the same
output directory. When I rerun it succe
Hi again:
Good news!! What I posted before is not an error. I guess I was tired and
didn't realize... It came down on me by chance. There are no errors, it does
not get stuck, it's the process itself running. I supposed that some kind of
message would come up to let the user know that the fuse-dfs
Or if that does not work for any reason (haven't tried it really), try
writing your own InputFormat wrapper where in you can have direct
access to the InputSplit object to do what you want to (open two
record readers, and manage them separately).
On Wed, Jun 8, 2011 at 1:48 PM, Stefan Wienert wro
Try input.clone()...
2011/6/8 Mark question :
> Hi,
>
> I'm trying to read the inputSplit over and over using following function
> in MapperRunner:
>
> @Override
> public void run(RecordReader input, OutputCollector output, Reporter
> reporter) throws IOException {
>
> RecordReader copyInpu
I tried to run the wordcount example in C++ using Hadoop-0.20.203 on two
nodes cluster. My two nodes are on Fedora. Compilation works fine, I got no
errors and no warnings. But when I tried to run my program, I get an error
like this :
bin/hadoop pipes -conf ../dev/word.xml -input testtile.txt -o
I am in a same situation as you. I tried to run simple Wordcount example
using Hadoop-0.20.203. I did instructions from Hadoop tutorial to compile
example with ant and at the configure step, an error appeared :
configure: error: Cannot find libssl.so
I do not know what is going wrong, while I ha
16 matches
Mail list logo