Re: Hbase startup error: NoNode for /hbase/master after running out of space

2011-06-08 Thread Joey Echeverria
Hey Andy, You're correct that 0.20.203 doesn't have append. Your best bet is to build a version of the append branch or switch to CDH3u0. -Joey On Tue, Jun 7, 2011 at 6:31 PM, Zhong, Sheng wrote: > Thanks! The issue has been resolved by removing some bad blks... > > But St.Ack, > > We do want a

Re: re-reading

2011-06-08 Thread Mark question
I assumed before reading the split API that it is the actual split, my bad. Thanks alot Harsh, it's working great! Mark

Re: Linear scalability question

2011-06-08 Thread Shantian Purkad
any comments? From: Shantian Purkad To: "common-user@hadoop.apache.org" Sent: Tuesday, June 7, 2011 3:53 PM Subject: Linear scalability question Hi, I have a question on the linear scalability of Hadoop. We have a situation where we have to do reduce side jo

New API for TupleWritable / Mapside join

2011-06-08 Thread Shi Yu
Hi, I am trying to rewrite and improve some old code using Map side join such as TupleWritable, KeyValueTextInputFormat, etc. The reference materials I have are based on old API (0.19.x). Since Hadoop is updating rapidly, I am wondering is there any new functions / API / framework about Map

Re: MultipleTextOutputFormat - temp dir does not exist error

2011-06-08 Thread Harsh J
Or not. I misread your error, that shouldn't be it, sorry. On Wed, Jun 8, 2011 at 10:21 PM, Harsh J wrote: > Madhu, > > Are you multi-threading in your Reducer code by any chance? MTOF may > not be thread-safe in the release you're using. Using MultipleOutputs > is recommended right now, if this

Re: MultipleTextOutputFormat - temp dir does not exist error

2011-06-08 Thread Harsh J
Madhu, Are you multi-threading in your Reducer code by any chance? MTOF may not be thread-safe in the release you're using. Using MultipleOutputs is recommended right now, if this is the cause/case. On Wed, Jun 8, 2011 at 7:58 PM, Madhu Ramanna wrote: > Hello, > > We're using CDH3b3 0.20.2 hadoo

Re: re-reading

2011-06-08 Thread Harsh J
Mark, The InputSplit is something of a meta class you ought to use to get path, offset and length information from. Your RecordReader implementation in the InputFormat would ideally be wrapping two instantiated RecordReaders made from the same InputSplit meta information. The InputSplit object doe

Re: re-reading

2011-06-08 Thread Mark question
I have a question though for Harsh case... I wrote my custom inputFormat which will create an array of recordReaders and give them to the MapRunner. Will that mean multiple copies of the inputSplit are all in memory? or will there be one copy pointed by all of them .. as if they were pointers ? T

/etc/hosts related error?

2011-06-08 Thread bikash sharma
Hi I am experiencing a lot of tasks failures while running any Hadoop application. In particular, I get the following warnings: Error initializing attempt_201106081500_0018_r_00_0: java.io.IOException: Could not obtain block: blk_-7386162385184325734_1214 file=/home/hadoop/data/mapred/system/jo

Re: re-reading

2011-06-08 Thread Mark question
Thanks for the replies, but input doesn't have 'clone' I don't know why ... so I'll have to write my custom inputFormat ... I was hoping for an easier way though. Thank you, Mark On Wed, Jun 8, 2011 at 1:58 AM, Harsh J wrote: > Or if that does not work for any reason (haven't tried it really),

MultipleTextOutputFormat - temp dir does not exist error

2011-06-08 Thread Madhu Ramanna
Hello, We're using CDH3b3 0.20.2 hadoop. In our map reduce jobs we've extended MultipleTextOutputFormat to override checkOutputSpecs() and generateFileNameForKeyValue() returning relative path based on key. I don't have multiple jobs running with the same output directory. When I rerun it succe

Re: error -2 (No such file or directory) when mounting fuse-dfs

2011-06-08 Thread elena.otero
Hi again: Good news!! What I posted before is not an error. I guess I was tired and didn't realize... It came down on me by chance. There are no errors, it does not get stuck, it's the process itself running. I supposed that some kind of message would come up to let the user know that the fuse-dfs

Re: re-reading

2011-06-08 Thread Harsh J
Or if that does not work for any reason (haven't tried it really), try writing your own InputFormat wrapper where in you can have direct access to the InputSplit object to do what you want to (open two record readers, and manage them separately). On Wed, Jun 8, 2011 at 1:48 PM, Stefan Wienert wro

Re: re-reading

2011-06-08 Thread Stefan Wienert
Try input.clone()... 2011/6/8 Mark question : > Hi, > >   I'm trying to read the inputSplit over and over using following function > in MapperRunner: > > @Override >    public void run(RecordReader input, OutputCollector output, Reporter > reporter) throws IOException { > >   RecordReader copyInpu

Server failed to authenticate, wordcount example

2011-06-08 Thread Oran Charles
I tried to run the wordcount example in C++ using Hadoop-0.20.203 on two nodes cluster. My two nodes are on Fedora. Compilation works fine, I got no errors and no warnings. But when I tried to run my program, I get an error like this : bin/hadoop pipes -conf ../dev/word.xml -input testtile.txt -o

Re: Linker errors with Hadoop pipes

2011-06-08 Thread Oran Charles
I am in a same situation as you. I tried to run simple Wordcount example using Hadoop-0.20.203. I did instructions from Hadoop tutorial to compile example with ant and at the configure step, an error appeared : configure: error: Cannot find libssl.so I do not know what is going wrong, while I ha