Re: Mystery, A Tale of Two Reducers

2011-06-17 Thread Geoffry Roberts
This is for the edification of the group. The clone solution worked. Here's how I handled it. Second Reducer (redux) : protected void reduce(Text key, Iterable visitors, Context ctx) throws IOException, InterruptedException { List list = new ArrayList(); for (Text visitor : visitors) {

Re: Mystery, A Tale of Two Reducers

2011-06-17 Thread Harsh J
Geoffry, The problem here is that the Reducer in Hadoop reuses the same container object to pass on all values and keys. Thus, what you're really holding in your second reducer's code are "References" to this object -> Which upon writing will all be a mess of duplicates and what not cause they are

Mystery, A Tale of Two Reducers

2011-06-17 Thread Geoffry Roberts
All, I have come across a situation that I don't understand. *First Reducer: *Behold the first of two reducers. A fragment of it's output follows. Simple no? It doesn't do anything. I've highlighted two records from the output. Keep them in mind. Now lets look at the second reducer. * *prot

Re: Query about "hadoop dfs -cat" in hadoop-0-0.20.2

2011-06-17 Thread Lemon Cheng
Nothing can see of slaves command. Am I missing something ? Background: When my first time installation of hadoop on last month, i followed the instructions of mapreduce wordcount example, and it works. And this is the second time i use, the computer is restarted, and call bin/s

Re: Query about "hadoop dfs -cat" in hadoop-0-0.20.2

2011-06-17 Thread Marcos Ortiz
On 06/17/2011 09:51 AM, Lemon Cheng wrote: Hi, Thanks for your reply. I am not sure that. How can I prove that? Which is your dfs.tmp.dir and dfs.data.dir values? You can check the DataNodes“s health with bin/slaves.sh jps | grep Datanode | sort Which is the output of bin/hadoop dfsadmin -r

Re: Query about "hadoop dfs -cat" in hadoop-0-0.20.2

2011-06-17 Thread Lemon Cheng
[appuser@localhost bin]$ ./hadoop dfs -ls Found 2 items drwxr-xr-x - appuser supergroup 0 2011-04-26 09:41 /user/appuser/input drwxr-xr-x - appuser supergroup 0 2011-04-26 09:42 /user/appuser/output [appuser@localhost bin]$ ./hadoop dfs -ls /usr/lemon/wordcount/input Found 2 i

Re: Query about "hadoop dfs -cat" in hadoop-0-0.20.2

2011-06-17 Thread Mostafa Gaber
Can you send us the output of "hadoop dfs -ls"? On Jun 17, 2011, at 10:21, Lemon Cheng wrote: > Hi, > > Thanks for your reply. > I am not sure that. How can I prove that? > > I checked the localhost:50070, it shows 1 live node and 0 dead node. > And the log "hadoop-appuser-datanode-localhos

Re: Single Reducer Times Out During Shuffle

2011-06-17 Thread Shaun Martinec
On Fri, Jun 17, 2011 at 10:29 AM, Marcos Ortiz wrote: > On 06/17/2011 08:50 AM, Shaun Martinec wrote: > > I have a MapReduce job that is failing occasionally at the reduce > phase. I believe it's during the shuffle phase, but am not positive. I > have copied the end of the job log below. As you ca

Re: Query about "hadoop dfs -cat" in hadoop-0-0.20.2

2011-06-17 Thread Lemon Cheng
Hi, Thanks for your reply. I am not sure that. How can I prove that? I checked the localhost:50070, it shows 1 live node and 0 dead node. And the log "hadoop-appuser-datanode-localhost.localdomain.log" shows: / 2011-06-17 19:59:38,658

Re: Single Reducer Times Out During Shuffle

2011-06-17 Thread Marcos Ortiz
On 06/17/2011 08:50 AM, Shaun Martinec wrote: I have a MapReduce job that is failing occasionally at the reduce phase. I believe it's during the shuffle phase, but am not positive. I have copied the end of the job log below. As you can see, I have a very large number of maps (2910) and only 1 red

Single Reducer Times Out During Shuffle

2011-06-17 Thread Shaun Martinec
I have a MapReduce job that is failing occasionally at the reduce phase. I believe it's during the shuffle phase, but am not positive. I have copied the end of the job log below. As you can see, I have a very large number of maps (2910) and only 1 reducer that is used to cat the results together (/

Re: Query about "hadoop dfs -cat" in hadoop-0-0.20.2

2011-06-17 Thread Marcos Ortiz
On 06/17/2011 07:41 AM, Lemon Cheng wrote: Hi, I am using the hadoop-0.20.2. After calling ./start-all.sh, i can type "hadoop dfs -ls". However, when i type "hadoop dfs -cat /usr/lemon/wordcount/input/file01", the error is shown as follow. I have searched the related problem in the web, but i

Query about "hadoop dfs -cat" in hadoop-0-0.20.2

2011-06-17 Thread Lemon Cheng
Hi, I am using the hadoop-0.20.2. After calling ./start-all.sh, i can type "hadoop dfs -ls". However, when i type "hadoop dfs -cat /usr/lemon/wordcount/input/file01", the error is shown as follow. I have searched the related problem in the web, but i can't find a solution for helping me to solve t