Build failed in Jenkins: Hadoop-Common-trunk #945

2013-11-07 Thread Apache Jenkins Server
See https://builds.apache.org/job/Hadoop-Common-trunk/945/changes Changes: [cnauroth] HADOOP-9660. Update CHANGES.txt. [cnauroth] MAPREDUCE-5451. MR uses LD_LIBRARY_PATH which doesn't mean anything in Windows. Contributed by Yingda Chen. --

Re: [DISCUSS] What is the purpose of merge vote threads?

2013-11-07 Thread Chris Nauroth
Thank you to everyone who replied. Even though it sounds like there is not complete consensus on some of the finer points, I think I have a clearer understanding on how to participate now. I do think posting all requirements in jira before calling the merge vote makes the process more effective.

Next releases

2013-11-07 Thread Arun C Murthy
Gang, Thinking through the next couple of releases here, appreciate f/b. # hadoop-2.2.1 I was looking through commit logs and there is a *lot* of content here (81 commits as on 11/7). Some are features/improvements and some are fixes - it's really hard to distinguish what is important and

release-2.2.0 tag?

2013-11-07 Thread Mohammad Islam
Hi, I could not find out any release tag for 2.2.0. Which branch should I use instead for latest released version? branch-2.2.0 or branch-2.2? Regards, Mohammad

Unable to perform terasort for 50GB of data

2013-11-07 Thread Khai Cher LIM (NYP)
Dear all, I have just started learning Hadoop setup and I am having problem with running terasort on my Hadoop cluster. My input folder contains 50 GB of data but when I run the terasort, the tasks failed and it gave me the error message as shown in the following screenshot.

Re: release-2.2.0 tag?

2013-11-07 Thread Tsuyoshi OZAWA
Hi Mohammad, IIUC, 2.2.0 is the latest released version. branch-2.2 is the latest branch under development. On Fri, Nov 8, 2013 at 11:52 AM, Mohammad Islam misla...@yahoo.com wrote: Hi, I could not find out any release tag for 2.2.0. Which branch should I use instead for latest released

Re: Unable to perform terasort for 50GB of data

2013-11-07 Thread inelu nagamallikarjuna
Hai, Check the individual data nodes usage: Hadoop dfsadmin -report And moreover override the config parameter mapred.local.dir to store intermediate data in some path rather than /tmp directory and don't use single reducer, increase no of reducers and use totalorderpartitioner Thanks