Error while building Hadoop-Yarn

2011-08-19 Thread rajesh putta
Hi,
I am using apache-maven-3.0.3 and i have set LD_LIBRARY_PATH=/usr/local/lib
which has google protocbuf library.
I am getting following error while building hadoop-yarn using mvn clean
install -DskipTests=true

[INFO] hadoop-yarn-api ... SUCCESS [14.904s]
[INFO] hadoop-yarn-common  SUCCESS [8.787s]
[INFO] hadoop-yarn-server-common . SUCCESS [4.691s]
[INFO] hadoop-yarn-server-nodemanager  FAILURE [6.051s]
[INFO] hadoop-yarn-server-resourcemanager  SKIPPED
[INFO] hadoop-yarn-server-tests .. SKIPPED
[INFO] hadoop-yarn-server  SKIPPED
[INFO] hadoop-yarn ... SKIPPED
[INFO]

[INFO] BUILD FAILURE
[INFO]

[INFO] Total time: 34.870s
[INFO] Finished at: Fri Aug 19 11:48:22 IST 2011
[INFO] Final Memory: 44M/107M

[ERROR] Failed to execute goal
org.codehaus.mojo:make-maven-plugin:1.0-beta-1:autoreconf (autoreconf) on
project hadoop-yarn-server-nodemanager: autoreconf command returned an exit
value != 0. Aborting build; see debug output for more information. - [Help
1]

Thanks in advance

Thanks  Regards
Rajesh Putta


Re: Error while building Hadoop-Yarn

2011-08-19 Thread rajesh putta
Thanks Arun,
 Now its working fine




Thanks  Regards
Rajesh Putta
Development Engineer
Pramati Technologies

On Fri, Aug 19, 2011 at 12:25 PM, Arun Murthy a...@hortonworks.com wrote:

 That means you don't have the autotool chain necessary for build the
 native code.

 For now pass -P-cbuild to skip them.

 Arun

 Sent from my iPhone

 On Aug 18, 2011, at 11:26 PM, rajesh putta rajesh.p...@gmail.com wrote:

  Hi,
  I am using apache-maven-3.0.3 and i have set
 LD_LIBRARY_PATH=/usr/local/lib
  which has google protocbuf library.
  I am getting following error while building hadoop-yarn using mvn clean
  install -DskipTests=true
 
  [INFO] hadoop-yarn-api ... SUCCESS
 [14.904s]
  [INFO] hadoop-yarn-common  SUCCESS
 [8.787s]
  [INFO] hadoop-yarn-server-common . SUCCESS
 [4.691s]
  [INFO] hadoop-yarn-server-nodemanager  FAILURE
 [6.051s]
  [INFO] hadoop-yarn-server-resourcemanager  SKIPPED
  [INFO] hadoop-yarn-server-tests .. SKIPPED
  [INFO] hadoop-yarn-server  SKIPPED
  [INFO] hadoop-yarn ... SKIPPED
  [INFO]
  
  [INFO] BUILD FAILURE
  [INFO]
  
  [INFO] Total time: 34.870s
  [INFO] Finished at: Fri Aug 19 11:48:22 IST 2011
  [INFO] Final Memory: 44M/107M
 
  [ERROR] Failed to execute goal
  org.codehaus.mojo:make-maven-plugin:1.0-beta-1:autoreconf (autoreconf) on
  project hadoop-yarn-server-nodemanager: autoreconf command returned an
 exit
  value != 0. Aborting build; see debug output for more information. -
 [Help
  1]
 
  Thanks in advance
 
  Thanks  Regards
  Rajesh Putta



RE: Wonky reduce progress

2011-08-19 Thread Vinayakumar B

Please check the defect in MAPREDUCE jira
https://issues.apache.org/jira/browse/MAPREDUCE-2264

This is because the compression is enabled for map outputs and statistics
are taken on compressed data instead of original data.

-Original Message-
From: Joey Echeverria [mailto:j...@cloudera.com] 
Sent: Friday, August 19, 2011 3:07 AM
To: d...@hive.apache.org
Subject: Wonky reduce progress

I'm seeing really weird numbers in the JobTracker when monitoring jobs
launched by hives. Some jobs who above 100% completion while running
and others show less than 100% after the job completed. I tried
searching for JIRAs but didn't see any matches.

Has anyone else run into this?

-Joey

-- 
Joseph Echeverria
Cloudera, Inc.
443.305.9434



SequenceFileInputFormat doesn't return whole records

2011-08-19 Thread Tim Fletcher
Hi all,

I am having issues using SequenceFileInputFormat to retrieve whole records

I have 1 job that is used to write to a SequenceFile

SequenceFileOutputFormat.setOutputPath(job, new Path(out/data));
 SequenceFileOutputFormat.setOutputCompressionType(job,
SequenceFile.CompressionType.NONE);

I then have a second job that is ment to read the file for processing

SequenceFileInputFormat.addInputPath(job, new Path(out/data));

However, the values that i get as the arguments to the Map part of my job
only seems to contain parts of the record. I am sure that i am missing
something rather fundamental as to how Hadoop splits inputs to the Mapper,
but can't seem to find a way to stop the records being split.

Any help (or a pointer to a specific page in the doc) would be greatly
appreciated

Regards,
Tim


Re: SequenceFileInputFormat doesn't return whole records

2011-08-19 Thread Harsh J
Tim,

Do you also set your I/O formats explicitly to SequenceFileInputFormat
and SequenceFileOutputFormat? Via job.setInputFormat/setOutputFormat I
mean.

Hadoop should not be splitting records across maps/mappers. There are
specific test cases that ensure this does not happen, so it would seem
strange if it does this.

On Fri, Aug 19, 2011 at 6:01 PM, Tim Fletcher zigomu...@gmail.com wrote:
 Hi all,
 I am having issues using SequenceFileInputFormat to retrieve whole records
 I have 1 job that is used to write to a SequenceFile
 SequenceFileOutputFormat.setOutputPath(job, new Path(out/data));
 SequenceFileOutputFormat.setOutputCompressionType(job,
 SequenceFile.CompressionType.NONE);
 I then have a second job that is ment to read the file for processing
 SequenceFileInputFormat.addInputPath(job, new Path(out/data));
 However, the values that i get as the arguments to the Map part of my job
 only seems to contain parts of the record. I am sure that i am missing
 something rather fundamental as to how Hadoop splits inputs to the Mapper,
 but can't seem to find a way to stop the records being split.
 Any help (or a pointer to a specific page in the doc) would be greatly
 appreciated
 Regards,
 Tim



-- 
Harsh J


Re: SequenceFileInputFormat doesn't return whole records

2011-08-19 Thread Tim Fletcher
Harsh, that was exactly the issue!

Thanks very much for your help
Tim

On 19 August 2011 15:15, Harsh J ha...@cloudera.com wrote:

 Tim,

 Do you also set your I/O formats explicitly to SequenceFileInputFormat
 and SequenceFileOutputFormat? Via job.setInputFormat/setOutputFormat I
 mean.

 Hadoop should not be splitting records across maps/mappers. There are
 specific test cases that ensure this does not happen, so it would seem
 strange if it does this.

 On Fri, Aug 19, 2011 at 6:01 PM, Tim Fletcher zigomu...@gmail.com wrote:
  Hi all,
  I am having issues using SequenceFileInputFormat to retrieve whole
 records
  I have 1 job that is used to write to a SequenceFile
  SequenceFileOutputFormat.setOutputPath(job, new Path(out/data));
  SequenceFileOutputFormat.setOutputCompressionType(job,
  SequenceFile.CompressionType.NONE);
  I then have a second job that is ment to read the file for processing
  SequenceFileInputFormat.addInputPath(job, new Path(out/data));
  However, the values that i get as the arguments to the Map part of my job
  only seems to contain parts of the record. I am sure that i am missing
  something rather fundamental as to how Hadoop splits inputs to the
 Mapper,
  but can't seem to find a way to stop the records being split.
  Any help (or a pointer to a specific page in the doc) would be greatly
  appreciated
  Regards,
  Tim



 --
 Harsh J



Re: Wonky reduce progress

2011-08-19 Thread Joey Echeverria
Yup, that would do it. Thanks.

-Joey

On Fri, Aug 19, 2011 at 4:14 AM, Vinayakumar B vinayakuma...@huawei.com wrote:

 Please check the defect in MAPREDUCE jira
 https://issues.apache.org/jira/browse/MAPREDUCE-2264

 This is because the compression is enabled for map outputs and statistics
 are taken on compressed data instead of original data.

 -Original Message-
 From: Joey Echeverria [mailto:j...@cloudera.com]
 Sent: Friday, August 19, 2011 3:07 AM
 To: d...@hive.apache.org
 Subject: Wonky reduce progress

 I'm seeing really weird numbers in the JobTracker when monitoring jobs
 launched by hives. Some jobs who above 100% completion while running
 and others show less than 100% after the job completed. I tried
 searching for JIRAs but didn't see any matches.

 Has anyone else run into this?

 -Joey

 --
 Joseph Echeverria
 Cloudera, Inc.
 443.305.9434





-- 
Joseph Echeverria
Cloudera, Inc.
443.305.9434