Re: Run hadoop Map/Reduce app from another machine

2011-10-05 Thread Yang Xiaoliang
Install hadoop on your local machine, copy the configuration files from the
remote
hadoop culuster server to your local machine(including the hosts file), then
you can
just submit a *.jar locally as before.


2011/10/5 oleksiy gayduk.a.s...@mail.ru


 Hello,

 I'm trying to find a way how to run hadoop MapReduce app from another
 machine.
 For instance I have *.jar file with MapReduce app it works ok when I run it
 from command line for instance using this command: hadoop jar
 /usr/joe/wordcount.jar org.myorg.WordCount /usr/joe/wordcount/input
 /usr/joe/wordcount/output
 But in situation when I have another server (simple web app) where user can
 upload jar file, specify configuration for the MapReduce app and so on. And
 this server should interact with hadoop server. I mean somehow to upload
 this jar file to the hadoop server and run it with attributes.

 So, right now I see only one way of how to do this, it's upload jar file to
 the hadoop server and run remotely command: hadoop jar
 /usr/joe/wordcount.jar org.myorg.WordCount /usr/joe/wordcount/input
 /usr/joe/wordcount/output.

 So may be hadoop has spatial API for doing this kind of tasks remotely?
 --
 View this message in context:
 http://old.nabble.com/Run-hadoop-Map-Reduce-app-from-another-machine-tp32595264p32595264.html
 Sent from the Hadoop core-user mailing list archive at Nabble.com.




Re: hadoop input buffer size

2011-10-05 Thread Yang Xiaoliang
Hi,

Hadoop neither read one line each time, nor fetching dfs.block.size of lines
into a buffer,
Actually, for the TextInputFormat, it read io.file.buffer.size bytes of text
into a buffer each time,
this can be seen from the hadoop source file LineReader.java



2011/10/5 Mark question markq2...@gmail.com

 Hello,

  Correct me if I'm wrong, but when a program opens n-files at the same time
 to read from, and start reading from each file at a time 1 line at a time.
 Isn't hadoop actually fetching dfs.block.size of lines into a buffer? and
 not actually one line.

  If this is correct, I set up my dfs.block.size = 3MB and each line takes
 about 650 bytes only, then I would assume the performance for reading
 1-4000
 lines would be the same, but it isn't !  Do you know a way to find #n of
 lines to be read at once?

 Thank you,
 Mark



TestDFSIO error: libhdfs.so.1 does not exist

2011-07-28 Thread Yang Xiaoliang
Hi all,

I am benchmarking a Hadoop Cluster with the hadoop-*-test.jar TestDFSIO

but the following error returns:
File /usr/hadoop-0.20.2/libhdfs/libhdfs.so.1 does not exist.

How to solve this problem?

Thanks!


Re: Current available Memory

2011-02-24 Thread Yang Xiaoliang
Thanks a lot!

Yang Xiaoliang

2011/2/25 maha m...@umail.ucsb.edu

 Hi Yang,

  The problem could be solved using the following link:
 http://www.roseindia.net/java/java-get-example/get-memory-usage.shtml
  You need to use other memory managers like the Garbage collector and its
 finalize method to measure memory accurately.

  Good Luck,
   Maha

 On Feb 23, 2011, at 10:11 PM, Yang Xiaoliang wrote:

  I had also encuntered the smae problem a few days ago.
 
  any one has another method?
 
  2011/2/24 maha m...@umail.ucsb.edu
 
  Based on the Java function documentation, it gives approximately the
  available memory, so I need to tweak it with other functions.
  So it's a Java issue not Hadoop.
 
  Thanks anyways,
  Maha
 
  On Feb 23, 2011, at 6:31 PM, maha wrote:
 
  Hello Everyone,
 
  I'm using   Runtime.getRuntime().freeMemory() to see current memory
  available before and after creation of an object, but this doesn't seem
 to
  work well with Hadoop?
 
  Why? and is there another alternative?
 
  Thank you,
 
  Maha
 
 
 




Re: Current available Memory

2011-02-23 Thread Yang Xiaoliang
I had also encuntered the smae problem a few days ago.

any one has another method?

2011/2/24 maha m...@umail.ucsb.edu

 Based on the Java function documentation, it gives approximately the
 available memory, so I need to tweak it with other functions.
 So it's a Java issue not Hadoop.

 Thanks anyways,
 Maha

 On Feb 23, 2011, at 6:31 PM, maha wrote:

  Hello Everyone,
 
   I'm using   Runtime.getRuntime().freeMemory() to see current memory
 available before and after creation of an object, but this doesn't seem to
 work well with Hadoop?
 
  Why? and is there another alternative?
 
  Thank you,
 
  Maha