Re: Probelms getting Eclipse Hadoop plugin to work.

2009-02-19 Thread Rasit OZDAS
Erik,
Try to add following properties into hadoop-site.xml:

property
namefs.default.name/name
valuehdfs://ip_address:9000/value
/property
property
namemapred.job.tracker/name
valuehdfs://ip_address:9001/value
/property

This way your ports become static. Then use port 9001 for MR, 9000 for HDFS
in your properties window.
If it still doesn't work, try to write ip address instead of host name as
target host.

Hope this helps,
Rasit

2009/2/18 Erik Holstad erikhols...@gmail.com

 I'm using Eclipse 3.3.2 and want to view my remote cluster using the Hadoop
 plugin.
 Everything shows up and I can see the map/reduce perspective but when
 trying
 to
 connect to a location I get:
 Error: Call failed on local exception

 I've set the host to for example xx0, where xx0 is a remote machine
 accessible from
 the terminal, and the ports to 50020/50040 for M/R master and
 DFS master respectively. Is there anything I'm missing to set for remote
 access to the
 Hadoop cluster?

 Regards Erik




-- 
M. Raşit ÖZDAŞ


where is the log when using eclipse plugin

2009-02-19 Thread Nick Cen
The log message can be found from the stdout log file when i am using
standard hadoop command. But when using the eclipse plugin and run from the
eclipse, no output mess ,age is capture, btw, i use System.out.println to
emit log message.

-- 
http://daily.appspot.com/food/


Re: Persistent completed jobs status not showing in jobtracker UI

2009-02-19 Thread Bill Au
The idea behind persistent completed jobs status is that they are still
available in case of a jobtracker failure, right?  The job history is stored
as files under the log directory on the jobtracker node.  So they won't be
available when the jobtraker node fails.  For folks who have enabled this
feature, how do you retrieve the presistended completed jobs status?

Bill

On Wed, Feb 18, 2009 at 10:48 PM, Amareshwari Sriramadasu 
amar...@yahoo-inc.com wrote:

 Bill Au wrote:

 I have enabled persistent completed jobs status and can see them in HDFS.
 However, they are not listed in the jobtracker's UI after the jobtracker
 is
 restarted.  I thought that jobtracker will automatically look in HDFS if
 it
 does not find a job in its memory cache.  What am I missing?  How to I
 retrieve the persistent completed job status?

 Bill



 JobTracker web ui doesn't look at persistence storage after a restart. You
 can access the old jobs from job history. History link is accesible from web
 ui.
 -Amareshwari




Re: Probelms getting Eclipse Hadoop plugin to work.

2009-02-19 Thread Norbert Burger
What platform are you running Eclipse on?  If Windows, see this thread
regarding Cygwin:

http://www.mail-archive.com/core-user@hadoop.apache.org/msg07669.html

For my case, I've never had to touch any of the plugin's advanced
parameters.  Usually, setting just the Map/Reduce Master and DFS Master (and
associated ports) has been sufficient.

Norbert

On 2/18/09, Erik Holstad erikhols...@gmail.com wrote:

 I'm using Eclipse 3.3.2 and want to view my remote cluster using the Hadoop
 plugin.
 Everything shows up and I can see the map/reduce perspective but when
 trying
 to
 connect to a location I get:
 Error: Call failed on local exception

 I've set the host to for example xx0, where xx0 is a remote machine
 accessible from
 the terminal, and the ports to 50020/50040 for M/R master and
 DFS master respectively. Is there anything I'm missing to set for remote
 access to the
 Hadoop cluster?

 Regards Erik



Re: Probelms getting Eclipse Hadoop plugin to work.

2009-02-19 Thread Erik Holstad
Thanks guys!
Running Linux and the remote cluster is also Linux.
I have the properties set up like that already on my remote cluster, but
not sure where to input this info into Eclipse.
And when changing the ports to 9000 and 9001 I get:

Error: java.io.IOException: Unknown protocol to job tracker:
org.apache.hadoop.dfs.ClientProtocol

Regards Erik


Re: Map/Recuce Job done locally?

2009-02-19 Thread Erik Holstad
Hey Philipp!
MR jobs are run locally if you just run the java file, to get it running in
distributed mode
you need to create a job jar and run that like ./bin/hadoop jar ...

Regards Erik


Re: Map/Recuce Job done locally?

2009-02-19 Thread Philipp Dobrigkeit
Hi Erik,

thank you, this list is really quick and helpful. I will try that.

Is it enough to just create a .jar file containing my Class and how would I 
start such a job from another program, when I want to track for example 
execution time? Do I run some kind of system.exec(path/hadoop jar MyClass 
Param)?

Best,
Philipp


 Original-Nachricht 
 Datum: Thu, 19 Feb 2009 13:42:41 -0800
 Von: Erik Holstad erikhols...@gmail.com
 An: core-user@hadoop.apache.org
 Betreff: Re: Map/Recuce Job done locally?

 Hey Philipp!
 MR jobs are run locally if you just run the java file, to get it running
 in
 distributed mode
 you need to create a job jar and run that like ./bin/hadoop jar ...
 
 Regards Erik

-- 
Pt! Schon vom neuen GMX MultiMessenger gehört? Der kann`s mit allen: 
http://www.gmx.net/de/go/multimessenger01


Re: Map/Recuce Job done locally?

2009-02-19 Thread Erik Holstad
Hey Philipp!
Not sure about your time tracking thing, probably works, I've just used a
bash script
to start the jar and then you can do the timing in the script.
About how to compile the jars, you need to include the dependencies too, but
you will see what you are missing when you run the job.

Regards Erik


empty log file...

2009-02-19 Thread zander1013

hi,
 
i am setting up hadoop for the first time on multi-node cluster. right now i
have two nodes. the two node cluster consists of two laptops connected via
ad-hoc wifi network. they they do not have access to the internet. i
formated the datanodes on both machines prior to startup...

output form the commands /usr/local/hadoop/bin/start-all.sh, jps (on both
machines), and /usr/local/hadoop/bin/stop-all.sh all appear normal. however
the file /usr/local/hadoop/logs/hadoop-hadoop-datanode-node1.log (the slave
node) is empty.

the same file for the master node shows the startup and shutdown events as
normal and without error.

is it okay that the log file on the slave is empty?

zander
-- 
View this message in context: 
http://www.nabble.com/empty-log-file...-tp22113398p22113398.html
Sent from the Hadoop core-user mailing list archive at Nabble.com.



How to use Hadoop API to submit job?

2009-02-19 Thread Wu Wei

Hi,

I used to submit Hadoop job with the utility RunJar.main() on hadoop 
0.18. On hadoop 0.19, because the commandLineConfig of JobClient was 
null, I got a NullPointerException error when RunJar.main() calls 
GenericOptionsParser to get libJars (0.18 didn't do this call). I also 
tried the class JobShell to submit job, but it catches all exceptions 
and sends to stderr so that I cann't handle the exceptions myself.


I noticed that if I can call JobClient's setCommandLineConfig method, 
everything goes easy. But this method has default package accessibility, 
I cann't see the method out of package org.apache.hadoop.mapred.


Any advices on using Java APIs to submit job?

Wei