It is definitely easier to build a jar and use the hadoop script.  You can
do it yourself, though.  Just duplicate the line in bin/hadoop that runs
java and prefix it with "echo" to see what is happening.


On 11/1/07 1:37 PM, "Jim the Standing Bear" <[EMAIL PROTECTED]> wrote:

> Hi Ted,
> 
> It is funny how I am having so much difficulties with hadoop... since
> i am on this subject, let me ask another stupid question:
> 
> In order to run a hadoop job, I must zip it into a jar, and then use
> $HADOOP_HOME/bin/hadoop to launch it, correct?  I cannot simply use
> "java" to directly launch it, right?
> 
> best wishes
> 
> Yiming
> 
> 
> On 11/1/07, Ted Dunning <[EMAIL PROTECTED]> wrote:
>> 
>> When running a job standalone (typically for debugging), it suffices for me
>> to simply add -Dmapred.job.tracker=local to the program command line.
>> 
>> If you simply want to run the program on a single node, then you can just
>> stop the other node.  Using local storage with distributed programs is not
>> recommended.
>> 
>> 
>> On 11/1/07 12:35 PM, "Jim the Standing Bear" <[EMAIL PROTECTED]> wrote:
>> 
>>> Hello,
>>> 
>>> I am in need of some clarifications on how to run a hadoop job locally.
>>> 
>>> The cluster was originally set up to have two nodes, where one of them
>>> also acts as the master node and job tracker.
>>> 
>>> According to the wiki, I can run a job locally by altering
>>> "mapred.job.tracker" and "fs.default.name" properties to "local" in
>>> hadoop-site.xml.  But when I start the server, it stack dumped:
>>> 
>>> localhost: starting secondarynamenode, logging to /home/blahblahblah
>>> localhost: Exception in thread "main" java.lang.RuntimeException: Not
>>> a host:port pair: local
>>> 
>>> Apparently it didn't like the value "local"?
>>> 
>>> Also, the wiki noted that all these XML configuration files should be
>>> included somewhere in the class path to the job, does it mean I need
>>> to include the XMLs as I do jars?
>>> 
>>> Thank
>>> 
>>> -- Jim
>> 
>> 
> 

Reply via email to