Ravishankar,

If you run $ jps, do you see a TaskTracker process running? Can you please
post the tasktracker logs as well?

On Thu, May 17, 2012 at 8:49 PM, Ravishankar Nair <
ravishankar.n...@gmail.com> wrote:

> Dear experts,
>
> Today is my tenth day working with Hadoop on installing on my windows
> machine. I am trying again and again because , some where someone has
> written that it works on Windows with CYGWIN.(And noone has written that
> Hadoop wont work on Windows). I am attaching my config files.
>
> Kindly help me, if anything can make this work.A feeble and humble request
> to all experts out there.
>
> Here is the error, if you search , you can see thousands have reported
> this and there is no solution I got yet, though I tried all ways possible.
> I am using Windows XP SP3, Hadoop (tried with five versions so far
> including 1.0.3).  I am running on a single node.(machine WSUSJXLHRN13067
> IP:- 192.168.0.16)
> When I start Hadoop, no issues in any of the versions
>
> rn13067@WSUSJXLHRN13067 /home/hadoop-1.0.3
> $ bin/start-all.sh
> starting namenode, logging to
> /home/hadoop-1.0.3/libexec/../logs/hadoop-SUNDOOP-namenode-WSUSJXLHRN13067.out
> localhost: starting datanode, logging to
> /home/hadoop-1.0.3/libexec/../logs/hadoop-SUNDOOP-datanode-WSUSJXLHRN13067.out
> localhost: starting secondarynamenode, logging to
> /home/hadoop-1.0.3/libexec/../logs/hadoop-SUNDOOP-secondarynamenode-WSUSJXLHRN13067.out
> starting jobtracker, logging to
> /home/hadoop-1.0.3/libexec/../logs/hadoop-SUNDOOP-jobtracker-WSUSJXLHRN13067.out
> localhost: starting tasktracker, logging to
> /home/hadoop-1.0.3/libexec/../logs/hadoop-SUNDOOP-tasktracker-WSUSJXLHRN13067.out
>
>
>
> When I run the example program, this is what is printed on my console:-
> $ bin/hadoop jar hadoop-examples-1.0.3.jar  grep input output 'dfs[a-z.]+'
> 12/05/17 21:44:46 WARN util.NativeCodeLoader: Unable to load native-hadoop
> library for your platform... using builtin-java classes where applicable
> 12/05/17 21:44:46 WARN snappy.LoadSnappy: Snappy native library not loaded
> 12/05/17 21:44:46 INFO mapred.FileInputFormat: Total input paths to
> process : 16
> 12/05/17 21:44:47 INFO mapred.JobClient: Running job: job_201205172141_0001
> 12/05/17 21:44:48 INFO mapred.JobClient:  map 0% reduce 0%
>
>
> Now it is HUNG!!. IN most of the versions this is the behaviour.
>
> Here is the log from JOBTRACKER:-
>
>
> 2012-05-17 21:41:28,037 INFO org.apache.hadoop.mapred.JobTracker:
> STARTUP_MSG:
> /************************************************************
> STARTUP_MSG: Starting JobTracker
> STARTUP_MSG:   host = WSUSJXLHRN13067/192.168.0.16
> STARTUP_MSG:   args = []
> STARTUP_MSG:   version = 1.0.3
> STARTUP_MSG:   build =
> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.0 -r
> 1335192; compiled by 'hortonfo' on Tue May  8 20:31:25 UTC 2012
> ************************************************************/
> 2012-05-17 21:41:28,147 INFO
> org.apache.hadoop.metrics2.impl.MetricsConfig: loaded properties from
> hadoop-metrics2.properties
> 2012-05-17 21:41:28,147 INFO
> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
> MetricsSystem,sub=Stats registered.
> 2012-05-17 21:41:28,162 INFO
> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot
> period at 10 second(s).
> 2012-05-17 21:41:28,162 INFO
> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: JobTracker metrics
> system started
> 2012-05-17 21:41:28,209 INFO
> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
> QueueMetrics,q=default registered.
> 2012-05-17 21:41:28,428 INFO
> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source ugi
> registered.
> 2012-05-17 21:41:28,428 WARN
> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Source name ugi already
> exists!
> 2012-05-17 21:41:28,428 INFO
> org.apache.hadoop.security.token.delegation.AbstractDelegationTokenSecretManager:
> Updating the current master key for generating delegation tokens
> 2012-05-17 21:41:28,428 INFO
> org.apache.hadoop.security.token.delegation.AbstractDelegationTokenSecretManager:
> Starting expired delegation token remover thread,
> tokenRemoverScanInterval=60 min(s)
> 2012-05-17 21:41:28,428 INFO
> org.apache.hadoop.security.token.delegation.AbstractDelegationTokenSecretManager:
> Updating the current master key for generating delegation tokens
> 2012-05-17 21:41:28,428 INFO org.apache.hadoop.mapred.JobTracker:
> Scheduler configured with (memSizeForMapSlotOnJT, memSizeForReduceSlotOnJT,
> limitMaxMemForMapTasks, limitMaxMemForReduceTasks) (-1, -1, -1, -1)
> 2012-05-17 21:41:28,428 INFO org.apache.hadoop.util.HostsFileReader:
> Refreshing hosts (include/exclude) list
> 2012-05-17 21:41:28,444 INFO org.apache.hadoop.mapred.JobTracker: Starting
> jobtracker with owner as rn13067
> 2012-05-17 21:41:28,475 INFO org.apache.hadoop.ipc.Server: Starting
> SocketReader
> 2012-05-17 21:41:28,475 INFO
> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
> RpcDetailedActivityForPort47111 registered.
> 2012-05-17 21:41:28,475 INFO
> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
> RpcActivityForPort47111 registered.
> 2012-05-17 21:41:28,522 INFO org.mortbay.log: Logging to
> org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via
> org.mortbay.log.Slf4jLog
> 2012-05-17 21:41:28,584 INFO org.apache.hadoop.http.HttpServer: Added
> global filtersafety
> (class=org.apache.hadoop.http.HttpServer$QuotingInputFilter)
> 2012-05-17 21:41:28,615 WARN org.apache.hadoop.util.NativeCodeLoader:
> Unable to load native-hadoop library for your platform... using
> builtin-java classes where applicable
> 2012-05-17 21:41:28,615 INFO org.apache.hadoop.http.HttpServer: Port
> returned by webServer.getConnectors()[0].getLocalPort() before open() is
> -1. Opening the listener on 50030
> 2012-05-17 21:41:28,615 INFO org.apache.hadoop.http.HttpServer:
> listener.getLocalPort() returned 50030
> webServer.getConnectors()[0].getLocalPort() returned 50030
> 2012-05-17 21:41:28,615 INFO org.apache.hadoop.http.HttpServer: Jetty
> bound to port 50030
> 2012-05-17 21:41:28,615 INFO org.mortbay.log: jetty-6.1.26
> 2012-05-17 21:41:28,834 INFO org.mortbay.log: Started
> SelectChannelConnector@0.0.0.0:50030
> 2012-05-17 21:41:28,834 INFO
> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source jvm
> registered.
> 2012-05-17 21:41:28,834 INFO
> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
> JobTrackerMetrics registered.
> 2012-05-17 21:41:28,850 INFO org.apache.hadoop.mapred.JobTracker:
> JobTracker up at: 47111
> 2012-05-17 21:41:28,850 INFO org.apache.hadoop.mapred.JobTracker:
> JobTracker webserver: 50030
> 2012-05-17 21:41:29,225 INFO org.apache.hadoop.mapred.JobTracker: Cleaning
> up the system directory
> 2012-05-17 21:41:29,772 INFO org.apache.hadoop.mapred.JobHistory: Creating
> DONE folder at file:/C:/cygwin/home/hadoop-1.0.3/logs/history/done
> 2012-05-17 21:41:29,787 INFO org.apache.hadoop.mapred.JobTracker: History
> server being initialized in embedded mode
> 2012-05-17 21:41:29,787 INFO org.apache.hadoop.mapred.JobHistoryServer:
> Started job history server at: localhost:50030
> 2012-05-17 21:41:29,787 INFO org.apache.hadoop.mapred.JobTracker: Job
> History Server web address: localhost:50030
> 2012-05-17 21:41:29,787 INFO
> org.apache.hadoop.mapred.CompletedJobStatusStore: Completed job store is
> inactive
> 2012-05-17 21:41:29,990 INFO org.apache.hadoop.mapred.JobTracker:
> Refreshing hosts information
> 2012-05-17 21:41:30,006 INFO org.apache.hadoop.util.HostsFileReader:
> Setting the includes file to
> 2012-05-17 21:41:30,006 INFO org.apache.hadoop.util.HostsFileReader:
> Setting the excludes file to
> 2012-05-17 21:41:30,006 INFO org.apache.hadoop.util.HostsFileReader:
> Refreshing hosts (include/exclude) list
> 2012-05-17 21:41:30,006 INFO org.apache.hadoop.mapred.JobTracker:
> Decommissioning 0 nodes
> 2012-05-17 21:41:30,006 INFO org.apache.hadoop.ipc.Server: IPC Server
> Responder: starting
> 2012-05-17 21:41:30,006 INFO org.apache.hadoop.ipc.Server: IPC Server
> listener on 47111: starting
> 2012-05-17 21:41:30,006 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 0 on 47111: starting
> 2012-05-17 21:41:30,006 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 1 on 47111: starting
> 2012-05-17 21:41:30,006 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 2 on 47111: starting
> 2012-05-17 21:41:30,006 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 3 on 47111: starting
> 2012-05-17 21:41:30,006 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 4 on 47111: starting
> 2012-05-17 21:41:30,006 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 5 on 47111: starting
> 2012-05-17 21:41:30,006 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 6 on 47111: starting
> 2012-05-17 21:41:30,006 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 7 on 47111: starting
> 2012-05-17 21:41:30,006 INFO org.apache.hadoop.mapred.JobTracker: Starting
> RUNNING
> 2012-05-17 21:41:30,006 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 8 on 47111: starting
> 2012-05-17 21:41:30,006 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 9 on 47111: starting
> 2012-05-17 21:44:47,069 INFO org.apache.hadoop.mapred.JobInProgress:
> job_201205172141_0001: nMaps=16 nReduces=1 max=-1
> 2012-05-17 21:44:47,069 INFO org.apache.hadoop.mapred.JobTracker: Job
> job_201205172141_0001 added successfully for user 'rn13067' to queue
> 'default'
> 2012-05-17 21:44:47,069 INFO org.apache.hadoop.mapred.JobTracker:
> Initializing job_201205172141_0001
> 2012-05-17 21:44:47,069 INFO org.apache.hadoop.mapred.JobInProgress:
> Initializing job_201205172141_0001
> 2012-05-17 21:44:47,069 INFO org.apache.hadoop.mapred.AuditLogger:
> USER=rn13067    IP=192.168.0.16    OPERATION=SUBMIT_JOB
> TARGET=job_201205172141_0001    RESULT=SUCCESS
> *2012-05-17 21:44:47,084 ERROR org.apache.hadoop.mapred.JobHistory:
> Failed creating job history log file for job job_201205172141_0001
> java.io.IOException: Failed to set permissions of path:
> C:\cygwin\home\hadoop-1.0.3\logs\history\job_201205172141_0001_1337305487022_rn13067_grep-search
> to 0744*
>     at org.apache.hadoop.fs.FileUtil.checkReturnValue(FileUtil.java:689)
>     at org.apache.hadoop.fs.FileUtil.setPermission(FileUtil.java:678)
>     at
> org.apache.hadoop.fs.RawLocalFileSystem.setPermission(RawLocalFileSystem.java:509)
>     at
> org.apache.hadoop.fs.FilterFileSystem.setPermission(FilterFileSystem.java:286)
>     at
> org.apache.hadoop.fs.ChecksumFileSystem.create(ChecksumFileSystem.java:385)
>     at
> org.apache.hadoop.fs.ChecksumFileSystem.create(ChecksumFileSystem.java:364)
>     at
> org.apache.hadoop.mapred.JobHistory$JobInfo.logSubmitted(JobHistory.java:1696)
>     at org.apache.hadoop.mapred.JobInProgress$3.run(JobInProgress.java:681)
>     at java.security.AccessController.doPrivileged(Native Method)
>     at javax.security.auth.Subject.doAs(Subject.java:396)
>     at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1121)
>     at
> org.apache.hadoop.mapred.JobInProgress.initTasks(JobInProgress.java:678)
>     at org.apache.hadoop.mapred.JobTracker.initJob(JobTracker.java:4207)
>     at
> org.apache.hadoop.mapred.EagerTaskInitializationListener$InitJob.run(EagerTaskInitializationListener.java:79)
>     at
> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
>     at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
>     at java.lang.Thread.run(Thread.java:662)
>
> Kindly help me. CYGWIN is in the path. (As some one suggested in some
> thread).
>
>
>
>
> --
> Warmest Regards,
>
> Ravi
>

Reply via email to