Re: cygwin single node setup

2012-04-28 Thread Onder SEZGIN
EXT0125622/input
-rw-r--r--   1 EXT0125622 supergroup   7457 2012-04-29 02:13
/user/EXT0125622/input/capacity-scheduler.xml
drwxr-xr-x   - EXT0125622 supergroup  0 2012-04-29 03:05
/user/EXT0125622/input/conf
-rw-r--r--   1 EXT0125622 supergroup   7457 2012-04-29 03:05
/user/EXT0125622/input/conf/capacity-scheduler.xml
-rw-r--r--   1 EXT0125622 supergroup535 2012-04-29 03:05
/user/EXT0125622/input/conf/configuration.xsl
-rw-r--r--   1 EXT0125622 supergroup438 2012-04-29 03:05
/user/EXT0125622/input/conf/core-site.xml
-rw-r--r--   1 EXT0125622 supergroup327 2012-04-29 03:05
/user/EXT0125622/input/conf/fair-scheduler.xml
-rw-r--r--   1 EXT0125622 supergroup   2292 2012-04-29 03:05
/user/EXT0125622/input/conf/hadoop-env.sh
-rw-r--r--   1 EXT0125622 supergroup   1488 2012-04-29 03:05
/user/EXT0125622/input/conf/hadoop-metrics2.properties
-rw-r--r--   1 EXT0125622 supergroup   4644 2012-04-29 03:05
/user/EXT0125622/input/conf/hadoop-policy.xml
-rw-r--r--   1 EXT0125622 supergroup274 2012-04-29 03:05
/user/EXT0125622/input/conf/hdfs-site.xml
-rw-r--r--   1 EXT0125622 supergroup   4441 2012-04-29 03:05
/user/EXT0125622/input/conf/log4j.properties
-rw-r--r--   1 EXT0125622 supergroup   2033 2012-04-29 03:05
/user/EXT0125622/input/conf/mapred-queue-acls.xml
-rw-r--r--   1 EXT0125622 supergroup290 2012-04-29 03:05
/user/EXT0125622/input/conf/mapred-site.xml
-rw-r--r--   1 EXT0125622 supergroup 10 2012-04-29 03:05
/user/EXT0125622/input/conf/masters
-rw-r--r--   1 EXT0125622 supergroup 10 2012-04-29 03:05
/user/EXT0125622/input/conf/slaves
-rw-r--r--   1 EXT0125622 supergroup   1243 2012-04-29 03:05
/user/EXT0125622/input/conf/ssl-client.xml.example
-rw-r--r--   1 EXT0125622 supergroup   1195 2012-04-29 03:05
/user/EXT0125622/input/conf/ssl-server.xml.example
-rw-r--r--   1 EXT0125622 supergroup382 2012-04-29 03:05
/user/EXT0125622/input/conf/taskcontroller.cfg
-rw-r--r--   1 EXT0125622 supergroup535 2012-04-29 02:13
/user/EXT0125622/input/configuration.xsl
-rw-r--r--   1 EXT0125622 supergroup438 2012-04-29 02:13
/user/EXT0125622/input/core-site.xml
-rw-r--r--   1 EXT0125622 supergroup327 2012-04-29 02:13
/user/EXT0125622/input/fair-scheduler.xml
-rw-r--r--   1 EXT0125622 supergroup   2292 2012-04-29 02:13
/user/EXT0125622/input/hadoop-env.sh
-rw-r--r--   1 EXT0125622 supergroup   1488 2012-04-29 02:13
/user/EXT0125622/input/hadoop-metrics2.properties
-rw-r--r--   1 EXT0125622 supergroup   4644 2012-04-29 02:13
/user/EXT0125622/input/hadoop-policy.xml
-rw-r--r--   1 EXT0125622 supergroup274 2012-04-29 02:13
/user/EXT0125622/input/hdfs-site.xml
-rw-r--r--   1 EXT0125622 supergroup   4441 2012-04-29 02:13
/user/EXT0125622/input/log4j.properties
-rw-r--r--   1 EXT0125622 supergroup   2033 2012-04-29 02:13
/user/EXT0125622/input/mapred-queue-acls.xml
-rw-r--r--   1 EXT0125622 supergroup290 2012-04-29 02:13
/user/EXT0125622/input/mapred-site.xml
-rw-r--r--   1 EXT0125622 supergroup 10 2012-04-29 02:13
/user/EXT0125622/input/masters
-rw-r--r--   1 EXT0125622 supergroup 10 2012-04-29 02:13
/user/EXT0125622/input/slaves
-rw-r--r--   1 EXT0125622 supergroup   1243 2012-04-29 02:13
/user/EXT0125622/input/ssl-client.xml.example
-rw-r--r--   1 EXT0125622 supergroup   1195 2012-04-29 02:13
/user/EXT0125622/input/ssl-server.xml.example
-rw-r--r--   1 EXT0125622 supergroup382 2012-04-29 02:13
/user/EXT0125622/input/taskcontroller.cfg

any help?


On Sat, Apr 28, 2012 at 3:35 PM, kasi subrahmanyam
wrote:

> Hi Onder,
> You could try to format the namenode and restart the daemons,
> That solved my problem most number of times.
> May be the running daemons where not able to pickup the all the datanodes
> configurations
>
> On Sat, Apr 28, 2012 at 4:23 PM, Onder SEZGIN 
> wrote:
>
> > Hi,
> >
> > I am pretty a newbie and i am following the quick start guide for single
> > node set up on windows using cygwin.
> >
> > In this step,
> >
> > $ bin/hadoop fs -put conf input
> >
> > I am getting the following errors.
> >
> > I have got no files
> > under /user/EXT0125622/input/conf/capacity-scheduler.xml. That might be a
> > reason for the errors i get but why does hadoop look for such directory
> as
> > i have not configured anything like that. so supposably, hadoop is making
> > up and looking for such file and directory?
> >
> > Any idea and help is welcome.
> >
> > Cheers
> > Onder
> >
> > 12/04/27 13:44:37 WARN hdfs.DFSClient: DataStreamer Exception:
> > org.apache.hadoop.ipc.RemoteException: java.io.IOException: File
> > /user/EXT0125622/input/conf/capacity-scheduler.xml could only be
> replicated
> > to 0 nodes, instead of 1
> >at
> >

cygwin single node setup

2012-04-28 Thread Onder SEZGIN
Hi,

I am pretty a newbie and i am following the quick start guide for single
node set up on windows using cygwin.

In this step,

$ bin/hadoop fs -put conf input

I am getting the following errors.

I have got no files
under /user/EXT0125622/input/conf/capacity-scheduler.xml. That might be a
reason for the errors i get but why does hadoop look for such directory as
i have not configured anything like that. so supposably, hadoop is making
up and looking for such file and directory?

Any idea and help is welcome.

Cheers
Onder

12/04/27 13:44:37 WARN hdfs.DFSClient: DataStreamer Exception:
org.apache.hadoop.ipc.RemoteException: java.io.IOException: File
/user/EXT0125622/input/conf/capacity-scheduler.xml could only be replicated
to 0 nodes, instead of 1
at
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1558)
at
org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:696)
at sun.reflect.GeneratedMethodAccessor6.invoke(Unknown Source)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:601)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:563)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1388)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1384)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:415)
at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1093)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1382)

at org.apache.hadoop.ipc.Client.call(Client.java:1066)
at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:225)
at $Proxy1.addBlock(Unknown Source)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:601)
at
org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:82)
at
org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59)
at $Proxy1.addBlock(Unknown Source)
at
org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.locateFollowingBlock(DFSClient.java:3507)
at
org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.nextBlockOutputStream(DFSClient.java:3370)
at
org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.access$2700(DFSClient.java:2586)
at
org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2826)

 12/04/27 13:44:37 WARN hdfs.DFSClient: Error Recovery for block null bad
datanode[0] nodes == null
12/04/27 13:44:37 WARN hdfs.DFSClient: Could not get block locations.
Source file "/user/EXT0125622/input/conf/capacity-scheduler.xml" -
Aborting...
put: java.io.IOException: File
/user/EXT0125622/input/conf/capacity-scheduler.xml could only be replicated
to 0 nodes, instead of 1
12/04/27 13:44:37 ERROR hdfs.DFSClient: Exception closing file
/user/EXT0125622/input/conf/capacity-scheduler.xml :
org.apache.hadoop.ipc.RemoteException: java.io.IOException: File
/user/EXT0125622/input/conf/capacity-scheduler.xml could only be replicated
to 0 nodes, instead of 1
at
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1558)
at
org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:696)
at sun.reflect.GeneratedMethodAccessor6.invoke(Unknown Source)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:601)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:563)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1388)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1384)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:415)
at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1093)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1382)

org.apache.hadoop.ipc.RemoteException: java.io.IOException: File
/user/EXT0125622/input/conf/capacity-scheduler.xml could only be replicated
to 0 nodes, instead of 1
at
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1558)
at
org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:696)
at sun.reflect.GeneratedMethodAccessor6.invoke(Unknown Source)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:6

cygwin single node setup

2012-04-27 Thread Onder SEZGIN
Hi,

I am pretty a newbie and i am following the quick start guide for single
node set up on windows using cygwin.

In this step,

$ bin/hadoop fs -put conf input

I am getting the following errors.

I have got no files
under /user/EXT0125622/input/conf/capacity-scheduler.xml. That might be a
reason for the errors i get but why does hadoop look for such directory as
i have not configured anything like that. so supposably, hadoop is making
up and looking for such file and directory?

Any idea and help is welcome.

Cheers
Onder

12/04/27 13:44:37 WARN hdfs.DFSClient: DataStreamer Exception:
org.apache.hadoop.ipc.RemoteException: java.io.IOException: File
/user/EXT0125622/input/conf/capacity-scheduler.xml could only be replicated
to 0 nodes, instead of 1
at
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1558)
at
org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:696)
at sun.reflect.GeneratedMethodAccessor6.invoke(Unknown Source)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:601)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:563)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1388)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1384)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:415)
at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1093)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1382)

at org.apache.hadoop.ipc.Client.call(Client.java:1066)
at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:225)
at $Proxy1.addBlock(Unknown Source)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:601)
at
org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:82)
at
org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59)
at $Proxy1.addBlock(Unknown Source)
at
org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.locateFollowingBlock(DFSClient.java:3507)
at
org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.nextBlockOutputStream(DFSClient.java:3370)
at
org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.access$2700(DFSClient.java:2586)
at
org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2826)

 12/04/27 13:44:37 WARN hdfs.DFSClient: Error Recovery for block null bad
datanode[0] nodes == null
12/04/27 13:44:37 WARN hdfs.DFSClient: Could not get block locations.
Source file "/user/EXT0125622/input/conf/capacity-scheduler.xml" -
Aborting...
put: java.io.IOException: File
/user/EXT0125622/input/conf/capacity-scheduler.xml could only be replicated
to 0 nodes, instead of 1
12/04/27 13:44:37 ERROR hdfs.DFSClient: Exception closing file
/user/EXT0125622/input/conf/capacity-scheduler.xml :
org.apache.hadoop.ipc.RemoteException: java.io.IOException: File
/user/EXT0125622/input/conf/capacity-scheduler.xml could only be replicated
to 0 nodes, instead of 1
at
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1558)
at
org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:696)
at sun.reflect.GeneratedMethodAccessor6.invoke(Unknown Source)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:601)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:563)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1388)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1384)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:415)
at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1093)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1382)

org.apache.hadoop.ipc.RemoteException: java.io.IOException: File
/user/EXT0125622/input/conf/capacity-scheduler.xml could only be replicated
to 0 nodes, instead of 1
at
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1558)
at
org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:696)
at sun.reflect.GeneratedMethodAccessor6.invoke(Unknown Source)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:6

cygwin single node setup

2012-04-27 Thread Onder SEZGIN
Hi,

I am pretty a newbie and i am following the quick start guide for single
node set up on windows using cygwin.

In this step,

$ bin/hadoop fs -put conf input

I am getting the following errors.

I have got no files
under /user/EXT0125622/input/conf/capacity-scheduler.xml. That might be a
reason for the errors i get but why does hadoop look for such directory as
i have not configured anything like that. so supposably, hadoop is making
up and looking for such file and directory?

Any idea and help is welcome.

Cheers
Onder

12/04/27 13:44:37 WARN hdfs.DFSClient: DataStreamer Exception:
org.apache.hadoop.ipc.RemoteException: java.io.IOException: File
/user/EXT0125622/input/conf/capacity-scheduler.xml could only be replicated
to 0 nodes, instead of 1
at
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1558)
at
org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:696)
at sun.reflect.GeneratedMethodAccessor6.invoke(Unknown Source)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:601)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:563)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1388)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1384)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:415)
at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1093)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1382)

at org.apache.hadoop.ipc.Client.call(Client.java:1066)
at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:225)
at $Proxy1.addBlock(Unknown Source)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:601)
at
org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:82)
at
org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59)
at $Proxy1.addBlock(Unknown Source)
at
org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.locateFollowingBlock(DFSClient.java:3507)
at
org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.nextBlockOutputStream(DFSClient.java:3370)
at
org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.access$2700(DFSClient.java:2586)
at
org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2826)

12/04/27 13:44:37 WARN hdfs.DFSClient: Error Recovery for block null bad
datanode[0] nodes == null
12/04/27 13:44:37 WARN hdfs.DFSClient: Could not get block locations.
Source file "/user/EXT0125622/input/conf/capacity-scheduler.xml" -
Aborting...
put: java.io.IOException: File
/user/EXT0125622/input/conf/capacity-scheduler.xml could only be replicated
to 0 nodes, instead of 1
12/04/27 13:44:37 ERROR hdfs.DFSClient: Exception closing file
/user/EXT0125622/input/conf/capacity-scheduler.xml :
org.apache.hadoop.ipc.RemoteException: java.io.IOException: File
/user/EXT0125622/input/conf/capacity-scheduler.xml could only be replicated
to 0 nodes, instead of 1
at
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1558)
at
org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:696)
at sun.reflect.GeneratedMethodAccessor6.invoke(Unknown Source)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:601)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:563)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1388)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1384)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:415)
at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1093)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1382)

org.apache.hadoop.ipc.RemoteException: java.io.IOException: File
/user/EXT0125622/input/conf/capacity-scheduler.xml could only be replicated
to 0 nodes, instead of 1
at
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1558)
at
org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:696)
at sun.reflect.GeneratedMethodAccessor6.invoke(Unknown Source)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:60