Looks like your DataNode isn't properly up. Wipe your dfs.data.dir
directory and restart your DN (might be cause of the formatting
troubles you had earlier). Take a look at your DN's logs though, to
confirm and understand what's going wrong.

On Thu, Aug 11, 2011 at 3:03 PM, A Df <abbey_dragonfor...@yahoo.com> wrote:
> Hi again:
>
> I did format the namenode and it had a problem with a folder being locked. I
> tried again and it formatted but still unable to work. I tried to copy input
> files and run example jar. It gives:
> my-user@ngs:~/hadoop-0.20.2_pseudo> bin/hadoop fs -put input input
> 11/08/11 10:25:11 WARN hdfs.DFSClient: DataStreamer Exception:
> org.apache.hadoop.ipc.RemoteException: java.io.IOException: File
> /user/my-user/input/HadoopInputFile_Request_2011-08-05_162106_1.txt could
> only be replicated to 0 nodes, instead of 1
>         at
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1271)
>         at
> org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:422)
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>         at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>         at java.lang.reflect.Method.invoke(Method.java:597)
>         at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508)
>         at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:959)
>         at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:955)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at javax.security.auth.Subject.doAs(Subject.java:396)
>         at org.apache.hadoop.ipc.Server$Handler.run(Server.java:953)
>
>         at org.apache.hadoop.ipc.Client.call(Client.java:740)
>         at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220)
>         at $Proxy0.addBlock(Unknown Source)
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>         at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>         at java.lang.reflect.Method.invoke(Method.java:597)
>         at
> org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:82)
>         at
> org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59)
>         at $Proxy0.addBlock(Unknown Source)
>         at
> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.locateFollowingBlock(DFSClient.java:2937)
>         at
> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.nextBlockOutputStream(DFSClient.java:2819)
>         at
> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.access$2000(DFSClient.java:2102)
>         at
> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2288)
>
> 11/08/11 10:25:11 WARN hdfs.DFSClient: Error Recovery for block null bad
> datanode[0] nodes == null
> 11/08/11 10:25:11 WARN hdfs.DFSClient: Could not get block locations. Source
> file "/user/my-user/input/HadoopInputFile_Request_2011-08-05_162106_1.txt" -
> Aborting...
> put: java.io.IOException: File
> /user/my-user/input/HadoopInputFile_Request_2011-08-05_162106_1.txt could
> only be replicated to 0 nodes, instead of 1
> 11/08/11 10:25:11 ERROR hdfs.DFSClient: Exception closing file
> /user/my-user/input/HadoopInputFile_Request_2011-08-05_162106_1.txt :
> org.apache.hadoop.ipc.RemoteException: java.io.IOException: File
> /user/my-user/input/HadoopInputFile_Request_2011-08-05_162106_1.txt could
> only be replicated to 0 nodes, instead of 1
>         at
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1271)
>         at
> org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:422)
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>         at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>         at java.lang.reflect.Method.invoke(Method.java:597)
>         at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508)
>         at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:959)
>         at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:955)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at javax.security.auth.Subject.doAs(Subject.java:396)
>         at org.apache.hadoop.ipc.Server$Handler.run(Server.java:953)
>
> org.apache.hadoop.ipc.RemoteException: java.io.IOException: File
> /user/my-user/input/HadoopInputFile_Request_2011-08-05_162106_1.txt could
> only be replicated to 0 nodes, instead of 1
>         at
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1271)
>         at
> org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:422)
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>         at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>         at java.lang.reflect.Method.invoke(Method.java:597)
>         at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508)
>         at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:959)
>         at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:955)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at javax.security.auth.Subject.doAs(Subject.java:396)
>         at org.apache.hadoop.ipc.Server$Handler.run(Server.java:953)
>
>         at org.apache.hadoop.ipc.Client.call(Client.java:740)
>         at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220)
>         at $Proxy0.addBlock(Unknown Source)
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>         at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>         at java.lang.reflect.Method.invoke(Method.java:597)
>         at
> org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:82)
>         at
> org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59)
>         at $Proxy0.addBlock(Unknown Source)
>         at
> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.locateFollowingBlock(DFSClient.java:2937)
>         at
> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.nextBlockOutputStream(DFSClient.java:2819)
>         at
> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.access$2000(DFSClient.java:2102)
>         at
> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2288)
> my-user@ngs:~/hadoop-0.20.2_pseudo> bin/hadoop fs -ls
> Found 1 items
> drwxr-xr-x   - my-user supergroup          0 2011-08-11 10:25
> /user/my-user/input
> my-user@ngs:~/hadoop-0.20.2_pseudo> ls
> bin          docs                        input        logs
> build.xml    hadoop-0.20.2-ant.jar       ivy          NOTICE.txt
> c++          hadoop-0.20.2-core.jar      ivy.xml      README.txt
> CHANGES.txt  hadoop-0.20.2-examples.jar  lib          src
> conf         hadoop-0.20.2-test.jar      librecordio  webapps
> contrib      hadoop-0.20.2-tools.jar     LICENSE.txt
> my-user@ngs:~/hadoop-0.20.2_pseudo> bin/hadoop hadoop-0.20.2-examples.jar
> wordcount input output
> Exception in thread "main" java.lang.NoClassDefFoundError:
> hadoop-0/20/2-examples/jar
> my-user@ngs:~/hadoop-0.20.2_pseudo> bin/hadoop hadoop-0.20.2-examples.jar
> grep input output 'dfs[a-z.]+'
> Exception in thread "main" java.lang.NoClassDefFoundError:
> hadoop-0/20/2-examples/jar
>
>
> ________________________________
> From: Harsh J <ha...@cloudera.com>
> To: common-user@hadoop.apache.org
> Sent: Thursday, 11 August 2011, 6:28
> Subject: Re: Where is web interface in stand alone operation?
>
> Note: NameNode format affects the directory specified by "dfs.name.dir"
>
> On Thu, Aug 11, 2011 at 10:57 AM, Harsh J <ha...@cloudera.com> wrote:
>> Have you done the following?
>>
>> bin/hadoop namenode -format
>>
>> On Thu, Aug 11, 2011 at 10:50 AM, A Df <abbey_dragonfor...@yahoo.com>
>> wrote:
>>> Hello Again:
>>> I extracted hadoop and changed the xml as shwon in the tutorial but now
>>> it
>>> seems it connect get a connection. I am using putty to ssh to the server
>>> and
>>> I change the config files to set it up in pseudo mode as shown
>>> conf/core-site.xml:
>>> <configuration>
>>>      <property>
>>>          <name>fs.default.name</name>
>>>          <value>hdfs://localhost:9000</
>>> value>
>>>      </property>
>>> </configuration>
>>> hdfs-site.xml:
>>> <configuration>
>>>      <property>
>>>          <name>dfs.replication</name>
>>>          <value>1</value>
>>>      </property>
>>>      <property>
>>>          <name>dfs.http.address</name>
>>>          <value>0.0.0.0:3500</value>
>>>      </property>
>>> </configuration>
>>>
>>> conf/mapred-site.xml:
>>> <configuration>
>>>      <property>
>>>          <name>mapred.job.tracker</name>
>>>          <value>localhost:9001</value>
>>>      </property>
>>>      <property>
>>>          <name>mapred.job.tracker.http.address</name>
>>>          <value>0.0.0.0:3501</value>
>>>      </property>
>>> </configuration>
>>>
>>> I tried to format the namenode, started all processes but I notice that
>>> when
>>> I stop it, it said that the namenode was not running.When I try to run
>>> the
>>> example jar, it keeps timing out when connecting to 127.0.0.1:port#. I
>>> used
>>> various port numbers and tried replacing localhost with the name for the
>>> server but it still times out. It also has a long ip address for
>>> name.server.ac.uk/161.74.12.97:3000 which seems to repeating itself since
>>> name.server.ac.uk already has the ip address of 161.74.12.97. The console
>>> message is shown below. I was also having problems where it did not want
>>> to
>>> format the namenode.
>>>
>>> Is there something is wrong with connecting to the namenode and what
>>> cause
>>> it to not format?
>>>
>>>
>>> 2011-08-11 05:49:13,529 INFO
>>> org.apache.hadoop.hdfs.server.namenode.NameNode: STARTUP_MSG:
>>> /************************************************************
>>> STARTUP_MSG: Starting NameNode
>>> STARTUP_MSG:   host = name.server.ac.uk/161.74.12.97
>>> STARTUP_MSG:   args = []
>>> STARTUP_MSG:   version = 0.20.2
>>> STARTUP_MSG:   build =
>>> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20 -r
>>> 911707; compiled by 'chrisdo' on Fri Feb 19 08:07:34 UTC 2010
>>> ************************************************************/
>>> 2011-08-11 05:49:13,663 INFO org.apache.hadoop.ipc.metrics.RpcMetrics:
>>> Initializing RPC Metrics with hostName=NameNode, port=3000
>>> 2011-08-11 05:49:13,669 INFO
>>> org.apache.hadoop.hdfs.server.namenode.NameNode: Namenode up at:
>>> name.server.ac.uk/161.74.12.97:3000
>>> 2011-08-11 05:49:13,672 INFO org.apache.hadoop.metrics.jvm.JvmMetrics:
>>> Initializing JVM Metrics with processName=NameNode, sessionId=null
>>> 2011-08-11 05:49:13,674 INFO
>>> org.apache.hadoop.hdfs.server.namenode.metrics.NameNodeMetrics:
>>> Initializing
>>> NameNodeMeterics using context
>>> object:org.apache.hadoop.metrics.spi.NullContext
>>> 2011-08-11 05:49:13,755 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
>>> fsOwner=my-user,users,cluster_login
>>> 2011-08-11 05:49:13,755 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
>>> supergroup=supergroup
>>> 2011-08-11 05:49:13,756 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
>>> isPermissionEnabled=true
>>> 2011-08-11 05:49:13,768 INFO
>>> org.apache.hadoop.hdfs.server.namenode.metrics.FSNamesystemMetrics:
>>> Initializing FSNamesystemMetrics using context
>>> object:org.apache.hadoop.metrics.spi.NullContext
>>> 2011-08-11 05:49:13,770 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Registered
>>> FSNamesystemStatusMBean
>>> 2011-08-11 05:49:13,812 ERROR
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: FSNamesystem
>>> initialization failed.
>>> java.io.IOException: NameNode is not formatted.
>>>     at
>>>
>>> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:317)
>>>     at
>>>
>>> org.apache.hadoop.hdfs.server.namenode.FSDirectory.loadFSImage(FSDirectory.java:87)
>>>     at
>>>
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.initialize(FSNamesystem.java:311)
>>>     at
>>>
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.<init>(FSNamesystem.java:292)
>>>     at
>>>
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:201)
>>>     at
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:279)
>>>     at
>>>
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:956)
>>>     at
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:965)
>>> 2011-08-11 05:49:13,813 INFO org.apache.hadoop.ipc.Server: Stopping
>>> server
>>> on 3000
>>> 2011-08-11 05:49:13,814 ERROR
>>> org.apache.hadoop.hdfs.server.namenode.NameNode: java.io.IOException:
>>> NameNode is not formatted.
>>>     at
>>>
>>> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:317)
>>>     at
>>>
>>> org.apache.hadoop.hdfs.server.namenode.FSDirectory.loadFSImage(FSDirectory.java:87)
>>>     at
>>>
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.initialize(FSNamesystem.java:311)
>>>     at
>>>
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.<init>(FSNamesystem.java:292)
>>>     at
>>>
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:201)
>>>     at
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:279)
>>>     at
>>>
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:956)
>>>     at
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:965)
>>>
>>> 2011-08-11 05:49:13,814 INFO
>>> org.apache.hadoop.hdfs.server.namenode.NameNode: SHUTDOWN_MSG:
>>> /************************************************************
>>> SHUTDOWN_MSG: Shutting down NameNode at name.server.ac.uk/161.74.12.97
>>> ************************************************************/
>>>
>>> Thank you,
>>> A Df
>>>
>>> ________________________________
>>> From: Harsh J <ha...@cloudera.com>
>>> To: A Df <abbey_dragonfor...@yahoo.com>
>>> Sent: Wednesday, 10 August 2011, 15:13
>>> Subject: Re: Where is web interface in stand alone operation?
>>>
>>> A Df,
>>>
>>> On Wed, Aug 10, 2011 at 7:28 PM, A Df <abbey_dragonfor...@yahoo.com>
>>> wrote:
>>>>
>>>> Hello Harsh:
>>>> See inline at *
>>>>
>>>> ________________________________
>>>> From: Harsh J <ha...@cloudera.com>
>>>> To: common-user@hadoop.apache.org; A Df <abbey_dragonfor...@yahoo.com>
>>>> Sent: Wednesday, 10 August 2011, 14:44
>>>> Subject: Re: Where is web interface in stand alone operation?
>>>>
>>>> A Df,
>>>>
>>>> The web UIs are a feature of the daemons JobTracker and NameNode. In
>>>> standalone/'local'/'file:///' modes, these daemons aren't run
>>>> (actually, no daemon is run at all), and hence there would be no 'web'
>>>> interface.
>>>>
>>>> *ok, but is there any other way to check the performance in this mode
>>>> such
>>>> as time to complete etc? I am trying to compare performance between the
>>>> two.
>>>> And also for the pseudo mode how would I change the ports for the web
>>>> interface because I have to connect to a remote server which only allows
>>>> certain ports to be accessed from the web?
>>>
>>> The ports Kai mentioned above are sourced from the configs:
>>> dfs.http.address (hdfs-site.xml) and mapred.job.tracker.http.address
>>> (mapred-site.xml). You can change them to bind to a host:port of your
>>> preference.
>>>
>>>
>>> --
>>> Harsh J
>>>
>>>
>>>
>>
>>
>>
>> --
>> Harsh J
>>
>
>
>
> --
> Harsh J
>
>
>



-- 
Harsh J

Reply via email to