Sandy, as far as I remember, there were some threads about the same
problem (I don't know if it's solved). Searching the mailing list for
this error: "could only be replicated to 0 nodes, instead of 1" may
help.

Cheers,
Rasit

2009/2/16 Sandy <snickerdoodl...@gmail.com>:
> just some more information:
> hadoop fsck produces:
> Status: HEALTHY
>  Total size: 0 B
>  Total dirs: 9
>  Total files: 0 (Files currently being written: 1)
>  Total blocks (validated): 0
>  Minimally replicated blocks: 0
>  Over-replicated blocks: 0
>  Under-replicated blocks: 0
>  Mis-replicated blocks: 0
>  Default replication factor: 1
>  Average block replication: 0.0
>  Corrupt blocks: 0
>  Missing replicas: 0
>  Number of data-nodes: 0
>  Number of racks: 0
>
>
> The filesystem under path '/' is HEALTHY
>
> on the newly formatted hdfs.
>
> jps says:
> 4723 Jps
> 4527 NameNode
> 4653 JobTracker
>
>
> I can't copy files onto the dfs since I get "NotReplicatedYetExceptions",
> which I suspect has to do with the fact that there are no datanodes. My
> "cluster" is a single MacPro with 8 cores. I haven't had to do anything
> extra before in order to get datanodes to be generated.
>
> 09/02/15 15:56:27 WARN dfs.DFSClient: Error Recovery for block null bad
> datanode[0]
> copyFromLocal: Could not get block locations. Aborting...
>
>
> The corresponding error in the logs is:
>
> 2009-02-15 15:56:27,123 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 1 on 9000, call addBlock(/user/hadoop/input/.DS_Store,
> DFSClient_755366230) from 127.0.0.1:49796: error: java.io.IOException: File
> /user/hadoop/input/.DS_Store could only be replicated to 0 nodes, instead of
> 1
> java.io.IOException: File /user/hadoop/input/.DS_Store could only be
> replicated to 0 nodes, instead of 1
> at
> org.apache.hadoop.dfs.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1120)
> at org.apache.hadoop.dfs.NameNode.addBlock(NameNode.java:330)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
> at java.lang.reflect.Method.invoke(Method.java:597)
> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:452)
> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:888)
>
> On Sun, Feb 15, 2009 at 3:26 PM, Sandy <snickerdoodl...@gmail.com> wrote:
>
>> Thanks for your responses.
>>
>> I checked in the namenode and jobtracker logs and both say:
>>
>> INFO org.apache.hadoop.ipc.Server: IPC Server handler 6 on 9000, call
>> delete(/Users/hadoop/hadoop-0.18.2/hadoop-hadoop/mapred/system, true) from
>> 127.0.0.1:61086: error: org.apache.hadoop.dfs.SafeModeException: Cannot
>> delete /Users/hadoop/hadoop-0.18.2/hadoop-hadoop/mapred/system. Name node
>> is in safe mode.
>> The ratio of reported blocks 0.0000 has not reached the threshold 0.9990.
>> Safe mode will be turned off automatically.
>> org.apache.hadoop.dfs.SafeModeException: Cannot delete
>> /Users/hadoop/hadoop-0.18.2/hadoop-hadoop/mapred/system. Name node is in
>> safe mode.
>> The ratio of reported blocks 0.0000 has not reached the threshold 0.9990.
>> Safe mode will be turned off automatically.
>>         at
>> org.apache.hadoop.dfs.FSNamesystem.deleteInternal(FSNamesystem.java:1505)
>>         at
>> org.apache.hadoop.dfs.FSNamesystem.delete(FSNamesystem.java:1477)
>>         at org.apache.hadoop.dfs.NameNode.delete(NameNode.java:425)
>>         at sun.reflect.GeneratedMethodAccessor2.invoke(Unknown Source)
>>         at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>>         at java.lang.reflect.Method.invoke(Method.java:597)
>>         at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:452)
>>         at org.apache.hadoop.ipc.Server$Handler.run(Server.java:888)
>>
>>
>> I think this is a continuation of my running problem. The nodes stay in
>> safe mode, but won't come out, even after several minutes. I believe this is
>> due to the fact that it keep trying to contact a datanode that does not
>> exist. Any suggestions on what I can do?
>>
>> I have recently tried to reformat the hdfs, using bin/hadoop namenode
>> -format. From the output directed to standard out, I thought this completed
>> correctly:
>>
>> Re-format filesystem in /Users/hadoop/hadoop-0.18.2/hadoop-hadoop/dfs/name
>> ? (Y or N) Y
>> 09/02/15 15:16:39 INFO fs.FSNamesystem:
>> fsOwner=hadoop,staff,_lpadmin,com.apple.sharepoint.group.8,com.apple.sharepoint.group.3,com.apple.sharepoint.group.4,com.apple.sharepoint.group.2,com.apple.sharepoint.group.6,com.apple.sharepoint.group.9,com.apple.sharepoint.group.1,com.apple.sharepoint.group.5
>> 09/02/15 15:16:39 INFO fs.FSNamesystem: supergroup=supergroup
>> 09/02/15 15:16:39 INFO fs.FSNamesystem: isPermissionEnabled=true
>> 09/02/15 15:16:39 INFO dfs.Storage: Image file of size 80 saved in 0
>> seconds.
>> 09/02/15 15:16:39 INFO dfs.Storage: Storage directory
>> /Users/hadoop/hadoop-0.18.2/hadoop-hadoop/dfs/name has been successfully
>> formatted.
>> 09/02/15 15:16:39 INFO dfs.NameNode: SHUTDOWN_MSG:
>> /************************************************************
>> SHUTDOWN_MSG: Shutting down NameNode at
>> loteria.cs.tamu.edu/128.194.143.170
>> ************************************************************/
>>
>> However, after reformatting, I find that I have the same problems.
>>
>> Thanks,
>> SM
>>
>> On Fri, Feb 13, 2009 at 5:39 PM, james warren <ja...@rockyou.com> wrote:
>>
>>> Sandy -
>>>
>>> I suggest you take a look into your NameNode and DataNode logs.  From the
>>> information posted, these likely would be at
>>>
>>>
>>> /Users/hadoop/hadoop-0.18.2/bin/../logs/hadoop-hadoop-namenode-loteria.cs.tamu.edu.log
>>>
>>> /Users/hadoop/hadoop-0.18.2/bin/../logs/hadoop-hadoop-jobtracker-loteria.cs.tamu.edu.log
>>>
>>> If the cause isn't obvious from what you see there, could you please post
>>> the last few lines from each log?
>>>
>>> -jw
>>>
>>> On Fri, Feb 13, 2009 at 3:28 PM, Sandy <snickerdoodl...@gmail.com> wrote:
>>>
>>> > Hello,
>>> >
>>> > I would really appreciate any help I can get on this! I've suddenly ran
>>> > into
>>> > a very strange error.
>>> >
>>> > when I do:
>>> > bin/start-all
>>> > I get:
>>> > hadoop$ bin/start-all.sh
>>> > starting namenode, logging to
>>> >
>>> >
>>> /Users/hadoop/hadoop-0.18.2/bin/../logs/hadoop-hadoop-namenode-loteria.cs.tamu.edu.out
>>> > starting jobtracker, logging to
>>> >
>>> >
>>> /Users/hadoop/hadoop-0.18.2/bin/../logs/hadoop-hadoop-jobtracker-loteria.cs.tamu.edu.out
>>> >
>>> > No datanode, secondary namenode or jobtracker are being started.
>>> >
>>> > When I try to upload anything on the dfs, I get a "node in safemode"
>>> error
>>> > (even after waiting 5 minutes), presumably because it's trying to reach
>>> a
>>> > datanode that does not exist.  The same "safemode" error occurs when I
>>> try
>>> > to run jobs.
>>> >
>>> > I have tried bin/stop-all and then bin/start-all again. I get the same
>>> > problem!
>>> >
>>> > This is incredibly strange, since I was previously able to start and run
>>> > jobs without any issue using this version on this machine. I am running
>>> > jobs
>>> > on a single Mac Pro running OS X 10.5
>>> >
>>> > I have tried updating to hadoop-0.19.0, and I get the same problem. I
>>> have
>>> > even tried this using previous versions, and I'm getting the same
>>> problem!
>>> >
>>> > Anyone have any idea why this suddenly could be happening? What am I
>>> doing
>>> > wrong?
>>> >
>>> > For convenience, I'm including portions of both conf/hadoop-env.sh and
>>> > conf/hadoop-site.xml:
>>> >
>>> > --- hadoop-env.sh ---
>>> >  # Set Hadoop-specific environment variables here.
>>> >
>>> > # The only required environment variable is JAVA_HOME.  All others are
>>> > # optional.  When running a distributed configuration it is best to
>>> > # set JAVA_HOME in this file, so that it is correctly defined on
>>> > # remote nodes.
>>> >
>>> > # The java implementation to use.  Required.
>>> >  export
>>> >
>>> JAVA_HOME=/System/Library/Frameworks/JavaVM.framework/Versions/1.6.0/Home
>>> >
>>> > # Extra Java CLASSPATH elements.  Optional.
>>> > # export HADOOP_CLASSPATH=
>>> >
>>> > # The maximum amount of heap to use, in MB. Default is 1000.
>>> >  export HADOOP_HEAPSIZE=3000
>>> > ...
>>> > --- hadoop-site.xml ---
>>> > <configuration>
>>> >
>>> > <property>
>>> >  <name>hadoop.tmp.dir</name>
>>> >  <value>/Users/hadoop/hadoop-0.18.2/hadoop-${user.name}</value>
>>> >  <description>A base for other temporary directories.</description>
>>> > </property>
>>> >
>>> > <property>
>>> >  <name>fs.default.name</name>
>>> >  <value>hdfs://localhost:9000</value>
>>> >  <description>The name of the default file system.  A URI whose
>>> >  scheme and authority determine the FileSystem implementation.  The
>>> >  uri's scheme determines the config property (fs.SCHEME.impl) naming
>>> >  the FileSystem implementation class.  The uri's authority is used to
>>> >  determine the host, port, etc. for a filesystem.</description>
>>> > </property>
>>> >
>>> > <property>
>>> >  <name>mapred.job.tracker</name>
>>> >  <value>localhost:9001</value>
>>> >  <description>The host and port that the MapReduce job tracker runs
>>> >  at.  If "local", then jobs are run in-process as a single map
>>> >  and reduce task.
>>> >  </description>
>>> > </property>
>>> >
>>> > <property>
>>> > <name>mapred.tasktracker.tasks.maximum</name>
>>> > <value>1</value>
>>> > <description>The maximum number of tasks that will be run simultaneously
>>> by
>>> > a
>>> > a task tracker
>>> > </description>
>>> > </property>
>>> > ...
>>> >
>>>
>>
>>
>



-- 
M. Raşit ÖZDAŞ

Reply via email to