Hi Nagarjuna,

 

What's is in your /etc/hosts file? I think the line in logs where it says
"DataNodeRegistration(0.0.0.0 [..]", should be the hostname or IP of the
datanode (124.123.215.187 since you said it's a pseudo-distributed setup)
and not 0.0.0.0.

 

By the way are you using the dfs.hosts parameter for specifying the
datanodes that can connect to the namenode?

 

Vijay

 

From: nagarjuna kanamarlapudi [mailto:nagarjuna.kanamarlap...@gmail.com] 
Sent: 20 February 2013 15:52
To: user@hadoop.apache.org
Subject: Re: In Compatible clusterIDs

 

 

Hi Jean Marc,

 

Yes, this is the cluster I am trying  to create and then will scale up.

 

As per your suggestion I deleted the folder
/Users/nagarjunak/Documents/hadoop-install/hadoop-2.0.3-alpha/tmp_20 an
formatted the cluster.

 

 

Now I get the following error.

 

 

2013-02-20 21:17:25,668 FATAL
org.apache.hadoop.hdfs.server.datanode.DataNode: Initialization failed for
block pool Block pool BP-811644675-124.123.215.187-1361375214801 (storage id
DS-1515823288-124.123.215.187-50010-1361375245435) service to
nagarjuna/124.123.215.187:9000

org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.hdfs.server.protocol
.DisallowedDatanodeException): Datanode denied communication with namenode:
DatanodeRegistration(0.0.0.0,
storageID=DS-1515823288-124.123.215.187-50010-1361375245435, infoPort=50075,
ipcPort=50020,
storageInfo=lv=-40;cid=CID-723b02a7-3441-41b5-8045-2a45a9cf96b0;nsid=1805451
571;c=0)

        at
org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager.registerDatano
de(DatanodeManager.java:629)

        at
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.registerDatanode(FSNames
ystem.java:3459)

        at
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.registerDatanode(Na
meNodeRpcServer.java:881)

        at
org.apache.hadoop.hdfs.protocolPB.DatanodeProtocolServerSideTranslatorPB.reg
isterDatanode(DatanodeProtocolServerSideTranslatorPB.java:90)

        at
org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos$DatanodeProtoco
lService$2.callBlockingMethod(DatanodeProtocolProtos.java:18295)

        at org.apache.hadoop.ipc.Protob

 

 

On Wed, Feb 20, 2013 at 9:10 PM, Jean-Marc Spaggiari
<jean-m...@spaggiari.org> wrote:

Hi Nagarjuna,

Is it a test cluster? Do you have another cluster running close-by?
Also, is it your first try?

It seems there is some previous data in the dfs directory which is not
in sync with the last installation.

Maybe you can remove the content of
/Users/nagarjunak/Documents/hadoop-install/hadoop-2.0.3-alpha/tmp_20
if it's not usefull for you, reformat your node and restart it?

JM

2013/2/20, nagarjuna kanamarlapudi <nagarjuna.kanamarlap...@gmail.com>:

> Hi,
>
> I am trying to setup single node cluster of hadop 2.0.*
>
> When trying to start datanode I got the following error. Could anyone help
> me out
>
> Block pool BP-1894309265-124.123.215.187-1361374377471 (storage id
> DS-1175433225-124.123.215.187-50010-1361374235895) service to nagarjuna/
> 124.123.215.187:9000
> java.io.IOException: Incompatible clusterIDs in
>
/Users/nagarjunak/Documents/hadoop-install/hadoop-2.0.3-alpha/tmp_20/dfs/dat
a:
> namenode clusterID = CID-800b7eb1-7a83-4649-86b7-617913e82ad8; datanode
> clusterID = CID-1740b490-8413-451c-926f-2f0676b217ec
>         at
>
org.apache.hadoop.hdfs.server.datanode.DataStorage.doTransition(DataStorage.
java:391)
>         at
>
org.apache.hadoop.hdfs.server.datanode.DataStorage.recoverTransitionRead(Dat
aStorage.java:191)
>         at
>
org.apache.hadoop.hdfs.server.datanode.DataStorage.recoverTransitionRead(Dat
aStorage.java:219)
>         at
>
org.apache.hadoop.hdfs.server.datanode.DataNode.initStorage(DataNode.java:85
0)
>         at
>
org.apache.hadoop.hdfs.server.datanode.DataNode.initBlockPool(DataNode.java:
821)
>         at
>
org.apache.hadoop.hdfs.server.datanode.BPOfferService.verifyAndSetNamespaceI
nfo(BPOfferService.java:280)
>         at
>
org.apache.hadoop.hdfs.server.datanode.BPServiceActor.connectToNNAndHandshak
e(BPServiceActor.java:222)
>         at
>
org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.jav
a:664)
>         at java.lang.Thread.run(Thread.java:680)
> 2013-02-20 21:03:39,856 WARN
> org.apache.hadoop.hdfs.server.datanode.DataNode: Ending block pool service
> for: Block pool BP-1894309265-124.123.215.187-1361374377471 (storage id
> DS-1175433225-124.123.215.187-50010-1361374235895) service to nagarjuna/
> 124.123.215.187:9000
> 2013-02-20 21:03:39,958 INFO
> org.apache.hadoop.hdfs.server.datanode.DataNode: Removed Block pool
> BP-1894309265-124.123.215.187-1361374377471 (storage id
> DS-1175433225-124.123.215.187-50010-1361374235895)
> 2013-02-20 21:03:41,959 WARN
> org.apache.hadoop.hdfs.server.datanode.DataNode: Exiting Datanode
> 2013-02-20 21:03:41,961 INFO org.apache.hadoop.util.ExitUtil: Exiting with
> status 0
> 2013-02-20 21:03:41,963 INFO
> org.apache.hadoop.hdfs.server.datanode.DataNode: SHUTDOWN_MSG:
> /************************************************************
> SHUTDOWN_MSG: Shutting down DataNode at nagarjuna/124.123.215.187
> ************************************************************/
>

 

Reply via email to