No, not ofcourse I blinded it.

On Wed, Mar 19, 2014 at 5:09 PM, praveenesh kumar <praveen...@gmail.com>wrote:

> Is this property correct ?
>
>
> <property>
>     <name>fs.default.name</name>
>     <value>-BLANKED</value>
>   </property>
>
> Regards
> Prav
>
>
> On Wed, Mar 19, 2014 at 12:58 PM, Fatih Haltas <fatih.hal...@nyu.edu>wrote:
>
>> Thanks for you helps, but still could not solve my problem.
>>
>>
>> On Tue, Mar 18, 2014 at 10:13 AM, Stanley Shi <s...@gopivotal.com> wrote:
>>
>>> Ah yes, I overlooked this. Then please check the file are there or not:
>>> "ls /home/hadoop/project/hadoop-data/dfs/name"?
>>>
>>> Regards,
>>> *Stanley Shi,*
>>>
>>>
>>>
>>> On Tue, Mar 18, 2014 at 2:06 PM, Azuryy Yu <azury...@gmail.com> wrote:
>>>
>>>> I don't think this is the case, because there is;
>>>>   <property>
>>>>     <name>hadoop.tmp.dir</name>
>>>>     <value>/home/hadoop/project/hadoop-data</value>
>>>>   </property>
>>>>
>>>>
>>>> On Tue, Mar 18, 2014 at 1:55 PM, Stanley Shi <s...@gopivotal.com>wrote:
>>>>
>>>>> one possible reason is that you didn't set the namenode working
>>>>> directory, by default it's in "/tmp" folder; and the "/tmp" folder might
>>>>> get deleted by the OS without any notification. If this is the case, I am
>>>>> afraid you have lost all your namenode data.
>>>>>
>>>>> *<property>
>>>>>   <name>dfs.name.dir</name>
>>>>>   <value>${hadoop.tmp.dir}/dfs/name</value>
>>>>>   <description>Determines where on the local filesystem the DFS name node
>>>>>       should store the name table(fsimage).  If this is a comma-delimited 
>>>>> list
>>>>>       of directories then the name table is replicated in all of the
>>>>>       directories, for redundancy. </description>
>>>>> </property>*
>>>>>
>>>>>
>>>>> Regards,
>>>>> *Stanley Shi,*
>>>>>
>>>>>
>>>>>
>>>>> On Sun, Mar 16, 2014 at 5:29 PM, Mirko Kämpf 
>>>>> <mirko.kae...@gmail.com>wrote:
>>>>>
>>>>>> Hi,
>>>>>>
>>>>>> what is the location of the namenodes fsimage and editlogs?
>>>>>> And how much memory has the NameNode.
>>>>>>
>>>>>> Did you work with a Secondary NameNode or a Standby NameNode for
>>>>>> checkpointing?
>>>>>>
>>>>>> Where are your HDFS blocks located, are those still safe?
>>>>>>
>>>>>> With this information at hand, one might be able to fix your setup,
>>>>>> but do not format the old namenode before
>>>>>> all is working with a fresh one.
>>>>>>
>>>>>> Grab a copy of the maintainance guide:
>>>>>> http://shop.oreilly.com/product/0636920025085.do?sortby=publicationDate
>>>>>> which helps solving such type of problems as well.
>>>>>>
>>>>>> Best wishes
>>>>>> Mirko
>>>>>>
>>>>>>
>>>>>> 2014-03-16 9:07 GMT+00:00 Fatih Haltas <fatih.hal...@nyu.edu>:
>>>>>>
>>>>>> Dear All,
>>>>>>>
>>>>>>> I have just restarted machines of my hadoop clusters. Now, I am
>>>>>>> trying to restart hadoop clusters again, but getting error on namenode
>>>>>>> restart. I am afraid of loosing my data as it was properly running for 
>>>>>>> more
>>>>>>> than 3 months. Currently, I believe if I do namenode formatting, it will
>>>>>>> work again, however, data will be lost. Is there anyway to solve this
>>>>>>> without losing the data.
>>>>>>>
>>>>>>> I will really appreciate any help.
>>>>>>>
>>>>>>> Thanks.
>>>>>>>
>>>>>>>
>>>>>>> =====================
>>>>>>> Here is the logs;
>>>>>>> ====================
>>>>>>> 2014-02-26 16:02:39,698 INFO
>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode: STARTUP_MSG:
>>>>>>> /************************************************************
>>>>>>> STARTUP_MSG: Starting NameNode
>>>>>>> STARTUP_MSG:   host = ADUAE042-LAP-V/127.0.0.1
>>>>>>> STARTUP_MSG:   args = []
>>>>>>> STARTUP_MSG:   version = 1.0.4
>>>>>>> STARTUP_MSG:   build =
>>>>>>> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.0-r 
>>>>>>> 1393290; compiled by 'hortonfo' on Wed Oct  3 05:13:58 UTC 2012
>>>>>>> ************************************************************/
>>>>>>> 2014-02-26 16:02:40,005 INFO
>>>>>>> org.apache.hadoop.metrics2.impl.MetricsConfig: loaded properties from
>>>>>>> hadoop-metrics2.properties
>>>>>>> 2014-02-26 16:02:40,019 INFO
>>>>>>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
>>>>>>> MetricsSystem,sub=Stats registered.
>>>>>>> 2014-02-26 16:02:40,021 INFO
>>>>>>> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot
>>>>>>> period at 10 second(s).
>>>>>>> 2014-02-26 16:02:40,021 INFO
>>>>>>> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics 
>>>>>>> system
>>>>>>> started
>>>>>>> 2014-02-26 16:02:40,169 INFO
>>>>>>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source 
>>>>>>> ugi
>>>>>>> registered.
>>>>>>> 2014-02-26 16:02:40,193 INFO
>>>>>>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source 
>>>>>>> jvm
>>>>>>> registered.
>>>>>>> 2014-02-26 16:02:40,194 INFO
>>>>>>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
>>>>>>> NameNode registered.
>>>>>>> 2014-02-26 16:02:40,242 INFO org.apache.hadoop.hdfs.util.GSet: VM
>>>>>>> type       = 64-bit
>>>>>>> 2014-02-26 16:02:40,242 INFO org.apache.hadoop.hdfs.util.GSet: 2%
>>>>>>> max memory = 17.77875 MB
>>>>>>> 2014-02-26 16:02:40,242 INFO org.apache.hadoop.hdfs.util.GSet:
>>>>>>> capacity      = 2^21 = 2097152 entries
>>>>>>> 2014-02-26 16:02:40,242 INFO org.apache.hadoop.hdfs.util.GSet:
>>>>>>> recommended=2097152, actual=2097152
>>>>>>> 2014-02-26 16:02:40,273 INFO
>>>>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: fsOwner=hadoop
>>>>>>> 2014-02-26 16:02:40,273 INFO
>>>>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: 
>>>>>>> supergroup=supergroup
>>>>>>> 2014-02-26 16:02:40,274 INFO
>>>>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
>>>>>>> isPermissionEnabled=true
>>>>>>> 2014-02-26 16:02:40,279 INFO
>>>>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
>>>>>>> dfs.block.invalidate.limit=100
>>>>>>> 2014-02-26 16:02:40,279 INFO
>>>>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
>>>>>>> isAccessTokenEnabled=false accessKeyUpdateInterval=0 min(s),
>>>>>>> accessTokenLifetime=0 min(s)
>>>>>>> 2014-02-26 16:02:40,724 INFO
>>>>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Registered
>>>>>>> FSNamesystemStateMBean and NameNodeMXBean
>>>>>>> 2014-02-26 16:02:40,749 INFO
>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode: Caching file names
>>>>>>> occuring more than 10 times
>>>>>>> 2014-02-26 16:02:40,780 ERROR
>>>>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: FSNamesystem
>>>>>>> initialization failed.
>>>>>>> java.io.IOException: NameNode is not formatted.
>>>>>>>         at
>>>>>>> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:330)
>>>>>>>         at
>>>>>>> org.apache.hadoop.hdfs.server.namenode.FSDirectory.loadFSImage(FSDirectory.java:100)
>>>>>>>         at
>>>>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.initialize(FSNamesystem.java:388)
>>>>>>>         at
>>>>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.<init>(FSNamesystem.java:362)
>>>>>>>         at
>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:276)
>>>>>>>         at
>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:496)
>>>>>>>         at
>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1279)
>>>>>>>         at
>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1288)
>>>>>>> 2014-02-26 16:02:40,781 ERROR
>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode: java.io.IOException:
>>>>>>> NameNode is not formatted.
>>>>>>>         at
>>>>>>> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:330)
>>>>>>>         at
>>>>>>> org.apache.hadoop.hdfs.server.namenode.FSDirectory.loadFSImage(FSDirectory.java:100)
>>>>>>>         at
>>>>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.initialize(FSNamesystem.java:388)
>>>>>>>         at
>>>>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.<init>(FSNamesystem.java:362)
>>>>>>>         at
>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:276)
>>>>>>>         at
>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:496)
>>>>>>>         at
>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1279)
>>>>>>>         at
>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1288)
>>>>>>>
>>>>>>> 2014-02-26 16:02:40,781 INFO
>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode: SHUTDOWN_MSG:
>>>>>>> /************************************************************
>>>>>>> SHUTDOWN_MSG: Shutting down NameNode at ADUAE042-LAP-V/127.0.0.1
>>>>>>> ************************************************************/
>>>>>>>
>>>>>>> ===========================
>>>>>>> Here is the core-site.xml
>>>>>>> ===========================
>>>>>>>  <?xml version="1.0"?>
>>>>>>> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
>>>>>>>
>>>>>>> <!-- Put site-specific property overrides in this file. -->
>>>>>>>
>>>>>>> <configuration>
>>>>>>> <property>
>>>>>>>     <name>fs.default.name</name>
>>>>>>>     <value>-BLANKED</value>
>>>>>>>   </property>
>>>>>>>   <property>
>>>>>>>     <name>hadoop.tmp.dir</name>
>>>>>>>     <value>/home/hadoop/project/hadoop-data</value>
>>>>>>>   </property>
>>>>>>> </configuration>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>

Reply via email to