Aseem,

I'm glad it solved your problem.

I'm not familiar with this plugin, maybe someone else on this mailing
list is or you should probably write to the Hadoop mailing list.

J-D

On Wed, Mar 18, 2009 at 11:03 AM, Puri, Aseem <[email protected]> wrote:
> Thanks Jean, my problem is solved now.
>
> I have removed all old data and then restarted my Hadoop server then my 
> datanode starts. Also after that my HBase master starts working.
>
> I am using Hadoop - HBase 0.18 bcoz my eclipse supports 
> hadoop-0.18.0-eclipse-plugin.
>        When I switch to Hadoop-HBase 0.19 and use 
> hadoop-0.19.0-eclipse-plugin then my eclipse doesn't show mapreduce 
> perspective. I am using Eclipse Platform (GANYMEDE), Version: 3.4.1.
>
> Can you tell which version of eclipse supports Hadoop - HBase 0.19 and can 
> use hadoop-0.19.0-eclipse-plugin?
>
> Thanks & Regards
> Aseem Puri
>
> -----Original Message-----
> From: [email protected] [mailto:[email protected]] On Behalf Of Jean-Daniel 
> Cryans
> Sent: Wednesday, March 18, 2009 6:11 PM
> To: [email protected]
> Subject: Re: problem in configuring hbase with hdfs
>
> Aseem,
>
> What happened here is that your master was able to create the
> namespace on the namenode but then wasn't able to write the files
> because no datanodes are alive (then it tries to read and it failed
> because the file was empty).
>
> That storage version exception means that you probably at some point
> formatted your namenode but didn't delete the datanode's data. Do
> that, start dfs, look for any exceptions (should be cleared but just
> in case) then start HBase. You should probably reformat your namenode
> first too.
>
> Any reason why you are using 0.18?
>
> J-D
>
> On Wed, Mar 18, 2009 at 12:24 AM, Puri, Aseem <[email protected]> 
> wrote:
>> Hi
>>        There is exception while staring HBase master. The exceptions are:
>>
>> Wed Mar 18 09:10:51 IST 2009 Starting master on ie11dtxpficbfise
>> java version "1.6.0_11"
>> Java(TM) SE Runtime Environment (build 1.6.0_11-b03)
>> Java HotSpot(TM) Client VM (build 11.0-b16, mixed mode, sharing)
>> ulimit -n 256
>> 2009-03-18 09:10:55,650 INFO org.apache.hadoop.hbase.master.HMaster: Root 
>> region dir: hdfs://localhost:9000/hbase/-ROOT-/70236052
>> 2009-03-18 09:10:55,806 FATAL org.apache.hadoop.hbase.master.HMaster: Not 
>> starting HMaster because:
>> java.io.EOFException
>>        at java.io.DataInputStream.readUnsignedShort(DataInputStream.java:323)
>>        at java.io.DataInputStream.readUTF(DataInputStream.java:572)
>>        at org.apache.hadoop.hbase.util.FSUtils.getVersion(FSUtils.java:101)
>>        at org.apache.hadoop.hbase.util.FSUtils.checkVersion(FSUtils.java:120)
>>        at org.apache.hadoop.hbase.master.HMaster.<init>(HMaster.java:203)
>>        at org.apache.hadoop.hbase.master.HMaster.<init>(HMaster.java:147)
>>        at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native 
>> Method)
>>        at 
>> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39)
>>        at 
>> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
>>        at java.lang.reflect.Constructor.newInstance(Constructor.java:513)
>>        at org.apache.hadoop.hbase.master.HMaster.doMain(HMaster.java:784)
>>        at org.apache.hadoop.hbase.master.HMaster.main(HMaster.java:818)
>> 2009-03-18 09:10:55,806 ERROR org.apache.hadoop.hbase.master.HMaster: Can 
>> not start master
>> java.lang.reflect.InvocationTargetException
>>        at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native 
>> Method)
>>        at 
>> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39)
>>        at 
>> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
>>        at java.lang.reflect.Constructor.newInstance(Constructor.java:513)
>>        at org.apache.hadoop.hbase.master.HMaster.doMain(HMaster.java:784)
>>        at org.apache.hadoop.hbase.master.HMaster.main(HMaster.java:818)
>> Caused by: java.io.EOFException
>>        at java.io.DataInputStream.readUnsignedShort(DataInputStream.java:323)
>>        at java.io.DataInputStream.readUTF(DataInputStream.java:572)
>>        at org.apache.hadoop.hbase.util.FSUtils.getVersion(FSUtils.java:101)
>>        at org.apache.hadoop.hbase.util.FSUtils.checkVersion(FSUtils.java:120)
>>        at org.apache.hadoop.hbase.master.HMaster.<init>(HMaster.java:203)
>>        at org.apache.hadoop.hbase.master.HMaster.<init>(HMaster.java:147)
>>        ... 6 more
>>
>> Also when I start my start my hadoop server datanode not started it also
>> throw exceptions. Datanode exceptions are:
>>
>> 2009-03-18 09:08:47,354 INFO org.apache.hadoop.dfs.DataNode: STARTUP_MSG:
>> /************************************************************
>> STARTUP_MSG: Starting DataNode
>> STARTUP_MSG:   host = ie11dtxpficbfise/199.63.66.65
>> STARTUP_MSG:   args = []
>> STARTUP_MSG:   version = 0.18.0
>> STARTUP_MSG:   build = 
>> http://svn.apache.org/repos/asf/hadoop/core/branches/branch-0.18 -r 686010; 
>> compiled by 'hadoopqa' on Thu Aug 14 19:48:33 UTC 2008
>> ************************************************************/
>> 2009-03-18 09:08:56,667 ERROR org.apache.hadoop.dfs.DataNode: 
>> org.apache.hadoop.dfs.IncorrectVersionException: Unexpected version of 
>> storage directory C:\tmp\hadoop-HadoopAdmin\dfs\data. Reported: -18. 
>> Expecting = -16.
>>        at org.apache.hadoop.dfs.Storage.getFields(Storage.java:584)
>>        at org.apache.hadoop.dfs.DataStorage.getFields(DataStorage.java:171)
>>        at 
>> org.apache.hadoop.dfs.Storage$StorageDirectory.read(Storage.java:164)
>>        at 
>> org.apache.hadoop.dfs.Storage$StorageDirectory.read(Storage.java:153)
>>        at 
>> org.apache.hadoop.dfs.DataStorage.doTransition(DataStorage.java:221)
>>        at 
>> org.apache.hadoop.dfs.DataStorage.recoverTransitionRead(DataStorage.java:141)
>>        at org.apache.hadoop.dfs.DataNode.startDataNode(DataNode.java:273)
>>        at org.apache.hadoop.dfs.DataNode.<init>(DataNode.java:190)
>>        at org.apache.hadoop.dfs.DataNode.makeInstance(DataNode.java:2987)
>>        at 
>> org.apache.hadoop.dfs.DataNode.instantiateDataNode(DataNode.java:2942)
>>        at org.apache.hadoop.dfs.DataNode.createDataNode(DataNode.java:2950)
>>        at org.apache.hadoop.dfs.DataNode.main(DataNode.java:3072)
>>
>> 2009-03-18 09:08:56,682 INFO org.apache.hadoop.dfs.DataNode: SHUTDOWN_MSG:
>> /************************************************************
>> SHUTDOWN_MSG: Shutting down DataNode at ie11dtxpficbfise/199.63.66.65
>> ************************************************************/
>>
>> So is there any relation of Hbase master with hadoop datanode exceptions 
>> that's why it do not start?
>>
>> Please tell where is the problem bcoz of it Hbase master do not start
>>
>> Thanks & Regards
>> Aseem Puri
>>
>>
>> -----Original Message-----
>> From: [email protected] [mailto:[email protected]] On Behalf Of 
>> Jean-Daniel Cryans
>> Sent: Tuesday, March 17, 2009 5:26 PM
>> To: [email protected]
>> Subject: Re: problem in configuring hbase with hdfs
>>
>> Aseem,
>>
>> It tells you that there is no master to stop so it means that
>> something went wrong when it got started and shut down by itself. Can
>> you look in your master log and see if there are any exceptions
>> thrown?
>>
>> Thx,
>>
>> J-D
>>
>> On Tue, Mar 17, 2009 at 3:32 AM, Puri, Aseem <[email protected]> 
>> wrote:
>>> Hi
>>>
>>>
>>>
>>> I am newbie working on Hadoop - HBase. I am using Hadoop-0.18.0 and
>>> HBase-0.18.1. There is some problem with using HBase master when HBase
>>> uses HDFS. My Hadoop-site configuration is:
>>>
>>>
>>>
>>> <configuration>
>>>
>>>  <property>
>>>
>>>    <name>fs.default.name</name>
>>>
>>>    <value>hdfs://localhost:9000</value>
>>>
>>>  </property>
>>>
>>>  <property>
>>>
>>>    <name>mapred.job.tracker</name>
>>>
>>>    <value>localhost:9001</value>
>>>
>>>  </property>
>>>
>>>  <property>
>>>
>>>    <name>dfs.replication</name>
>>>
>>>    <value>1</value>
>>>
>>>  </property>
>>>
>>> </configuration>
>>>
>>>
>>>
>>> And HBase-site configuration is:
>>>
>>>
>>>
>>> <configuration>
>>>
>>>  <property>
>>>
>>>    <name>hbase.rootdir</name>
>>>
>>>    <value>hdfs://localhost:9000/hbase</value>
>>>
>>>    <description>The directory shared by region servers.
>>>
>>>    </description>
>>>
>>>  </property>
>>>
>>>  <property>
>>>
>>>    <name>hbase.master</name>
>>>
>>>    <value>localhost:60000</value>
>>>
>>>    <description>The host and port that the HBase master runs at.
>>>
>>>    </description>
>>>
>>>  </property>
>>>
>>> <property>
>>>
>>>    <name>hbase.regionserver</name>
>>>
>>>    <value>localhost:60020</value>
>>>
>>>    <description>The host and port a HBase region server runs at.
>>>
>>>    </description>
>>>
>>>  </property>
>>>
>>>  </configuration>
>>>
>>>
>>>
>>> When I try command $ bin/hadoop dfs -ls / I got follwing result:
>>>
>>>
>>>
>>> $ bin/hadoop dfs -ls /
>>>
>>> Found 2 items
>>>
>>> drwxr-xr-x   - HadoopAdmin supergroup          0 2009-03-17 12:22 /hbase
>>>
>>> drwxr-xr-x   - HadoopAdmin supergroup          0 2009-03-17 12:21 /tmp
>>>
>>>
>>>
>>> But when I use HBase command list in $ bin/hbase shell I got following
>>> exception:
>>>
>>>
>>>
>>> hbase(main):001:0> list
>>>
>>> 09/03/17 12:24:24 INFO ipc.Client: Retrying connect to server:
>>> localhost/127.0.0
>>>
>>> .1:60000. Already tried 0 time(s).
>>>
>>> 09/03/17 12:24:26 INFO ipc.Client: Retrying connect to server:
>>> localhost/127.0.0
>>>
>>> .1:60000. Already tried 1 time(s).
>>>
>>> 09/03/17 12:24:28 INFO ipc.Client: Retrying connect to server:
>>> localhost/127.0.0
>>>
>>> .1:60000. Already tried 2 time(s).
>>>
>>> 09/03/17 12:24:29 INFO client.HConnectionManager$TableServers: Attempt 0
>>> of 5 fa
>>>
>>> iled with <java.io.IOException: Call failed on local exception>.
>>> Retrying after
>>>
>>> sleep of 2000
>>>
>>> 09/03/17 12:24:33 INFO ipc.Client: Retrying connect to server:
>>> localhost/127.0.0
>>>
>>> .1:60000. Already tried 0 time(s).
>>>
>>> 09/03/17 12:24:35 INFO ipc.Client: Retrying connect to server:
>>> localhost/127.0.0
>>>
>>> .1:60000. Already tried 1 time(s).
>>>
>>> 09/03/17 12:24:37 INFO ipc.Client: Retrying connect to server:
>>> localhost/127.0.0
>>>
>>> .1:60000. Already tried 2 time(s).
>>>
>>> 09/03/17 12:24:38 INFO client.HConnectionManager$TableServers: Attempt 1
>>> of 5 fa
>>>
>>> iled with <java.io.IOException: Call failed on local exception>.
>>> Retrying after
>>>
>>> sleep of 2000
>>>
>>> 09/03/17 12:24:42 INFO ipc.Client: Retrying connect to server:
>>> localhost/127.0.0
>>>
>>> .1:60000. Already tried 0 time(s).
>>>
>>> 09/03/17 12:24:44 INFO ipc.Client: Retrying connect to server:
>>> localhost/127.0.0
>>>
>>> .1:60000. Already tried 1 time(s).
>>>
>>> 09/03/17 12:24:46 INFO ipc.Client: Retrying connect to server:
>>> localhost/127.0.0
>>>
>>> .1:60000. Already tried 2 time(s).
>>>
>>> 09/03/17 12:24:48 INFO client.HConnectionManager$TableServers: Attempt 2
>>> of 5 fa
>>>
>>> iled with <java.io.IOException: Call failed on local exception>.
>>> Retrying after
>>>
>>> sleep of 2000
>>>
>>> 09/03/17 12:24:52 INFO ipc.Client: Retrying connect to server:
>>> localhost/127.0.0
>>>
>>> .1:60000. Already tried 0 time(s).
>>>
>>> 09/03/17 12:24:54 INFO ipc.Client: Retrying connect to server:
>>> localhost/127.0.0
>>>
>>> .1:60000. Already tried 1 time(s).
>>>
>>> 09/03/17 12:24:56 INFO ipc.Client: Retrying connect to server:
>>> localhost/127.0.0
>>>
>>> .1:60000. Already tried 2 time(s).
>>>
>>> 09/03/17 12:24:57 INFO client.HConnectionManager$TableServers: Attempt 3
>>> of 5 fa
>>>
>>> iled with <java.io.IOException: Call failed on local exception>.
>>> Retrying after
>>>
>>> sleep of 4000
>>>
>>> 09/03/17 12:25:03 INFO ipc.Client: Retrying connect to server:
>>> localhost/127.0.0
>>>
>>> .1:60000. Already tried 0 time(s).
>>>
>>> 09/03/17 12:25:05 INFO ipc.Client: Retrying connect to server:
>>> localhost/127.0.0
>>>
>>> .1:60000. Already tried 1 time(s).
>>>
>>> 09/03/17 12:25:07 INFO ipc.Client: Retrying connect to server:
>>> localhost/127.0.0
>>>
>>> .1:60000. Already tried 2 time(s).
>>>
>>> NativeException: org.apache.hadoop.hbase.MasterNotRunningException:
>>> localhost:60
>>>
>>> 000
>>>
>>>        from
>>> org/apache/hadoop/hbase/client/HConnectionManager.java:221:in `getM
>>>
>>> aster'
>>>
>>>        from org/apache/hadoop/hbase/client/HBaseAdmin.java:67:in
>>> `<init>'
>>>
>>>        from sun/reflect/NativeConstructorAccessorImpl.java:-2:in
>>> `newInstance0'
>>>
>>>
>>>
>>>        from sun/reflect/NativeConstructorAccessorImpl.java:39:in
>>> `newInstance'
>>>
>>>        from sun/reflect/DelegatingConstructorAccessorImpl.java:27:in
>>> `newInstan
>>>
>>> ce'
>>>
>>>        from java/lang/reflect/Constructor.java:513:in `newInstance'
>>>
>>>        from org/jruby/javasupport/JavaConstructor.java:195:in
>>> `new_instance'
>>>
>>>        from
>>> org.jruby.javasupport.JavaConstructorInvoker$new_instance_method_0_
>>>
>>> 0:-1:in `call'
>>>
>>>        from org/jruby/runtime/CallSite.java:261:in `call'
>>>
>>>        from org/jruby/evaluator/ASTInterpreter.java:670:in `callNode'
>>>
>>>        from org/jruby/evaluator/ASTInterpreter.java:324:in
>>> `evalInternal'
>>>
>>>        from org/jruby/evaluator/ASTInterpreter.java:2173:in `setupArgs'
>>>
>>>        from org/jruby/evaluator/ASTInterpreter.java:571:in
>>> `attrAssignNode'
>>>
>>>        from org/jruby/evaluator/ASTInterpreter.java:309:in
>>> `evalInternal'
>>>
>>>        from org/jruby/evaluator/ASTInterpreter.java:620:in `blockNode'
>>>
>>>        from org/jruby/evaluator/ASTInterpreter.java:318:in
>>> `evalInternal'
>>>
>>> ... 178 levels...
>>>
>>>        from
>>> ruby/C_3a_/Documents_20_and_20_Settings/HadoopAdmin/hbase/bin/C:\DO
>>>
>>> CUME~1\HADOOP~1\hbase\/bin/hirb.rb:-1:in `__file__'
>>>
>>>        from
>>> ruby/C_3a_/Documents_20_and_20_Settings/HadoopAdmin/hbase/bin/C:\DO
>>>
>>> CUME~1\HADOOP~1\hbase\/bin/hirb.rb:-1:in `load'
>>>
>>>        from org/jruby/Ruby.java:512:in `runScript'
>>>
>>>        from org/jruby/Ruby.java:432:in `runNormally'
>>>
>>>        from org/jruby/Ruby.java:312:in `runFromMain'
>>>
>>>        from org/jruby/Main.java:144:in `run'
>>>
>>>        from org/jruby/Main.java:89:in `run'
>>>
>>>        from org/jruby/Main.java:80:in `main'
>>>
>>>        from file:/C:/Documents and
>>> Settings/HadoopAdmin/hbase/lib/jruby-complet
>>>
>>> e-1.1.2.jar!/builtin/javasupport/proxy/concrete.rb:23:in `initialize'
>>>
>>>        from file:/C:/Documents and
>>> Settings/HadoopAdmin/hbase/lib/jruby-complet
>>>
>>> e-1.1.2.jar!/builtin/javasupport/proxy/concrete.rb:6:in `new'
>>>
>>>        from file:/C:/Documents and
>>> Settings/HadoopAdmin/hbase/lib/jruby-complet
>>>
>>> e-1.1.2.jar!/builtin/javasupport/proxy/concrete.rb:6:in `new'
>>>
>>>        from C:/DOCUME~1/HADOOP~1/hbase/bin/HBase.rb:37:in `initialize'
>>>
>>>        from C:\DOCUME~1\HADOOP~1\hbase\/bin/hirb.rb:218:in `new'
>>>
>>>        from C:\DOCUME~1\HADOOP~1\hbase\/bin/hirb.rb:218:in `admin'
>>>
>>>        from C:\DOCUME~1\HADOOP~1\hbase\/bin/hirb.rb:242:in `list'
>>>
>>>        from (hbase):2:in `binding'hbase(main):002:0>
>>>
>>>
>>>
>>>
>>>
>>> Also when I try to stop HBase with the command $ bin/stop-hbase.sh I got
>>> the message:
>>>
>>> no master to stop
>>>
>>>
>>>
>>> When I change my HBase-site configuration to:
>>>
>>>
>>>
>>> <configuration>
>>>
>>> <property>
>>>
>>>    <name>hbase.master</name>
>>>
>>>    <value>localhost:60000</value>
>>>
>>>    <description>The host and port that the HBase master runs at.
>>>
>>>    </description>
>>>
>>>  </property>
>>>
>>> <property>
>>>
>>>    <name>hbase.regionserver</name>
>>>
>>>    <value>localhost:60020</value>
>>>
>>>    <description>The host and port a HBase region server runs at.
>>>
>>>    </description>
>>>
>>>  </property>
>>>
>>>  </configuration>
>>>
>>>
>>>
>>> My master starts working as HBase is now using local file system.
>>>
>>>
>>>
>>> But when I want to use HDFS and change HBase-site configuration then my
>>> master do not starts. Please tell me how I should configure so my HBase
>>> master starts working and HBase use HDFS. Hope you people help me in
>>> this.
>>>
>>>
>>>
>>> -Aseem
>>>
>>>
>>>
>>>
>>
>

Reply via email to