Which port does your HDFS listen to? It is not 9000, right?

<name>hbase.rootdir</name>
    <value>hdfs://localhost:9000/hbase</value>

You need to fix this and make sure your HDFS is working, for example,
the following command should work for you.

hadoop fs -ls /



On Mon, Feb 13, 2012 at 11:44 AM, Bing Li <lbl...@gmail.com> wrote:

> Dear Jimmy,
>
> I configured the standalone mode successfully. But I wonder why the pseudo
> distributed one does work.
>
> I checked in logs and got the following exceptions. Does the information
> give you some hints?
>
> Thanks so much for your help again!
>
> Best,
> Bing
>
> 2012-02-13 18:25:49,782 FATAL org.apache.hadoop.hbase.master.HMaster:
> Unhandled exception. Starting shutdown.
> java.net.ConnectException: Call to localhost/127.0.0.1:9000 failed on
> connection exception: java.net.ConnectException: Connection refuse
> d
> at org.apache.hadoop.ipc.Client.wrapException(Client.java:1095)
>  at org.apache.hadoop.ipc.Client.call(Client.java:1071)
> at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:225)
>  at $Proxy10.getProtocolVersion(Unknown Source)
> at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:396)
>  at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:379)
> at org.apache.hadoop.hdfs.DFSClient.createRPCNamenode(DFSClient.java:119)
>  at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:238)
> at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:203)
>  at
> org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:89)
> at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1386)
>  at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:66)
> at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1404)
>  at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:254)
> at org.apache.hadoop.fs.Path.getFileSystem(Path.java:187)
>  at org.apache.hadoop.hbase.util.FSUtils.getRootDir(FSUtils.java:471)
> at
> org.apache.hadoop.hbase.master.MasterFileSystem.<init>(MasterFileSystem.java:94)
>  at
> org.apache.hadoop.hbase.master.HMaster.finishInitialization(HMaster.java:448)
> at org.apache.hadoop.hbase.master.HMaster.run(HMaster.java:326)
>  at java.lang.Thread.run(Thread.java:662)
> Caused by: java.net.ConnectException: Connection refused
> at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
>  at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567)
> at
> org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
>  at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:656)
> at org.apache.hadoop.ipc.Client$Connection.setupConnection(Client.java:434)
>  at
> org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:560)
> at org.apache.hadoop.ipc.Client$Connection.access$2000(Client.java:184)
>  at org.apache.hadoop.ipc.Client.getConnection(Client.java:1202)
> at org.apache.hadoop.ipc.Client.call(Client.java:1046)
>  ... 18 more
> 2012-02-13 18:25:49,787 INFO org.apache.hadoop.hbase.master.HMaster:
> Aborting
> 2012-02-13 18:25:49,787 DEBUG org.apache.hadoop.hbase.master.HMaster:
> Stopping service threads
>
>
> Thanks so much!
> Bing
>
>
> On Tue, Feb 14, 2012 at 3:35 AM, Jimmy Xiang <jxi...@cloudera.com> wrote:
>
>> In this case, you may just use the standalone mode.  You can follow the
>> quick start step by step.
>>
>> The default zookeeper port is 2181, you don't need to configure it.
>>
>>
>>
>> On Mon, Feb 13, 2012 at 11:28 AM, Bing Li <lbl...@gmail.com> wrote:
>>
>>> Dear Jimmy,
>>>
>>> I am a new user of HBase. My experiences in HBase and Hadoop is very
>>> limited. I just tried to follow some books, such as Hadoop/HBase the
>>> Definitive Guide. However, I still got some problems.
>>>
>>> What I am trying to do is just to set up a pseudo distributed HBase
>>> environment on a single node. After that, I will start my system
>>> programming in Java. I hope I could deploy the system in fully distributed
>>> mode when my system is done.
>>>
>>> So what I am configuring is very simple. Do I need to set up the
>>> zookeeper port in hbase-site.xml?
>>>
>>> Thanks so much!
>>>
>>> Best,
>>> Bing
>>>
>>>
>>> On Tue, Feb 14, 2012 at 3:16 AM, Jimmy Xiang <jxi...@cloudera.com>wrote:
>>>
>>>> Have you restarted your HBase after the change?  What's the zookeeper
>>>> port does your HMaster use?
>>>>
>>>> Can you run the following to checkout where is your HMaster as below?
>>>>
>>>> hbase zkcli
>>>>   then:  get /hbase/master
>>>>  It should show you master location.
>>>>
>>>> It seems you have a distributed installation.  How many regionservers
>>>> do you have?  Can you check your
>>>> master web UI to make sure all look fine.
>>>>
>>>> Thanks,
>>>> Jimmy
>>>>
>>>>
>>>> On Mon, Feb 13, 2012 at 10:51 AM, Bing Li <lbl...@gmail.com> wrote:
>>>>
>>>>> Dear Jimmy,
>>>>>
>>>>> Thanks so much for your reply!
>>>>>
>>>>> I didn't set up the zookeeper.quorom. After getting your email, I made
>>>>> a change. Now my hbase-site.xml is as follows.
>>>>>
>>>>> <configuration>
>>>>>   <property>
>>>>>     <name>hbase.rootdir</name>
>>>>>     <value>hdfs://localhost:9000/hbase</value>
>>>>>   </property>
>>>>>   <property>
>>>>>     <name>dfs.replication</name>
>>>>>     <value>1</value>
>>>>>   </property>
>>>>>   <property>
>>>>>     <name>hbase.cluster.distributed</name>
>>>>>     <value>true</value>
>>>>>   </property>
>>>>>   <property>
>>>>>     <name>hbase.zookeeper.quorum</name>
>>>>>     <value>localhost</value>
>>>>>   </property>
>>>>> </configuration>
>>>>>
>>>>> The previous error is still existed. I feel weird why HBase developers
>>>>> cannot provide a reliable description about their work.
>>>>>
>>>>> Best,
>>>>> Bing
>>>>>
>>>>>
>>>>> On Tue, Feb 14, 2012 at 2:16 AM, Jimmy Xiang <jxi...@cloudera.com>wrote:
>>>>>
>>>>>> What's your hbase.zookeeper.quorom configuration?   You can check out
>>>>>> this quick start guide:
>>>>>>
>>>>>> http://hbase.apache.org/book/quickstart.html
>>>>>>
>>>>>> Thanks,
>>>>>> Jimmy
>>>>>>
>>>>>>
>>>>>> On Mon, Feb 13, 2012 at 10:09 AM, Bing Li <lbl...@gmail.com> wrote:
>>>>>>
>>>>>>> Dear all,
>>>>>>>
>>>>>>> After searching on the Web and asking for help from friends, I
>>>>>>> noticed that
>>>>>>> the pseudo distributed configuration in the book, HBase the
>>>>>>> Definitive
>>>>>>> Guide, was not complete. Now the ZooKeeper related exception is
>>>>>>> fixed.
>>>>>>> However, I got another error when typing "status" in the HBase shell.
>>>>>>>
>>>>>>>    ERROR: org.apache.hadoop.hbase.MasterNotRunningException: Retried
>>>>>>> 7
>>>>>>> Times
>>>>>>>
>>>>>>> I am trying to fix it myself. Your help is highly appreciated.
>>>>>>>
>>>>>>> Thanks so much!
>>>>>>> Bing Li
>>>>>>>
>>>>>>> On Mon, Feb 13, 2012 at 5:00 AM, Bing Li <lbl...@gmail.com> wrote:
>>>>>>>
>>>>>>> > Dear all,
>>>>>>> >
>>>>>>> > I am a new learner of HBase. I tried to set up my HBase on a
>>>>>>> > pseudo-distributed HDFS.
>>>>>>> >
>>>>>>> > After starting HDFS by running ./start-dfs.sh and
>>>>>>> ./start-hbase.sh, I
>>>>>>> > started the HBase shell.
>>>>>>> >
>>>>>>> >    ./hbase shell
>>>>>>> >
>>>>>>> > It was started properly. However, when I typed the command,
>>>>>>> status, as
>>>>>>> > follows.
>>>>>>> >
>>>>>>> >    hbase(main):001:0> status
>>>>>>> >
>>>>>>> > It got the following exception. Since I had very limited
>>>>>>> experiences to
>>>>>>> > use HBase, I could not figure out what the problem was.
>>>>>>> >
>>>>>>> > SLF4J: Class path contains multiple SLF4J bindings.
>>>>>>> > SLF4J: Found binding in
>>>>>>> >
>>>>>>> [jar:file:/opt/hbase-0.92.0/lib/slf4j-log4j12-1.5.8.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>>>>>>> > SLF4J: Found binding in
>>>>>>> >
>>>>>>> [jar:file:/opt/hadoop-1.0.0/lib/slf4j-log4j12-1.4.3.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>>>>>>> > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for
>>>>>>> an
>>>>>>> > explanation.
>>>>>>> > 12/02/13 04:34:01 ERROR zookeeper.RecoverableZooKeeper: ZooKeeper
>>>>>>> exists
>>>>>>> > failed after 3 retries
>>>>>>> > 12/02/13 04:34:01 WARN zookeeper.ZKUtil: hconnection Unable to set
>>>>>>> watcher
>>>>>>> > on znode /hbase/master
>>>>>>> > org.apache.zookeeper.KeeperException$ConnectionLossException:
>>>>>>> > KeeperErrorCode = ConnectionLoss for /hbase/master
>>>>>>> >         at
>>>>>>> >
>>>>>>> org.apache.zookeeper.KeeperException.create(KeeperException.java:99)
>>>>>>> >         at
>>>>>>> >
>>>>>>> org.apache.zookeeper.KeeperException.create(KeeperException.java:51)
>>>>>>> >         at
>>>>>>> org.apache.zookeeper.ZooKeeper.exists(ZooKeeper.java:1003)
>>>>>>> >         at
>>>>>>> >
>>>>>>> org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.exists(RecoverableZooKeeper.java:154)
>>>>>>> >         at
>>>>>>> >
>>>>>>> org.apache.hadoop.hbase.zookeeper.ZKUtil.watchAndCheckExists(ZKUtil.java:226)
>>>>>>> >         at
>>>>>>> >
>>>>>>> org.apache.hadoop.hbase.zookeeper.ZooKeeperNodeTracker.start(ZooKeeperNodeTracker.java:76)
>>>>>>> >         at
>>>>>>> >
>>>>>>> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.setupZookeeperTrackers(HConnectionManager.java:580)
>>>>>>> >         at
>>>>>>> >
>>>>>>> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.<init>(HConnectionManager.java:569)
>>>>>>> >         at
>>>>>>> >
>>>>>>> org.apache.hadoop.hbase.client.HConnectionManager.getConnection(HConnectionManager.java:186)
>>>>>>> >         at
>>>>>>> >
>>>>>>> org.apache.hadoop.hbase.client.HBaseAdmin.<init>(HBaseAdmin.java:98)
>>>>>>> >         at
>>>>>>> sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
>>>>>>> > Method)
>>>>>>> >         at
>>>>>>> >
>>>>>>> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39)
>>>>>>> >         at
>>>>>>> >
>>>>>>> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
>>>>>>> >         at
>>>>>>> java.lang.reflect.Constructor.newInstance(Constructor.java:513)
>>>>>>> >         at
>>>>>>> >
>>>>>>> org.jruby.javasupport.JavaConstructor.newInstanceDirect(JavaConstructor.java:275)
>>>>>>> >         at
>>>>>>> >
>>>>>>> org.jruby.java.invokers.ConstructorInvoker.call(ConstructorInvoker.java:91)
>>>>>>> >         at
>>>>>>> >
>>>>>>> org.jruby.java.invokers.ConstructorInvoker.call(ConstructorInvoker.java:178)
>>>>>>> >         at
>>>>>>> >
>>>>>>> org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:322)
>>>>>>> >         at
>>>>>>> >
>>>>>>> org.jruby.runtime.callsite.CachingCallSite.callBlock(CachingCallSite.java:178)
>>>>>>> >         at
>>>>>>> >
>>>>>>> org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:182)
>>>>>>> >         at
>>>>>>> >
>>>>>>> org.jruby.java.proxies.ConcreteJavaProxy$2.call(ConcreteJavaProxy.java:47)
>>>>>>> >         at
>>>>>>> >
>>>>>>> org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:322)
>>>>>>> >
>>>>>>> > Could you please give me a hand? Thanks so much!
>>>>>>> >
>>>>>>> > Best regards,
>>>>>>> > Bing
>>>>>>> >
>>>>>>> >
>>>>>>> >
>>>>>>>
>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>

Reply via email to