Which port does your HDFS listen to? It is not 9000, right? <name>hbase.rootdir</name> <value>hdfs://localhost:9000/hbase</value>
You need to fix this and make sure your HDFS is working, for example, the following command should work for you. hadoop fs -ls / On Mon, Feb 13, 2012 at 11:44 AM, Bing Li <lbl...@gmail.com> wrote: > Dear Jimmy, > > I configured the standalone mode successfully. But I wonder why the pseudo > distributed one does work. > > I checked in logs and got the following exceptions. Does the information > give you some hints? > > Thanks so much for your help again! > > Best, > Bing > > 2012-02-13 18:25:49,782 FATAL org.apache.hadoop.hbase.master.HMaster: > Unhandled exception. Starting shutdown. > java.net.ConnectException: Call to localhost/127.0.0.1:9000 failed on > connection exception: java.net.ConnectException: Connection refuse > d > at org.apache.hadoop.ipc.Client.wrapException(Client.java:1095) > at org.apache.hadoop.ipc.Client.call(Client.java:1071) > at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:225) > at $Proxy10.getProtocolVersion(Unknown Source) > at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:396) > at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:379) > at org.apache.hadoop.hdfs.DFSClient.createRPCNamenode(DFSClient.java:119) > at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:238) > at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:203) > at > org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:89) > at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1386) > at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:66) > at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1404) > at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:254) > at org.apache.hadoop.fs.Path.getFileSystem(Path.java:187) > at org.apache.hadoop.hbase.util.FSUtils.getRootDir(FSUtils.java:471) > at > org.apache.hadoop.hbase.master.MasterFileSystem.<init>(MasterFileSystem.java:94) > at > org.apache.hadoop.hbase.master.HMaster.finishInitialization(HMaster.java:448) > at org.apache.hadoop.hbase.master.HMaster.run(HMaster.java:326) > at java.lang.Thread.run(Thread.java:662) > Caused by: java.net.ConnectException: Connection refused > at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) > at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567) > at > org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206) > at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:656) > at org.apache.hadoop.ipc.Client$Connection.setupConnection(Client.java:434) > at > org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:560) > at org.apache.hadoop.ipc.Client$Connection.access$2000(Client.java:184) > at org.apache.hadoop.ipc.Client.getConnection(Client.java:1202) > at org.apache.hadoop.ipc.Client.call(Client.java:1046) > ... 18 more > 2012-02-13 18:25:49,787 INFO org.apache.hadoop.hbase.master.HMaster: > Aborting > 2012-02-13 18:25:49,787 DEBUG org.apache.hadoop.hbase.master.HMaster: > Stopping service threads > > > Thanks so much! > Bing > > > On Tue, Feb 14, 2012 at 3:35 AM, Jimmy Xiang <jxi...@cloudera.com> wrote: > >> In this case, you may just use the standalone mode. You can follow the >> quick start step by step. >> >> The default zookeeper port is 2181, you don't need to configure it. >> >> >> >> On Mon, Feb 13, 2012 at 11:28 AM, Bing Li <lbl...@gmail.com> wrote: >> >>> Dear Jimmy, >>> >>> I am a new user of HBase. My experiences in HBase and Hadoop is very >>> limited. I just tried to follow some books, such as Hadoop/HBase the >>> Definitive Guide. However, I still got some problems. >>> >>> What I am trying to do is just to set up a pseudo distributed HBase >>> environment on a single node. After that, I will start my system >>> programming in Java. I hope I could deploy the system in fully distributed >>> mode when my system is done. >>> >>> So what I am configuring is very simple. Do I need to set up the >>> zookeeper port in hbase-site.xml? >>> >>> Thanks so much! >>> >>> Best, >>> Bing >>> >>> >>> On Tue, Feb 14, 2012 at 3:16 AM, Jimmy Xiang <jxi...@cloudera.com>wrote: >>> >>>> Have you restarted your HBase after the change? What's the zookeeper >>>> port does your HMaster use? >>>> >>>> Can you run the following to checkout where is your HMaster as below? >>>> >>>> hbase zkcli >>>> then: get /hbase/master >>>> It should show you master location. >>>> >>>> It seems you have a distributed installation. How many regionservers >>>> do you have? Can you check your >>>> master web UI to make sure all look fine. >>>> >>>> Thanks, >>>> Jimmy >>>> >>>> >>>> On Mon, Feb 13, 2012 at 10:51 AM, Bing Li <lbl...@gmail.com> wrote: >>>> >>>>> Dear Jimmy, >>>>> >>>>> Thanks so much for your reply! >>>>> >>>>> I didn't set up the zookeeper.quorom. After getting your email, I made >>>>> a change. Now my hbase-site.xml is as follows. >>>>> >>>>> <configuration> >>>>> <property> >>>>> <name>hbase.rootdir</name> >>>>> <value>hdfs://localhost:9000/hbase</value> >>>>> </property> >>>>> <property> >>>>> <name>dfs.replication</name> >>>>> <value>1</value> >>>>> </property> >>>>> <property> >>>>> <name>hbase.cluster.distributed</name> >>>>> <value>true</value> >>>>> </property> >>>>> <property> >>>>> <name>hbase.zookeeper.quorum</name> >>>>> <value>localhost</value> >>>>> </property> >>>>> </configuration> >>>>> >>>>> The previous error is still existed. I feel weird why HBase developers >>>>> cannot provide a reliable description about their work. >>>>> >>>>> Best, >>>>> Bing >>>>> >>>>> >>>>> On Tue, Feb 14, 2012 at 2:16 AM, Jimmy Xiang <jxi...@cloudera.com>wrote: >>>>> >>>>>> What's your hbase.zookeeper.quorom configuration? You can check out >>>>>> this quick start guide: >>>>>> >>>>>> http://hbase.apache.org/book/quickstart.html >>>>>> >>>>>> Thanks, >>>>>> Jimmy >>>>>> >>>>>> >>>>>> On Mon, Feb 13, 2012 at 10:09 AM, Bing Li <lbl...@gmail.com> wrote: >>>>>> >>>>>>> Dear all, >>>>>>> >>>>>>> After searching on the Web and asking for help from friends, I >>>>>>> noticed that >>>>>>> the pseudo distributed configuration in the book, HBase the >>>>>>> Definitive >>>>>>> Guide, was not complete. Now the ZooKeeper related exception is >>>>>>> fixed. >>>>>>> However, I got another error when typing "status" in the HBase shell. >>>>>>> >>>>>>> ERROR: org.apache.hadoop.hbase.MasterNotRunningException: Retried >>>>>>> 7 >>>>>>> Times >>>>>>> >>>>>>> I am trying to fix it myself. Your help is highly appreciated. >>>>>>> >>>>>>> Thanks so much! >>>>>>> Bing Li >>>>>>> >>>>>>> On Mon, Feb 13, 2012 at 5:00 AM, Bing Li <lbl...@gmail.com> wrote: >>>>>>> >>>>>>> > Dear all, >>>>>>> > >>>>>>> > I am a new learner of HBase. I tried to set up my HBase on a >>>>>>> > pseudo-distributed HDFS. >>>>>>> > >>>>>>> > After starting HDFS by running ./start-dfs.sh and >>>>>>> ./start-hbase.sh, I >>>>>>> > started the HBase shell. >>>>>>> > >>>>>>> > ./hbase shell >>>>>>> > >>>>>>> > It was started properly. However, when I typed the command, >>>>>>> status, as >>>>>>> > follows. >>>>>>> > >>>>>>> > hbase(main):001:0> status >>>>>>> > >>>>>>> > It got the following exception. Since I had very limited >>>>>>> experiences to >>>>>>> > use HBase, I could not figure out what the problem was. >>>>>>> > >>>>>>> > SLF4J: Class path contains multiple SLF4J bindings. >>>>>>> > SLF4J: Found binding in >>>>>>> > >>>>>>> [jar:file:/opt/hbase-0.92.0/lib/slf4j-log4j12-1.5.8.jar!/org/slf4j/impl/StaticLoggerBinder.class] >>>>>>> > SLF4J: Found binding in >>>>>>> > >>>>>>> [jar:file:/opt/hadoop-1.0.0/lib/slf4j-log4j12-1.4.3.jar!/org/slf4j/impl/StaticLoggerBinder.class] >>>>>>> > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for >>>>>>> an >>>>>>> > explanation. >>>>>>> > 12/02/13 04:34:01 ERROR zookeeper.RecoverableZooKeeper: ZooKeeper >>>>>>> exists >>>>>>> > failed after 3 retries >>>>>>> > 12/02/13 04:34:01 WARN zookeeper.ZKUtil: hconnection Unable to set >>>>>>> watcher >>>>>>> > on znode /hbase/master >>>>>>> > org.apache.zookeeper.KeeperException$ConnectionLossException: >>>>>>> > KeeperErrorCode = ConnectionLoss for /hbase/master >>>>>>> > at >>>>>>> > >>>>>>> org.apache.zookeeper.KeeperException.create(KeeperException.java:99) >>>>>>> > at >>>>>>> > >>>>>>> org.apache.zookeeper.KeeperException.create(KeeperException.java:51) >>>>>>> > at >>>>>>> org.apache.zookeeper.ZooKeeper.exists(ZooKeeper.java:1003) >>>>>>> > at >>>>>>> > >>>>>>> org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.exists(RecoverableZooKeeper.java:154) >>>>>>> > at >>>>>>> > >>>>>>> org.apache.hadoop.hbase.zookeeper.ZKUtil.watchAndCheckExists(ZKUtil.java:226) >>>>>>> > at >>>>>>> > >>>>>>> org.apache.hadoop.hbase.zookeeper.ZooKeeperNodeTracker.start(ZooKeeperNodeTracker.java:76) >>>>>>> > at >>>>>>> > >>>>>>> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.setupZookeeperTrackers(HConnectionManager.java:580) >>>>>>> > at >>>>>>> > >>>>>>> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.<init>(HConnectionManager.java:569) >>>>>>> > at >>>>>>> > >>>>>>> org.apache.hadoop.hbase.client.HConnectionManager.getConnection(HConnectionManager.java:186) >>>>>>> > at >>>>>>> > >>>>>>> org.apache.hadoop.hbase.client.HBaseAdmin.<init>(HBaseAdmin.java:98) >>>>>>> > at >>>>>>> sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native >>>>>>> > Method) >>>>>>> > at >>>>>>> > >>>>>>> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39) >>>>>>> > at >>>>>>> > >>>>>>> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27) >>>>>>> > at >>>>>>> java.lang.reflect.Constructor.newInstance(Constructor.java:513) >>>>>>> > at >>>>>>> > >>>>>>> org.jruby.javasupport.JavaConstructor.newInstanceDirect(JavaConstructor.java:275) >>>>>>> > at >>>>>>> > >>>>>>> org.jruby.java.invokers.ConstructorInvoker.call(ConstructorInvoker.java:91) >>>>>>> > at >>>>>>> > >>>>>>> org.jruby.java.invokers.ConstructorInvoker.call(ConstructorInvoker.java:178) >>>>>>> > at >>>>>>> > >>>>>>> org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:322) >>>>>>> > at >>>>>>> > >>>>>>> org.jruby.runtime.callsite.CachingCallSite.callBlock(CachingCallSite.java:178) >>>>>>> > at >>>>>>> > >>>>>>> org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:182) >>>>>>> > at >>>>>>> > >>>>>>> org.jruby.java.proxies.ConcreteJavaProxy$2.call(ConcreteJavaProxy.java:47) >>>>>>> > at >>>>>>> > >>>>>>> org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:322) >>>>>>> > >>>>>>> > Could you please give me a hand? Thanks so much! >>>>>>> > >>>>>>> > Best regards, >>>>>>> > Bing >>>>>>> > >>>>>>> > >>>>>>> > >>>>>>> >>>>>> >>>>>> >>>>> >>>> >>> >> >