You can also use netstat -a -t
or if you have netcat (nc) installed, try nc -v -z hostname port --- Jim Kellerman, Senior Engineer; Powerset > -----Original Message----- > From: stack [mailto:[EMAIL PROTECTED] > Sent: Saturday, August 02, 2008 10:06 AM > To: [email protected] > Subject: Re: Hbase single-Node cluster config problem > > That looks like it should work thought IIRC, its not needed. > The webui comes up by default. Have you done as J-D suggests > below, tried telneting to the port to see that there is > something listening? Look in the HDFS logs to see if its > registering putting up a webui. > > St.Ack > > > Yabo-Arber Xu wrote: > > Thanks J-D and St.Act for your help. I will try what you > suggested to > > expand the cluster. > > > > St.Act: I explicitly set the following property in hadoop-site.xml > > > > <property> > > <name>dfs.http.address</name> > > <value>*my_host_name*:50070 <http://0.0.0.0:50070/></value> > > </property> > > > > I also can see there is a instance listening on 50070, but > when i type > > http://*my_host_name*:50070 <http://0.0.0.0:50070/> in the > firefox of > > the other computer, there is still no connection. Did I > miss anything? > > > > Thanks again. > > > > On Fri, Aug 1, 2008 at 3:05 PM, stack <[EMAIL PROTECTED]> wrote: > > > > > >> Default for HDFS webui is: > >> > >> <property> > >> <name>dfs.http.address</name> > >> <value>0.0.0.0:50070</value> > >> <description> > >> The address and the base port where the dfs namenode web ui will > >> listen on. > >> If the port is 0 then the server will start on a free port. > >> </description> > >> </property> > >> > >> I may not be reading the below properly but it looks like there is > >> something listening on 50070. > >> > >> St.Ack > >> > >> > >> Yabo-Arber Xu wrote: > >> > >> > >>> Hi J-D, > >>> > >>> Sorry that just now forgot to ask another question. Even though i > >>> have HDFS and HBase run well one one computer, but > strangely I can > >>> not connect to HDFS using WebUI. I run the following > command on my > >>> computer, and it seems the only two port active are for HDFS and > >>> HBase, but there is no such default port open for WebUI > connection. > >>> > >>> netstat -plten | grep java > >>> > >>> tcp 0 0 10.254.199.132:60000 0.0.0.0:* > >>> LISTEN 0 tcp 0 0 0.0.0.0:37669 > >>> 0.0.0.0:* LISTEN 0 tcp > 0 0 > >>> 10.254.199.132:54310 0.0.0.0:* LISTEN > >>> 0 tcp 0 0 0.0.0.0:49769 > 0.0.0.0:* > >>> LISTEN 0 tcp 0 0 0.0.0.0:60010 > >>> 0.0.0.0:* LISTEN 0 tcp > 0 0 > >>> 0.0.0.0:50090 0.0.0.0:* LISTEN > >>> 0 tcp 0 0 0.0.0.0:60020 > 0.0.0.0:* > >>> LISTEN 0 tcp 0 0 0.0.0.0:50070 > >>> 0.0.0.0:* LISTEN 0 tcp > 0 0 > >>> 0.0.0.0:41625 0.0.0.0:* LISTEN > >>> 0 tcp 0 0 0.0.0.0:50010 > 0.0.0.0:* > >>> LISTEN 0 tcp 0 0 0.0.0.0:50075 > >>> 0.0.0.0:* LISTEN 0 tcp > 0 0 > >>> 0.0.0.0:60030 0.0.0.0:* LISTEN > >>> 0 > >>> > >>> > >>> Thanks, > >>> Arber > >>> > >>> On Fri, Aug 1, 2008 at 2:48 PM, Yabo-Arber Xu > >>> <[EMAIL PROTECTED] > >>> > >>>> wrote: > >>>> > >>> > >>> > >>>> Hi J-D, > >>>> > >>>> Thanks, J-D. I cleaned HDFS directory and re-run it. > It's fine now. > >>>> > >>>> Wondered if there is any documents out there showing how > to expand > >>>> such > >>>> one- computer-with all-servers structure to a truely distributed > >>>> one but without re-importing all the data? > >>>> > >>>> Thanks again, > >>>> Arber > >>>> > >>>> > >>>> On Fri, Aug 1, 2008 at 6:13 AM, Jean-Daniel Cryans > >>>> <[EMAIL PROTECTED] > >>>> > >>>>> wrote: > >>>>> > >>>> > >>>> > >>>>> Yair, > >>>>> > >>>>> It seems that your master is unable to communicate with HDFS > >>>>> (that's the SocketTimeoutException). To correct this, I would > >>>>> check that HDFS is running by looking at the web UI, I > would make > >>>>> sure that the ports are open (using telnet for example) and I > >>>>> would also check that HDFS uses the default ports. > >>>>> > >>>>> J-D > >>>>> > >>>>> On Fri, Aug 1, 2008 at 5:40 AM, Yabo-Arber Xu > >>>>> <[EMAIL PROTECTED] > >>>>> > >>>>> > >>>>> > >>>>>> wrote: > >>>>>> Greetings, > >>>>>> > >>>>>> I am trying to set up a hbase cluster. To simplify the > setting, i > >>>>>> first tried the single node cluster, where HDFS name/data node > >>>>>> are set on one computer, and hbase master/regionserver > are also > >>>>>> set on the same > >>>>>> > >>>>>> > >>>>>> > >>>>> computer. > >>>>> > >>>>> > >>>>> > >>>>>> The HDFS passed the test and works well. But, for > hbase, when I > >>>>>> try to create a table using hbase shell. It keeps popping the > >>>>>> following > >>>>>> > >>>>>> > >>>>>> > >>>>> message: > >>>>> > >>>>> > >>>>> > >>>>>> 08/08/01 02:30:29 INFO ipc.Client: Retrying connect to server: > >>>>>> ec2-67-202-24-167.compute-1.amazonaws.com/10.254.199.132:60000. > >>>>>> Already > >>>>>> tried 1 time(s). > >>>>>> > >>>>>> I checked the hbase log, and it has the following error: > >>>>>> > >>>>>> 2008-08-01 02:30:24,337 ERROR org.apache.hadoop.hbase.HMaster: > >>>>>> Can not start master > java.lang.reflect.InvocationTargetException > >>>>>> at > >>>>>> sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native > >>>>>> Method) > >>>>>> at > >>>>>> > >>>>>> > >>>>>> > >>>>>> > >>>>>> > >>>>> > sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstr > >>>>> uctorAccessorImpl.java:39) > >>>>> > >>>>> > >>>>> > >>>>>> at > >>>>>> > >>>>>> > >>>>>> > >>>>>> > >>>>>> > >>>>> > sun.reflect.DelegatingConstructorAccessorImpl.newInstance(Delegati > >>>>> ngConstructorAccessorImpl.java:27) > >>>>> > >>>>> > >>>>> > >>>>>> at > >>>>>> > >>>>>> > >>>>>> > >>>>> java.lang.reflect.Constructor.newInstance(Constructor.java:513) > >>>>> > >>>>> > >>>>> > >>>>>> at > org.apache.hadoop.hbase.HMaster.doMain(HMaster.java:3313) > >>>>>> at > org.apache.hadoop.hbase.HMaster.main(HMaster.java:3347) > >>>>>> Caused by: java.net.SocketTimeoutException: timed out > waiting for > >>>>>> rpc response > >>>>>> at org.apache.hadoop.ipc.Client.call(Client.java:514) > >>>>>> at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:198) > >>>>>> at > org.apache.hadoop.dfs.$Proxy0.getProtocolVersion(Unknown > >>>>>> > >>>>>> > >>>>>> > >>>>> Source) > >>>>> > >>>>> > >>>>> > >>>>>> at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:291) > >>>>>> at > >>>>>> > org.apache.hadoop.dfs.DFSClient.createNamenode(DFSClient.java:128) > >>>>>> at > org.apache.hadoop.dfs.DFSClient.<init>(DFSClient.java:151) > >>>>>> at > >>>>>> > >>>>>> > >>>>>> > >>>>>> > >>>>>> > >>>>> > org.apache.hadoop.dfs.DistributedFileSystem.initialize(Distributed > >>>>> FileSystem.java:65) > >>>>> > >>>>> > >>>>> > >>>>>> at > >>>>>> > org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1182) > >>>>>> at > org.apache.hadoop.fs.FileSystem.access$300(FileSystem.java:55) > >>>>>> at > >>>>>> > >>>>>> > >>>>>> > >>>>> org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1193) > >>>>> > >>>>> > >>>>> > >>>>>> at > org.apache.hadoop.fs.FileSystem.get(FileSystem.java:150) > >>>>>> > >>>>>> > >>>>>> For your information, i also attach the hbase-site.xml: > >>>>>> > >>>>>> <property> > >>>>>> <name>hbase.master</name> > >>>>>> > <value>ec2-67-202-24-167.compute-1.amazonaws.com:60000</value> > >>>>>> <description>The host and port that the HBase master runs at. > >>>>>> </description> > >>>>>> </property> > >>>>>> > >>>>>> <property> > >>>>>> <name>hbase.rootdir</name> > >>>>>> > >>>>>> > <value>hdfs://ec2-67-202-24-167.compute-1.amazonaws.com:9000/hbas > >>>>>> e > >>>>>> </value> > >>>>>> <description>The directory shared by region servers. > >>>>>> </description> > >>>>>> </property> > >>>>>> > >>>>>> Can anybody point out what i did wrong? > >>>>>> > >>>>>> Thanks in advance > >>>>>> > >>>>>> -Arber > >>>>>> > >>>>>> > >>>>>> > >>>>>> > >>> > >> > > > > > > > > > No virus found in this incoming message. > Checked by AVG - http://www.avg.com > Version: 8.0.138 / Virus Database: 270.5.10/1586 - Release > Date: 8/1/2008 6:59 PM >
