In core-file.xml, do you have this? <configuration> <property> <name>fs.default.name</name> <value>hdfs://namenode:8020/hbase</value> </property>
If you want hbase to connect to 8020 you must have hdfs listening on 8020 as well. On Mon, May 14, 2012 at 5:17 PM, Dalia Sobhy <dalia.mohso...@hotmail.com> wrote: > Hiiii > > I have tried to make both ports the same. > But the prob is the hbase cannot connect to port 8020. > When i run nmap hostname, port 8020 wasnt with the list of open ports. > I have tried what harsh told me abt. > I used the same port he used but same error occurred. > Another aspect in cloudera doc it says that i have to canonical name for the > host ex: namenode.example.com as the hostname, but i didnt find it in any > tutorial. No one makes it. > Note that i am deploying my cluster in fully distributed mode i.e am using 4 > machines.. > > So any ideas??!! > > Sent from my iPhone > > On 2012-05-14, at 4:07 PM, "N Keywal" <nkey...@gmail.com> wrote: > >> Hi, >> >> There could be multiple issues, but it's strange to have in hbase-site.xml >> >> <value>hdfs://namenode:9000/hbase</value> >> >> while the core-site.xml says: >> >> <value>hdfs://namenode:54310/</value> >> >> The two entries should match. >> >> I would recommend to: >> - use netstat to check the ports (netstat -l) >> - do the check recommended by Harsh J previously. >> >> N. >> >> >> On Mon, May 14, 2012 at 3:21 PM, Dalia Sobhy <dalia.mohso...@hotmail.com> >> wrote: >>> >>> >>> pleaseeeeeeeeeeee helpppppppppppppppppppp >>> >>>> From: dalia.mohso...@hotmail.com >>>> To: user@hbase.apache.org >>>> Subject: RE: Important "Undefined Error" >>>> Date: Mon, 14 May 2012 12:20:18 +0200 >>>> >>>> >>>> >>>> Hi, >>>> I tried what you told me, but nothing worked:((( >>>> First when I run this command:dalia@namenode:~$ host -v -t A >>>> `hostname`Output:Trying "namenode"Host namenode not found: >>>> 3(NXDOMAIN)Received 101 bytes from 10.0.2.1#53 in 13 ms My >>>> core-site.xml:<configuration><property> >>>> <name>fs.default.name</name> >>>> <!--<value>hdfs://namenode:8020</value>--> >>>> <value>hdfs://namenode:54310/</value></property></configuration> >>>> My >>>> hdfs-site.xml<configuration><property><name>dfs.name.dir</name><value>/data/1/dfs/nn,/nfsmount/dfs/nn</value></property><!--<property><name>dfs.data.dir</name><value>/data/1/dfs/dn,/data/2/dfs/dn,/data/3/dfs/dn</value></property>--><property><name>dfs.datanode.max.xcievers</name><value>4096</value></property><property><name>dfs.replication</name><value>3</value></property><property> >>>> <name>dfs.permissions.superusergroup</name> >>>> <value>hadoop</value></property> >>>> My >>>> Mapred-site.xml<configuration><name>mapred.local.dir</name><value>/data/1/mapred/local,/data/2/mapred/local,/data/3/mapred/local</value></configuration> >>>> My >>>> Hbase-site.xml<configuration><property><name>hbase.cluster.distributed</name> >>>> <value>true</value></property><property> <name>hbase.rootdir</name> >>>> <value>hdfs://namenode:9000/hbase</value></property><property><name>hbase.zookeeper.quorun</name> >>>> >>>> <value>namenode</value></property><property><name>hbase.regionserver.port</name><value>60020</value><description>The >>>> host and port that the HBase master runs >>>> at.</description></property><property><name>dfs.replication</name><value>1</value></property><property><name>hbase.zookeeper.property.clientPort</name><value>2181</value><description>Property >>>> from ZooKeeper's config zoo.cfg.The port at which the clients will >>>> connect.</description></property></configuration> >>>> Please Help I am really disappointed I have been through all that for two >>>> weeks !!!! >>>> >>>> >>>> >>>>> From: dwivedishash...@gmail.com >>>>> To: user@hbase.apache.org >>>>> Subject: RE: Important "Undefined Error" >>>>> Date: Sat, 12 May 2012 23:31:49 +0530 >>>>> >>>>> The problem is your hbase is not able to connect to Hadoop, can you put >>>>> your >>>>> hbase-site.xml content >> here.. have you specified localhost somewhere, >>>>> if >>>>> so remove localhost from everywhere and put your hdfsl namenode address >>>>> suppose your namenode is running on master:9000 then put your hbase file >>>>> system setting as master:9000/hbase here I am sending you the >>>>> configuration >>>>> which I am using in hbase and is working >>>>> >>>>> >>>>> My hbase-site.xml content is >>>>> >>>>> <?xml version="1.0"?> >>>>> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> >>>>> <!-- >>>>> /** >>>>> * Copyright 2010 The Apache Software Foundation >>>>> * >>>>> * Licensed to the Apache Software Foundation (ASF) under one >>>>> * or more contributor license agreements. See the NOTICE file >>>>> * distributed with this work for additional information >>>>> * regarding copyright ownership. The ASF licenses this file >>>>> * to you under the Apache License, Version 2.0 (the >>>>> * "License"); you may not use this file except in compliance >>>>> * with the License. You may obtain a copy of the License at >>>>> * >>>>> * http://www.apache.org/licenses/LICENSE-2.0 >>>>> * >>>>> * Unless required by applicable law or agreed to in writing, software >>>>> * distributed under the License is distributed on an "AS IS" BASIS, >>>>> * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or >>>>> implied. >>>>> * See the License for the specific language governing permissions and >>>>> * limitations under the License. >>>>> */ >>>>> --> >>>>> <configuration> >>>>> <property> >>>>> <name>hbase.rootdir</name> >>>>> <value>hdfs://master:9000/hbase</value> >>>>> </property> >>>>> <property> >>>>> <name>hbase.master</name> >>>>> <value>master:60000</value> >>>>> <description>The host and port that the HBase master runs >>>>> at.</description> >>>>> </property> >>>>> <property> >>>>> <name>hbase.regionserver.port</name> >>>>> <value>60020</value> >>>>> <description>The host and port that the HBase master runs >>>>> at.</description> >>>>> </property> >>>>> <!--<property> >>>>> <name>hbase.master.port</name> >>>>> <value>60000</value> >>>>> <description>The host and port that the HBase master runs >>>>> at.</description> >>>>> </property>--> >>>>> <property> >>>>> <name>hbase.cluster.distributed</name> >>>>> <value>true</value> >>>>> </property> >>>>> <property> >>>>> <name>hbase.tmp.dir</name> >>>>> <value>/home/shashwat/Hadoop/hbase-0.90.4/temp</value> >>>>> </property> >>>>> <property> >>>>> <name>hbase.zookeeper.quorum</name> >>>>> <value>master</value> >>>>> </property> >>>>> <property> >>>>> <name>dfs.replication</name> >>>>> <value>1</value> >>>>> </property> >>>>> <property> >>>>> <name>hbase.zookeeper.property.clientPort</name> >>>>> <value>2181</value> >>>>> <description>Property from ZooKeeper's config zoo.cfg. >>>>> The port at which the clients will connect. >>>>> </description> >>>>> </property> >>>>> <property> >>>>> <name>hbase.zookeeper.property.dataDir</name> >>>>> <value>/home/shashwat/zookeeper</value> >>>>> <description>Property from ZooKeeper's config zoo.cfg. >>>>> The directory where the snapshot is stored. >>>>> </description> >>>>> </property> >>>>> >>>>> </configuration> >>>>> >>>>> >>>>> >>>>> >>>>> Check this out, and also stop hbase, If its not stopping kill all the >>>>> processes, and after putting your hdfs-site.xml, mapred-site.xml and >>>>> core-site.sml to hbase conf directory try to restart, and also delete the >>>>> folders created by hbase ,,, like temp directory or other then try to >>>>> start. >>>>> >>>>> Regards >>>>> ∞ >>>>> Shashwat Shriparv >>>>> >>>>> >>>>> -----Original Message----- >>>>> From: Dalia Sobhy [mailto:dalia.mohso...@hotmail.com] >>>>> Sent: 12 May 2012 22:48 >>>>> To: user@hbase.apache.org >>>>> Subject: RE: Important "Undefined Error" >>>>> >>>>> >>>>> Hi Shashwat, >>>>> I want to tell you about my configurations: >>>>> I am using 4 nodesOne "Master": Namenode, SecondaryNamenode, Job Tracker, >>>>> Zookeeper, HMasterThree "Slaves": datanodes, tasktrackers, regionservers >>>>> In >>>>> both master and slaves, all the hadoop daemons are working well, but as >>>>> for >>>>> the hbase master service it is not working.. >>>>> As for region server here is the error:12/05/12 14:42:13 INFO >>>>> util.ServerCommandLine: vmName=Java HotSpot(TM) 64-Bit Server VM, >>>>> vmVendor=Sun Microsystems Inc., vmVersion=20.1-b0212/05/12 14:42:13 INFO >>>>> util.ServerCommandLine: vmInputArguments=[-Xmx1000m, -ea, >>>>> -XX:+UseConcMarkSweepGC, -XX:+CMSIncrementalMode, >>>>> -Dhbase.log.dir=/usr/lib/hbase/bin/../logs, -Dhbase.log.file=hbase.log, >>>>> -Dhbase.home.dir=/usr/lib/hbase/bin/.., -Dhbase.id.str=, >>>>> -Dhbase.root.logger=INFO,console, >>>>> -Djava.library.path=/usr/lib/hadoop-0.20/lib/native/Linux-amd64-64:/usr/lib/ >>>>> hbase/bin/../lib/native/Linux-amd64-64]12/05/12 14:42:13 INFO >>>>> ipc.HBaseRpcMetrics: Initializing RPC Metrics with hostName=HRegionServer, >>>>> port=6002012/05/12 14:42:14 FATAL zookeeper.ZKConfig: The server in >>>>> zoo.cfg >>>>> cannot be set to localhost in a fully-distributed setup because it won't >>>>> be >>>>> reachable. See "Getting Started" for more information.12/05/12 14:42:14 >>>>> WARN >>>>> zookeeper.ZKConfig: Cannot read zoo.cfg, loading from XML >>>>> filesjava.io.IOException: >>>>> The server in zoo.cfg cannot be set to localhost in a fully-distributed >>>>> setup because it won't be reachable. See "Getting Started" for more >>>>> information. at >>>>> org.apache.hadoop.hbase.zookeeper.ZKConfig.parseZooCfg(ZKConfig.java:172) >>>>> at >>>>> org.apache.hadoop.hbase.zookeeper.ZKConfig.makeZKProps(ZKConfig.java:68) >>>>> at >>>>> org.apache.hadoop.hbase.zookeeper.ZKConfig.getZKQuorumServersString(ZKConfig >>>>> .java:249) at >>>>> org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher.<init>(ZooKeeperWatcher.j >>>>> ava:117) at >>>>> org.apache.hadoop.hbase.regionserver.HRegionServer.initializeZooKeeper(HRegi >>>>> onServer.java:489) at >>>>> org.apache.hadoop.hbase.regionserver.HRegionServer.initialize(HRegionServer. >>>>> java:465) at >>>>> org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:56 >>>>> 4) at java.lang.Thread.run(Thread.java:662)12/05/12 14:42:14 INFO >>>>> zookeeper.ZooKeeper: Client environment:zookeeper.version=3.3.5-cdh3u4--1, >>>>> built on 05/07/2012 21:12 GMT12/05/12 14:42:14 INFO zookeeper.ZooKeeper: >>>>> Client environment:host.name=data >>>>> node212/05/12 14:42:14 INFO zookeeper.ZooKeeper: Client >>>>> environment:java.version=1.6.0_2612/05/12 14:42:14 INFO >>>>> zookeeper.ZooKeeper: >>>>> Client environment:java.vendor=Sun Microsystems Inc.12/05/12 14:42:14 INFO >>>>> zookeeper.ZooKeeper: Client >>>>> environment:java.home=/usr/lib/jvm/java-6-sun-1. >>>>> 6.0.26/jre12/05/12 14:42:14 INFO zookeeper.ZooKeeper: Client >>>>> environment:java.class.path=/usr/lib/hbase/bin/../conf:/usr/lib/jvm/java-6-s >>>>> un/lib/tools.jar:/usr/lib/hbase/bin/..:/usr/lib/hbase/bin/../hbase-0.90.6-cd >>>>> h3u4.jar:/usr/lib/hbase/bin/../hbase-0.90.6-cdh3u4-tests.jar:/usr/lib/hbase/ >>>>> bin/../lib/activation-1.1.jar:/usr/lib/hbase/bin/../lib/asm-3.1.jar:/usr/lib >>>>> /hbase/bin/../lib/avro-1.5.4.jar:/usr/lib/hbase/bin/../lib/avro-ipc-1.5.4.ja >>>>> r:/usr/lib/hbase/bin/../lib/commons-cli-1.2.jar:/usr/lib/hbase/bin/../lib/co >>>>> mmons-codec-1.4.jar:/usr/lib/hbase/bin/../lib/commons-el-1.0.jar:/usr/lib/hb >>>>> ase/bin/../lib/commons-httpclient-3.1.jar:/usr/lib/hbase/bin/../lib/commons- >>>>> lang-2.5.jar:/usr/lib/hbase/bin/../lib/commo >>>>> >>>>> ns-logging-1.1.1.jar:/usr/lib/hbase/bin/../lib/commons-net-1.4.1.jar:/usr/li >>>>> b/hbase/bin/../lib/core-3.1.1.jar:/usr/lib/hbase/bin/../lib/guava-r06.jar:/u >>>>> sr/lib/hbase/bin/../lib/guava-r09-jarjar.jar:/usr/lib/hbase/bin/../lib/hadoo >>>>> p-core.jar:/usr/lib/hbase/bin/../lib/jackson-core-asl-1.5.2.jar:/usr/lib/hba >>>>> se/bin/../lib/jackson-jaxrs-1.5.5.jar:/usr/lib/hbase/bin/../lib/jackson-mapp >>>>> er-asl-1.5.2.jar:/usr/lib/hbase/bin/../lib/jackson-xc-1.5.5.jar:/usr/lib/hba >>>>> se/bin/../lib/jamon-runtime-2.3.1.jar:/usr/lib/hbase/bin/../lib/jasper-compi >>>>> ler-5.5.23.jar:/usr/lib/hbase/bin/../lib/jasper-runtime-5.5.23.jar:/usr/lib/ >>>>> hbase/bin/../lib/jaxb-api-2.1.jar:/usr/lib/hbase/bin/../lib/jaxb-impl-2.1.12 >>>>> .jar:/usr/lib/hbase/bin/../lib/jersey-core-1.4.jar:/usr/lib/hbase/bin/../lib >>>>> /jersey-json-1.4.jar:/usr/lib/hbase/bin/../lib/jersey-server-1.4.jar:/usr/li >>>>> b/hbase/bin/../lib/jettison-1.1.jar:/usr/lib/hbase/bin/../lib/jetty-6.1.26.j >>>>> ar:/usr/lib/hbase/bin/../lib/jetty-util-6.1.26.jar:/usr/lib/hbase/bin/../lib >>>>> /jruby-co >>>>> >>>>> mplete-1.6.0.jar:/usr/lib/hbase/bin/../lib/jsp-2.1-6.1.14.jar:/usr/lib/hbase >>>>> /bin/../lib/jsp-api-2.1-6.1.14.jar:/usr/lib/hbase/bin/../lib/jsp-api-2.1.jar >>>>> :/usr/lib/hbase/bin/../lib/jsr311-api-1.1.1.jar:/usr/lib/hbase/bin/../lib/lo >>>>> g4j-1.2.16.jar:/usr/lib/hbase/bin/../lib/netty-3.2.4.Final.jar:/usr/lib/hbas >>>>> e/bin/../lib/protobuf-java-2.3.0.jar:/usr/lib/hbase/bin/../lib/servlet-api-2 >>>>> .5-6.1.14.jar:/usr/lib/hbase/bin/../lib/servlet-api-2.5.jar:/usr/lib/hbase/b >>>>> in/../lib/slf4j-api-1.5.8.jar:/usr/lib/hbase/bin/../lib/slf4j-log4j12-1.5.8. >>>>> jar:/usr/lib/hbase/bin/../lib/snappy-java-1.0.3.2.jar:/usr/lib/hbase/bin/../ >>>>> lib/stax-api-1.0.1.jar:/usr/lib/hbase/bin/../lib/thrift-0.2.0.jar:/usr/lib/h >>>>> base/bin/../lib/velocity-1.5.jar:/usr/lib/hbase/bin/../lib/xmlenc-0.52.jar:/ >>>>> usr/lib/hbase/bin/../lib/zookeeper.jar:/etc/zookeeper:/etc/hadoop-0.20/conf: >>>>> /usr/lib/hadoop-0.20/hadoop-0.20.2-cdh3u4-ant.jar:/usr/lib/hadoop-0.20/hadoo >>>>> p-0.20.2-cdh3u4-tools.jar:/usr/lib/hadoop-0.20/hadoop-tools.jar:/usr/lib/had >>>>> oop-0.20/ >>>>> >>>>> hadoop-examples-0.20.2-cdh3u4.jar:/usr/lib/hadoop-0.20/hadoop-ant-0.20.2-cdh >>>>> 3u4.jar:/usr/lib/hadoop-0.20/hadoop-0.20.2-cdh3u4-examples.jar:/usr/lib/hado >>>>> op-0.20/hadoop-ant.jar:/usr/lib/hadoop-0.20/hadoop-core-0.20.2-cdh3u4.jar:/u >>>>> sr/lib/hadoop-0.20/hadoop-core.jar:/usr/lib/hadoop-0.20/hadoop-tools-0.20.2- >>>>> cdh3u4.jar:/usr/lib/hadoop-0.20/hadoop-examples.jar:/usr/lib/hadoop-0.20/had >>>>> oop-0.20.2-cdh3u4-core.jar:/usr/lib/hadoop-0.20/hadoop-test-0.20.2-cdh3u4.ja >>>>> r:/usr/lib/hadoop-0.20/hadoop-0.20.2-cdh3u4-test.jar:/usr/lib/hadoop-0.20/ha >>>>> doop-test.jar:/usr/lib/hadoop-0.20/lib/jetty-6.1.26.cloudera.1.jar:/usr/lib/ >>>>> hadoop-0.20/lib/jets3t-0.6.1.jar:/usr/lib/hadoop-0.20/lib/junit-4.5.jar:/usr >>>>> /lib/hadoop-0.20/lib/commons-daemon-1.0.1.jar:/usr/lib/hadoop-0.20/lib/slf4j >>>>> -api-1.4.3.jar:/usr/lib/hadoop-0.20/lib/commons-codec-1.4.jar:/usr/lib/hadoo >>>>> p-0.20/lib/log4j-1.2.15.jar:/usr/lib/hadoop-0.20/lib/jasper-compiler-5.5.12. >>>>> jar:/usr/lib/hadoop-0.20/lib/guava-r09-jarjar.jar:/usr/lib/hadoop-0.20/lib/j >>>>> ackson-ma >>>>> >>>>> pper-asl-1.5.2.jar:/usr/lib/hadoop-0.20/lib/commons-logging-1.0.4.jar:/usr/l >>>>> ib/hadoop-0.20/lib/ant-contrib-1.0b3.jar:/usr/lib/hadoop-0.20/lib/commons-la >>>>> ng-2.4.jar:/usr/lib/hadoop-0.20/lib/kfs-0.2.2.jar:/usr/lib/hadoop-0.20/lib/j >>>>> etty-util-6.1.26.cloudera.1.jar:/usr/lib/hadoop-0.20/lib/commons-httpclient- >>>>> 3.1.jar:/usr/lib/hadoop-0.20/lib/jsch-0.1.42.jar:/usr/lib/hadoop-0.20/lib/sl >>>>> f4j-log4j12-1.4.3.jar:/usr/lib/hadoop-0.20/lib/servlet-api-2.5-20081211.jar: >>>>> /usr/lib/hadoop-0.20/lib/jasper-runtime-5.5.12.jar:/usr/lib/hadoop-0.20/lib/ >>>>> servlet-api-2.5-6.1.14.jar:/usr/lib/hadoop-0.20/lib/jetty-servlet-tester-6.1 >>>>> .26.cloudera.1.jar:/usr/lib/hadoop-0.20/lib/aspectjtools-1.6.5.jar:/usr/lib/ >>>>> hadoop-0.20/lib/xmlenc-0.52.jar:/usr/lib/hadoop-0.20/lib/hadoop-fairschedule >>>>> r-0.20.2-cdh3u4.jar:/usr/lib/hadoop-0.20/lib/jackson-core-asl-1.5.2.jar:/usr >>>>> /lib/hadoop-0.20/lib/mockito-all-1.8.2.jar:/usr/lib/hadoop-0.20/lib/commons- >>>>> el-1.0.jar:/usr/lib/hadoop-0.20/lib/commons-logging-api-1.0.4.jar:/usr/lib/h >>>>> adoop-0.2 >>>>> >>>>> 0/lib/commons-net-3.1.jar:/usr/lib/hadoop-0.20/lib/commons-cli-1.2.jar:/usr/ >>>>> lib/hadoop-0.20/lib/aspectjrt-1.6.5.jar:/usr/lib/hadoop-0.20/lib/core-3.1.1. >>>>> jar:/usr/lib/hadoop-0.20/lib/hsqldb-1.8.0.10.jar:/usr/lib/hadoop-0.20/lib/or >>>>> o-2.0.8.jar:/usr/lib/zookeeper/zookeeper.jar:/usr/lib/zookeeper/zookeeper-3. >>>>> 3.5-cdh3u4.jar:/usr/lib/zookeeper/lib/log4j-1.2.15.jar:/usr/lib/zookeeper/li >>>>> b/jline-0.9.94.jar::/usr/lib/hadoop-0.20/conf:/usr/lib/hadoop-0.20/hadoop-co >>>>> re-0.20.2-cdh3u4.jar:/usr/lib/hadoop-0.20/lib/ant-contrib-1.0b3.jar:/usr/lib >>>>> /hadoop-0.20/lib/aspectjrt-1.6.5.jar:/usr/lib/hadoop-0.20/lib/aspectjtools-1 >>>>> .6.5.jar:/usr/lib/hadoop-0.20/lib/commons-cli-1.2.jar:/usr/lib/hadoop-0.20/l >>>>> ib/commons-codec-1.4.jar:/usr/lib/hadoop-0.20/lib/commons-daemon-1.0.1.jar:/ >>>>> usr/lib/hadoop-0.20/lib/commons-el-1.0.jar:/usr/lib/hadoop-0.20/lib/commons- >>>>> httpclient-3.1.jar:/usr/lib/hadoop-0.20/lib/commons-lang-2.4.jar:/usr/lib/ha >>>>> doop-0.20/lib/commons-logging-1.0.4.jar:/usr/lib/hadoop-0.20/lib/commons-log >>>>> ging-api- >>>>> >>>>> 1.0.4.jar:/usr/lib/hadoop-0.20/lib/commons-net-3.1.jar:/usr/lib/hadoop-0.20/ >>>>> lib/core-3.1.1.jar:/usr/lib/hadoop-0.20/lib/guava-r09-jarjar.jar:/usr/lib/ha >>>>> doop-0.20/lib/hadoop-fairscheduler-0.20.2-cdh3u4.jar:/usr/lib/hadoop-0.20/li >>>>> b/hsqldb-1.8.0.10.jar:/usr/lib/hadoop-0.20/lib/jackson-core-asl-1.5.2.jar:/u >>>>> sr/lib/hadoop-0.20/lib/jackson-mapper-asl-1.5.2.jar:/usr/lib/hadoop-0.20/lib >>>>> /jasper-compiler-5.5.12.jar:/usr/lib/hadoop-0.20/lib/jasper-runtime-5.5.12.j >>>>> ar:/usr/lib/hadoop-0.20/lib/jets3t-0.6.1.jar:/usr/lib/hadoop-0.20/lib/jetty- >>>>> 6.1.26.cloudera.1.jar:/usr/lib/hadoop-0.20/lib/jetty-servlet-tester-6.1.26.c >>>>> loudera.1.jar:/usr/lib/hadoop-0.20/lib/jetty-util-6.1.26.cloudera.1.jar:/usr >>>>> /lib/hadoop-0.20/lib/jsch-0.1.42.jar:/usr/lib/hadoop-0.20/lib/junit-4.5.jar: >>>>> /usr/lib/hadoop-0.20/lib/kfs-0.2.2.jar:/usr/lib/hadoop-0.20/lib/log4j-1.2.15 >>>>> .jar:/usr/lib/hadoop-0.20/lib/mockito-all-1.8.2.jar:/usr/lib/hadoop-0.20/lib >>>>> /oro-2.0.8.jar:/usr/lib/hadoop-0.20/lib/servlet-api-2.5-20081211.jar:/usr/li >>>>> b/hadoop- >>>>> 0.20/lib/servlet-api-2.5-6.1.14.jar:/usr/lib/hadoop-0.20/lib/slf4j-api-1.4. >>>>> 3.jar:/usr/lib/hadoop-0.20/lib/slf4j-log4j12-1.4.3.jar:/usr/lib/hadoop-0.20/ >>>>> lib/xmlenc-0.52.jar12/05/12 14:42:14 INFO zookeeper.ZooKeeper: Client >>>>> environment:java.library.path=/usr/lib/hadoop-0.20/lib/native/Linux-amd64-64 >>>>> :/usr/lib/hbase/bin/../lib/native/Linux-amd64-6412/05/12 14:42:14 INFO >>>>> zookeeper.ZooKeeper: Client environment:java.io.tmpdir=/tmp12/05/12 >>>>> 14:42:14 >>>>> INFO zookeeper.ZooKeeper: Client environment:java.compiler=<NA>12/05/12 >>>>> 14:42:14 INFO zookeeper.ZooKeeper: Client >>>>> environment:os.name=Linux12/05/12 >>>>> 14:42:14 INFO zookeeper.ZooKeeper: Client >>>>> environment:os.arch=amd6412/05/12 >>>>> 14:42:14 INFO zookeeper.ZooKeeper: Client >>>>> environment:os.version=2.6.35-22-server12/05/12 14:42:14 INFO >>>>> zookeeper.ZooKeeper: Client environment:user.name=dalia12/05/12 14:42:14 >>>>> INFO zookeeper.ZooKeeper: Client environment:user.home=/home/dalia12/05/12 >>>>> 14:42:14 INFO zookeeper.ZooKeeper: Client >>>>> environment:user.dir=/home/dalia12 >>>>> /05/12 14:42:14 INFO zookeeper.ZooKeeper: Initiating client connection, >>>>> connectString=localhost:2181 sessionTimeout=180000 >>>>> watcher=regionserver:6002012/05/12 14:42:14 INFO zookeeper.ClientCnxn: >>>>> Opening socket connection to server localhost/0:0:0:0:0:0:0:1:218112/05/12 >>>>> 14:42:14 WARN zookeeper.ClientCnxn: Session 0x0 for server null, >>>>> unexpected >>>>> error, closing socket connection and attempting >>>>> reconnectjava.net.ConnectException: Connection refused at >>>>> sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) at >>>>> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567) >>>>> at >>>>> org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1143)12/05/12 >>>>> 14:42:14 INFO zookeeper.ClientCnxn: Opening socket connection to server >>>>> localhost/127.0.0.1:218112/05/12 14:42:14 WARN zookeeper.ClientCnxn: >>>>> Session >>>>> 0x0 for server null, unexpected error, closing socket connection and >>>>> attempting reconnectjava.net.ConnectException: Connection refused at >>>>> sun.nio.ch.SocketChannelImpl.checkConnect(Native >>>>> Method) at >>>>> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567) >>>>> at >>>>> org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1143)12/05/12 >>>>> 14:42:15 INFO ipc.Client: Retrying connect to server: >>>>> namenode/10.0.2.3:8020. Already tried 0 time(s).12/05/12 14:42:16 INFO >>>>> zookeeper.ClientCnxn: Opening socket connection to server >>>>> localhost/0:0:0:0: >>>>> 0:0:0:1:218112/05/12 14:42:16 WARN zookeeper.ClientCnxn: Session 0x0 for >>>>> server null, unexpected error, closing socket connection and attempting >>>>> reconnectjava.net.ConnectException: Connection refused at >>>>> sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) at >>>>> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567) >>>>> at >>>>> org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1143)12/05/12 >>>>> 14:42:16 INFO ipc.Client: Retrying connect to server: >>>>> namenode/10.0.2.3:8020. Already tried 1 time(s).12/05/12 14:42:16 INFO >>>>> zookeeper.ClientCnxn: Opening socket connection to server >>>>> localhost/127.0.0. >>>>> 1:218112/05/12 14: >>>>> 42:16 WARN zookeeper.ClientCnxn: Session 0x0 for server null, unexpected >>>>> error, closing socket connection and attempting >>>>> reconnectjava.net.ConnectException: Connection refused at >>>>> sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) at >>>>> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567) >>>>> at >>>>> org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1143)12/05/12 >>>>> 14:42:17 INFO ipc.Client: Retrying connect to server: >>>>> namenode/10.0.2.3:8020. Already tried 2 time(s).12/05/12 14:42:18 INFO >>>>> ipc.Client: Retrying connect to server: namenode/10.0.2.3:8020. Already >>>>> tried 3 time(s).12/05/12 14:42:18 INFO zookeeper.ClientCnxn: Opening >>>>> socket >>>>> connection to server localhost/0:0:0:0:0:0:0:1:218112/05/12 14:42:18 WARN >>>>> zookeeper.ClientCnxn: Session 0x0 for server null, unexpected error, >>>>> closing >>>>> socket connection and attempting reconnectjava.net.ConnectException: >>>>> Connection refused at sun.nio.ch.SocketChannelImpl.checkConnect(Native >>>>> Method) at sun.nio.ch.SocketChannelImpl.fin >>>>> ishConnect(SocketChannelImpl.java:567) at >>>>> org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1143)12/05/12 >>>>> 14:42:18 INFO zookeeper.ClientCnxn: Opening socket connection to server >>>>> localhost/127.0.0.1:218112/05/12 14:42:18 WARN zookeeper.ClientCnxn: >>>>> Session >>>>> 0x0 for server null, unexpected error, closing socket connection and >>>>> attempting reconnectjava.net.ConnectException: Connection refused at >>>>> sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) at >>>>> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567) >>>>> at >>>>> org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1143)12/05/12 >>>>> 14:42:19 INFO ipc.Client: Retrying connect to server: >>>>> namenode/10.0.2.3:8020. Already tried 4 time(s).12/05/12 14:42:19 INFO >>>>> zookeeper.ClientCnxn: Opening socket connection to server >>>>> localhost/0:0:0:0: >>>>> 0:0:0:1:218112/05/12 14:42:19 WARN zookeeper.ClientCnxn: Session 0x0 for >>>>> server null, unexpected error, closing socket connection and attempting >>>>> reconnectjava.net.ConnectException >>>>> : Connection refused at >>>>> sun.nio.ch.SocketChannelImpl.checkConnect(Native >>>>> Method) at >>>>> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567) >>>>> at >>>>> org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1143)12/05/12 >>>>> 14:42:20 INFO ipc.Client: Retrying connect to server: >>>>> namenode/10.0.2.3:8020. Already tried 5 time(s).12/05/12 14:42:20 INFO >>>>> zookeeper.ClientCnxn: Opening socket connection to server >>>>> localhost/127.0.0. >>>>> 1:218112/05/12 14:42:20 WARN zookeeper.ClientCnxn: Session 0x0 for server >>>>> null, unexpected error, closing socket connection and attempting >>>>> reconnectjava.net.ConnectException: Connection refused at >>>>> sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) at >>>>> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567) >>>>> at >>>>> org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1143)12/05/12 >>>>> 14:42:21 INFO ipc.Client: Retrying connect to server: >>>>> namenode/10.0.2.3:8020. Already tried 6 time(s).12/05/12 14:42:22 INFO >>>>> zookeeper.ClientCnxn: Openin >>>>> g socket connection to server localhost/0:0:0:0:0:0:0:1:218112/05/12 >>>>> 14:42: >>>>> 22 WARN zookeeper.ClientCnxn: Session 0x0 for server null, unexpected >>>>> error, >>>>> closing socket connection and attempting >>>>> reconnectjava.net.ConnectException: >>>>> Connection refused at sun.nio.ch.SocketChannelImpl.checkConnect(Native >>>>> Method) at >>>>> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567) >>>>> at >>>>> org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1143)12/05/12 >>>>> 14:42:22 INFO ipc.Client: Retrying connect to server: >>>>> namenode/10.0.2.3:8020. Already tried 7 time(s).12/05/12 14:42:22 INFO >>>>> zookeeper.ClientCnxn: Opening socket connection to server >>>>> localhost/127.0.0. >>>>> 1:218112/05/12 14:42:22 WARN zookeeper.ClientCnxn: Session 0x0 for server >>>>> null, unexpected error, closing socket connection and attempting >>>>> reconnectjava.net.ConnectException: Connection refused at >>>>> sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) at >>>>> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567) >>>>> at >>>>> org >>>>> .apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1143)12/05/12 >>>>> 14:42:23 INFO ipc.Client: Retrying connect to server: >>>>> namenode/10.0.2.3:8020. Already tried 8 time(s).12/05/12 14:42:24 INFO >>>>> ipc.Client: Retrying connect to server: namenode/10.0.2.3:8020. Already >>>>> tried 9 time(s).Exception in thread "main" java.net.ConnectException: Call >>>>> to namenode/10.0.2.3:8020 failed on connection exception: >>>>> java.net.ConnectException: Connection refused at >>>>> org.apache.hadoop.ipc.Client.wrapException(Client.java:1134) at >>>>> org.apache.hadoop.ipc.Client.call(Client.java:1110) at >>>>> org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:226) at >>>>> $Proxy5.getProtocolVersion(Unknown Source) at >>>>> org.apache.hadoop.ipc.RPC.getProxy(RPC.java:398) at >>>>> org.apache.hadoop.ipc.RPC.getProxy(RPC.java:384) at >>>>> org.apache.hadoop.hdfs.DFSClient.createRPCNamenode(DFSClient.java:129) >>>>> at >>>>> org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:255) at >>>>> org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:217) at >>>>> org.apache.hadoop >>>>> .hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:89) >>>>> at >>>>> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1563) >>>>> at >>>>> org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:67) at >>>>> org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:1597) >>>>> at >>>>> org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1579) at >>>>> org.apache.hadoop.fs.FileSystem.get(FileSystem.java:228) at >>>>> org.apache.hadoop.fs.FileSystem.get(FileSystem.java:111) at >>>>> org.apache.hadoop.hbase.regionserver.HRegionServer.startRegionServer(HRegion >>>>> Server.java:2785) at >>>>> org.apache.hadoop.hbase.regionserver.HRegionServer.startRegionServer(HRegion >>>>> Server.java:2768) at >>>>> org.apache.hadoop.hbase.regionserver.HRegionServerCommandLine.start(HRegionS >>>>> erverCommandLine.java:61) at >>>>> org.apache.hadoop.hbase.regionserver.HRegionServerCommandLine.run(HRegionSer >>>>> verCommandLine.java:75) at >>>>> org.apache.hadoop.util.ToolRunner.run(ToolRunner. >>>>> java:65) at >>>>> org.apache.hadoop.hbase.util.ServerCommandLine.doMain(Serve >>>>> rCommandLine.java:76) at >>>>> org.apache.hadoop.hbase.regionserver.HRegionServer.main(HRegionServer.java:2 >>>>> 829)Caused by: java.net.ConnectException: Connection refused at >>>>> sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) at >>>>> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567) >>>>> at >>>>> org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:2 >>>>> 06) at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:429) at >>>>> org.apache.hadoop.net.NetUtils.connect(NetUtils.java:394) at >>>>> org.apache.hadoop.ipc.Client$Connection.setupConnection(Client.java:425) >>>>> at org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:532) >>>>> at org.apache.hadoop.ipc.Client$Connection.access$2300(Client.java:210) >>>>> at >>>>> org.apache.hadoop.ipc.Client.getConnection(Client.java:1247) at >>>>> org.apache.hadoop.ipc.Client.call(Client.java:1078) ... 21 more12/05/12 >>>>> 14:42:24 INFO zookeeper.ClientCnxn: Opening socket connection to server >>>>> localhost/0:0:0:0:0:0:0:1:218112/05/12 14:42:24 WARN zookeeper >>>>> .ClientCnxn: Session 0x0 for server null, unexpected error, closing >>>>> socket >>>>> connection and attempting reconnectjava.net.ConnectException: Connection >>>>> refused at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) >>>>> at >>>>> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567) >>>>> at >>>>> org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1143)12/05/12 >>>>> 14:42:25 INFO zookeeper.ClientCnxn: Opening socket connection to server >>>>> localhost/127.0.0.1:218112/05/12 14:42:25 INFO zookeeper.ZooKeeper: >>>>> Session: >>>>> 0x0 closed12/05/12 14:42:25 INFO zookeeper.ClientCnxn: EventThread shut >>>>> down12/05/12 14:42:25 INFO ipc.HBaseServer: Stopping server on >>>>> 6002012/05/12 >>>>> 14:42:25 FATAL regionserver.HRegionServer: ABORTING region server >>>>> serverName=datanode2,60020,1336826533870, load=(requests=0, regions=0, >>>>> usedHeap=0, maxHeap=0): Initialization of RS failed. Hence aborting >>>>> RS.org. >>>>> apache.hadoop.hbase.ZooKeeperConnectionException: HBase is able to connect >>>>> to ZooKeeper but the connection closes im >>>>> mediately. This could be a sign that the server has too many connections >>>>> (30 is the default). Consider inspecting your ZK server logs for that >>>>> error >>>>> and then make sure you are reusing HBaseConfiguration as often as you can. >>>>> See HTable's javadoc for more information. at >>>>> org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher.<init>(ZooKeeperWatcher.j >>>>> ava:160) at >>>>> org.apache.hadoop.hbase.regionserver.HRegionServer.initializeZooKeeper(HRegi >>>>> onServer.java:489) at >>>>> org.apache.hadoop.hbase.regionserver.HRegionServer.initialize(HRegionServer. >>>>> java:465) at >>>>> org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:56 >>>>> 4) at java.lang.Thread.run(Thread.java:662)Caused by: >>>>> org.apache.zookeeper.KeeperException$ConnectionLossException: >>>>> KeeperErrorCode = ConnectionLoss for /hbase at >>>>> org.apache.zookeeper.KeeperException.create(KeeperException.java:90) >>>>> at >>>>> org.apache.zookeeper.KeeperException.create(KeeperException.java:42) >>>>> at >>>>> org.apache.zookeeper.ZooKeeper.exists(ZooKeeper.java:815) at org.apac >>>>> he.zookeeper.ZooKeeper.exists(ZooKeeper.java:843) at >>>>> org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndFailSilent(ZKUtil.java:930 >>>>> ) at >>>>> org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher.<init>(ZooKeeperWatcher.j >>>>> ava:138) ... 4 more12/05/12 14:42:25 INFO regionserver.HRegionServer: >>>>> STOPPED: Initialization of RS failed. Hence aborting RS.Exception in >>>>> thread >>>>> "regionserver60020" java.lang.NullPointerException at >>>>> org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:63 >>>>> 3) at java.lang.Thread.run(Thread.java:662) >>>>> So any help??? >>>>>> Date: Sat, 12 May 2012 20:22:03 +0530 >>>>>> Subject: Re: Important "Undefined Error" >>>>>> From: dwivedishash...@gmail.com >>>>>> To: user@hbase.apache.org >>>>>> >>>>>> you can turn off hadoop safe mode uisng *hadoop dfsadmin -safemode >>>>>> leave* >>>>>> >>>>>> On Sat, May 12, 2012 at 8:15 PM, shashwat shriparv < >>>>>> dwivedishash...@gmail.com> wrote: >>>>>> >>>>>>> First thing copy core-site.xml, dfs xml from hadoop conf directory >>>>>>> to hbase conf dirctory. and turn of hadoop save mode and then try... >>>>>>> >>>>>>> >>>>>>> On Sat, May 12, 2012 at 6:27 PM, Harsh J <ha...@cloudera.com> wrote: >>>>>>> >>>>>>>> Dalia, >>>>>>>> >>>>>>>> Is your NameNode running fine? The issue is that HBase Master has >>>>>>>> been asked to talk to HDFS, but it can't connect to the HDFS >>>>>>>> NameNode. Does "hadoop dfs -touchz foobar" pass or fail with similar >>>>> retry issues? >>>>>>>> >>>>>>>> What's your fs.default.name's value in Hadoop's core-site.xml? And >>>>>>>> whats the output of that fixed host command I'd posted before? >>>>>>>> >>>>>>>> On Sat, May 12, 2012 at 6:06 PM, Dalia Sobhy >>>>>>>> <dalia.mohso...@hotmail.com> >>>>>>>> wrote: >>>>>>>>> >>>>>>>>> >>>>>>>>> Dear Harsh >>>>>>>>> When I run $hbase master start >>>>>>>>> I found the following errors:12/05/12 08:32:42 INFO >>>>>>>> ipc.HBaseRpcMetrics: Initializing RPC Metrics with >>>>>>>> hostName=HMaster, >>>>>>>> port=6000012/05/12 08:32:42 INFO security.UserGroupInformation: >>>>>>>> JAAS Configuration already set up for Hadoop, not >>>>>>>> re-installing.12/05/12 >>>>>>>> 08:32:42 INFO ipc.HBaseServer: IPC Server Responder: >>>>>>>> starting12/05/12 >>>>>>>> 08:32:42 INFO ipc.HBaseServer: IPC Server listener on 60000: >>>>>>>> starting12/05/12 08:32:42 INFO ipc.HBaseServer: IPC Server handler >>>>>>>> 0 on >>>>>>>> 60000: starting12/05/12 08:32:42 INFO ipc.HBaseServer: IPC Server >>>>>>>> handler 1 on 60000: starting12/05/12 08:32:42 INFO ipc.HBaseServer: >>>>>>>> IPC Server handler 2 on 60000: starting12/05/12 08:32:42 INFO >>>>>>>> ipc.HBaseServer: IPC Server handler 3 on 60000: starting12/05/12 08:32: >>>>> 42 INFO ipc.HBaseServer: >>>>>>>> IPC Server handler 5 on 60000: starting12/05/12 08:32:42 INFO >>>>>>>> ipc.HBaseServer: IPC Server handler 4 on 60000: starting12/05/12 >>>>>>>> 08:32:42 INFO ipc.HBaseServer: IPC Server handler 7 on 60000: >>>>>>>> starting12/05/12 >>>>>>>> 08:32:42 INFO ipc.HBaseServer: IPC Serv >>>>>>>>> er handler 6 on 60000: starting12/05/12 08:32:42 INFO >>>>> ipc.HBaseServer: >>>>>>>> IPC Server handler 8 on 60000: starting12/05/12 08:32:42 INFO >>>>>>>> ipc.HBaseServer: IPC Server handler 9 on 60000: starting12/05/12 >>>>>>>> 08:32:42 INFO zookeeper.ZooKeeper: Client >>>>>>>> environment:zookeeper.version=3.3.5-cdh3u4--1, built on 05/07/2012 >>>>>>>> 21:12 >>>>>>>> GMT12/05/12 08:32:42 INFO zookeeper.ZooKeeper: Client environment: >>>>>>>> host.name=namenode12/05/12 08:32:42 INFO zookeeper.ZooKeeper: >>>>>>>> Client >>>>>>>> environment:java.version=1.6.0_3012/05/12 08:32:42 INFO >>>>>>>> zookeeper.ZooKeeper: Client environment:java.vendor=Sun >>>>>>>> Microsystems >>>>>>>> Inc.12/05/12 08:32:42 INFO zookeeper.ZooKeeper: Client >>>>>>>> environment:java.home=/usr/lib/jvm/java-6-sun-1.6.0.30/jre12/05/12 >>>>>>>> 08:32:42 INFO zookeeper.ZooKeeper: Client >>>>>>>> environment:java.class.path=/usr/lib/hbase/bin/../conf:/usr/lib/jvm >>>>>>>> /java-6-sun/lib/tools.jar:/usr/lib/hbase/bin/..:/usr/lib/hbase/bin/ >>>>>>>> ../hbase-0.90.4-cdh3u3.jar:/usr/lib/hbase/bin/../hbase-0.90.4-cdh3u >>>>>>>> 3-tests.jar:/usr/lib/hbase/bin/../lib/activation-1.1.jar:/u >>>>>>>>> >>>>>>>> >>>>>>>> sr/lib/hbase/bin/../lib/asm-3.1.jar:/usr/lib/hbase/bin/../lib/avro- >>>>>>>> 1.5.4.jar:/usr/lib/hbase/bin/../lib/avro-ipc-1.5.4.jar:/usr/lib/hba >>>>>>>> se/bin/../lib/commons-cli-1.2.jar:/usr/lib/hbase/bin/../lib/commons >>>>>>>> -codec-1.4.jar:/usr/lib/hbase/bin/../lib/commons-el-1.0.jar:/usr/li >>>>>>>> b/hbase/bin/../lib/commons-httpclient-3.1.jar:/usr/lib/hbase/bin/.. >>>>>>>> /lib/commons-lang-2.5.jar:/usr/lib/hbase/bin/../lib/commons-logging >>>>>>>> -1.1.1.jar:/usr/lib/hbase/bin/../lib/commons-net-1.4.1.jar:/usr/lib >>>>>>>> /hbase/bin/../lib/core-3.1.1.jar:/usr/lib/hbase/bin/../lib/guava-r0 >>>>>>>> 6.jar:/usr/lib/hbase/bin/../lib/guava-r09-jarjar.jar:/usr/lib/hbase >>>>>>>> /bin/../lib/hadoop-core.jar:/usr/lib/hbase/bin/../lib/jackson-core- >>>>>>>> asl-1.5.2.jar:/usr/lib/hbase/bin/../lib/jackson-jaxrs-1.5.5.jar:/us >>>>>>>> r/lib/hbase/bin/../lib/jackson-mapper-asl-1.5.2.jar:/usr/lib/hbase/ >>>>>>>> bin/../lib/jackson-xc-1.5.5.jar:/usr/lib/hbase/bin/../lib/jamon-run >>>>>>>> time-2.3.1.jar:/usr/lib/hbase/bin/../lib/jasper-compiler-5.5.23.jar >>>>>>>> :/usr/lib/hbase/bin/../lib/jasper-runtime-5.5.23.jar:/ >>>>> usr/l >>>>>>>>> >>>>>>>> >>>>>>>> ib/hbase/bin/../lib/jaxb-api-2.1.jar:/usr/lib/hbase/bin/../lib/jaxb >>>>>>>> -impl-2.1.12.jar:/usr/lib/hbase/bin/../lib/jersey-core-1.4.jar:/usr >>>>>>>> /lib/hbase/bin/../lib/jersey-json-1.4.jar:/usr/lib/hbase/bin/../lib >>>>>>>> /jersey-server-1.4.jar:/usr/lib/hbase/bin/../lib/jettison-1.1.jar:/ >>>>>>>> usr/lib/hbase/bin/../lib/jetty-6.1.26.jar:/usr/lib/hbase/bin/../lib >>>>>>>> /jetty-util-6.1.26.jar:/usr/lib/hbase/bin/../lib/jruby-complete-1.6 >>>>>>>> .0.jar:/usr/lib/hbase/bin/../lib/jsp-2.1-6.1.14.jar:/usr/lib/hbase/ >>>>>>>> bin/../lib/jsp-api-2.1-6.1.14.jar:/usr/lib/hbase/bin/../lib/jsp-api >>>>>>>> -2.1.jar:/usr/lib/hbase/bin/../lib/jsr311-api-1.1.1.jar:/usr/lib/hb >>>>>>>> ase/bin/../lib/log4j-1.2.16.jar:/usr/lib/hbase/bin/../lib/netty-3.2 >>>>>>>> .4.Final.jar:/usr/lib/hbase/bin/../lib/protobuf-java-2.3.0.jar:/usr >>>>>>>> /lib/hbase/bin/../lib/servlet-api-2.5-6.1.14.jar:/usr/lib/hbase/bin >>>>>>>> /../lib/servlet-api-2.5.jar:/usr/lib/hbase/bin/../lib/slf4j-api-1.5 >>>>>>>> .8.jar:/usr/lib/hbase/bin/../lib/slf4j-log4j12-1.5.8.jar:/usr/lib/h >>>>>>>> base/bin/../lib/snappy-java-1.0.3.2.jar:/usr/lib/hbase >>>>> /bin/ >>>>>>>>> >>>>>>>> >>>>>>>> ../lib/stax-api-1.0.1.jar:/usr/lib/hbase/bin/../lib/thrift-0.2.0.ja >>>>>>>> r:/usr/lib/hbase/bin/../lib/velocity-1.5.jar:/usr/lib/hbase/bin/../ >>>>>>>> lib/xmlenc-0.52.jar:/usr/lib/hbase/bin/../lib/zookeeper.jar:/etc/zo >>>>>>>> okeeper:/etc/hadoop-0.20/conf:/usr/lib/hadoop-0.20/hadoop-examples. >>>>>>>> jar:/usr/lib/hadoop-0.20/hadoop-0.20.2-cdh3u3-core.jar:/usr/lib/had >>>>>>>> oop-0.20/hadoop-0.20.2-cdh3u3-ant.jar:/usr/lib/hadoop-0.20/hadoop-c >>>>>>>> ore-0.20.2-cdh3u3.jar:/usr/lib/hadoop-0.20/hadoop-test.jar:/usr/lib >>>>>>>> /hadoop-0.20/hadoop-ant-0.20.2-cdh3u3.jar:/usr/lib/hadoop-0.20/hado >>>>>>>> op-tools.jar:/usr/lib/hadoop-0.20/hadoop-tools-0.20.2-cdh3u3.jar:/u >>>>>>>> sr/lib/hadoop-0.20/hadoop-test-0.20.2-cdh3u3.jar:/usr/lib/hadoop-0. >>>>>>>> 20/hadoop-core.jar:/usr/lib/hadoop-0.20/hadoop-0.20.2-cdh3u3-exampl >>>>>>>> es.jar:/usr/lib/hadoop-0.20/hadoop-0.20.2-cdh3u3-test.jar:/usr/lib/ >>>>>>>> hadoop-0.20/hadoop-ant.jar:/usr/lib/hadoop-0.20/hadoop-examples-0.2 >>>>>>>> 0.2-cdh3u3.jar:/usr/lib/hadoop-0.20/hadoop-0.20.2-cdh3u3-tools.jar: >>>>>>>> /usr/lib/hadoop-0.20/lib/jasper-runtime-5.5.12.jar:/us >>>>> r/lib >>>>>>>>> >>>>>>>> >>>>>>>> /hadoop-0.20/lib/hsqldb-1.8.0.10.jar:/usr/lib/hadoop-0.20/lib/jacks >>>>>>>> on-mapper-asl-1.5.2.jar:/usr/lib/hadoop-0.20/lib/jets3t-0.6.1.jar:/ >>>>>>>> usr/lib/hadoop-0.20/lib/jetty-servlet-tester-6.1.26.cloudera.1.jar: >>>>>>>> /usr/lib/hadoop-0.20/lib/jackson-core-asl-1.5.2.jar:/usr/lib/hadoop >>>>>>>> -0.20/lib/oro-2.0.8.jar:/usr/lib/hadoop-0.20/lib/ant-contrib-1.0b3. >>>>>>>> jar:/usr/lib/hadoop-0.20/lib/commons-daemon-1.0.1.jar:/usr/lib/hado >>>>>>>> op-0.20/lib/mockito-all-1.8.2.jar:/usr/lib/hadoop-0.20/lib/aspectjr >>>>>>>> t-1.6.5.jar:/usr/lib/hadoop-0.20/lib/commons-lang-2.4.jar:/usr/lib/ >>>>>>>> hadoop-0.20/lib/junit-4.5.jar:/usr/lib/hadoop-0.20/lib/commons-code >>>>>>>> c-1.4.jar:/usr/lib/hadoop-0.20/lib/servlet-api-2.5-6.1.14.jar:/usr/ >>>>>>>> lib/hadoop-0.20/lib/log4j-1.2.15.jar:/usr/lib/hadoop-0.20/lib/jsch- >>>>>>>> 0.1.42.jar:/usr/lib/hadoop-0.20/lib/core-3.1.1.jar:/usr/lib/hadoop- >>>>>>>> 0.20/lib/jetty-6.1.26.cloudera.1.jar:/usr/lib/hadoop-0.20/lib/commo >>>>>>>> ns-logging-1.0.4.jar:/usr/lib/hadoop-0.20/lib/jetty-util-6.1.26.clo >>>>>>>> udera.1.jar:/usr/lib/hadoop-0.20/lib/servlet-api-2.5-2 >>>>> 00812 >>>>>>>>> >>>>>>>> >>>>>>>> 11.jar:/usr/lib/hadoop-0.20/lib/jasper-compiler-5.5.12.jar:/usr/lib >>>>>>>> /hadoop-0.20/lib/kfs-0.2.2.jar:/usr/lib/hadoop-0.20/lib/commons-cli >>>>>>>> -1.2.jar:/usr/lib/hadoop-0.20/lib/commons-net-1.4.1.jar:/usr/lib/ha >>>>>>>> doop-0.20/lib/commons-httpclient-3.1.jar:/usr/lib/hadoop-0.20/lib/s >>>>>>>> lf4j-api-1.4.3.jar:/usr/lib/hadoop-0.20/lib/xmlenc-0.52.jar:/usr/li >>>>>>>> b/hadoop-0.20/lib/commons-logging-api-1.0.4.jar:/usr/lib/hadoop-0.2 >>>>>>>> 0/lib/commons-el-1.0.jar:/usr/lib/hadoop-0.20/lib/slf4j-log4j12-1.4 >>>>>>>> .3.jar:/usr/lib/hadoop-0.20/lib/aspectjtools-1.6.5.jar:/usr/lib/had >>>>>>>> oop-0.20/lib/guava-r09-jarjar.jar:/usr/lib/hadoop-0.20/lib/hadoop-f >>>>>>>> airscheduler-0.20.2-cdh3u3.jar:/usr/lib/zookeeper/zookeeper.jar:/us >>>>>>>> r/lib/zookeeper/zookeeper-3.3.5-cdh3u4.jar:/usr/lib/zookeeper/lib/l >>>>>>>> og4j-1.2.15.jar:/usr/lib/zookeeper/lib/jline-0.9.94.jar::/usr/lib/h >>>>>>>> adoop-0.20/conf:/usr/lib/hadoop-0.20/hadoop-core-0.20.2-cdh3u3.jar: >>>>>>>> /usr/lib/hadoop-0.20/lib/ant-contrib-1.0b3.jar:/usr/lib/hadoop-0.20 >>>>>>>> /lib/aspectjrt-1.6.5.jar:/usr/lib/hadoop-0.20/lib/aspe >>>>> ctjto >>>>>>>>> >>>>>>>> >>>>>>>> ols-1.6.5.jar:/usr/lib/hadoop-0.20/lib/commons-cli-1.2.jar:/usr/lib >>>>>>>> /hadoop-0.20/lib/commons-codec-1.4.jar:/usr/lib/hadoop-0.20/lib/com >>>>>>>> mons-daemon-1.0.1.jar:/usr/lib/hadoop-0.20/lib/commons-el-1.0.jar:/ >>>>>>>> usr/lib/hadoop-0.20/lib/commons-httpclient-3.1.jar:/usr/lib/hadoop- >>>>>>>> 0.20/lib/commons-lang-2.4.jar:/usr/lib/hadoop-0.20/lib/commons-logg >>>>>>>> ing-1.0.4.jar:/usr/lib/hadoop-0.20/lib/commons-logging-api-1.0.4.ja >>>>>>>> r:/usr/lib/hadoop-0.20/lib/commons-net-1.4.1.jar:/usr/lib/hadoop-0. >>>>>>>> 20/lib/core-3.1.1.jar:/usr/lib/hadoop-0.20/lib/guava-r09-jarjar.jar >>>>>>>> :/usr/lib/hadoop-0.20/lib/hadoop-fairscheduler-0.20.2-cdh3u3.jar:/u >>>>>>>> sr/lib/hadoop-0.20/lib/hsqldb-1.8.0.10.jar:/usr/lib/hadoop-0.20/lib >>>>>>>> /jackson-core-asl-1.5.2.jar:/usr/lib/hadoop-0.20/lib/jackson-mapper >>>>>>>> -asl-1.5.2.jar:/usr/lib/hadoop-0.20/lib/jasper-compiler-5.5.12.jar: >>>>>>>> /usr/lib/hadoop-0.20/lib/jasper-runtime-5.5.12.jar:/usr/lib/hadoop- >>>>>>>> 0.20/lib/jets3t-0.6.1.jar:/usr/lib/hadoop-0.20/lib/jetty-6.1.26.clo >>>>>>>> udera.1.jar:/usr/lib/hadoop-0.20/lib/jetty-servlet-tes >>>>> ter-6 >>>>>>>>> >>>>>>>> >>>>>>>> .1.26.cloudera.1.jar:/usr/lib/hadoop-0.20/lib/jetty-util-6.1.26.clo >>>>>>>> udera.1.jar:/usr/lib/hadoop-0.20/lib/jsch-0.1.42.jar:/usr/lib/hadoo >>>>>>>> p-0.20/lib/junit-4.5.jar:/usr/lib/hadoop-0.20/lib/kfs-0.2.2.jar:/us >>>>>>>> r/lib/hadoop-0.20/lib/log4j-1.2.15.jar:/usr/lib/hadoop-0.20/lib/moc >>>>>>>> kito-all-1.8.2.jar:/usr/lib/hadoop-0.20/lib/oro-2.0.8.jar:/usr/lib/ >>>>>>>> hadoop-0.20/lib/servlet-api-2.5-20081211.jar:/usr/lib/hadoop-0.20/l >>>>>>>> ib/servlet-api-2.5-6.1.14.jar:/usr/lib/hadoop-0.20/lib/slf4j-api-1. >>>>>>>> 4.3.jar:/usr/lib/hadoop-0.20/lib/slf4j-log4j12-1.4.3.jar:/usr/lib/h >>>>>>>> adoop-0.20/lib/xmlenc-0.52.jar12/05/12 >>>>>>>> 08:32:42 INFO zookeeper.ZooKeeper: Client >>>>>>>> environment:java.library.path=/usr/lib/hadoop-0.20/lib/native/Linux >>>>>>>> -amd64-64:/usr/lib/hbase/bin/../lib/native/Linux-amd64-6412/05/12 >>>>>>>> 08:32:42 INFO zookeeper.ZooKeeper: Client >>>>>>>> environment:java.io.tmpdir=/tmp12/05/12 08:32:42 INFO >>>>> zookeeper.ZooKeeper: >>>>>>>> Client environment:java.compiler=<NA>12/05/12 08:32:42 INFO >>>>>>>> zookeeper.ZooKeeper: Client environment:os.name=Linux12/05/12 >>>>>>>> 08:32:42 >>>>>>>>> INFO zookeeper.ZooKeeper: Client >>>>>>>>> environment:os.arch=amd6412/05/12 >>>>>>>> 08:32:42 INFO zookeeper.ZooKeeper: Client >>>>>>>> environment:os.version=2.6.35-22-server12/05/12 08:32:42 INFO >>>>>>>> zookeeper.ZooKeeper: Client environment:user.name=dalia12/05/12 >>>>>>>> 08:32:42 INFO zookeeper.ZooKeeper: Client >>>>>>>> environment:user.home=/home/dalia12/05/12 >>>>>>>> 08:32:42 INFO zookeeper.ZooKeeper: Client >>>>>>>> environment:user.dir=/home/dalia12/05/12 08:32:42 INFO >>>>> zookeeper.ZooKeeper: >>>>>>>> Initiating client connection, connectString=namenode:2181 >>>>>>>> sessionTimeout=180000 watcher=master:6000012/05/12 08:32:42 INFO >>>>>>>> zookeeper.ClientCnxn: Opening socket connection to server namenode/ >>>>>>>> 10.0.2.3:218112/05/12 08:32:42 INFO zookeeper.ClientCnxn: Socket >>>>>>>> connection established to namenode/10.0.2.3:2181, initiating >>>>>>>> session12/05/12 08:32:42 INFO zookeeper.ClientCnxn: Session >>>>>>>> establishment complete on server namenode/10.0.2.3:2181, sessionid >>>>>>>> = 0x13740bc4f70000c, negotiated timeout = 4000012/05/12 08:32:42 >>>>>>>> INFO >>>>>>>> jvm.JvmMetrics: Initializing JVM Metrics with >>>>>>>>> processName=Master, sessionId=namenode:6000012/05/12 08:32:42 >>>>>>>>> INFO >>>>>>>> hbase.metrics: MetricsString added: revision12/05/12 08:32:42 INFO >>>>>>>> hbase.metrics: MetricsString added: hdfsUser12/05/12 08:32:42 INFO >>>>>>>> hbase.metrics: MetricsString added: hdfsDate12/05/12 08:32:42 INFO >>>>>>>> hbase.metrics: MetricsString added: hdfsUrl12/05/12 08:32:42 INFO >>>>>>>> hbase.metrics: MetricsString added: date12/05/12 08:32:42 INFO >>>>>>>> hbase.metrics: MetricsString added: hdfsRevision12/05/12 08:32:42 >>>>>>>> INFO >>>>>>>> hbase.metrics: MetricsString added: user12/05/12 08:32:42 INFO >>>>>>>> hbase.metrics: MetricsString added: hdfsVersion12/05/12 08:32:42 >>>>>>>> INFO >>>>>>>> hbase.metrics: MetricsString added: url12/05/12 08:32:42 INFO >>>>>>>> hbase.metrics: MetricsString added: version12/05/12 08:32:42 INFO >>>>>>>> hbase.metrics: new MBeanInfo12/05/12 08:32:42 INFO hbase.metrics: >>>>>>>> new >>>>>>>> MBeanInfo12/05/12 08:32:42 INFO metrics.MasterMetrics: >>>>>>>> Initialized12/05/12 >>>>>>>> 08:32:42 INFO master.ActiveMasterManager: >>>>>>>> Master=namenode:6000012/05/12 >>>>>>>> 08:32:44 INFO ipc.Client: Retrying connect to serve >>>>>>>>> r: namenode/10.0.2.3:8020. Already tried 0 time(s).12/05/12 >>>>>>>>> 08:32:45 >>>>>>>> INFO ipc.Client: Retrying connect to server: namenode/10.0.2.3:8020. >>>>>>>> Already tried 1 time(s).12/05/12 08:32:46 INFO ipc.Client: Retrying >>>>>>>> connect to server: namenode/10.0.2.3:8020. Already tried 2 >>>>>>>> time(s).12/05/12 >>>>>>>> 08:32:47 INFO ipc.Client: Retrying connect to server: namenode/ >>>>>>>> 10.0.2.3:8020. Already tried 3 time(s).12/05/12 08:32:48 INFO >>>>>>>> ipc.Client: Retrying connect to server: namenode/10.0.2.3:8020. >>>>>>>> Already tried 4 time(s).12/05/12 08:32:49 INFO ipc.Client: Retrying >>>>>>>> connect to >>>>>>>> server: namenode/10.0.2.3:8020. Already tried 5 time(s).12/05/12 >>>>>>>> 08:32:50 INFO ipc.Client: Retrying connect to server: namenode/ >>>>>>>> 10.0.2.3:8020. Already tried 6 time(s).12/05/12 08:32:51 INFO >>>>>>>> ipc.Client: Retrying connect to server: namenode/10.0.2.3:8020. >>>>>>>> Already tried 7 time(s).12/05/12 08:32:52 INFO ipc.Client: Retrying >>>>>>>> connect to >>>>>>>> server: namenode/10.0.2.3:8020. Already tried 8 time(s).12/05/12 >>>>>>>> 08:32:53 INFO ipc.Client: Retrying connect to ser >>>>>>>>> ver: namenode/10.0.2.3:8020. Already tried 9 time(s).12/05/12 >>>>>>>> 08:32:53 FATAL master.HMaster: Unhandled exception. Starting >>>>>>>> shutdown.java.net.ConnectException: Call to >>>>> namenode/10.0.2.3:8020failed on connection exception: >>>>> java.net.ConnectException: Connection >>>>>>>> refused at >>>>> org.apache.hadoop.ipc.Client.wrapException(Client.java:1134) >>>>>>>> at org.apache.hadoop.ipc.Client.call(Client.java:1110) at >>>>>>>> org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:226) at >>>>>>>> $Proxy6.getProtocolVersion(Unknown Source) at >>>>>>>> org.apache.hadoop.ipc.RPC.getProxy(RPC.java:398) at >>>>>>>> org.apache.hadoop.ipc.RPC.getProxy(RPC.java:384) at >>>>>>>> org.apache.hadoop.hdfs.DFSClient.createRPCNamenode(DFSClient.java:123) >>>>>>>> at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:246) >>>>>>>> at >>>>>>>> org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:208) at >>>>>>>> >>>>> org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSyste >>>>> m.java:89) >>>>>>>> at >>>>>>>> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1563) >>>>>>>> at org.apache.hadoop.fs.FileSystem.acc >>>>>>>>> ess$200(FileSystem.java:67) at >>>>>>>> org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:1597) >>>>>>>> at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1579) >>>>> at >>>>>>>> org.apache.hadoop.fs.FileSystem.get(FileSystem.java:228) at >>>>>>>> org.apache.hadoop.fs.Path.getFileSystem(Path.java:183) at >>>>>>>> org.apache.hadoop.hbase.util.FSUtils.getRootDir(FSUtils.java:364) at >>>>>>>> >>>>> org.apache.hadoop.hbase.master.MasterFileSystem.<init>(MasterFileSystem.java >>>>> :86) >>>>>>>> at >>>>>>>> >>>>> org.apache.hadoop.hbase.master.HMaster.finishInitialization(HMaster.java:360 >>>>> ) >>>>>>>> at >>>>>>>> org.apache.hadoop.hbase.master.HMaster.run(HMaster.java:293)Caused by: >>>>>>>> java.net.ConnectException: Connection refused at >>>>>>>> sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) at >>>>>>>> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567) >>>>>>>> at >>>>>>>> >>>>> org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:2 >>>>> 06) >>>>>>>> at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:408) at >>>>>>>> >>>>> org.apache.hadoop.ipc.Client$Connection.setupConnection(Client.java:425) >>>>>>>> a >>>>>>>>> t >>>>>>>> org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java: >>>>>>>> 532) at >>>>>>>> org.apache.hadoop.ipc.Client$Connection.access$2300(Client.java:210 >>>>>>>> ) at >>>>>>>> org.apache.hadoop.ipc.Client.getConnection(Client.java:1247) at >>>>>>>> org.apache.hadoop.ipc.Client.call(Client.java:1078) ... 18 >>>>>>>> more12/05/12 >>>>>>>> 08:32:53 INFO master.HMaster: Aborting12/05/12 08:32:53 DEBUG >>>>>>>> master.HMaster: Stopping service threads12/05/12 08:32:53 INFO >>>>>>>> ipc.HBaseServer: Stopping server on 6000012/05/12 08:32:53 INFO >>>>>>>> ipc.HBaseServer: IPC Server handler 5 on 60000: exiting12/05/12 >>>>>>>> 08:32:53 INFO ipc.HBaseServer: Stopping IPC Server listener on >>>>>>>> 6000012/05/12 >>>>>>>> 08:32:53 INFO ipc.HBaseServer: IPC Server handler 1 on 60000: >>>>>>>> exiting12/05/12 08:32:53 INFO ipc.HBaseServer: IPC Server handler 0 >>>>>>>> on >>>>>>>> 60000: exiting12/05/12 08:32:53 INFO ipc.HBaseServer: IPC Server >>>>>>>> handler 3 on 60000: exiting12/05/12 08:32:53 INFO ipc.HBaseServer: >>>>>>>> IPC Server handler >>>>>>>> 7 on 60000: exiting12/05/12 08:32:53 INFO ipc.HBaseServer: IPC >>>>>>>> Server handler 9 on 60000: exiting1 >>>>>>>>> 2/05/12 08:32:53 INFO ipc.HBaseServer: IPC Server handler 6 on >>>>> 60000: >>>>>>>> exiting12/05/12 08:32:53 INFO ipc.HBaseServer: IPC Server handler 4 >>>>>>>> on >>>>>>>> 60000: exiting12/05/12 08:32:53 INFO ipc.HBaseServer: IPC Server >>>>>>>> handler 2 on 60000: exiting12/05/12 08:32:53 INFO ipc.HBaseServer: >>>>>>>> IPC Server handler >>>>>>>> 8 on 60000: exiting12/05/12 08:32:53 INFO ipc.HBaseServer: Stopping >>>>>>>> IPC Server Responder12/05/12 08:32:53 INFO zookeeper.ZooKeeper: >>>>> Session: >>>>>>>> 0x13740bc4f70000c closed12/05/12 08:32:53 INFO zookeeper.ClientCnxn: >>>>>>>> EventThread shut down12/05/12 08:32:53 INFO master.HMaster: HMaster >>>>>>>> main thread exiting> From: ha...@cloudera.com >>>>>>>>>> Date: Sat, 12 May 2012 17:28:29 +0530 >>>>>>>>>> Subject: Re: Important "Undefined Error" >>>>>>>>>> To: user@hbase.apache.org >>>>>>>>>> >>>>>>>>>> Hi Dalia, >>>>>>>>>> >>>>>>>>>> On Sat, May 12, 2012 at 5:14 PM, Dalia Sobhy < >>>>>>>> dalia.mohso...@hotmail.com> wrote: >>>>>>>>>>> >>>>>>>>>>> Dear all, >>>>>>>>>>> I have first a problem with Hbase I am trying to install it on >>>>>>>>>>> a >>>>>>>> distributed/multinode cluster.. >>>>>>>>>>> I am using the cloudera >>>>>>>> https://ccp.cloudera.com/display/CDH4B2/HBase+Installation#HBaseIns >>>>>>>> tallation-StartingtheHBaseMaster >>>>>>>>>>> But when I write this command >>>>>>>>>>> Creating the /hbase Directory in HDFS $sudo -u hdfs hadoop fs >>>>>>>>>>> -mkdir >>>>>>>> /hbase >>>>>>>>>>> I get the following error:12/05/12 07:20:42 INFO >>>>>>>> security.UserGroupInformation: JAAS Configuration already set up >>>>>>>> for Hadoop, not re-installing. >>>>>>>>>> >>>>>>>>>> This is not an error and you shouldn't be worried. It is rather >>>>>>>>>> a noisy INFO log that should be fixed (as a DEBUG level instead) >>>>>>>>>> in subsequent releases (Are you using CDH3 or CDH4? IIRC only >>>>>>>>>> CDH3u3 printed these, not in anything above that.) >>>>>>>>>> >>>>>>>>>>> 2. Another Aspect is when I start the hbase master it closes >>>>>>>> automatically after a while. >>>>>>>>>> >>>>>>>>>> Could you post us your HMaster start->crash log? You can use a >>>>>>>>>> service like pastebin.com to send us the output. >>>>>>>>>> >>>>>>>>>>> 3. Also this command is not working$host -v -t A >>>>> `namenode`namenode: >>>>>>>> command not found >>>>>>>>>> >>>>>>>>>> The right command is perhaps just: >>>>>>>>>> >>>>>>>>>> $host -v -t A `hostname` >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> Harsh J >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> Harsh J >>>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> -- >>>>>>> >>>>>>> >>>>>>> ∞ >>>>>>> Shashwat Shriparv >>>>>>> >>>>>>> >>>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> >>>>>> >>>>>> ∞ >>>>>> Shashwat Shriparv >>>>> >>>>> >>>> >>>