Now my hbase is working fine now, but i am still getting the same error

[127.0.0.1:10000] hive> CREATE TABLE hbase_table_1(key int, value string)
                      > STORED BY 'org.apache.hadoop.hive.hbase.
HBaseStorageHandler'
                      > WITH SERDEPROPERTIES ("hbase.columns.mapping" =
":key,cf1:val")
                      > TBLPROPERTIES ("hbase.table.name" = "xyz");



[Hive Error]: Query returned non-zero code: 1, cause: FAILED: Execution
Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask.
MetaException(message:org.apache.hadoop.hbase.client.RetriesExhaustedException:
Can't get the locations
    at org.apache.hadoop.hbase.client.RpcRetryingCallerWithReadRepli
cas.getRegionLocations(RpcRetryingCallerWithReadReplicas.java:305)

On Thu, May 14, 2015 at 1:18 AM, Ibrar Ahmed <ibrar.ah...@gmail.com> wrote:

> Seems you are right, Sometime I got this error while running hbase shell
> command.
>
>
> ibrar@ibrar-virtual-machine:/usr/local/hbase/bin$ ./hbase shell
>
> SLF4J: Class path contains multiple SLF4J bindings.
> SLF4J: Found binding in
> [jar:file:/usr/local/hbase/lib/slf4j-log4j12-1.7.7.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in
> [jar:file:/usr/local/hadoop/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
> explanation.
> SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
> 2015-05-14 01:14:27,063 WARN  [main] util.NativeCodeLoader: Unable to load
> native-hadoop library for your platform... using builtin-java classes where
> applicable
> 2015-05-14 01:14:43,982 ERROR [main] zookeeper.RecoverableZooKeeper:
> ZooKeeper exists failed after 4 attempts
> 2015-05-14 01:14:43,983 WARN  [main] zookeeper.ZKUtil:
> hconnection-0x4d980c0x0, quorum=localhost:2181, baseZNode=/hbase Unable to
> set watcher on znode (/hbase/hbaseid)
> org.apache.zookeeper.KeeperException$ConnectionLossException:
> KeeperErrorCode = ConnectionLoss for /hbase/hbaseid
>     at org.apache.zookeeper.KeeperException.create(KeeperException.java:99)
>     at org.apache.zookeeper.KeeperException.create(KeeperException.java:51)
>     at org.apache.zookeeper.ZooKeeper.exists(ZooKeeper.java:1045)
>     at
> org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.exists(RecoverableZooKeeper.java:222)
>     at
> org.apache.hadoop.hbase.zookeeper.ZKUtil.checkExists(ZKUtil.java:481)
>     at
> org.apache.hadoop.hbase.zookeeper.ZKClusterId.readClusterIdZNode(ZKClusterId.java:65)
>     at
> org.apache.hadoop.hbase.client.ZooKeeperRegistry.getClusterId(ZooKeeperRegistry.java:86)
>     at
> org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.retrieveClusterId(ConnectionManager.java:833)
>     at
> org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.<init>(ConnectionManager.java:623)
>     at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
> Method)
>     at
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
>     at
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
>     at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
>     at
> org.apache.hadoop.hbase.client.ConnectionFactory.createConnection(ConnectionFactory.java:238)
>     at
> org.apache.hadoop.hbase.client.ConnectionFactory.createConnection(ConnectionFactory.java:218)
>     at
> org.apache.hadoop.hbase.client.ConnectionFactory.createConnection(ConnectionFactory.java:119)
>     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>     at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>     at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>     at java.lang.reflect.Method.invoke(Method.java:606)
>     at
> org.jruby.javasupport.JavaMethod.invokeDirectWithExceptionHandling(JavaMethod.java:450)
>     at
> org.jruby.javasupport.JavaMethod.invokeStaticDirect(JavaMethod.java:362)
>     at
> org.jruby.java.invokers.StaticMethodInvoker.call(StaticMethodInvoker.java:58)
>     at
> org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:312)
>     at
> org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:169)
>     at org.jruby.ast.CallOneArgNode.interpret(CallOneArgNode.java:57)
>     at org.jruby.ast.InstAsgnNode.interpret(InstAsgnNode.java:95)
>     at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104)
>     at org.jruby.ast.BlockNode.interpret(BlockNode.java:71)
>     at
> org.jruby.evaluator.ASTInterpreter.INTERPRET_METHOD(ASTInterpreter.java:74)
>     at
> org.jruby.internal.runtime.methods.InterpretedMethod.call(InterpretedMethod.java:169)
>     at
> org.jruby.internal.runtime.methods.DefaultMethod.call(DefaultMethod.java:191)
>     at
> org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:302)
>     at
> org.jruby.runtime.callsite.CachingCallSite.callBlock(CachingCallSite.java:144)
>     at
> org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:148)
>     at org.jruby.RubyClass.newInstance(RubyClass.java:822)
>     at
> org.jruby.RubyClass$i$newInstance.call(RubyClass$i$newInstance.gen:65535)
>     at
> org.jruby.internal.runtime.methods.JavaMethod$JavaMethodZeroOrNBlock.call(JavaMethod.java:249)
>     at
> org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:292)
>     at
> org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:135)
>     at
> usr.local.hbase.bin.$_dot_dot_.bin.hirb.__file__(/usr/local/hbase/bin/../bin/hirb.rb:118)
>     at
> usr.local.hbase.bin.$_dot_dot_.bin.hirb.load(/usr/local/hbase/bin/../bin/hirb.rb)
>     at org.jruby.Ruby.runScript(Ruby.java:697)
>     at org.jruby.Ruby.runScript(Ruby.java:690)
>     at org.jruby.Ruby.runNormally(Ruby.java:597)
>     at org.jruby.Ruby.runFromMain(Ruby.java:446)
>     at org.jruby.Main.doRunFromMain(Main.java:369)
>     at org.jruby.Main.internalRun(Main.java:258)
>     at org.jruby.Main.run(Main.java:224)
>     at org.jruby.Main.run(Main.java:208)
>     at org.jruby.Main.main(Main.java:188)
>
>
> On Thu, May 14, 2015 at 1:11 AM, kulkarni.swar...@gmail.com <
> kulkarni.swar...@gmail.com> wrote:
>
>> Ibrar,
>>
>> This seems to be an issue with the cluster rather than the integration
>> itself. Can you make sure that HBase is happy and healthy and all RS are up
>> and running?
>>
>> On Wed, May 13, 2015 at 1:58 PM, Ibrar Ahmed <ibrar.ah...@gmail.com>
>> wrote:
>>
>>> Hi,
>>>
>>> I am creating a table using hive and getting this error.
>>>
>>> [127.0.0.1:10000] hive> CREATE TABLE hbase_table_1(key int, value
>>> string)
>>>                       > STORED BY
>>> 'org.apache.hadoop.hive.hbase.HBaseStorageHandler'
>>>                       > WITH SERDEPROPERTIES ("hbase.columns.mapping" =
>>> ":key,cf1:val")
>>>                       > TBLPROPERTIES ("hbase.table.name" = "xyz");
>>>
>>>
>>>
>>> [Hive Error]: Query returned non-zero code: 1, cause: FAILED: Execution
>>> Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask.
>>> MetaException(message:org.apache.hadoop.hbase.client.RetriesExhaustedException:
>>> Can't get the locations
>>>     at
>>> org.apache.hadoop.hbase.client.RpcRetryingCallerWithReadReplicas.getRegionLocations(RpcRetryingCallerWithReadReplicas.java:305)
>>>     at
>>> org.apache.hadoop.hbase.client.ScannerCallableWithReplicas.call(ScannerCallableWithReplicas.java:147)
>>>     at
>>> org.apache.hadoop.hbase.client.ScannerCallableWithReplicas.call(ScannerCallableWithReplicas.java:56)
>>>     at
>>> org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithoutRetries(RpcRetryingCaller.java:200)
>>>     at
>>> org.apache.hadoop.hbase.client.ClientScanner.call(ClientScanner.java:288)
>>>     at
>>> org.apache.hadoop.hbase.client.ClientScanner.nextScanner(ClientScanner.java:267)
>>>     at
>>> org.apache.hadoop.hbase.client.ClientScanner.initializeScannerInConstruction(ClientScanner.java:139)
>>>     at
>>> org.apache.hadoop.hbase.client.ClientScanner.<init>(ClientScanner.java:134)
>>>     at org.apache.hadoop.hbase.client.HTable.getScanner(HTable.java:823)
>>>     at
>>> org.apache.hadoop.hbase.MetaTableAccessor.fullScan(MetaTableAccessor.java:601)
>>>     at
>>> org.apache.hadoop.hbase.MetaTableAccessor.tableExists(MetaTableAccessor.java:365)
>>>     at
>>> org.apache.hadoop.hbase.client.HBaseAdmin.tableExists(HBaseAdmin.java:281)
>>>     at
>>> org.apache.hadoop.hbase.client.HBaseAdmin.tableExists(HBaseAdmin.java:291)
>>>     at
>>> org.apache.hadoop.hive.hbase.HBaseStorageHandler.preCreateTable(HBaseStorageHandler.java:162)
>>>     at
>>> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.createTable(HiveMetaStoreClient.java:554)
>>>     at
>>> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.createTable(HiveMetaStoreClient.java:547)
>>>     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>     at
>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>>>     at
>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>>     at java.lang.reflect.Method.invoke(Method.java:606)
>>>     at
>>> org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:89)
>>>     at com.sun.proxy.$Proxy7.createTable(Unknown Source)
>>>     at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:613)
>>>     at
>>> org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:4194)
>>>     at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:281)
>>>     at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:153)
>>>     at
>>> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:85)
>>>     at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1472)
>>>     at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1239)
>>>     at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1057)
>>>     at org.apache.hadoop.hive.ql.Driver.run(Driver.java:880)
>>>     at org.apache.hadoop.hive.ql.Driver.run(Driver.java:870)
>>>     at
>>> org.apache.hadoop.hive.service.HiveServer$HiveServerHandler.execute(HiveServer.java:198)
>>>     at
>>> org.apache.hadoop.hive.service.ThriftHive$Processor$execute.getResult(ThriftHive.java:644)
>>>     at
>>> org.apache.hadoop.hive.service.ThriftHive$Processor$execute.getResult(ThriftHive.java:628)
>>>     at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
>>>     at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
>>>     at
>>> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:206)
>>>     at
>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>>>     at
>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>>>     at java.lang.Thread.run(Thread.java:745)
>>> )
>>>
>>>
>>> Any help/clue can help.
>>>
>>>
>>
>>
>> --
>> Swarnim
>>
>
>
>
> --
>
>


-- 
Ibrar Ahmed

Reply via email to