Hi,
I am getting error when I am trying to connect hive table (which is being
created through HbaseIntegration) in spark

Steps I followed :
*Hive Table creation code  *:
CREATE EXTERNAL TABLE IF NOT EXISTS TEST(NAME STRING,AGE INT)
STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler'
WITH SERDEPROPERTIES ("hbase.columns.mapping" = ":key,0:AGE")
TBLPROPERTIES ("hbase.table.name" = "TEST",
"hbase.mapred.output.outputtable" = "TEST");


*DESCRIBE TEST ;*
col_name    data_type    comment
name            string         from deserializer
age               int             from deserializer


*Spark Code :*
import org.apache.spark._
import org.apache.spark.sql._

val hiveContext = new org.apache.spark.sql.hive.HiveContext(sc)
hiveContext.sql("from TEST SELECT  NAME").collect.foreach(println)


*Starting Spark shell*
spark-shell --jars
/usr/hdp/2.3.4.0-3485/hive/lib/guava-14.0.1.jar,/usr/hdp/2.3.4.0-3485/hive/lib/hive-hbase-handler.jar,/usr/hdp/2.3.4.0-3485/hbase/lib/hbase-client.jar,/usr/hdp/2.3.4.0-3485/hbase/lib/hbase-common.jar,/usr/hdp/2.3.4.0-3485/hbase/lib/hbase-protocol.jar,/usr/hdp/2.3.4.0-3485/hbase/lib/hbase-server.jar,/usr/hdp/2.3.4.0-3485/hive/lib/hive-hbase-handler.jar,/usr/hdp/2.3.4.0-3485/hive/lib/htrace-core-3.1.0-incubating.jar,/usr/hdp/2.3.4.0-3485/hive/lib/zookeeper-3.4.6.2.3.4.0-3485.jar,/usr/hdp/2.3.4.0-3485/hbase/lib/hbase-server.jar
--driver-class-path
/usr/hdp/2.3.4.0-3485/hive/lib/guava-14.0.1.jar,/usr/hdp/2.3.4.0-3485/hbase/lib/hbase-client.jar,/usr/hdp/2.3.4.0-3485/hbase/lib/hbase-common.jar,/usr/hdp/2.3.4.0-3485/hbase/lib/hbase-protocol.jar,/usr/hdp/2.3.4.0-3485/hbase/lib/hbase-server.jar,/usr/hdp/2.3.4.0-3485/hive/lib/hive-hbase-handler.jar,/usr/hdp/2.3.4.0-3485/hive/lib/htrace-core-3.1.0-incubating.jar,/usr/hdp/2.3.4.0-3485/hive/lib/zookeeper-3.4.6.2.3.4.0-3485.jar,/usr/hdp/2.3.4.0-3485/hive/lib/hive-hbase-handler.jar,/usr/hdp/2.3.4.0-3485/hbase/lib/hbase-server.jar
--packages com.databricks:spark-csv_2.10:1.3.0  --master yarn-client -i
/TestDivya/Spark/InstrumentCopyToHDFSHive.scala

*Stack Trace* :

Stack SQL context available as sqlContext.
> Loading /TestDivya/Spark/InstrumentCopyToHDFSHive.scala...
> import org.apache.spark._
> import org.apache.spark.sql._
> 16/02/29 23:09:29 INFO HiveContext: Initializing execution hive, version
> 1.2.1
> 16/02/29 23:09:29 INFO ClientWrapper: Inspected Hadoop version:
> 2.7.1.2.3.4.0-3485
> 16/02/29 23:09:29 INFO ClientWrapper: Loaded
> org.apache.hadoop.hive.shims.Hadoop23Shims for Hadoop version
> 2.7.1.2.3.4.0-3485
> 16/02/29 23:09:29 INFO HiveContext: default warehouse location is
> /user/hive/warehouse
> 16/02/29 23:09:29 INFO HiveContext: Initializing HiveMetastoreConnection
> version 1.2.1 using Spark classes.
> 16/02/29 23:09:29 INFO ClientWrapper: Inspected Hadoop version:
> 2.7.1.2.3.4.0-3485
> 16/02/29 23:09:29 INFO ClientWrapper: Loaded
> org.apache.hadoop.hive.shims.Hadoop23Shims for Hadoop version
> 2.7.1.2.3.4.0-3485
> 16/02/29 23:09:30 WARN NativeCodeLoader: Unable to load native-hadoop
> library for your platform... using builtin-java classes where applicable
> 16/02/29 23:09:30 INFO metastore: Trying to connect to metastore with URI
> thrift://ip-xxx-xx-xx-xxx.ap-southeast-1.compute.internal:9083
> 16/02/29 23:09:30 INFO metastore: Connected to metastore.
> 16/02/29 23:09:30 WARN DomainSocketFactory: The short-circuit local reads
> feature cannot be used because libhadoop cannot be loaded.
> 16/02/29 23:09:31 INFO SessionState: Created local directory:
> /tmp/1bf53785-f7c8-406d-a733-a5858ccb2d16_resources
> 16/02/29 23:09:31 INFO SessionState: Created HDFS directory:
> /tmp/hive/hdfs/1bf53785-f7c8-406d-a733-a5858ccb2d16
> 16/02/29 23:09:31 INFO SessionState: Created local directory:
> /tmp/hdfs/1bf53785-f7c8-406d-a733-a5858ccb2d16
> 16/02/29 23:09:31 INFO SessionState: Created HDFS directory:
> /tmp/hive/hdfs/1bf53785-f7c8-406d-a733-a5858ccb2d16/_tmp_space.db
> hiveContext: org.apache.spark.sql.hive.HiveContext =
> org.apache.spark.sql.hive.HiveContext@10b14f32
> 16/02/29 23:09:32 INFO ParseDriver: Parsing command: from TEST SELECT  NAME
> 16/02/29 23:09:32 INFO ParseDriver: Parse Completed
> 16/02/29 23:09:33 INFO deprecation: mapred.map.tasks is deprecated.
> Instead, use mapreduce.job.maps
> 16/02/29 23:09:33 INFO MemoryStore: ensureFreeSpace(468352) called with
> curMem=0, maxMem=556038881
> 16/02/29 23:09:33 INFO MemoryStore: Block broadcast_0 stored as values in
> memory (estimated size 457.4 KB, free 529.8 MB)
> 16/02/29 23:09:33 INFO MemoryStore: ensureFreeSpace(49454) called with
> curMem=468352, maxMem=556038881
> 16/02/29 23:09:33 INFO MemoryStore: Block broadcast_0_piece0 stored as
> bytes in memory (estimated size 48.3 KB, free 529.8 MB)
> 16/02/29 23:09:33 INFO BlockManagerInfo: Added broadcast_0_piece0 in
> memory on xxx.xx.xx.xxx:37784 (size: 48.3 KB, free: 530.2 MB)
> 16/02/29 23:09:33 INFO SparkContext: Created broadcast 0 from collect at
> <console>:30
> 16/02/29 23:09:34 INFO HBaseStorageHandler: Configuring input job
> properties
> 16/02/29 23:09:34 INFO RecoverableZooKeeper: Process
> identifier=hconnection-0x26fa89a2 connecting to ZooKeeper
> ensemble=localhost:2181
> 16/02/29 23:09:34 INFO ZooKeeper: Client
> environment:zookeeper.version=3.4.6-3485--1, built on 12/16/2015 02:35 GMT
> 16/02/29 23:09:34 INFO ZooKeeper: Client environment:host.name
> =ip-xxx-xx-xx-xxx.ap-southeast-1.compute.internal
> 16/02/29 23:09:34 INFO ZooKeeper: Client environment:java.version=1.7.0_67
> 16/02/29 23:09:34 INFO ZooKeeper: Client environment:java.vendor=Oracle
> Corporation
> 16/02/29 23:09:34 INFO ZooKeeper: Client
> environment:java.home=/usr/jdk64/jdk1.7.0_67/jre
> 16/02/29 23:09:34 INFO ZooKeeper: Client
> environment:java.class.path=/usr/hdp/2.3.4.0-3485/hive/lib/guava-14.0.1.jar,/usr/hdp/2.3.4.0-3485/hbase/lib/hbase-client.jar,/usr/hdp/2.3.4.0-3485/hbase/lib/hbase-common.jar,/usr/hdp/2.3.4.0-3485/hbase/lib/hbase-protocol.jar,/usr/hdp/2.3.4.0-3485/hbase/lib/hbase-server.jar,/usr/hdp/2.3.4.0-3485/hive/lib/hive-hbase-handler.jar,/usr/hdp/2.3.4.0-3485/hive/lib/htrace-core-3.1.0-incubating.jar,/usr/hdp/2.3.4.0-3485/hive/lib/zookeeper-3.4.6.2.3.4.0-3485.jar,/usr/hdp/2.3.4.0-3485/hive/lib/hive-hbase-handler.jar,/usr/hdp/2.3.4.0-3485/hbase/lib/hbase-server.jar:/usr/hdp/current/spark-thriftserver/conf/:/usr/hdp/2.3.4.0-3485/spark/lib/spark-assembly-1.5.2.2.3.4.0-3485-hadoop2.7.1.2.3.4.0-3485.jar:/usr/hdp/2.3.4.0-3485/spark/lib/datanucleus-api-jdo-3.2.6.jar:/usr/hdp/2.3.4.0-3485/spark/lib/datanucleus-core-3.2.10.jar:/usr/hdp/2.3.4.0-3485/spark/lib/datanucleus-rdbms-3.2.9.jar:/usr/hdp/current/hadoop-client/conf/
> 16/02/29 23:09:34 INFO ZooKeeper: Client
> environment:java.library.path=/usr/java/packages/lib/amd64:/usr/lib64:/lib64:/lib:/usr/lib
> 16/02/29 23:09:34 INFO ZooKeeper: Client environment:java.io.tmpdir=/tmp
> 16/02/29 23:09:34 INFO ZooKeeper: Client environment:java.compiler=<NA>
> 16/02/29 23:09:34 INFO ZooKeeper: Client environment:os.name=Linux
> 16/02/29 23:09:34 INFO ZooKeeper: Client environment:os.arch=amd64
> 16/02/29 23:09:34 INFO ZooKeeper: Client
> environment:os.version=3.10.0-229.el7.x86_64
> 16/02/29 23:09:34 INFO ZooKeeper: Client environment:user.name=hdfs
> 16/02/29 23:09:34 INFO ZooKeeper: Client environment:user.home=/home/hdfs
> 16/02/29 23:09:34 INFO ZooKeeper: Client environment:user.dir=/home/hdfs
> 16/02/29 23:09:34 INFO ZooKeeper: Initiating client connection,
> connectString=localhost:2181 sessionTimeout=90000
> watcher=hconnection-0x26fa89a20x0, quorum=localhost:2181, baseZNode=/hbase
> 16/02/29 23:09:34 INFO ClientCnxn: Opening socket connection to server
> localhost/0:0:0:0:0:0:0:1:2181. Will not attempt to authenticate using SASL
> (unknown error)
> 16/02/29 23:09:34 INFO ClientCnxn: Socket connection established to
> localhost/0:0:0:0:0:0:0:1:2181, initiating session
> 16/02/29 23:09:34 INFO ClientCnxn: Session establishment complete on
> server localhost/0:0:0:0:0:0:0:1:2181, sessionid = 0x3532fb70ba20034,
> negotiated timeout = 40000
> 16/02/29 23:09:34 WARN TableInputFormatBase: You are using an HTable
> instance that relies on an HBase-managed Connection. This is usually due to
> directly creating an HTable, which is deprecated. Instead, you should
> create a Connection object and then request a Table instance from it. If
> you don't need the Table instance for your own use, you should instead use
> the TableInputFormatBase.initalizeTable method directly.
> 16/02/29 23:09:34 INFO TableInputFormatBase: Creating an additional
> unmanaged connection because user provided one can't be used for
> administrative actions. We'll close it when we close out the table.
> 16/02/29 23:09:34 INFO RecoverableZooKeeper: Process
> identifier=hconnection-0x6fd74d35 connecting to ZooKeeper
> ensemble=localhost:2181
> 16/02/29 23:09:34 INFO ZooKeeper: Initiating client connection,
> connectString=localhost:2181 sessionTimeout=90000
> watcher=hconnection-0x6fd74d350x0, quorum=localhost:2181, baseZNode=/hbase
> 16/02/29 23:09:34 INFO ClientCnxn: Opening socket connection to server
> localhost/0:0:0:0:0:0:0:1:2181. Will not attempt to authenticate using SASL
> (unknown error)
> 16/02/29 23:09:34 INFO ClientCnxn: Socket connection established to
> localhost/0:0:0:0:0:0:0:1:2181, initiating session
> 16/02/29 23:09:34 INFO ClientCnxn: Session establishment complete on
> server localhost/0:0:0:0:0:0:0:1:2181, sessionid = 0x3532fb70ba20035,
> negotiated timeout = 40000
> 16/02/29 23:09:34 INFO RegionSizeCalculator: Calculating region sizes for
> table "TEST".
> 16/02/29 23:10:23 INFO RpcRetryingCaller: Call exception, tries=10,
> retries=35, started=48318 ms ago, cancelled=false, msg=
> 16/02/29 23:10:43 INFO RpcRetryingCaller: Call exception, tries=11,
> retries=35, started=68524 ms ago, cancelled=false, msg=
> 16/02/29 23:11:03 INFO RpcRetryingCaller: Call exception, tries=12,
> retries=35, started=88617 ms ago, cancelled=false, msg=
> 16/02/29 23:11:23 INFO RpcRetryingCaller: Call exception, tries=13,
> retries=35, started=108676 ms ago, cancelled=false, msg=
> 16/02/29 23:11:43 INFO RpcRetryingCaller: Call exception, tries=14,
> retries=35, started=128747 ms ago, cancelled=false, msg=
> 16/02/29 23:12:03 INFO RpcRetryingCaller: Call exception, tries=15,
> retries=35, started=148938 ms ago, cancelled=false, msg=
> 16/02/29 23:12:23 INFO RpcRetryingCaller: Call exception, tries=16,
> retries=35, started=168942 ms ago, cancelled=false, msg=
> 16/02/29 23:12:43 INFO RpcRetryingCaller: Call exception, tries=17,
> retries=35, started=188975 ms ago, cancelled=false, msg=
> Trace :



Could somebody help me in resolving the error.
Would really appreciate the help .


Thanks,
Divya

Reply via email to