Now able to solve the issue by setting

SparkConf sconf = *new* SparkConf().setAppName(“App").setMaster("local")
and

conf.set(“zookeeper.znode.parent”, “/hbase-unsecure”)

Standalone hbase has a table 'test'
hbase(main):001:0> scan 'test'
ROW                                  COLUMN+CELL
row1                                column=cf:a, timestamp=1430234895637,
value=value1
row2                                column=cf:b, timestamp=1430234907537,
value=value2
row3                                column=cf:c, timestamp=1430234918284,
value=value3

Now facing this issue:

ERROR TableInputFormat: java.io.IOException:
java.lang.reflect.InvocationTargetException
at
org.apache.hadoop.hbase.client.HConnectionManager.createConnection(HConnectionManager.java:416)
at
org.apache.hadoop.hbase.client.HConnectionManager.createConnection(HConnectionManager.java:393)
at
org.apache.hadoop.hbase.client.HConnectionManager.getConnection(HConnectionManager.java:274)
at org.apache.hadoop.hbase.client.HTable.<init>(HTable.java:194)
at org.apache.hadoop.hbase.client.HTable.<init>(HTable.java:156)
at
org.apache.hadoop.hbase.mapreduce.TableInputFormat.setConf(TableInputFormat.java:101)
at org.apache.spark.rdd.NewHadoopRDD.getPartitions(NewHadoopRDD.scala:91)
at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:219)
at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:217)
at scala.Option.getOrElse(Option.scala:120)
at org.apache.spark.rdd.RDD.partitions(RDD.scala:217)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:1632)
at org.apache.spark.rdd.RDD.count(RDD.scala:1012)
at org.apache.spark.examples.HBaseTest$.main(HBaseTest.scala:58)
at org.apache.spark.examples.HBaseTest.main(HBaseTest.scala)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at
org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:607)
at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:167)
at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:190)
at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:111)
at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
Caused by: java.lang.reflect.InvocationTargetException
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
at
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
at
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
at
org.apache.hadoop.hbase.client.HConnectionManager.createConnection(HConnectionManager.java:414)
... 23 more
Caused by: java.lang.VerifyError: class
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$Result overrides
final method getUnknownFields.()Lcom/google/protobuf/UnknownFieldSet;
at java.lang.ClassLoader.defineClass1(Native Method)
at java.lang.ClassLoader.defineClass(ClassLoader.java:800)
at java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142)
at java.net.URLClassLoader.defineClass(URLClassLoader.java:449)
at java.net.URLClassLoader.access$100(URLClassLoader.java:71)
at java.net.URLClassLoader$1.run(URLClassLoader.java:361)
at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
at java.security.AccessController.doPrivileged(Native Method)
at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
at java.lang.ClassLoader.loadClass(ClassLoader.java:425)
at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)
at java.lang.ClassLoader.loadClass(ClassLoader.java:358)
at
org.apache.hadoop.hbase.protobuf.ProtobufUtil.<clinit>(ProtobufUtil.java:176)
at org.apache.hadoop.hbase.ClusterId.parseFrom(ClusterId.java:64)
at
org.apache.hadoop.hbase.zookeeper.ZKClusterId.readClusterIdZNode(ZKClusterId.java:69)
at
org.apache.hadoop.hbase.client.ZooKeeperRegistry.getClusterId(ZooKeeperRegistry.java:83)
at
org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.retrieveClusterId(HConnectionManager.java:857)
at
org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.<init>(HConnectionManager.java:662)
... 28 more

Exception in thread "main" java.io.IOException: No table was provided.
at
org.apache.hadoop.hbase.mapreduce.TableInputFormatBase.getSplits(TableInputFormatBase.java:154)
at org.apache.spark.rdd.NewHadoopRDD.getPartitions(NewHadoopRDD.scala:95)
at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:219)
at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:217)
at scala.Option.getOrElse(Option.scala:120)
at org.apache.spark.rdd.RDD.partitions(RDD.scala:217)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:1632)
at org.apache.spark.rdd.RDD.count(RDD.scala:1012)
at org.apache.spark.examples.HBaseTest$.main(HBaseTest.scala:58)
at org.apache.spark.examples.HBaseTest.main(HBaseTest.scala)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at
org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:607)
at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:167)
at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:190)
at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:111)
at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)

On Thu, Apr 30, 2015 at 1:25 PM, Saurabh Gupta <[email protected]>
wrote:

> I am using hbase -0.94.8.
>
> On Wed, Apr 29, 2015 at 11:56 PM, Ted Yu <[email protected]> wrote:
>
>> Can you enable HBase DEBUG logging in log4j.properties so that we can
>> have more clue ?
>>
>> What hbase release are you using ?
>>
>> Cheers
>>
>> On Wed, Apr 29, 2015 at 4:27 AM, Saurabh Gupta <[email protected]>
>> wrote:
>>
>>> Hi,
>>>
>>> I am working with standalone HBase. And I want to execute
>>> HBaseTest.scala (in scala examples) .
>>>
>>> I have created a test table with three rows and I just want to get the
>>> count using HBaseTest.scala
>>>
>>> I am getting this issue:
>>>
>>> 15/04/29 11:17:10 INFO BlockManagerMaster: Registered BlockManager
>>> 15/04/29 11:17:11 INFO ZooKeeper: Client
>>> environment:zookeeper.version=3.4.5-1392090, built on 09/30/2012 17:52 GMT
>>> 15/04/29 11:17:11 INFO ZooKeeper: Client environment:host.name
>>> =ip-10-144-185-113
>>> 15/04/29 11:17:11 INFO ZooKeeper: Client
>>> environment:java.version=1.7.0_79
>>> 15/04/29 11:17:11 INFO ZooKeeper: Client environment:java.vendor=Oracle
>>> Corporation
>>> 15/04/29 11:17:11 INFO ZooKeeper: Client
>>> environment:java.home=/usr/lib/jvm/java-7-openjdk-amd64/jre
>>> 15/04/29 11:17:11 INFO ZooKeeper: Client
>>> environment:java.class.path=/home/ubuntu/sparkfolder/conf/:/home/ubuntu/sparkfolder/assembly/target/scala-2.10/spark-assembly-1.4.0-SNAPSHOT-hadoop2.2.0.jar:/home/ubuntu/sparkfolder/lib_managed/jars/datanucleus-core-3.2.10.jar:/home/ubuntu/sparkfolder/lib_managed/jars/datanucleus-api-jdo-3.2.6.jar:/home/ubuntu/sparkfolder/lib_managed/jars/datanucleus-rdbms-3.2.9.jar
>>> 15/04/29 11:17:11 INFO ZooKeeper: Client
>>> environment:java.library.path=/usr/java/packages/lib/amd64:/usr/lib/x86_64-linux-gnu/jni:/lib/x86_64-linux-gnu:/usr/lib/x86_64-linux-gnu:/usr/lib/jni:/lib:/usr/lib
>>> 15/04/29 11:17:11 INFO ZooKeeper: Client environment:java.io.tmpdir=/tmp
>>> 15/04/29 11:17:11 INFO ZooKeeper: Client environment:java.compiler=<NA>
>>> 15/04/29 11:17:11 INFO ZooKeeper: Client environment:os.name=Linux
>>> 15/04/29 11:17:11 INFO ZooKeeper: Client environment:os.arch=amd64
>>> 15/04/29 11:17:11 INFO ZooKeeper: Client
>>> environment:os.version=3.13.0-49-generic
>>> 15/04/29 11:17:11 INFO ZooKeeper: Client environment:user.name=root
>>> 15/04/29 11:17:11 INFO ZooKeeper: Client environment:user.home=/root
>>> 15/04/29 11:17:11 INFO ZooKeeper: Client
>>> environment:user.dir=/home/ubuntu/sparkfolder
>>> 15/04/29 11:17:11 INFO ZooKeeper: Initiating client connection,
>>> connectString=localhost:2181 sessionTimeout=90000
>>> watcher=hconnection-0x2711025f, quorum=localhost:2181, baseZNode=/hbase
>>> 15/04/29 11:17:11 INFO RecoverableZooKeeper: Process
>>> identifier=hconnection-0x2711025f connecting to ZooKeeper
>>> ensemble=localhost:2181
>>> 15/04/29 11:17:11 INFO ClientCnxn: Opening socket connection to server
>>> ip-10-144-185-113/10.144.185.113:2181. Will not attempt to authenticate
>>> using SASL (unknown error)
>>> 15/04/29 11:17:11 INFO ClientCnxn: Socket connection established to
>>> ip-10-144-185-113/10.144.185.113:2181, initiating session
>>> 15/04/29 11:17:11 INFO ClientCnxn: Session establishment complete on
>>> server ip-10-144-185-113/10.144.185.113:2181, sessionid =
>>> 0x14d04d506da0005, negotiated timeout = 40000
>>> 15/04/29 11:17:11 INFO ZooKeeperRegistry: ClusterId read in ZooKeeper is
>>> null
>>>
>>> Its just stuck Not showing any error. There is no Hadoop on my machine.
>>> What could be the issue?
>>>
>>> here is hbase-site.xml:
>>>
>>> <configuration>
>>>     <property>
>>>        <name>hbase.zookeeper.quorum</name>
>>>       <value>localhost</value>
>>>     </property>
>>>
>>>    <property>
>>>       <name>hbase.zookeeper.property.clientPort</name>
>>>       <value>2181</value>
>>>    </property>
>>>         <property>
>>>                 <name>zookeeper.znode.parent</name>
>>>                <value>/hbase</value>
>>>         </property>
>>> </configuration>
>>>
>>>
>>
>

Reply via email to