Have you verified that this class is in the fat jar? It looks that it misses 
some of the Hbase libraries ... 

> On 21 Oct 2016, at 11:45, Mich Talebzadeh <mich.talebza...@gmail.com> wrote:
> 
> Still does not work with Spark 2.0.0 on apache-phoenix-4.8.1-HBase-1.2-bin
> 
> thanks
> 
> Dr Mich Talebzadeh
>  
> LinkedIn  
> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw
>  
> http://talebzadehmich.wordpress.com
> 
> Disclaimer: Use it at your own risk. Any and all responsibility for any loss, 
> damage or destruction of data or any other property which may arise from 
> relying on this email's technical content is explicitly disclaimed. The 
> author will in no case be liable for any monetary damages arising from such 
> loss, damage or destruction.
>  
> 
>> On 7 October 2016 at 09:27, Mich Talebzadeh <mich.talebza...@gmail.com> 
>> wrote:
>> Hi,
>> 
>> my code is trying to load a phoenix table built on an Hbase table.
>> 
>> import org.apache.spark.SparkContext
>> import org.apache.spark.SparkConf
>> import org.apache.hadoop.conf.Configuration
>> import org.apache.hadoop.hbase.HBaseConfiguration
>> import org.apache.hadoop.hbase.HColumnDescriptor
>> import org.apache.hadoop.hbase.HTableDescriptor
>> import org.apache.hadoop.hbase.{ HBaseConfiguration, HColumnDescriptor, 
>> HTableDescriptor }
>> import org.apache.hadoop.mapred.JobConf
>> import org.apache.hadoop.hbase.client.HBaseAdmin
>> import org.apache.spark.sql.types._
>> import org.apache.phoenix.spark._
>> 
>> 
>> The code line is from https://phoenix.apache.org/phoenix_spark.html
>> 
>> scala> val HiveContext = new org.apache.spark.sql.hive.HiveContext(sc)
>> warning: there was one deprecation warning; re-run with -deprecation for 
>> details
>> HiveContext: org.apache.spark.sql.hive.HiveContext = 
>> org.apache.spark.sql.hive.HiveContext@533e8807
>> 
>> scala> val df = HiveContext.load(
>>      | "org.apache.phoenix.spark",
>>      | Map("table" -> "temptable", "zkUrl" -> "rhes564:2181")
>>      |   )
>> 
>> warning: there was one deprecation warning; re-run with -deprecation for 
>> details
>> java.lang.NoClassDefFoundError: org/apache/hadoop/hbase/HBaseConfiguration
>>   at 
>> org.apache.phoenix.spark.PhoenixRDD.getPhoenixConfiguration(PhoenixRDD.scala:71)
>>   at 
>> org.apache.phoenix.spark.PhoenixRDD.phoenixConf$lzycompute(PhoenixRDD.scala:39)
>>   at org.apache.phoenix.spark.PhoenixRDD.phoenixConf(PhoenixRDD.scala:38)
>>   at org.apache.phoenix.spark.PhoenixRDD.<init>(PhoenixRDD.scala:42)
>>   at 
>> org.apache.phoenix.spark.PhoenixRelation.schema(PhoenixRelation.scala:50)
>>   at 
>> org.apache.spark.sql.execution.datasources.LogicalRelation.<init>(LogicalRelation.scala:40)
>>   at 
>> org.apache.spark.sql.SparkSession.baseRelationToDataFrame(SparkSession.scala:382)
>>   at org.apache.spark.sql.DataFrameReader.load(DataFrameReader.scala:143)
>>   at org.apache.spark.sql.DataFrameReader.load(DataFrameReader.scala:122)
>>   at org.apache.spark.sql.SQLContext.load(SQLContext.scala:958)
>>   ... 55 elided
>> I tried this as a fat jar file building it with Maven but I still get the 
>> same error.
>> 
>> Also the original code looks like this
>> 
>> val df = sqlContext.load(
>>   "org.apache.phoenix.spark",
>>   Map("table" -> "TABLE1", "zkUrl" -> "phoenix-server:2181")
>> )
>> 
>> Thanks
>> 
>> 
>> 
>> 
>> 
>> 
>> 
>> 
>> 
>> 
>> 
>> 
>> 
>> Dr Mich Talebzadeh
>>  
>> LinkedIn  
>> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw
>>  
>> http://talebzadehmich.wordpress.com
>> 
>> Disclaimer: Use it at your own risk. Any and all responsibility for any 
>> loss, damage or destruction of data or any other property which may arise 
>> from relying on this email's technical content is explicitly disclaimed. The 
>> author will in no case be liable for any monetary damages arising from such 
>> loss, damage or destruction.
>>  
> 

Reply via email to