Looks like you have an incompatible hbase-default.xml in some place. You can use the following code to find the location of "hbase-default.xml"
println(Thread.currentThread().getContextClassLoader().getResource("hbase-default.xml")) Best Regards, Shixiong Zhu 2015-09-21 15:46 GMT+08:00 Siva <sbhavan...@gmail.com>: > Hi, > > I m seeing some strange error while inserting data from spark streaming to > hbase. > > I can able to write the data from spark (without streaming) to hbase > successfully, but when i use the same code to write dstream I m seeing the > below error. > > I tried setting the below parameters, still didnt help. Did any face the > similar issue? > > conf.set("hbase.defaults.for.version.skip", "true") > conf.set("hbase.defaults.for.version", "0.98.4.2.2.4.2-2-hadoop2") > > 15/09/20 22:39:10 ERROR Executor: Exception in task 0.0 in stage 14.0 (TID > 16) > java.lang.RuntimeException: hbase-default.xml file seems to be for and old > version of HBase (null), this version is 0.98.4.2.2.4.2-2-hadoop2 > at > org.apache.hadoop.hbase.HBaseConfiguration.checkDefaultsVersion(HBaseConfiguration.java:73) > at > org.apache.hadoop.hbase.HBaseConfiguration.addHbaseResources(HBaseConfiguration.java:105) > at > org.apache.hadoop.hbase.HBaseConfiguration.create(HBaseConfiguration.java:116) > at > org.apache.hadoop.hbase.HBaseConfiguration.create(HBaseConfiguration.java:125) > at > $line51.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$HBaseConn$.hbaseConnection(<console>:49) > at > $line52.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$TestHbaseSpark$$anonfun$run$1$$anonfun$apply$1.apply(<console>:73) > at > $line52.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$TestHbaseSpark$$anonfun$run$1$$anonfun$apply$1.apply(<console>:73) > at scala.collection.Iterator$class.foreach(Iterator.scala:727) > at > org.apache.spark.InterruptibleIterator.foreach(InterruptibleIterator.scala:28) > at org.apache.spark.rdd.RDD$$anonfun$foreach$1.apply(RDD.scala:782) > at org.apache.spark.rdd.RDD$$anonfun$foreach$1.apply(RDD.scala:782) > at > org.apache.spark.SparkContext$$anonfun$runJob$4.apply(SparkContext.scala:1353) > at > org.apache.spark.SparkContext$$anonfun$runJob$4.apply(SparkContext.scala:1353) > at > org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:61) > at org.apache.spark.scheduler.Task.run(Task.scala:56) > at > org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:200) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) > at java.lang.Thread.run(Thread.java:745) > 15/09/20 22:39:10 WARN TaskSetManager: Lost task 0.0 in stage 14.0 (TID > 16, localhost): java.lang.RuntimeException: hbase-default.xml file seems to > be for and old version of HBase (null), this version is > 0.98.4.2.2.4.2-2-hadoop2 > > > Thanks, > Siva. >