Anyone know when Phoenix 4.7 will be officially released? And what Cloudera distribution versions will it be compatible with?
Thanks, Ben > On Feb 10, 2016, at 11:03 AM, Benjamin Kim <bbuil...@gmail.com> wrote: > > Hi Pierre, > > I am getting this error now. > > Error: org.apache.phoenix.exception.PhoenixIOException: > org.apache.hadoop.hbase.DoNotRetryIOException: > SYSTEM.CATALOG,,1453397732623.8af7b44f3d7609eb301ad98641ff2611.: > org.apache.hadoop.hbase.client.Delete.setAttribute(Ljava/lang/String;[B)Lorg/apache/hadoop/hbase/client/Delete; > > I even tried to use sqlline.py to do some queries too. It resulted in the > same error. I followed the installation instructions. Is there something > missing? > > Thanks, > Ben > > >> On Feb 9, 2016, at 10:20 AM, Ravi Kiran <maghamraviki...@gmail.com >> <mailto:maghamraviki...@gmail.com>> wrote: >> >> Hi Pierre, >> >> Try your luck for building the artifacts from >> https://github.com/chiastic-security/phoenix-for-cloudera >> <https://github.com/chiastic-security/phoenix-for-cloudera>. Hopefully it >> helps. >> >> Regards >> Ravi . >> >> On Tue, Feb 9, 2016 at 10:04 AM, Benjamin Kim <bbuil...@gmail.com >> <mailto:bbuil...@gmail.com>> wrote: >> Hi Pierre, >> >> I found this article about how Cloudera’s version of HBase is very different >> than Apache HBase so it must be compiled using Cloudera’s repo and versions. >> But, I’m not having any success with it. >> >> http://stackoverflow.com/questions/31849454/using-phoenix-with-cloudera-hbase-installed-from-repo >> >> <http://stackoverflow.com/questions/31849454/using-phoenix-with-cloudera-hbase-installed-from-repo> >> >> There’s also a Chinese site that does the same thing. >> >> https://www.zybuluo.com/xtccc/note/205739 >> <https://www.zybuluo.com/xtccc/note/205739> >> >> I keep getting errors like the one’s below. >> >> [ERROR] >> /opt/tools/phoenix/phoenix-core/src/main/java/org/apache/hadoop/hbase/regionserver/LocalIndexMerger.java:[110,29] >> cannot find symbol >> [ERROR] symbol: class Region >> [ERROR] location: class org.apache.hadoop.hbase.regionserver.LocalIndexMerger >> … >> >> Have you tried this also? >> >> As a last resort, we will have to abandon Cloudera’s HBase for Apache’s >> HBase. >> >> Thanks, >> Ben >> >> >>> On Feb 8, 2016, at 11:04 PM, pierre lacave <pie...@lacave.me >>> <mailto:pie...@lacave.me>> wrote: >>> >>> Havent met that one. >>> >>> According to SPARK-1867, the real issue is hidden. >>> >>> I d process by elimination, maybe try in local[*] mode first >>> >>> https://issues.apache.org/jira/plugins/servlet/mobile#issue/SPARK-1867 >>> <https://issues.apache.org/jira/plugins/servlet/mobile#issue/SPARK-1867> >>> On Tue, 9 Feb 2016, 04:58 Benjamin Kim <bbuil...@gmail.com >>> <mailto:bbuil...@gmail.com>> wrote: >>> Pierre, >>> >>> I got it to work using phoenix-4.7.0-HBase-1.0-client-spark.jar. But, now, >>> I get this error: >>> >>> org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 >>> in stage 0.0 failed 4 times, most recent failure: Lost task 0.3 in stage >>> 0.0 (TID 3, prod-dc1-datanode151.pdc1i.gradientx.com >>> <http://prod-dc1-datanode151.pdc1i.gradientx.com/>): >>> java.lang.IllegalStateException: unread block data >>> >>> It happens when I do: >>> >>> df.show() >>> >>> Getting closer… >>> >>> Thanks, >>> Ben >>> >>> >>> >>>> On Feb 8, 2016, at 2:57 PM, pierre lacave <pie...@lacave.me >>>> <mailto:pie...@lacave.me>> wrote: >>>> >>>> This is the wrong client jar try with the one named >>>> phoenix-4.7.0-HBase-1.1-client-spark.jar >>>> >>>> >>>> On Mon, 8 Feb 2016, 22:29 Benjamin Kim <bbuil...@gmail.com >>>> <mailto:bbuil...@gmail.com>> wrote: >>>> Hi Josh, >>>> >>>> I tried again by putting the settings within the spark-default.conf. >>>> >>>> spark.driver.extraClassPath=/opt/tools/phoenix/phoenix-4.7.0-HBase-1.0-client.jar >>>> spark.executor.extraClassPath=/opt/tools/phoenix/phoenix-4.7.0-HBase-1.0-client.jar >>>> >>>> I still get the same error using the code below. >>>> >>>> import org.apache.phoenix.spark._ >>>> val df = sqlContext.load("org.apache.phoenix.spark", Map("table" -> >>>> "TEST.MY_TEST", "zkUrl" -> “zk1,zk2,zk3:2181")) >>>> >>>> Can you tell me what else you’re doing? >>>> >>>> Thanks, >>>> Ben >>>> >>>> >>>>> On Feb 8, 2016, at 1:44 PM, Josh Mahonin <jmaho...@gmail.com >>>>> <mailto:jmaho...@gmail.com>> wrote: >>>>> >>>>> Hi Ben, >>>>> >>>>> I'm not sure about the format of those command line options you're >>>>> passing. I've had success with spark-shell just by setting the >>>>> 'spark.executor.extraClassPath' and 'spark.driver.extraClassPath' options >>>>> on the spark config, as per the docs [1]. >>>>> >>>>> I'm not sure if there's anything special needed for CDH or not though. I >>>>> also have a docker image I've been toying with which has a working >>>>> Spark/Phoenix setup using the Phoenix 4.7.0 RC and Spark 1.6.0. It might >>>>> be a useful reference for you as well [2]. >>>>> >>>>> Good luck, >>>>> >>>>> Josh >>>>> >>>>> [1] https://phoenix.apache.org/phoenix_spark.html >>>>> <https://phoenix.apache.org/phoenix_spark.html> >>>>> [2] https://github.com/jmahonin/docker-phoenix/tree/phoenix_spark >>>>> <https://github.com/jmahonin/docker-phoenix/tree/phoenix_spark> >>>>> >>>>> On Mon, Feb 8, 2016 at 4:29 PM, Benjamin Kim <bbuil...@gmail.com >>>>> <mailto:bbuil...@gmail.com>> wrote: >>>>> Hi Pierre, >>>>> >>>>> I tried to run in spark-shell using spark 1.6.0 by running this: >>>>> >>>>> spark-shell --master yarn-client --driver-class-path >>>>> /opt/tools/phoenix/phoenix-4.7.0-HBase-1.0-client.jar >>>>> --driver-java-options >>>>> "-Dspark.executor.extraClassPath=/opt/tools/phoenix/phoenix-4.7.0-HBase-1.0-client.jar” >>>>> >>>>> The version of HBase is the one in CDH5.4.8, which is 1.0.0-cdh5.4.8. >>>>> >>>>> When I get to the line: >>>>> >>>>> val df = sqlContext.load("org.apache.phoenix.spark", Map("table" -> >>>>> “TEST.MY_TEST", "zkUrl" -> “zk1,zk2,zk3:2181”)) >>>>> >>>>> I get this error: >>>>> >>>>> java.lang.NoClassDefFoundError: Could not initialize class >>>>> org.apache.spark.rdd.RDDOperationScope$ >>>>> >>>>> Any ideas? >>>>> >>>>> Thanks, >>>>> Ben >>>>> >>>>> >>>>>> On Feb 5, 2016, at 1:36 PM, pierre lacave <pie...@lacave.me >>>>>> <mailto:pie...@lacave.me>> wrote: >>>>>> >>>>>> I don't know when the full release will be, RC1 just got pulled out, and >>>>>> expecting RC2 soon >>>>>> >>>>>> you can find them here >>>>>> >>>>>> https://dist.apache.org/repos/dist/dev/phoenix/ >>>>>> <https://dist.apache.org/repos/dist/dev/phoenix/> >>>>>> >>>>>> >>>>>> there is a new phoenix-4.7.0-HBase-1.1-client-spark.jar that is all you >>>>>> need to have in spark classpath >>>>>> >>>>>> >>>>>> Pierre Lacave >>>>>> 171 Skellig House, Custom House, Lower Mayor street, Dublin 1, Ireland >>>>>> Phone : +353879128708 <tel:%2B353879128708> >>>>>> >>>>>> On Fri, Feb 5, 2016 at 9:28 PM, Benjamin Kim <bbuil...@gmail.com >>>>>> <mailto:bbuil...@gmail.com>> wrote: >>>>>> Hi Pierre, >>>>>> >>>>>> When will I be able to download this version? >>>>>> >>>>>> Thanks, >>>>>> Ben >>>>>> >>>>>> >>>>>> On Friday, February 5, 2016, pierre lacave <pie...@lacave.me >>>>>> <mailto:pie...@lacave.me>> wrote: >>>>>> This was addressed in Phoenix 4.7 (currently in RC) >>>>>> https://issues.apache.org/jira/browse/PHOENIX-2503 >>>>>> <https://issues.apache.org/jira/browse/PHOENIX-2503> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> Pierre Lacave >>>>>> 171 Skellig House, Custom House, Lower Mayor street, Dublin 1, Ireland >>>>>> Phone : +353879128708 <tel:%2B353879128708> >>>>>> >>>>>> On Fri, Feb 5, 2016 at 6:17 PM, Benjamin Kim <bbuil...@gmail.com <>> >>>>>> wrote: >>>>>> I cannot get this plugin to work in CDH 5.4.8 using Phoenix 4.5.2 and >>>>>> Spark 1.6. When I try to launch spark-shell, I get: >>>>>> >>>>>> java.lang.RuntimeException: java.lang.RuntimeException: Unable >>>>>> to instantiate >>>>>> org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient >>>>>> >>>>>> I continue on and run the example code. When I get tot the line below: >>>>>> >>>>>> val df = sqlContext.load("org.apache.phoenix.spark", Map("table" >>>>>> -> "TEST.MY_TEST", "zkUrl" -> "zookeeper1,zookeeper2,zookeeper3:2181") >>>>>> >>>>>> I get this error: >>>>>> >>>>>> java.lang.NoSuchMethodError: >>>>>> com.fasterxml.jackson.module.scala.deser.BigDecimalDeserializer$.handledType()Ljava/lang/Class; >>>>>> >>>>>> Can someone help? >>>>>> >>>>>> Thanks, >>>>>> Ben >>>>>> >>>>>> >>>>> >>>>> >>>> >>> >> >> >