Hi Pierre, Try your luck for building the artifacts from https://github.com/chiastic-security/phoenix-for-cloudera. Hopefully it helps.
Regards Ravi . On Tue, Feb 9, 2016 at 10:04 AM, Benjamin Kim <bbuil...@gmail.com> wrote: > Hi Pierre, > > I found this article about how Cloudera’s version of HBase is very > different than Apache HBase so it must be compiled using Cloudera’s repo > and versions. But, I’m not having any success with it. > > > http://stackoverflow.com/questions/31849454/using-phoenix-with-cloudera-hbase-installed-from-repo > > There’s also a Chinese site that does the same thing. > > https://www.zybuluo.com/xtccc/note/205739 > > I keep getting errors like the one’s below. > > [ERROR] > /opt/tools/phoenix/phoenix-core/src/main/java/org/apache/hadoop/hbase/regionserver/LocalIndexMerger.java:[110,29] > cannot find symbol > [ERROR] symbol: class Region > [ERROR] location: class > org.apache.hadoop.hbase.regionserver.LocalIndexMerger > … > > Have you tried this also? > > As a last resort, we will have to abandon Cloudera’s HBase for Apache’s > HBase. > > Thanks, > Ben > > > On Feb 8, 2016, at 11:04 PM, pierre lacave <pie...@lacave.me> wrote: > > Havent met that one. > > According to SPARK-1867, the real issue is hidden. > > I d process by elimination, maybe try in local[*] mode first > > https://issues.apache.org/jira/plugins/servlet/mobile#issue/SPARK-1867 > > On Tue, 9 Feb 2016, 04:58 Benjamin Kim <bbuil...@gmail.com> wrote: > >> Pierre, >> >> I got it to work using phoenix-4.7.0-HBase-1.0-client-spark.jar. But, >> now, I get this error: >> >> org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 >> in stage 0.0 failed 4 times, most recent failure: Lost task 0.3 in stage >> 0.0 (TID 3, prod-dc1-datanode151.pdc1i.gradientx.com): >> java.lang.IllegalStateException: unread block data >> >> It happens when I do: >> >> df.show() >> >> Getting closer… >> >> Thanks, >> Ben >> >> >> >> On Feb 8, 2016, at 2:57 PM, pierre lacave <pie...@lacave.me> wrote: >> >> This is the wrong client jar try with the one named >> phoenix-4.7.0-HBase-1.1-client-spark.jar >> >> On Mon, 8 Feb 2016, 22:29 Benjamin Kim <bbuil...@gmail.com> wrote: >> >>> Hi Josh, >>> >>> I tried again by putting the settings within the spark-default.conf. >>> >>> >>> spark.driver.extraClassPath=/opt/tools/phoenix/phoenix-4.7.0-HBase-1.0-client.jar >>> >>> spark.executor.extraClassPath=/opt/tools/phoenix/phoenix-4.7.0-HBase-1.0-client.jar >>> >>> I still get the same error using the code below. >>> >>> import org.apache.phoenix.spark._ >>> val df = sqlContext.load("org.apache.phoenix.spark", Map("table" -> >>> "TEST.MY_TEST", "zkUrl" -> “zk1,zk2,zk3:2181")) >>> >>> Can you tell me what else you’re doing? >>> >>> Thanks, >>> Ben >>> >>> >>> On Feb 8, 2016, at 1:44 PM, Josh Mahonin <jmaho...@gmail.com> wrote: >>> >>> Hi Ben, >>> >>> I'm not sure about the format of those command line options you're >>> passing. I've had success with spark-shell just by setting the >>> 'spark.executor.extraClassPath' and 'spark.driver.extraClassPath' options >>> on the spark config, as per the docs [1]. >>> >>> I'm not sure if there's anything special needed for CDH or not though. I >>> also have a docker image I've been toying with which has a working >>> Spark/Phoenix setup using the Phoenix 4.7.0 RC and Spark 1.6.0. It might be >>> a useful reference for you as well [2]. >>> >>> Good luck, >>> >>> Josh >>> >>> [1] https://phoenix.apache.org/phoenix_spark.html >>> [2] https://github.com/jmahonin/docker-phoenix/tree/phoenix_spark >>> >>> On Mon, Feb 8, 2016 at 4:29 PM, Benjamin Kim <bbuil...@gmail.com> wrote: >>> >>>> Hi Pierre, >>>> >>>> I tried to run in spark-shell using spark 1.6.0 by running this: >>>> >>>> spark-shell --master yarn-client --driver-class-path >>>> /opt/tools/phoenix/phoenix-4.7.0-HBase-1.0-client.jar --driver-java-options >>>> "-Dspark.executor.extraClassPath=/opt/tools/phoenix/phoenix-4.7.0-HBase-1.0-client.jar” >>>> >>>> The version of HBase is the one in CDH5.4.8, which is 1.0.0-cdh5.4.8. >>>> >>>> When I get to the line: >>>> >>>> val df = sqlContext.load("org.apache.phoenix.spark", Map("table" -> >>>> “TEST.MY_TEST", "zkUrl" -> “zk1,zk2,zk3:2181”)) >>>> >>>> I get this error: >>>> >>>> java.lang.NoClassDefFoundError: Could not initialize class >>>> org.apache.spark.rdd.RDDOperationScope$ >>>> >>>> Any ideas? >>>> >>>> Thanks, >>>> Ben >>>> >>>> >>>> On Feb 5, 2016, at 1:36 PM, pierre lacave <pie...@lacave.me> wrote: >>>> >>>> I don't know when the full release will be, RC1 just got pulled out, >>>> and expecting RC2 soon >>>> >>>> you can find them here >>>> >>>> https://dist.apache.org/repos/dist/dev/phoenix/ >>>> >>>> >>>> there is a new phoenix-4.7.0-HBase-1.1-client-spark.jar that is all you >>>> need to have in spark classpath >>>> >>>> >>>> *Pierre Lacave* >>>> 171 Skellig House, Custom House, Lower Mayor street, Dublin 1, Ireland >>>> Phone : +353879128708 >>>> >>>> On Fri, Feb 5, 2016 at 9:28 PM, Benjamin Kim <bbuil...@gmail.com> >>>> wrote: >>>> >>>>> Hi Pierre, >>>>> >>>>> When will I be able to download this version? >>>>> >>>>> Thanks, >>>>> Ben >>>>> >>>>> >>>>> On Friday, February 5, 2016, pierre lacave <pie...@lacave.me> wrote: >>>>> >>>>>> This was addressed in Phoenix 4.7 (currently in RC) >>>>>> https://issues.apache.org/jira/browse/PHOENIX-2503 >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> *Pierre Lacave* >>>>>> 171 Skellig House, Custom House, Lower Mayor street, Dublin 1, Ireland >>>>>> Phone : +353879128708 >>>>>> >>>>>> On Fri, Feb 5, 2016 at 6:17 PM, Benjamin Kim <bbuil...@gmail.com> >>>>>> wrote: >>>>>> >>>>>>> I cannot get this plugin to work in CDH 5.4.8 using Phoenix 4.5.2 >>>>>>> and Spark 1.6. When I try to launch spark-shell, I get: >>>>>>> >>>>>>> java.lang.RuntimeException: java.lang.RuntimeException: >>>>>>> Unable to instantiate >>>>>>> org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient >>>>>>> >>>>>>> I continue on and run the example code. When I get tot the line >>>>>>> below: >>>>>>> >>>>>>> val df = sqlContext.load("org.apache.phoenix.spark", >>>>>>> Map("table" -> "TEST.MY_TEST", "zkUrl" -> >>>>>>> "zookeeper1,zookeeper2,zookeeper3:2181") >>>>>>> >>>>>>> I get this error: >>>>>>> >>>>>>> java.lang.NoSuchMethodError: >>>>>>> com.fasterxml.jackson.module.scala.deser.BigDecimalDeserializer$.handledType()Ljava/lang/Class; >>>>>>> >>>>>>> Can someone help? >>>>>>> >>>>>>> Thanks, >>>>>>> Ben >>>>>> >>>>>> >>>>>> >>>> >>>> >>> >>> >> >