James,

Understood.
Thanks for the reply,
Dor

From: James Taylor [mailto:jamestay...@apache.org]
Sent: יום א 21 פברואר 2016 21:16
To: user
Subject: Re: Cloudera and Phoenix

Hi Dor,

Whether or not Phoenix becomes part of CDH is not under our control. It *is* 
under your control, though (assuming you're a customer of CDH). The *only* way 
Phoenix will transition from being in Cloudera Labs to being part of the 
official CDH distro is if you and other customers demand it.

Thanks,
James

On Sun, Feb 21, 2016 at 10:03 AM, Dor Ben Dov 
<dor.ben-...@amdocs.com<mailto:dor.ben-...@amdocs.com>> wrote:
Stephen

Any plans or do you or anyone where see the possibility that it will be 
although all below as official release ?

Dor

From: Stephen Wilcoxon [mailto:wilco...@gmail.com<mailto:wilco...@gmail.com>]
Sent: יום א 21 פברואר 2016 19:37
To: user@phoenix.apache.org<mailto:user@phoenix.apache.org>
Subject: Re: Cloudera and Phoenix

As of a few months ago, Cloudera includes Phoenix as a "lab" (basically beta) 
but it was out-of-date.  From what I gather, the official Phoenix releases will 
not run on Cloudera without modifications (someone was doing unofficial 
Phoenix/Cloudera releases but I'm not sure if they still are or not).

On Sun, Feb 21, 2016 at 6:39 AM, Dor Ben Dov 
<dor.ben-...@amdocs.com<mailto:dor.ben-...@amdocs.com>> wrote:
Hi All,

Do we have Phoenix release officially in Cloudera ? any plan to if not ?

Regards,

Dor ben Dov

From: Benjamin Kim [mailto:bbuil...@gmail.com<mailto:bbuil...@gmail.com>]
Sent: יום ו 19 פברואר 2016 19:41
To: user@phoenix.apache.org<mailto:user@phoenix.apache.org>
Subject: Re: Spark Phoenix Plugin

All,

Thanks for the help. I have switched out Cloudera’s HBase 1.0.0 with the 
current Apache HBase 1.1.3. Also, I installed Phoenix 4.7.0, and everything 
works fine except for the Phoenix Spark Plugin. I wonder if it’s a version 
incompatibility issue with Spark 1.6. Has anyone tried compiling 4.7.0 using 
Spark 1.6?

Thanks,
Ben

On Feb 12, 2016, at 6:33 AM, Benjamin Kim 
<bbuil...@gmail.com<mailto:bbuil...@gmail.com>> wrote:

Anyone know when Phoenix 4.7 will be officially released? And what Cloudera 
distribution versions will it be compatible with?

Thanks,
Ben

On Feb 10, 2016, at 11:03 AM, Benjamin Kim 
<bbuil...@gmail.com<mailto:bbuil...@gmail.com>> wrote:

Hi Pierre,

I am getting this error now.

Error: org.apache.phoenix.exception.PhoenixIOException: 
org.apache.hadoop.hbase.DoNotRetryIOException: 
SYSTEM.CATALOG,,1453397732623.8af7b44f3d7609eb301ad98641ff2611.: 
org.apache.hadoop.hbase.client.Delete.setAttribute(Ljava/lang/String;[B)Lorg/apache/hadoop/hbase/client/Delete;

I even tried to use sqlline.py to do some queries too. It resulted in the same 
error. I followed the installation instructions. Is there something missing?

Thanks,
Ben


On Feb 9, 2016, at 10:20 AM, Ravi Kiran 
<maghamraviki...@gmail.com<mailto:maghamraviki...@gmail.com>> wrote:

Hi Pierre,

  Try your luck for building the artifacts from 
https://github.com/chiastic-security/phoenix-for-cloudera. Hopefully it helps.

Regards
Ravi .

On Tue, Feb 9, 2016 at 10:04 AM, Benjamin Kim 
<bbuil...@gmail.com<mailto:bbuil...@gmail.com>> wrote:
Hi Pierre,

I found this article about how Cloudera’s version of HBase is very different 
than Apache HBase so it must be compiled using Cloudera’s repo and versions. 
But, I’m not having any success with it.

http://stackoverflow.com/questions/31849454/using-phoenix-with-cloudera-hbase-installed-from-repo

There’s also a Chinese site that does the same thing.

https://www.zybuluo.com/xtccc/note/205739

I keep getting errors like the one’s below.

[ERROR] 
/opt/tools/phoenix/phoenix-core/src/main/java/org/apache/hadoop/hbase/regionserver/LocalIndexMerger.java:[110,29]
 cannot find symbol
[ERROR] symbol:   class Region
[ERROR] location: class org.apache.hadoop.hbase.regionserver.LocalIndexMerger
…

Have you tried this also?

As a last resort, we will have to abandon Cloudera’s HBase for Apache’s HBase.

Thanks,
Ben


On Feb 8, 2016, at 11:04 PM, pierre lacave 
<pie...@lacave.me<mailto:pie...@lacave.me>> wrote:

Havent met that one.
According to SPARK-1867, the real issue is hidden.
I d process by elimination, maybe try in local[*] mode first
https://issues.apache.org/jira/plugins/servlet/mobile#issue/SPARK-1867

On Tue, 9 Feb 2016, 04:58 Benjamin Kim 
<bbuil...@gmail.com<mailto:bbuil...@gmail.com>> wrote:
Pierre,

I got it to work using phoenix-4.7.0-HBase-1.0-client-spark.jar. But, now, I 
get this error:

org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in 
stage 0.0 failed 4 times, most recent failure: Lost task 0.3 in stage 0.0 (TID 
3, 
prod-dc1-datanode151.pdc1i.gradientx.com<http://prod-dc1-datanode151.pdc1i.gradientx.com/>):
 java.lang.IllegalStateException: unread block data

It happens when I do:

df.show()

Getting closer…

Thanks,
Ben



On Feb 8, 2016, at 2:57 PM, pierre lacave 
<pie...@lacave.me<mailto:pie...@lacave.me>> wrote:

This is the wrong client jar try with the one named 
phoenix-4.7.0-HBase-1.1-client-spark.jar

On Mon, 8 Feb 2016, 22:29 Benjamin Kim 
<bbuil...@gmail.com<mailto:bbuil...@gmail.com>> wrote:
Hi Josh,

I tried again by putting the settings within the spark-default.conf.

spark.driver.extraClassPath=/opt/tools/phoenix/phoenix-4.7.0-HBase-1.0-client.jar
spark.executor.extraClassPath=/opt/tools/phoenix/phoenix-4.7.0-HBase-1.0-client.jar

I still get the same error using the code below.

import org.apache.phoenix.spark._
val df = sqlContext.load("org.apache.phoenix.spark", Map("table" -> 
"TEST.MY_TEST", "zkUrl" -> “zk1,zk2,zk3:2181"))

Can you tell me what else you’re doing?

Thanks,
Ben


On Feb 8, 2016, at 1:44 PM, Josh Mahonin 
<jmaho...@gmail.com<mailto:jmaho...@gmail.com>> wrote:

Hi Ben,

I'm not sure about the format of those command line options you're passing. 
I've had success with spark-shell just by setting the 
'spark.executor.extraClassPath' and 'spark.driver.extraClassPath' options on 
the spark config, as per the docs [1].

I'm not sure if there's anything special needed for CDH or not though. I also 
have a docker image I've been toying with which has a working Spark/Phoenix 
setup using the Phoenix 4.7.0 RC and Spark 1.6.0. It might be a useful 
reference for you as well [2].

Good luck,

Josh

[1] https://phoenix.apache.org/phoenix_spark.html
[2] https://github.com/jmahonin/docker-phoenix/tree/phoenix_spark

On Mon, Feb 8, 2016 at 4:29 PM, Benjamin Kim 
<bbuil...@gmail.com<mailto:bbuil...@gmail.com>> wrote:
Hi Pierre,

I tried to run in spark-shell using spark 1.6.0 by running this:

spark-shell --master yarn-client --driver-class-path 
/opt/tools/phoenix/phoenix-4.7.0-HBase-1.0-client.jar --driver-java-options 
"-Dspark.executor.extraClassPath=/opt/tools/phoenix/phoenix-4.7.0-HBase-1.0-client.jar”

The version of HBase is the one in CDH5.4.8, which is 1.0.0-cdh5.4.8.

When I get to the line:

val df = sqlContext.load("org.apache.phoenix.spark", Map("table" -> 
“TEST.MY_TEST", "zkUrl" -> “zk1,zk2,zk3:2181”))

I get this error:

java.lang.NoClassDefFoundError: Could not initialize class 
org.apache.spark.rdd.RDDOperationScope$

Any ideas?

Thanks,
Ben


On Feb 5, 2016, at 1:36 PM, pierre lacave 
<pie...@lacave.me<mailto:pie...@lacave.me>> wrote:

I don't know when the full release will be, RC1 just got pulled out, and 
expecting RC2 soon

you can find them here

https://dist.apache.org/repos/dist/dev/phoenix/


there is a new phoenix-4.7.0-HBase-1.1-client-spark.jar that is all you need to 
have in spark classpath


Pierre Lacave
171 Skellig House, Custom House, Lower Mayor street, Dublin 1, Ireland
Phone :       +353879128708<tel:%2B353879128708>

On Fri, Feb 5, 2016 at 9:28 PM, Benjamin Kim 
<bbuil...@gmail.com<mailto:bbuil...@gmail.com>> wrote:
Hi Pierre,

When will I be able to download this version?

Thanks,
Ben


On Friday, February 5, 2016, pierre lacave 
<pie...@lacave.me<mailto:pie...@lacave.me>> wrote:
This was addressed in Phoenix 4.7 (currently in RC)
https://issues.apache.org/jira/browse/PHOENIX-2503




Pierre Lacave
171 Skellig House, Custom House, Lower Mayor street, Dublin 1, Ireland
Phone :       +353879128708<tel:%2B353879128708>

On Fri, Feb 5, 2016 at 6:17 PM, Benjamin Kim 
<bbuil...@gmail.com<mailto:bbuil...@gmail.com>> wrote:
I cannot get this plugin to work in CDH 5.4.8 using Phoenix 4.5.2 and Spark 
1.6. When I try to launch spark-shell, I get:

        java.lang.RuntimeException: java.lang.RuntimeException: Unable to 
instantiate org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient

I continue on and run the example code. When I get tot the line below:

        val df = sqlContext.load("org.apache.phoenix.spark", Map("table" -> 
"TEST.MY_TEST", "zkUrl" -> "zookeeper1,zookeeper2,zookeeper3:2181")

I get this error:

        java.lang.NoSuchMethodError: 
com.fasterxml.jackson.module.scala.deser.BigDecimalDeserializer$.handledType()Ljava/lang/Class;

Can someone help?

Thanks,
Ben











This message and the information contained herein is proprietary and 
confidential and subject to the Amdocs policy statement, you may review at 
http://www.amdocs.com/email_disclaimer.asp


Reply via email to