Hi ,
   With both PHOENIX-1184 , PHOENIX-1183 tickets fixed, I happened to do a
quick test of Pig scripts on a CDH 5.1.0 backed cluster and things are
looking good. Below are the steps I followed,
a) Downloaded the binaries from
https://dist.apache.org/repos/dist/dev/phoenix/phoenix-4.1.0-rc1/bin/
b) Copied phoenix-4.1.0-server-hadoop2.jar to HBase Region Server lib path
and restarted.
c) Copied phoenix-4.1.0-client-hadoop2.jar and phoenix-pig-4.1.0-hadoop2.jar
on to the gateway node where I planned to run my Pig scripts.
d) Registered the following jars within the Pig script
       zookeeper.jar
       hbase-hadoop2-compat.jar
       hbase-client.jar
       hbase-protocol-0.98.1-cdh5.1.0.jar
       phoenix-4.1.0-client-hadoop2.jar
       phoenix-pig-4.1.0-hadoop2.jar


Regards
Ravi



On Tue, Aug 19, 2014 at 5:43 PM, Russell Jurney <russell.jur...@gmail.com>
wrote:

> I agree the vendor should resolve these issues. Hortonworks has already
> included Phoenix in HDP. Cloudera is behind the curve here. I'm told
> they'll include Phoenix when they feel they can support it well.
>
> That being said, wouldn't adding CDH/HDP options in pom.xml make the
> project easier to use, and result in more people trying to use CDH/HDP with
> Phoenix (and more peiople using Phoenix in general), which would bring up
> bugs like the ones here? Ideally the vendors would fix these JIRAs. That
> would seem to be a good thing.
>
> ᐧ
>
>
> On Tue, Aug 19, 2014 at 5:34 PM, Andrew Purtell <apurt...@apache.org>
> wrote:
>
>> Maybe "pick on" didn't get close enough to what I was after.
>>
>> > Maybe this is something I can fix. If I were to add the
>> cloudera/hortonworks maven repos, and then add some supported options for
>> hadoop beyond 1/2, that would pretty much do it, right?
>>
>> I doubt it, because v4 and master branches probably won't compile against
>> either, certainly not against CDH 5.1, their HBase releases are out of step
>> and stale with respect to the latest Apache HBase 0.98 and Apache Phoenix 4
>> releases.
>>
>> Getting back to my point, it's unfair in my opinion to expect the
>> upstream volunteer Apache projects to track all of the commercial options
>> and the vagaries of their arbitrary code freezes and "curated" additional
>> patches. It's unfair to expect Salesforce to fund such an effort, unless
>> Salesforce has somehow gone into the Hadoop distribution business.
>> Certainly I am not speaking on behalf of Salesforce or anyone else here. On
>> the other hand, I think it would be totally reasonable to request your
>> favorite vendor address Phoenix related issues with *their* derivative
>> distributions.
>>
>>
>>
>> On Tue, Aug 19, 2014 at 3:53 PM, Russell Jurney <russell.jur...@gmail.com
>> > wrote:
>>
>>> Maybe this is something I can fix. If I were to add the
>>> cloudera/hortonworks maven repos, and then add some supported options for
>>> hadoop beyond 1/2, that would pretty much do it, right?
>>> ᐧ
>>>
>>>
>>> On Tue, Aug 19, 2014 at 3:49 PM, Jesse Yates <jesse.k.ya...@gmail.com>
>>> wrote:
>>>
>>>> FWIW internally at Salesforce we also patch the HBase and Hadoop poms
>>>> to support our own internal 'light forks'. Its really not a big deal to
>>>> manage - a couple of jenkins jobs (one to automate, one to track open
>>>> source changes and ensure your patch(es) still work, etc) and you are good
>>>> to go.
>>>>
>>>> I imagine this is also what various distributors are doing for their
>>>> forks as well.
>>>>
>>>> -------------------
>>>> Jesse Yates
>>>> @jesse_yates
>>>> jyates.github.com
>>>>
>>>>
>>>> On Tue, Aug 19, 2014 at 3:36 PM, Russell Jurney <
>>>> russell.jur...@gmail.com> wrote:
>>>>
>>>>> First of all, I apologize if you feel like I was picking on you. I was
>>>>> not trying to do that.
>>>>>
>>>>> My understanding is that Salesforce pays people to work on Phoenix. Is
>>>>> that not the case? I'm hoping one of them will add spark-like support for
>>>>> CDH and HDP to advance the project.
>>>>>
>>>>> And I don't mention the POM thing to pick on someone. The majority of
>>>>> HBase users are not going to be able to use Phoenix because they run a
>>>>> commercial distribution of Hadoop and aren't pom wizards. That seems kind
>>>>> of important for the well being of the project.
>>>>> ᐧ
>>>>>
>>>>>
>>>>> On Tue, Aug 19, 2014 at 3:26 PM, Andrew Purtell <apurt...@apache.org>
>>>>> wrote:
>>>>>
>>>>>> I don't think an Apache project should spend precious bandwidth
>>>>>> tracking the various and sundry redistributors of Apache ecosystem
>>>>>> projects. This is putting the cart before the horse. The horse is the
>>>>>> Apache upstream projects. The cart is the commercial distributions
>>>>>> leveraging the Apache ecosystem for profit. Spark is not a good example, 
>>>>>> it
>>>>>> is supported by a commercial concern, Databricks. What commercial company
>>>>>> supports Phoenix? Why do you think it is appropriate to pick on 
>>>>>> volunteers
>>>>>> because editing POM files is too much work?
>>>>>>
>>>>>>
>>>>>> On Tue, Aug 19, 2014 at 3:09 PM, Russell Jurney <
>>>>>> russell.jur...@gmail.com> wrote:
>>>>>>
>>>>>>> I also created https://issues.apache.org/jira/browse/PHOENIX-1185
>>>>>>> because requiring users to hand-edit the pom.xml just to build against 
>>>>>>> CDH
>>>>>>> and HDP is nuts.
>>>>>>> ᐧ
>>>>>>>
>>>>>>>
>>>>>>> On Tue, Aug 19, 2014 at 3:03 PM, Russell Jurney <
>>>>>>> russell.jur...@gmail.com> wrote:
>>>>>>>
>>>>>>>> I have created https://issues.apache.org/jira/browse/PHOENIX-1183
>>>>>>>>  and https://issues.apache.org/jira/browse/PHOENIX-1184
>>>>>>>>
>>>>>>>> I hope this can be resolved soon, we are pretty screwed right now.
>>>>>>>> ᐧ
>>>>>>>>
>>>>>>>>
>>>>>>>> On Tue, Aug 19, 2014 at 2:08 PM, Jesse Yates <
>>>>>>>> jesse.k.ya...@gmail.com> wrote:
>>>>>>>>
>>>>>>>>> Yup, that looks like an issue to me :-/
>>>>>>>>>
>>>>>>>>> -------------------
>>>>>>>>> Jesse Yates
>>>>>>>>> @jesse_yates
>>>>>>>>> jyates.github.com
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Tue, Aug 19, 2014 at 2:06 PM, Russell Jurney <
>>>>>>>>> russell.jur...@gmail.com> wrote:
>>>>>>>>>
>>>>>>>>>> Running against any version would be ok, but it does not work. I
>>>>>>>>>> get this error:
>>>>>>>>>>
>>>>>>>>>> 2014-08-19 14:03:46,904 FATAL org.apache.hadoop.mapred.Child: Error 
>>>>>>>>>> running child : java.lang.IncompatibleClassChangeError: Found 
>>>>>>>>>> interface org.apache.hadoop.mapreduce.TaskAttemptContext, but class 
>>>>>>>>>> was expected
>>>>>>>>>>
>>>>>>>>>> ᐧ
>>>>>>>>>> I went back and checked out the Phoenix commit before the
>>>>>>>>>> pom.xml was upgraded to 0.98.4
>>>>>>>>>> <https://github.com/apache/phoenix/commit/36a41c86a4cdb16ab91b95882dbdb02e6006831d>,
>>>>>>>>>> and I get the same error.
>>>>>>>>>>
>>>>>>>>>> Not sure what to do. Stuck. I imagine most users must be too, as
>>>>>>>>>> CDH 5 is probably the most popular installation there is?
>>>>>>>>>>
>>>>>>>>>> On Tue, Aug 19, 2014 at 11:50 AM, James Taylor <
>>>>>>>>>> jamestay...@apache.org> wrote:
>>>>>>>>>>
>>>>>>>>>>>  The dependencies on HBase 0.98.4 are *compile time*
>>>>>>>>>>> dependencies. Is it necessary for you to compile against CDH 5.1 or 
>>>>>>>>>>> just
>>>>>>>>>>> run against it?
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> On Tuesday, August 19, 2014, Russell Jurney <
>>>>>>>>>>> russell.jur...@gmail.com> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> Thats really bad. That means... CDH 5.x can't run Phoenix? How
>>>>>>>>>>>> can this be fixed? I'm not sure what to do. We're in limbo on our 
>>>>>>>>>>>> new
>>>>>>>>>>>> cluster now.
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> On Mon, Aug 18, 2014 at 11:57 PM, Ravi Kiran <
>>>>>>>>>>>> maghamraviki...@gmail.com> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>> Hi Russel,
>>>>>>>>>>>>>     Apparently, Phoenix 4.0.0 leverages few API methods of
>>>>>>>>>>>>> HBase 0.98.4 v  which aren't present within 0.98.1 that comes 
>>>>>>>>>>>>> with CDH 5.1
>>>>>>>>>>>>> . That's the primary cause for the build issues.
>>>>>>>>>>>>>
>>>>>>>>>>>>> Regards
>>>>>>>>>>>>> Ravi
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> On Mon, Aug 18, 2014 at 5:56 PM, Russell Jurney <
>>>>>>>>>>>>> russell.jur...@gmail.com> wrote:
>>>>>>>>>>>>>
>>>>>>>>>>>>>> Talking to myself, but hopefully creating good docs.
>>>>>>>>>>>>>> Replacing the previous hadoop version with one I found here:
>>>>>>>>>>>>>> https://repository.cloudera.com/artifactory/cloudera-repos/org/apache/hadoop/hadoop-core/,
>>>>>>>>>>>>>>  2.3.0-mr1-cdh5.1.0,
>>>>>>>>>>>>>> makes things get a little further.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> I can't get past some build errors, however. Has anyone done
>>>>>>>>>>>>>> this before me who can help?
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> [ERROR]
>>>>>>>>>>>>>> /Users/rjurney/Software/phoenix4/phoenix-4.0.0-incubating-src/phoenix-core/src/main/java/org/apache/phoenix/query/ConnectionlessQueryServicesImpl.java:[143,27]
>>>>>>>>>>>>>> cannot find symbol
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>   symbol:   method valueOf(java.lang.String,int,int)
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>   location: class org.apache.hadoop.hbase.ServerName
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> [ERROR] Failed to execute goal
>>>>>>>>>>>>>> org.apache.maven.plugins:maven-compiler-plugin:3.0:compile
>>>>>>>>>>>>>> (default-compile) on project phoenix-core: Compilation failure
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> [ERROR]
>>>>>>>>>>>>>> /Users/rjurney/Software/phoenix4/phoenix-4.0.0-incubating-src/phoenix-core/src/main/java/org/apache/phoenix/query/ConnectionlessQueryServicesImpl.java:[143,27]
>>>>>>>>>>>>>> cannot find symbol
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> [ERROR] symbol:   method valueOf(java.lang.String,int,int)
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> [ERROR] location: class org.apache.hadoop.hbase.ServerName
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> [ERROR] -> [Help 1]
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> On Mon, Aug 18, 2014 at 5:41 PM, Russell Jurney <
>>>>>>>>>>>>>> russell.jur...@gmail.com> wrote:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Ok, so it is clear to me what I have to do. I have to edit
>>>>>>>>>>>>>>> my pom.xml to point at CDH 5.1, which translates into:
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Add the cloudera repo:
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>     <repository>
>>>>>>>>>>>>>>>       <id>cloudera</id>
>>>>>>>>>>>>>>>       <url>
>>>>>>>>>>>>>>> https://repository.cloudera.com/artifactory/cloudera-repos/
>>>>>>>>>>>>>>> </url>
>>>>>>>>>>>>>>>     </repository>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Then change the hadoop and hbase versions:
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>      <!-- Hadoop Versions -->
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> <hbase-hadoop1.version>0.98.1-cdh5.1.0</hbase-hadoop1.version>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> <hbase-hadoop2.version>0.98.1-cdh5.1.0</hbase-hadoop2.version>
>>>>>>>>>>>>>>>     <hadoop-one.version>2.3.0-cdh5.1.0</hadoop-one.version>
>>>>>>>>>>>>>>>     <hadoop-two.version>2.3.0-cdh5.1.0</hadoop-two.version>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> However, I get this error when I build, which tells me there
>>>>>>>>>>>>>>> is more complex POM surgery required.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> [ERROR] Failed to execute goal on project phoenix-core:
>>>>>>>>>>>>>>> Could not resolve dependencies for project
>>>>>>>>>>>>>>> org.apache.phoenix:phoenix-core:jar:4.0.0-incubating: The 
>>>>>>>>>>>>>>> following
>>>>>>>>>>>>>>> artifacts could not be resolved:
>>>>>>>>>>>>>>> org.apache.hadoop:hadoop-core:jar:2.3.0-cdh5.1.0,
>>>>>>>>>>>>>>> org.apache.hadoop:hadoop-test:jar:2.3.0-cdh5.1.0: Could not 
>>>>>>>>>>>>>>> find artifact
>>>>>>>>>>>>>>> org.apache.hadoop:hadoop-core:jar:2.3.0-cdh5.1.0 in apache 
>>>>>>>>>>>>>>> release (
>>>>>>>>>>>>>>> https://repository.apache.org/content/repositories/releases/)
>>>>>>>>>>>>>>> -> [Help 1]
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Beyond changing the versions, I do not know how to fix this.
>>>>>>>>>>>>>>> Can anyone help?
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> In general, is it possible to actually handle different CDH
>>>>>>>>>>>>>>> versions in this project? One shouldn't have to do pom surgery 
>>>>>>>>>>>>>>> to build
>>>>>>>>>>>>>>> Phoenix for the most common platform.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> On Mon, Aug 18, 2014 at 5:15 PM, Russell Jurney <
>>>>>>>>>>>>>>> russell.jur...@gmail.com> wrote:
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> When I try to store data into Phoenix from Pig, I get this
>>>>>>>>>>>>>>>> error. I am on CDH 5.1, and Phoenix 4.0.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Anyone know how to resolve this issue?
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> 2014-08-18 17:11:25,165 INFO 
>>>>>>>>>>>>>>>> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigRecordReader:
>>>>>>>>>>>>>>>>  Current split being processed 
>>>>>>>>>>>>>>>> hdfs://cluster1-srv1.e8.com:8020/e8/prod/web_behavior/anomaly_profile.txt/2014/07/15/00/part-r-00000:0+845
>>>>>>>>>>>>>>>> 2014-08-18 
>>>>>>>>>>>>>>>> <http://cluster1-srv1.e8.com:8020/e8/prod/web_behavior/anomaly_profile.txt/2014/07/15/00/part-r-00000:0+8452014-08-18>
>>>>>>>>>>>>>>>>  17:11:25,173 INFO org.apache.hadoop.mapred.TaskLogsTruncater: 
>>>>>>>>>>>>>>>> Initializing logs' truncater with mapRetainSize=-1 and 
>>>>>>>>>>>>>>>> reduceRetainSize=-1
>>>>>>>>>>>>>>>> 2014-08-18 17:11:25,175 FATAL org.apache.hadoop.mapred.Child: 
>>>>>>>>>>>>>>>> Error running child : java.lang.IncompatibleClassChangeError: 
>>>>>>>>>>>>>>>> Found interface 
>>>>>>>>>>>>>>>> org.apache.hadoop.mapreduce.TaskAttemptContext, but class was 
>>>>>>>>>>>>>>>> expected
>>>>>>>>>>>>>>>>        at 
>>>>>>>>>>>>>>>> org.apache.phoenix.pig.hadoop.PhoenixOutputFormat.getRecordWriter(PhoenixOutputFormat.java:65)
>>>>>>>>>>>>>>>>        at 
>>>>>>>>>>>>>>>> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigOutputFormat.getRecordWriter(PigOutputFormat.java:84)
>>>>>>>>>>>>>>>>        at 
>>>>>>>>>>>>>>>> org.apache.hadoop.mapred.MapTask$NewDirectOutputCollector.<init>(MapTask.java:548)
>>>>>>>>>>>>>>>>        at 
>>>>>>>>>>>>>>>> org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:653)
>>>>>>>>>>>>>>>>        at 
>>>>>>>>>>>>>>>> org.apache.hadoop.mapred.MapTask.run(MapTask.java:330)
>>>>>>>>>>>>>>>>        at org.apache.hadoop.mapred.Child$4.run(Child.java:268)
>>>>>>>>>>>>>>>>        at java.security.AccessController.doPrivileged(Native 
>>>>>>>>>>>>>>>> Method)
>>>>>>>>>>>>>>>>        at javax.security.auth.Subject.doAs(Subject.java:415)
>>>>>>>>>>>>>>>>        at 
>>>>>>>>>>>>>>>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1554)
>>>>>>>>>>>>>>>>        at org.apache.hadoop.mapred.Child.main(Child.java:262)
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>> Russell Jurney twitter.com/rjurney russell.jur...@gmail.com
>>>>>>>>>>>>>>>>  datasyndrome.com
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>> Russell Jurney twitter.com/rjurney russell.jur...@gmail.com
>>>>>>>>>>>>>>> datasyndrome.com
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> --
>>>>>>>>>>>>>> Russell Jurney twitter.com/rjurney russell.jur...@gmail.com
>>>>>>>>>>>>>> datasyndrome.com
>>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> --
>>>>>>>>>>>> Russell Jurney twitter.com/rjurney russell.jur...@gmail.com
>>>>>>>>>>>> datasyndrome.com
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> --
>>>>>>>>>> Russell Jurney twitter.com/rjurney russell.jur...@gmail.com
>>>>>>>>>> datasyndrome.com
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> --
>>>>>>>> Russell Jurney twitter.com/rjurney russell.jur...@gmail.com
>>>>>>>> datasyndrome.com
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>> Russell Jurney twitter.com/rjurney russell.jur...@gmail.com
>>>>>>> datasyndrome.com
>>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> --
>>>>>> Best regards,
>>>>>>
>>>>>>    - Andy
>>>>>>
>>>>>> Problems worthy of attack prove their worth by hitting back. - Piet
>>>>>> Hein (via Tom White)
>>>>>>
>>>>>
>>>>>
>>>>>
>>>>> --
>>>>> Russell Jurney twitter.com/rjurney russell.jur...@gmail.com
>>>>> datasyndrome.com
>>>>>
>>>>
>>>>
>>>
>>>
>>> --
>>> Russell Jurney twitter.com/rjurney russell.jur...@gmail.com datasyndrome
>>> .com
>>>
>>
>>
>>
>> --
>> Best regards,
>>
>>    - Andy
>>
>> Problems worthy of attack prove their worth by hitting back. - Piet Hein
>> (via Tom White)
>>
>
>
>
> --
> Russell Jurney twitter.com/rjurney russell.jur...@gmail.com datasyndrome.
> com
>

Reply via email to