Hi Russel, Apparently, Phoenix 4.0.0 leverages few API methods of HBase 0.98.4 v which aren't present within 0.98.1 that comes with CDH 5.1 . That's the primary cause for the build issues.
Regards Ravi On Mon, Aug 18, 2014 at 5:56 PM, Russell Jurney <russell.jur...@gmail.com> wrote: > Talking to myself, but hopefully creating good docs. Replacing the > previous hadoop version with one I found here: > https://repository.cloudera.com/artifactory/cloudera-repos/org/apache/hadoop/hadoop-core/, > 2.3.0-mr1-cdh5.1.0, > makes things get a little further. > > I can't get past some build errors, however. Has anyone done this before > me who can help? > > [ERROR] > /Users/rjurney/Software/phoenix4/phoenix-4.0.0-incubating-src/phoenix-core/src/main/java/org/apache/phoenix/query/ConnectionlessQueryServicesImpl.java:[143,27] > cannot find symbol > > symbol: method valueOf(java.lang.String,int,int) > > location: class org.apache.hadoop.hbase.ServerName > > > [ERROR] Failed to execute goal > org.apache.maven.plugins:maven-compiler-plugin:3.0:compile > (default-compile) on project phoenix-core: Compilation failure > > [ERROR] > /Users/rjurney/Software/phoenix4/phoenix-4.0.0-incubating-src/phoenix-core/src/main/java/org/apache/phoenix/query/ConnectionlessQueryServicesImpl.java:[143,27] > cannot find symbol > > [ERROR] symbol: method valueOf(java.lang.String,int,int) > > [ERROR] location: class org.apache.hadoop.hbase.ServerName > > [ERROR] -> [Help 1] > > ᐧ > > > On Mon, Aug 18, 2014 at 5:41 PM, Russell Jurney <russell.jur...@gmail.com> > wrote: > >> Ok, so it is clear to me what I have to do. I have to edit my pom.xml to >> point at CDH 5.1, which translates into: >> >> Add the cloudera repo: >> >> <repository> >> <id>cloudera</id> >> <url>https://repository.cloudera.com/artifactory/cloudera-repos/ >> </url> >> </repository> >> >> >> Then change the hadoop and hbase versions: >> >> <!-- Hadoop Versions --> >> <hbase-hadoop1.version>0.98.1-cdh5.1.0</hbase-hadoop1.version> >> <hbase-hadoop2.version>0.98.1-cdh5.1.0</hbase-hadoop2.version> >> <hadoop-one.version>2.3.0-cdh5.1.0</hadoop-one.version> >> <hadoop-two.version>2.3.0-cdh5.1.0</hadoop-two.version> >> >> >> However, I get this error when I build, which tells me there is more >> complex POM surgery required. >> >> [ERROR] Failed to execute goal on project phoenix-core: Could not resolve >> dependencies for project >> org.apache.phoenix:phoenix-core:jar:4.0.0-incubating: The following >> artifacts could not be resolved: >> org.apache.hadoop:hadoop-core:jar:2.3.0-cdh5.1.0, >> org.apache.hadoop:hadoop-test:jar:2.3.0-cdh5.1.0: Could not find artifact >> org.apache.hadoop:hadoop-core:jar:2.3.0-cdh5.1.0 in apache release ( >> https://repository.apache.org/content/repositories/releases/) -> [Help 1] >> >> Beyond changing the versions, I do not know how to fix this. Can anyone >> help? >> >> In general, is it possible to actually handle different CDH versions in >> this project? One shouldn't have to do pom surgery to build Phoenix for the >> most common platform. >> ᐧ >> >> >> On Mon, Aug 18, 2014 at 5:15 PM, Russell Jurney <russell.jur...@gmail.com >> > wrote: >> >>> When I try to store data into Phoenix from Pig, I get this error. I am >>> on CDH 5.1, and Phoenix 4.0. >>> >>> Anyone know how to resolve this issue? >>> >>> 2014-08-18 17:11:25,165 INFO >>> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigRecordReader: >>> Current split being processed >>> hdfs://cluster1-srv1.e8.com:8020/e8/prod/web_behavior/anomaly_profile.txt/2014/07/15/00/part-r-00000:0+845 >>> 2014-08-18 >>> <http://cluster1-srv1.e8.com:8020/e8/prod/web_behavior/anomaly_profile.txt/2014/07/15/00/part-r-00000:0+8452014-08-18> >>> 17:11:25,173 INFO org.apache.hadoop.mapred.TaskLogsTruncater: Initializing >>> logs' truncater with mapRetainSize=-1 and reduceRetainSize=-1 >>> 2014-08-18 17:11:25,175 FATAL org.apache.hadoop.mapred.Child: Error running >>> child : java.lang.IncompatibleClassChangeError: Found interface >>> org.apache.hadoop.mapreduce.TaskAttemptContext, but class was expected >>> at >>> org.apache.phoenix.pig.hadoop.PhoenixOutputFormat.getRecordWriter(PhoenixOutputFormat.java:65) >>> at >>> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigOutputFormat.getRecordWriter(PigOutputFormat.java:84) >>> at >>> org.apache.hadoop.mapred.MapTask$NewDirectOutputCollector.<init>(MapTask.java:548) >>> at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:653) >>> at org.apache.hadoop.mapred.MapTask.run(MapTask.java:330) >>> at org.apache.hadoop.mapred.Child$4.run(Child.java:268) >>> at java.security.AccessController.doPrivileged(Native Method) >>> at javax.security.auth.Subject.doAs(Subject.java:415) >>> at >>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1554) >>> at org.apache.hadoop.mapred.Child.main(Child.java:262) >>> >>> >>> >>> -- >>> Russell Jurney twitter.com/rjurney russell.jur...@gmail.com datasyndrome >>> .com >>> ᐧ >>> >> >> >> >> -- >> Russell Jurney twitter.com/rjurney russell.jur...@gmail.com datasyndrome. >> com >> > > > > -- > Russell Jurney twitter.com/rjurney russell.jur...@gmail.com datasyndrome. > com >