I am using Hortonworks distribution and It comes with phoenix :) No idea about the patch. You try posting the error in CDH forum . It might help.
All the Best !! Cheers, Divya On 1 March 2016 at 13:15, Amit Shah <amits...@gmail.com> wrote: > No, it doesn't work for phoenix 4.6. Attached is the error I get when I > execute 'sqlline.py <zookeeper-server-name>:2181' > > Can you please give more details about the patch? > > Thanks, > Amit. > > On Tue, Mar 1, 2016 at 10:39 AM, Divya Gehlot <divya.htco...@gmail.com> > wrote: > >> Hi Amit, >> Is it working ? >> No , Mine is phoenix 4.4 . >> >> Thanks, >> Divya >> >> On 1 March 2016 at 13:00, Amit Shah <amits...@gmail.com> wrote: >> >>> Hi Divya, >>> >>> Thanks for the patch. Is this for phoenix version 4.6 ? Are the changes >>> made to make phoenix work with CDH 5.5.2? >>> >>> Thanks, >>> Amit. >>> >>> On Tue, Mar 1, 2016 at 10:08 AM, Divya Gehlot <divya.htco...@gmail.com> >>> wrote: >>> >>>> Hi Amit, >>>> Extract attached jar and try placing it in your hbase classpath >>>> >>>> P.S. Please remove the 'x' from the jar extension >>>> Hope this helps. >>>> >>>> >>>> Thanks, >>>> Divya >>>> >>>> On 26 February 2016 at 20:44, Amit Shah <amits...@gmail.com> wrote: >>>> >>>>> Hello, >>>>> >>>>> I have been trying to install phoenix on my cloudera hbase cluster. >>>>> Cloudera version is CDH5.5.2 while HBase version is 1.0. >>>>> >>>>> I copied the server & core jar (version 4.6-HBase-1.0) on the master >>>>> and region servers and restarted the hbase cluster. I copied the >>>>> corresponding client jar on my SQuirrel client but I get an exception on >>>>> connect. Pasted below. The connection url is >>>>> “jdbc:phoenix:<zookeeper-server-name>:2181". >>>>> I even tried compiling the source by adding cloudera dependencies as >>>>> suggested on this post >>>>> <http://stackoverflow.com/questions/31849454/using-phoenix-with-cloudera-hbase-installed-from-repo>but >>>>> didn't succeed. >>>>> >>>>> Any suggestions to make this work? >>>>> >>>>> Thanks, >>>>> Amit. >>>>> >>>>> ________________________________________________________________ >>>>> >>>>> Caused by: >>>>> org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.DoNotRetryIOException): >>>>> org.apache.hadoop.hbase.DoNotRetryIOException: SYSTEM.CATALOG: >>>>> org.apache.hadoop.hbase.client.Scan.setRaw(Z)Lorg/apache/hadoop/hbase/client/Scan; >>>>> at >>>>> org.apache.phoenix.util.ServerUtil.createIOException(ServerUtil.java:87) >>>>> at >>>>> org.apache.phoenix.coprocessor.MetaDataEndpointImpl.createTable(MetaDataEndpointImpl.java:1319) >>>>> at >>>>> org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:11715) >>>>> at >>>>> org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:7388) >>>>> at >>>>> org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:1776) >>>>> at >>>>> org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:1758) >>>>> at >>>>> org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:32209) >>>>> at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2034) >>>>> at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:107) >>>>> at >>>>> org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130) >>>>> at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107) >>>>> at java.lang.Thread.run(Thread.java:745) >>>>> Caused by: java.lang.NoSuchMethodError: >>>>> org.apache.hadoop.hbase.client.Scan.setRaw(Z)Lorg/apache/hadoop/hbase/client/Scan; >>>>> at >>>>> org.apache.phoenix.coprocessor.MetaDataEndpointImpl.buildDeletedTable(MetaDataEndpointImpl.java:1016) >>>>> at >>>>> org.apache.phoenix.coprocessor.MetaDataEndpointImpl.loadTable(MetaDataEndpointImpl.java:1092) >>>>> at >>>>> org.apache.phoenix.coprocessor.MetaDataEndpointImpl.createTable(MetaDataEndpointImpl.java:1266) >>>>> ... 10 more >>>>> >>>>> P.S - The full stacktrace is attached in the mail. >>>>> >>>> >>>> >>> >> >