Hi Amit, Is it working ? No , Mine is phoenix 4.4 . Thanks, Divya
On 1 March 2016 at 13:00, Amit Shah <amits...@gmail.com> wrote: > Hi Divya, > > Thanks for the patch. Is this for phoenix version 4.6 ? Are the changes > made to make phoenix work with CDH 5.5.2? > > Thanks, > Amit. > > On Tue, Mar 1, 2016 at 10:08 AM, Divya Gehlot <divya.htco...@gmail.com> > wrote: > >> Hi Amit, >> Extract attached jar and try placing it in your hbase classpath >> >> P.S. Please remove the 'x' from the jar extension >> Hope this helps. >> >> >> Thanks, >> Divya >> >> On 26 February 2016 at 20:44, Amit Shah <amits...@gmail.com> wrote: >> >>> Hello, >>> >>> I have been trying to install phoenix on my cloudera hbase cluster. >>> Cloudera version is CDH5.5.2 while HBase version is 1.0. >>> >>> I copied the server & core jar (version 4.6-HBase-1.0) on the master and >>> region servers and restarted the hbase cluster. I copied the corresponding >>> client jar on my SQuirrel client but I get an exception on connect. Pasted >>> below. The connection url is >>> “jdbc:phoenix:<zookeeper-server-name>:2181". >>> I even tried compiling the source by adding cloudera dependencies as >>> suggested on this post >>> <http://stackoverflow.com/questions/31849454/using-phoenix-with-cloudera-hbase-installed-from-repo>but >>> didn't succeed. >>> >>> Any suggestions to make this work? >>> >>> Thanks, >>> Amit. >>> >>> ________________________________________________________________ >>> >>> Caused by: >>> org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.DoNotRetryIOException): >>> org.apache.hadoop.hbase.DoNotRetryIOException: SYSTEM.CATALOG: >>> org.apache.hadoop.hbase.client.Scan.setRaw(Z)Lorg/apache/hadoop/hbase/client/Scan; >>> at >>> org.apache.phoenix.util.ServerUtil.createIOException(ServerUtil.java:87) >>> at >>> org.apache.phoenix.coprocessor.MetaDataEndpointImpl.createTable(MetaDataEndpointImpl.java:1319) >>> at >>> org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:11715) >>> at >>> org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:7388) >>> at >>> org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:1776) >>> at >>> org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:1758) >>> at >>> org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:32209) >>> at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2034) >>> at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:107) >>> at >>> org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130) >>> at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107) >>> at java.lang.Thread.run(Thread.java:745) >>> Caused by: java.lang.NoSuchMethodError: >>> org.apache.hadoop.hbase.client.Scan.setRaw(Z)Lorg/apache/hadoop/hbase/client/Scan; >>> at >>> org.apache.phoenix.coprocessor.MetaDataEndpointImpl.buildDeletedTable(MetaDataEndpointImpl.java:1016) >>> at >>> org.apache.phoenix.coprocessor.MetaDataEndpointImpl.loadTable(MetaDataEndpointImpl.java:1092) >>> at >>> org.apache.phoenix.coprocessor.MetaDataEndpointImpl.createTable(MetaDataEndpointImpl.java:1266) >>> ... 10 more >>> >>> P.S - The full stacktrace is attached in the mail. >>> >> >> >