Hi Siddhi, I remember the fix was done and tested as part of https://issues.apache.org/jira/browse/PHOENIX-1078 . If possible, can you go a bit deeper in explaining how you are calling PhoenixHBaseStorage from a map task.
Regards Ravi On Thu, Aug 20, 2015 at 6:54 PM, Siddhi Mehta <siddhi....@gmail.com> wrote: > Hey Guys, > > Just wanted to ping once again and see if anyone has tried phoenix-pig > integration job against the secure hbase cluster. > Pig job started from within a map task. > > > I see the following exception in the HMaster logs > ipc.RpcServer - RpcServer.listener,port=60000: count of bytes read: 0 > org.apache.hadoop.hbase.security.AccessDeniedException: Authentication is > required > at > > org.apache.hadoop.hbase.ipc.RpcServer$Connection.readAndProcess(RpcServer.java:1516) > at > org.apache.hadoop.hbase.ipc.RpcServer$Listener.doRead(RpcServer.java:856) > at > > org.apache.hadoop.hbase.ipc.RpcServer$Listener$Reader.doRunLoop(RpcServer.java:647) > at > > org.apache.hadoop.hbase.ipc.RpcServer$Listener$Reader.run(RpcServer.java:622) > at > > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) > at > > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) > > Somewhere in the flow my HBASE_AUTH_TOKEN is being messed up. > > --Siddhi > > On Wed, Aug 19, 2015 at 7:39 PM, Siddhi Mehta <sm26...@gmail.com> wrote: > > > Hey Guys > > > > > > I am trying to make use of the PhoenixHbaseStorage to write to Hbase > Table. > > > > > > The way we start this pig job is from within a map task(Similar to oozie) > > > > > > I run TableMapReduceUtil.initCredentials(job) on the client to get the > > correct AuthTokens for my map task > > > > > > I have ensured that hbase-site.xml is on the classpath for the pigjob and > > also hbase-client and hbase-server jars. > > > > > > Any ideas on what could I be missing? > > > > > > I am using Phoenix4.5 version and hbase 0.98.13 > > > > > > I see the following exception in the the logs of the pig job that tries > > writing to hbase > > > > > > > > Aug 20, 2015 12:04:31 AM > > org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper <init> > > INFO: Process identifier=hconnection-0x3c1e23ff connecting to ZooKeeper > > ensemble=hmaster1:2181,hmaster2:2181,hmaster3:2181 > > Aug 20, 2015 12:04:31 AM > > > org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation > > makeStub > > INFO: getMaster attempt 1 of 35 failed; retrying after sleep of 100, > > exception=com.google.protobuf.ServiceException: > > java.lang.NullPointerException > > Aug 20, 2015 12:04:31 AM > > > org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation > > makeStub > > INFO: getMaster attempt 2 of 35 failed; retrying after sleep of 200, > > exception=com.google.protobuf.ServiceException: java.io.IOException: Call > > to blitz2-mnds1-3-sfm.ops.sfdc.net/{IPAddress}:60000 > <http://blitz2-mnds1-3-sfm.ops.sfdc.net/%7BIPAddress%7D:60000> failed on > local > > exception: java.io.EOFException > > Aug 20, 2015 12:04:31 AM > > > org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation > > makeStub > > INFO: getMaster attempt 3 of 35 failed; retrying after sleep of 300, > > exception=com.google.protobuf.ServiceException: java.io.IOException: Call > > to blitz2-mnds1-3-sfm.ops.sfdc.net/{IPAddress}:60000 > <http://blitz2-mnds1-3-sfm.ops.sfdc.net/%7BIPAddress%7D:60000> failed on > local > > exception: java.io.EOFException > > Aug 20, 2015 12:04:31 AM > > > org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation > > makeStub > > INFO: getMaster attempt 4 of 35 failed; retrying after sleep of 500, > > exception=com.google.protobuf.ServiceException: java.io.IOException: Call > > to blitz2-mnds1-3-sfm.ops.sfdc.net/{IPAddress}:60000 > <http://blitz2-mnds1-3-sfm.ops.sfdc.net/%7BIPAddress%7D:60000> failed on > local > > exception: java.io.EOFException > > Aug 20, 2015 12:04:32 AM: > > > > > > -- > Regards, > Siddhi >