Hey Guys,

Just wanted to ping once again and see if anyone has tried phoenix-pig
integration job against the secure hbase cluster.
Pig job started from within a map task.


I see the following exception in the HMaster logs
ipc.RpcServer - RpcServer.listener,port=60000: count of bytes read: 0
org.apache.hadoop.hbase.security.AccessDeniedException: Authentication is
required
        at
org.apache.hadoop.hbase.ipc.RpcServer$Connection.readAndProcess(RpcServer.java:1516)
        at
org.apache.hadoop.hbase.ipc.RpcServer$Listener.doRead(RpcServer.java:856)
        at
org.apache.hadoop.hbase.ipc.RpcServer$Listener$Reader.doRunLoop(RpcServer.java:647)
        at
org.apache.hadoop.hbase.ipc.RpcServer$Listener$Reader.run(RpcServer.java:622)
        at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
        at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)

Somewhere in the flow my HBASE_AUTH_TOKEN is being messed up.

--Siddhi

On Wed, Aug 19, 2015 at 7:39 PM, Siddhi Mehta <sm26...@gmail.com> wrote:

> Hey Guys
>
>
> I am trying to make use of the PhoenixHbaseStorage to write to Hbase Table.
>
>
> The way we start this pig job is from within a map task(Similar to oozie)
>
>
> I run TableMapReduceUtil.initCredentials(job) on the client to get the
> correct AuthTokens for my map task
>
>
> I have ensured that hbase-site.xml is on the classpath for the pigjob and
> also hbase-client and hbase-server jars.
>
>
> Any ideas on what could I be missing?
>
>
> I am using Phoenix4.5 version and hbase 0.98.13
>
>
> I see the following exception in the the logs of the pig job that tries
> writing to hbase
>
>
>
> Aug 20, 2015 12:04:31 AM
> org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper <init>
> INFO: Process identifier=hconnection-0x3c1e23ff connecting to ZooKeeper
> ensemble=hmaster1:2181,hmaster2:2181,hmaster3:2181
> Aug 20, 2015 12:04:31 AM
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation
> makeStub
> INFO: getMaster attempt 1 of 35 failed; retrying after sleep of 100,
> exception=com.google.protobuf.ServiceException:
> java.lang.NullPointerException
> Aug 20, 2015 12:04:31 AM
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation
> makeStub
> INFO: getMaster attempt 2 of 35 failed; retrying after sleep of 200,
> exception=com.google.protobuf.ServiceException: java.io.IOException: Call
> to blitz2-mnds1-3-sfm.ops.sfdc.net/{IPAddress}:60000 failed on local
> exception: java.io.EOFException
> Aug 20, 2015 12:04:31 AM
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation
> makeStub
> INFO: getMaster attempt 3 of 35 failed; retrying after sleep of 300,
> exception=com.google.protobuf.ServiceException: java.io.IOException: Call
> to blitz2-mnds1-3-sfm.ops.sfdc.net/{IPAddress}:60000 failed on local
> exception: java.io.EOFException
> Aug 20, 2015 12:04:31 AM
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation
> makeStub
> INFO: getMaster attempt 4 of 35 failed; retrying after sleep of 500,
> exception=com.google.protobuf.ServiceException: java.io.IOException: Call
> to blitz2-mnds1-3-sfm.ops.sfdc.net/{IPAddress}:60000 failed on local
> exception: java.io.EOFException
> Aug 20, 2015 12:04:32 AM:
>



-- 
Regards,
Siddhi

Reply via email to