Laxman, The current HBase integration does not support Kerberized HBase installs at the moment.
I created a JIRA to track this: https://issues.apache.org/jira/browse/NIFI-1488 -Bryan On Mon, Feb 8, 2016 at 10:36 AM, <laxman.siy...@thomsonreuters.com> wrote: > Hi, > > I have configured a Hbase client and kinit on the machine the that NIFI is > running on, but I get the following error: > > 016-02-08 15:30:12,475 WARN [pool-26-thread-1] > o.a.hadoop.hbase.ipc.AbstractRpcClient Exception encountered while > connecting to the server : javax.security.sasl.SaslException: GSS initiate > failed [Caused by GSSException: No valid credentials provided (Mechanism > level: Failed to find any Kerberos tgt)] > 2016-02-08 15:30:12,476 ERROR [pool-26-thread-1] > o.a.hadoop.hbase.ipc.AbstractRpcClient SASL authentication failed. The most > likely cause is missing or invalid credentials. Consider 'kinit'. > javax.security.sasl.SaslException: GSS initiate failed > at > com.sun.security.sasl.gsskerb.GssKrb5Client.evaluateChallenge(Unknown > Source) ~[na:1.8.0_71] > at > org.apache.hadoop.hbase.security.HBaseSaslRpcClient.saslConnect(HBaseSaslRpcClient.java:179) > ~[hbase-client-1.1.2.jar:1.1.2] > at > org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupSaslConnection(RpcClientImpl.java:642) > [hbase-client-1.1.2.jar:1.1.2] > at > org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.access$600(RpcClientImpl.java:166) > [hbase-client-1.1.2.jar:1.1.2] > at > org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection$2.run(RpcClientImpl.java:769) > ~[hbase-client-1.1.2.jar:1.1.2] > at > org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection$2.run(RpcClientImpl.java:766) > ~[hbase-client-1.1.2.jar:1.1.2] > at java.security.AccessController.doPrivileged(Native Method) > ~[na:1.8.0_71] > at javax.security.auth.Subject.doAs(Unknown Source) ~[na:1.8.0_71] > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1656) > ~[hadoop-common-2.6.2.jar:na] > at > org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:766) > [hbase-client-1.1.2.jar:1.1.2] > at > org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:920) > [hbase-client-1.1.2.jar:1.1.2] > at > org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:889) > [hbase-client-1.1.2.jar:1.1.2] > at > org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1222) > [hbase-client-1.1.2.jar:1.1.2] > at > org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213) > [hbase-client-1.1.2.jar:1.1.2] > at > org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287) > [hbase-client-1.1.2.jar:1.1.2] > at > org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:50918) > [hbase-protocol-1.1.2.jar:1.1.2] > at > org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceStubMaker.isMasterRunning(ConnectionManager.java:1564) > [hbase-client-1.1.2.jar:1.1.2] > at > org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$StubMaker.makeStubNoRetries(ConnectionManager.java:1502) > [hbase-client-1.1.2.jar:1.1.2] > at > org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$StubMaker.makeStub(ConnectionManager.java:1524) > [hbase-client-1.1.2.jar:1.1.2] > at > org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceStubMaker.makeStub(ConnectionManager.java:1553) > [hbase-client-1.1.2.jar:1.1.2] > at > org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1704) > [hbase-client-1.1.2.jar:1.1.2] > at > org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38) > [hbase-client-1.1.2.jar:1.1.2] > at > org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124) > [hbase-client-1.1.2.jar:1.1.2] > at > org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:3917) > [hbase-client-1.1.2.jar:1.1.2] > at > org.apache.hadoop.hbase.client.HBaseAdmin.listTableNames(HBaseAdmin.java:413) > [hbase-client-1.1.2.jar:1.1.2] > at > org.apache.hadoop.hbase.client.HBaseAdmin.listTableNames(HBaseAdmin.java:397) > [hbase-client-1.1.2.jar:1.1.2] > at > org.apache.nifi.hbase.HBase_1_1_2_ClientService.onEnabled(HBase_1_1_2_ClientService.java:137) > [nifi-hbase_1_1_2-client-service-0.4.1.jar:0.4.1] > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > ~[na:1.8.0_71] > at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source) > ~[na:1.8.0_71] > at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source) > ~[na:1.8.0_71] > at java.lang.reflect.Method.invoke(Unknown Source) ~[na:1.8.0_71] > at > org.apache.nifi.util.ReflectionUtils.invokeMethodsWithAnnotations(ReflectionUtils.java:102) > [nifi-framework-core-0.4.1.jar:0.4.1] > at > org.apache.nifi.util.ReflectionUtils.invokeMethodsWithAnnotation(ReflectionUtils.java:47) > [nifi-framework-core-0.4.1.jar:0.4.1] > at > org.apache.nifi.controller.scheduling.StandardProcessScheduler$6.run(StandardProcessScheduler.java:652) > [nifi-framework-core-0.4.1.jar:0.4.1] > at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown > Source) [na:1.8.0_71] > at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown > Source) [na:1.8.0_71] > at java.lang.Thread.run(Unknown Source) [na:1.8.0_71] > Caused by: org.ietf.jgss.GSSException: No valid credentials provided > (Mechanism level: Failed to find any Kerberos tgt) > at sun.security.jgss.krb5.Krb5InitCredential.getInstance(Unknown > Source) ~[na:1.8.0_71] > at > sun.security.jgss.krb5.Krb5MechFactory.getCredentialElement(Unknown Source) > ~[na:1.8.0_71] > at > sun.security.jgss.krb5.Krb5MechFactory.getMechanismContext(Unknown Source) > ~[na:1.8.0_71] > at sun.security.jgss.GSSManagerImpl.getMechanismContext(Unknown > Source) ~[na:1.8.0_71] > at sun.security.jgss.GSSContextImpl.initSecContext(Unknown Source) > ~[na:1.8.0_71] > at sun.security.jgss.GSSContextImpl.initSecContext(Unknown Source) > ~[na:1.8.0_71] > ... 37 common frames omitted > > > Does anyone have any experience with NIFI accessing a Secured HDFS/HBASE > setup. > > I can use the shell on the server fine to go to the hbase shell or use the > hadoop fs, so I know kinit works. But is there more to configure in the > nifi.properties? > > Regards, > Laxman > > > ________________________________ > > This e-mail is for the sole use of the intended recipient and contains > information that may be privileged and/or confidential. If you are not an > intended recipient, please notify the sender by return e-mail and delete > this e-mail and any attachments. Certain required legal entity disclosures > can be accessed on our website.< > http://site.thomsonreuters.com/site/disclosures/> >