Hi Shwetha, much appreciated. I missed /etc/atlas/conf in the class path. Now the hook works. Herman.
> On Jan 21, 2016, at 9:30 AM, Shwetha Shivalingamurthy > <sshivalingamur...@hortonworks.com> wrote: > > Hive-env.sh should contain > export > HADOOP_CLASSPATH=/etc/atlas/conf:/usr/hdp/current/atlas-server/hook/hive:${ > HADOOP_CLASSPATH} > > > So, /etc/atlas/conf/ should contain atlas’ application.properties > > > Regards, > Shwetha > > > > > > > On 21/01/16 7:46 pm, "Herman Yu" <herman...@teeupdata.com> wrote: > >> Thanks Shwetha. that’s the reason, the hook is getting application and >> client configurations from jar files under hook/hive. I must be missing >> some configuration steps? is it supposed to get those two config files >> from Hive configuration folder? >> >> Thanks >> Herman. >> >> >>> On Jan 20, 2016, at 11:27 PM, Shwetha Shivalingamurthy >>> <sshivalingamur...@hortonworks.com> wrote: >>> >>> Check if the atlas conf is picked up correctly. You can enable DEBUG >>> logs >>> for atlas by modifying the hive’s log4j.xml - add log level for package >>> org.apache.atlas. DEBUG log level will print the location of atlas’ >>> application.properties that’s picked up >>> >>> Regards, >>> Shwetha >>> >>> >>> >>> >>> >>> >>> On 21/01/16 12:41 am, "Herman Yu" <herman...@teeupdata.com> wrote: >>> >>>> Hi Shewtha, >>>> >>>> Thanks for the link. I followed the link, but am still getting the >>>> following error in hiveserver2.log. It seems to me that atlasclient is >>>> refused when trying to connect to atlas server via http. I already >>>> have >>>> atlas.http.authentication.enabled=false set to false in both >>>> application >>>> and client properties files. >>>> >>>> >>>> any other places I need to check? >>>> >>>> thanks >>>> Herman. >>>> >>>> >>>> 2016-01-20 13:55:22,568 INFO [Atlas Logger 3]: hook.HiveHook >>>> (HiveHook.java:fireAndForget(192)) - Entered Atlas hook for hook type >>>> POST_EXEC_HOOK operation CREATETABLE >>>> 2016-01-20 13:55:22,572 INFO [HiveServer2-Background-Pool: >>>> Thread-191]: >>>> log.PerfLogger (PerfLogger.java:PerfLogEnd(148)) - </PERFLOG >>>> method=releaseLocks start=1453316122568 end=1453316122572 duration=4 >>>> from=org.apache.hadoop.hive.ql.Driver> >>>> 2016-01-20 13:55:22,573 INFO [HiveServer2-Background-Pool: >>>> Thread-191]: >>>> log.PerfLogger (PerfLogger.java:PerfLogEnd(148)) - </PERFLOG >>>> method=Driver.run start=1453316122342 end=1453316122573 duration=231 >>>> from=org.apache.hadoop.hive.ql.Driver> >>>> 2016-01-20 13:55:22,634 INFO [Atlas Logger 3]: >>>> security.SecureClientUtils >>>> (SecureClientUtils.java:getClientConnectionHandler(90)) - Real User: >>>> hive >>>> (auth:SIMPLE), is from ticket cache? false >>>> 2016-01-20 13:55:22,635 INFO [Atlas Logger 3]: >>>> security.SecureClientUtils >>>> (SecureClientUtils.java:getClientConnectionHandler(93)) - doAsUser: hue >>>> 2016-01-20 13:55:22,635 INFO [Atlas Logger 2]: hook.HiveHook >>>> (HiveHook.java:run(182)) - Atlas hook failed >>>> com.sun.jersey.api.client.ClientHandlerException: java.io.IOException: >>>> java.net.ConnectException: Connection refused >>>> at >>>> >>>> com.sun.jersey.client.urlconnection.URLConnectionClientHandler.handle(UR >>>> LC >>>> onnectionClientHandler.java:149) >>>> at com.sun.jersey.api.client.Client.handle(Client.java:648) >>>> at >>>> com.sun.jersey.api.client.WebResource.handle(WebResource.java:670) >>>> at >>>> com.sun.jersey.api.client.WebResource.access$200(WebResource.java:74) >>>> at >>>> >>>> com.sun.jersey.api.client.WebResource$Builder.method(WebResource.java:62 >>>> 3) >>>> at >>>> org.apache.atlas.AtlasClient.callAPIWithResource(AtlasClient.java:584) >>>> at >>>> org.apache.atlas.AtlasClient.callAPIWithResource(AtlasClient.java:579) >>>> at org.apache.atlas.AtlasClient.getType(AtlasClient.java:257) >>>> at >>>> >>>> org.apache.atlas.hive.bridge.HiveMetaStoreBridge.registerHiveDataModel(H >>>> iv >>>> eMetaStoreBridge.java:487) >>>> at >>>> org.apache.atlas.hive.hook.HiveHook.fireAndForget(HiveHook.java:197) >>>> at >>>> org.apache.atlas.hive.hook.HiveHook.access$200(HiveHook.java:66) >>>> at org.apache.atlas.hive.hook.HiveHook$2.run(HiveHook.java:180) >>>> at >>>> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) >>>> at java.util.concurrent.FutureTask.run(FutureTask.java:262) >>>> at >>>> >>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.jav >>>> a: >>>> 1145) >>>> at >>>> >>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.ja >>>> va >>>> :615) >>>> at java.lang.Thread.run(Thread.java:745) >>>> Caused by: java.io.IOException: java.net.ConnectException: Connection >>>> refused >>>> at >>>> >>>> org.apache.atlas.security.SecureClientUtils$1$1.run(SecureClientUtils.ja >>>> va >>>> :106) >>>> at >>>> >>>> org.apache.atlas.security.SecureClientUtils$1$1.run(SecureClientUtils.ja >>>> va >>>> :98) >>>> at java.security.AccessController.doPrivileged(Native Method) >>>> at javax.security.auth.Subject.doAs(Subject.java:415) >>>> at >>>> >>>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformatio >>>> n. >>>> java:1657) >>>> at >>>> >>>> org.apache.atlas.security.SecureClientUtils$1.getHttpURLConnection(Secur >>>> eC >>>> lientUtils.java:98) >>>> at >>>> >>>> com.sun.jersey.client.urlconnection.URLConnectionClientHandler._invoke(U >>>> RL >>>> ConnectionClientHandler.java:159) >>>> at >>>> >>>> com.sun.jersey.client.urlconnection.URLConnectionClientHandler.handle(UR >>>> LC >>>> onnectionClientHandler.java:147) >>>> ... 16 more >>>> Caused by: java.net.ConnectException: Connection refused >>>> at java.net.PlainSocketImpl.socketConnect(Native Method) >>>> at >>>> >>>> java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java: >>>> 33 >>>> 9) >>>> at >>>> >>>> java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImp >>>> l. >>>> java:200) >>>> at >>>> >>>> java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:18 >>>> 2) >>>> at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:392) >>>> at java.net.Socket.connect(Socket.java:579) >>>> at sun.net.NetworkClient.doConnect(NetworkClient.java:175) >>>> at sun.net.www.http.HttpClient.openServer(HttpClient.java:432) >>>> at sun.net.www.http.HttpClient.openServer(HttpClient.java:527) >>>> at sun.net.www.http.HttpClient.<init>(HttpClient.java:211) >>>> at sun.net.www.http.HttpClient.New(HttpClient.java:308) >>>> at sun.net.www.http.HttpClient.New(HttpClient.java:326) >>>> at >>>> >>>> sun.net.www.protocol.http.HttpURLConnection.getNewHttpClient(HttpURLConn >>>> ec >>>> tion.java:998) >>>> at >>>> >>>> sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnecti >>>> on >>>> .java:934) >>>> at >>>> >>>> sun.net.www.protocol.http.HttpURLConnection.connect(HttpURLConnection.ja >>>> va >>>> :852) >>>> at >>>> >>>> org.apache.hadoop.security.authentication.client.PseudoAuthenticator.aut >>>> he >>>> nticate(PseudoAuthenticator.java:76) >>>> at >>>> >>>> org.apache.hadoop.security.token.delegation.web.DelegationTokenAuthentic >>>> at >>>> or.authenticate(DelegationTokenAuthenticator.java:128) >>>> at >>>> >>>> org.apache.hadoop.security.authentication.client.AuthenticatedURL.openCo >>>> nn >>>> ection(AuthenticatedURL.java:215) >>>> at >>>> >>>> org.apache.hadoop.security.token.delegation.web.DelegationTokenAuthentic >>>> at >>>> edURL.openConnection(DelegationTokenAuthenticatedURL.java:322) >>>> at >>>> >>>> org.apache.atlas.security.SecureClientUtils$1$1.run(SecureClientUtils.ja >>>> va >>>> :102) >>>> ... 23 more >>>> >>>> >>>>> On Jan 20, 2016, at 2:35 AM, Shwetha Shivalingamurthy >>>>> <sshivalingamur...@hortonworks.com> wrote: >>>>> >>>>> >>>>> >>>>> http://dev.hortonworks.com.s3.amazonaws.com/HDPDocuments/Ambari-2.2.0.0 >>>>> /b >>>>> k_ >>>>> ambari_reference_guide/content/ch_integrating_atlas.html has the >>>>> instructions on setting up 0.6 Atlas with Ambari >>>>> >>>>> >>>>> Regards, >>>>> Shwetha >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> On 20/01/16 2:26 am, "Herman Yu" <herman...@teeupdata.com> wrote: >>>>> >>>>>> Hi everyone, >>>>>> >>>>>> I am having some problems with installing and configuring v0.6 on HDP >>>>>> 2.3.2 sandbox. >>>>>> >>>>>> HDP 2.3.2 comes with v0.5 pre-configured with Ambari. After compiling >>>>>> v0.6 source code, and updated the link /usr/hdp/current/atlas-server >>>>>> to >>>>>> be associated with the new v0.6 folder. I also made necessary changes >>>>>> in >>>>>> amber¹s (e.g. those notification related parameters in >>>>>> application.properties). Atlas starts fine (even though with some >>>>>> errors >>>>>> in log file which are kafka related), however, Hive stopped working >>>>>> with >>>>>> the following errors with any operations. >>>>>> >>>>>> I suspect this is the Hive hook related, after several tries, I >>>>>> figured >>>>>> out that if I update atlas.hook.hive.synchronous from true to false, >>>>>> Hive >>>>>> starts working fine however the hook doesn¹t capture any hive queries >>>>>> to >>>>>> atlas. >>>>>> >>>>>> Did anyone experience the same problem? Also, how does hive know >>>>>> where >>>>>> the jar file of the hook is located? in HDP 2.3.2, I don¹t see >>>>>> HIVE_AUX_PATH is appended with <Atlas_home>/bridge/hive, I had to >>>>>> manually append this in Ambari. >>>>>> >>>>>> >>>>>> org.apache.hive.service.cli.HiveSQLException: Error while processing >>>>>> statement: FAILED: Hive Internal Error: >>>>>> com.sun.jersey.api.client.ClientHandlerException(java.io.IOException: >>>>>> java.net.ConnectException: Connection refused) >>>>>> at >>>>>> >>>>>> >>>>>> org.apache.hive.service.cli.operation.Operation.toSQLException(Operati >>>>>> on >>>>>> .j >>>>>> ava:315) >>>>>> at >>>>>> >>>>>> >>>>>> org.apache.hive.service.cli.operation.SQLOperation.runQuery(SQLOperati >>>>>> on >>>>>> .j >>>>>> ava:156) >>>>>> at >>>>>> >>>>>> >>>>>> org.apache.hive.service.cli.operation.SQLOperation.runInternal(SQLOper >>>>>> at >>>>>> io >>>>>> n.java:183) >>>>>> at >>>>>> >>>>>> org.apache.hive.service.cli.operation.Operation.run(Operation.java:257 >>>>>> ) >>>>>> at >>>>>> >>>>>> >>>>>> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementIn >>>>>> te >>>>>> rn >>>>>> al(HiveSessionImpl.java:388) >>>>>> at >>>>>> >>>>>> >>>>>> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatement(H >>>>>> iv >>>>>> eS >>>>>> essionImpl.java:369) >>>>>> at >>>>>> >>>>>> >>>>>> org.apache.hive.service.cli.CLIService.executeStatement(CLIService.jav >>>>>> a: >>>>>> 26 >>>>>> 1) >>>>>> at >>>>>> >>>>>> >>>>>> org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement(T >>>>>> hr >>>>>> if >>>>>> tCLIService.java:486) >>>>>> at >>>>>> >>>>>> >>>>>> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatem >>>>>> en >>>>>> t. >>>>>> getResult(TCLIService.java:1313) >>>>>> at >>>>>> >>>>>> >>>>>> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatem >>>>>> en >>>>>> t. >>>>>> getResult(TCLIService.java:1298) >>>>>> at >>>>>> org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39) >>>>>> at >>>>>> org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39) >>>>>> at >>>>>> >>>>>> >>>>>> org.apache.hive.service.auth.TSetIpAddressProcessor.process(TSetIpAddr >>>>>> es >>>>>> sP >>>>>> rocessor.java:56) >>>>>> at >>>>>> >>>>>> >>>>>> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPo >>>>>> ol >>>>>> Se >>>>>> rver.java:285) >>>>>> at >>>>>> >>>>>> >>>>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.j >>>>>> av >>>>>> a: >>>>>> 1145) >>>>>> at >>>>>> >>>>>> >>>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor. >>>>>> ja >>>>>> va >>>>>> :615) >>>>>> at java.lang.Thread.run(Thread.java:745) >>>>>> Caused by: com.sun.jersey.api.client.ClientHandlerException: >>>>>> java.io.IOException: java.net.ConnectException: Connection refused >>>>>> at >>>>>> >>>>>> >>>>>> com.sun.jersey.client.urlconnection.URLConnectionClientHandler.handle( >>>>>> UR >>>>>> LC >>>>>> onnectionClientHandler.java:149) >>>>>> at com.sun.jersey.api.client.Client.handle(Client.java:648) >>>>>> at >>>>>> com.sun.jersey.api.client.WebResource.handle(WebResource.java:670) >>>>>> at >>>>>> com.sun.jersey.api.client.WebResource.access$200(WebResource.java:74) >>>>>> at >>>>>> >>>>>> >>>>>> com.sun.jersey.api.client.WebResource$Builder.method(WebResource.java: >>>>>> 62 >>>>>> 3) >>>>>> at >>>>>> >>>>>> org.apache.atlas.AtlasClient.callAPIWithResource(AtlasClient.java:584) >>>>>> at >>>>>> >>>>>> org.apache.atlas.AtlasClient.callAPIWithResource(AtlasClient.java:579) >>>>>> at org.apache.atlas.AtlasClient.getType(AtlasClient.java:257) >>>>>> at >>>>>> >>>>>> >>>>>> org.apache.atlas.hive.bridge.HiveMetaStoreBridge.registerHiveDataModel >>>>>> (H >>>>>> iv >>>>>> eMetaStoreBridge.java:487) >>>>>> at >>>>>> org.apache.atlas.hive.hook.HiveHook.fireAndForget(HiveHook.java:197) >>>>>> at org.apache.atlas.hive.hook.HiveHook.run(HiveHook.java:174) >>>>>> at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1522) >>>>>> at >>>>>> org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1195) >>>>>> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1059) >>>>>> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1054) >>>>>> at >>>>>> >>>>>> >>>>>> org.apache.hive.service.cli.operation.SQLOperation.runQuery(SQLOperati >>>>>> on >>>>>> .j >>>>>> ava:154) >>>>>> ... 15 more >>>>>> >>>>>> >>>>>> Thanks >>>>>> Herman. >>>>>> >>>>> >>>> >>> >> >