Shaik, HDFS plugin built from ‘master’ branch uses different configuration files than the one built from 0.4 version. From the error details, it looks like HDFS plugin from the ‘master’ branch was installed in an environment that had earlier version installed.
To fix this issue, please try the following: * from your Hadoop installation lib directory (for example: /usr/local/hadoop/lib), remove Ranger libraries or symbolic links – filename of pattern ranger-*.jar * reinstall the plugin Hope this helps. Madhan From: Shaik M <[email protected]<mailto:[email protected]>> Reply-To: "[email protected]<mailto:[email protected]>" <[email protected]<mailto:[email protected]>> Date: Monday, March 9, 2015 at 3:45 AM To: "[email protected]<mailto:[email protected]>" <[email protected]<mailto:[email protected]>> Subject: Re: Ranger Hive Plugin issue HDFS Plugin Issue Hi Bosco, I am doing the HDFS Plugin-0.4 installation as per the Apache Ranger documentation. in that installation procedure it is pointing to xasecure-audit.xml file. But, after enabling the plugin it is generated following conf files in /etc/hadoop/conf ranger-hdfs-audit.xml, ranger-hdfs-security.xml & ranger-policymgr-ssl.xml. After restating NameNode, agents are not registered with Ranger-Admin. I have gone through with NameNode logs and i found following error log: 015-03-09 10:32:19,746 FATAL conf.Configuration (Configuration.java:loadResource(2512)) - error parsing conf file:/etc/hadoop/conf.empty/xasecure-audit.xml java.io.FileNotFoundException: /etc/hadoop/conf.empty/xasecure-audit.xml (No such file or directory) at java.io.FileInputStream.open(Native Method) at java.io.FileInputStream.<init>(FileInputStream.java:146) at java.io.FileInputStream.<init>(FileInputStream.java:101) at sun.net.www.protocol.file.FileURLConnection.connect(FileURLConnection.java:90) at sun.net.www.protocol.file.FileURLConnection.getInputStream(FileURLConnection.java:188) at java.net.URL.openStream(URL.java:1037) at org.apache.hadoop.conf.Configuration.parse(Configuration.java:2342) at org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:2410) at org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:2376) at org.apache.hadoop.conf.Configuration.getProps(Configuration.java:2283) at org.apache.hadoop.conf.Configuration.get(Configuration.java:1110) at org.apache.hadoop.hdfs.server.namenode.XaSecureFSPermissionChecker.<clinit>(XaSecureFSPermissionChecker.java:57) at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:6515) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getFileInfo(FSNamesystem.java:4143) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getFileInfo(NameNodeRpcServer.java:838) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getFileInfo(ClientNamenodeProtocolServerSideTranslatorPB.java:821) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:619) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:962) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2039) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2035) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2033) 2015-03-09 10:32:19,748 WARN ipc.Server (Server.java:run(2058)) - IPC Server handler 3 on 8020, call org.apache.hadoop.hdfs.protocol.ClientProtocol.getFileInfo from 10.193.153.223:56168<http://10.193.153.223:56168> Call#2579 Retry#10 java.lang.ExceptionInInitializerError Please let me know the solution for this. Regards, Shaik On 9 March 2015 at 18:01, Shaik M <[email protected]<mailto:[email protected]>> wrote: Sorry, Bosco... Some DB configuration issue...after fixing that issue has been resolved. Now all access audits are showing in Audit>Access tab. Thank you for your help! -Shaik On 9 March 2015 at 16:51, Shaik M <[email protected]<mailto:[email protected]>> wrote: Hi Bosco, Thank you again..:) I have successfully deployed Hive plugin as per Ranger documentation. All policies are working fine!!! But, schema access audits are not showing in Audit>Access tab. Please let me know if any configuration changes are required for this... Regards, Shaik On 9 March 2015 at 14:18, Don Bosco Durai <[email protected]<mailto:[email protected]>> wrote: Hi Shaik For HDFS and Hive plugins, you shouldn’t be needing to do any changes on the Ambari version 1.7 side. For HDFS, you might want to check the hdf.perm.enable property in Ambari and it is turned off, then you have to enable it. For Hbase you have to do couple of config changes on the Apache Ambari side. Thanks Bosco From: Shaik M <[email protected]<mailto:[email protected]>> Reply-To: "[email protected]<mailto:[email protected]>" <[email protected]<mailto:[email protected]>> Date: Sunday, March 8, 2015 at 8:42 PM To: "[email protected]<mailto:[email protected]>" <[email protected]<mailto:[email protected]>> Subject: Re: Ranger Hive Plugin issue Hi Bosco, Thank you so much for your suggestions. I am using HDP2.2 and using "Hive 0.14.0.2.2.0.0". I have reverted all changes based on XASecure 3.5 documentation. I have a question about: Is it required any changes for Apache Ranger 0.4 on HDP2.2+Ambari other than mentioned in Ranger 0.4 installation procedure? If any changes are required kindly let me know. Regards, Shaik M On 9 March 2015 at 13:20, Don Bosco Durai <[email protected]<mailto:[email protected]>> wrote: Hi Shaik There seems to be multiple issues here. The official released version of Apache Ranger is 0.4 and the instruction to build and install is available at: https://cwiki.apache.org/confluence/display/RANGER/Ranger+Installation+Guide Please note, Apache Ranger 0.4 release is compatible with Hive 0.14. So make sure you using the corresponding versions. The install and setup script assumes the folders as per Apache Hive install. Also note, you should follow the instructions as per documentation. Updating the install.properties and running enable-hive-plugin.sh helps in updating the appropriate configuration file. It seems you have been trying to update it manually, which is not advisable, unless your installation folders for Hive or other Hadoop components are customized. Another note, the error you are getting is if you are trying to configure Hive using the XASecure 3.5 version documentation. Note, XA Secure 3.5 is the erstwhile version of Apache Ranger 0.4 and it is compatible up to Hive 0.13 only. Thanks Bosco From: Hadoop Solutions <[email protected]<mailto:[email protected]>> Reply-To: "[email protected]<mailto:[email protected]>" <[email protected]<mailto:[email protected]>> Date: Sunday, March 8, 2015 at 7:03 PM To: "[email protected]<mailto:[email protected]>" <[email protected]<mailto:[email protected]>> Subject: Ranger Hive Plugin issue Hi, I have configured hive ranger plugin with the help of HDP doc. After configuring Ranger-Hive plugin, tries to run query from HUE. But I am getting following exception, unable to run any query. Error occurred executing hive query: Error while compiling statement: FAILED: ClassNotFoundException com.xasecure.authorization.hive.hooks.XaSecureSemanticAnalyzerHook Please help me to resolve the issue. Thanks, Shaik
