Shaik,

HDFS plugin built from ‘master’ branch uses different configuration files than 
the one built from 0.4 version. From the error details, it looks like HDFS 
plugin from the ‘master’ branch was installed in an environment that had 
earlier version installed.

To fix this issue, please try the following:

  *   from your Hadoop installation lib directory (for example: 
/usr/local/hadoop/lib), remove Ranger libraries or symbolic links – filename of 
pattern ranger-*.jar
  *   reinstall the plugin

Hope this helps.

Madhan

From: Shaik M <[email protected]<mailto:[email protected]>>
Reply-To: 
"[email protected]<mailto:[email protected]>" 
<[email protected]<mailto:[email protected]>>
Date: Monday, March 9, 2015 at 3:45 AM
To: "[email protected]<mailto:[email protected]>" 
<[email protected]<mailto:[email protected]>>
Subject: Re: Ranger Hive Plugin issue

HDFS Plugin Issue

Hi Bosco,

I am doing the HDFS Plugin-0.4 installation as per the Apache Ranger 
documentation. in that installation procedure it is pointing to 
xasecure-audit.xml file. But, after enabling the plugin it is generated 
following conf files in /etc/hadoop/conf ranger-hdfs-audit.xml, 
ranger-hdfs-security.xml & ranger-policymgr-ssl.xml.

After restating NameNode, agents are not registered with Ranger-Admin. I have 
gone through with NameNode logs and i found following error log:

015-03-09 10:32:19,746 FATAL conf.Configuration 
(Configuration.java:loadResource(2512)) - error parsing conf 
file:/etc/hadoop/conf.empty/xasecure-audit.xml
java.io.FileNotFoundException: /etc/hadoop/conf.empty/xasecure-audit.xml (No 
such file or directory)
        at java.io.FileInputStream.open(Native Method)
        at java.io.FileInputStream.<init>(FileInputStream.java:146)
        at java.io.FileInputStream.<init>(FileInputStream.java:101)
        at 
sun.net.www.protocol.file.FileURLConnection.connect(FileURLConnection.java:90)
        at 
sun.net.www.protocol.file.FileURLConnection.getInputStream(FileURLConnection.java:188)
        at java.net.URL.openStream(URL.java:1037)
        at org.apache.hadoop.conf.Configuration.parse(Configuration.java:2342)
        at 
org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:2410)
        at 
org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:2376)
        at 
org.apache.hadoop.conf.Configuration.getProps(Configuration.java:2283)
        at org.apache.hadoop.conf.Configuration.get(Configuration.java:1110)
        at 
org.apache.hadoop.hdfs.server.namenode.XaSecureFSPermissionChecker.<clinit>(XaSecureFSPermissionChecker.java:57)
        at 
org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java)
        at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:6515)
        at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getFileInfo(FSNamesystem.java:4143)
        at 
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getFileInfo(NameNodeRpcServer.java:838)
        at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getFileInfo(ClientNamenodeProtocolServerSideTranslatorPB.java:821)
        at 
org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
        at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:619)
        at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:962)
        at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2039)
        at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2035)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:415)
        at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628)
        at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2033)
2015-03-09 10:32:19,748 WARN  ipc.Server (Server.java:run(2058)) - IPC Server 
handler 3 on 8020, call 
org.apache.hadoop.hdfs.protocol.ClientProtocol.getFileInfo from 
10.193.153.223:56168<http://10.193.153.223:56168> Call#2579 Retry#10
java.lang.ExceptionInInitializerError


Please let me know the solution for this.

Regards,
Shaik

On 9 March 2015 at 18:01, Shaik M 
<[email protected]<mailto:[email protected]>> wrote:
Sorry, Bosco...

Some DB configuration issue...after fixing that issue has been resolved. Now 
all access audits are showing in Audit>Access tab.

Thank you for your help!

-Shaik

On 9 March 2015 at 16:51, Shaik M 
<[email protected]<mailto:[email protected]>> wrote:
Hi Bosco,

Thank you again..:)

I have successfully deployed Hive plugin as per Ranger documentation. All 
policies are working fine!!!

But, schema access audits are not showing in Audit>Access tab.

Please let me know if any configuration changes are required for this...

Regards,
Shaik

On 9 March 2015 at 14:18, Don Bosco Durai 
<[email protected]<mailto:[email protected]>> wrote:
Hi Shaik

For HDFS and Hive plugins, you shouldn’t be needing to do any changes on the 
Ambari version 1.7 side.  For HDFS, you might want to check the hdf.perm.enable 
property in Ambari and it is turned off, then you have to enable it. For Hbase 
you have to do couple of config changes on the Apache Ambari side.

Thanks

Bosco




From: Shaik M <[email protected]<mailto:[email protected]>>
Reply-To: 
"[email protected]<mailto:[email protected]>" 
<[email protected]<mailto:[email protected]>>
Date: Sunday, March 8, 2015 at 8:42 PM
To: "[email protected]<mailto:[email protected]>" 
<[email protected]<mailto:[email protected]>>
Subject: Re: Ranger Hive Plugin issue

Hi Bosco,

Thank you so much for your suggestions.

I am using HDP2.2 and using "Hive 0.14.0.2.2.0.0". I have reverted all changes 
based on XASecure 3.5 documentation.

I have a question about: Is it required any changes for Apache Ranger 0.4 on 
HDP2.2+Ambari  other than mentioned in Ranger 0.4 installation procedure?

If any changes are required kindly let me know.

Regards,
Shaik M


On 9 March 2015 at 13:20, Don Bosco Durai 
<[email protected]<mailto:[email protected]>> wrote:
Hi Shaik

There seems to be multiple issues here. The official released version of Apache 
Ranger is 0.4 and the instruction to build and install is available at:

https://cwiki.apache.org/confluence/display/RANGER/Ranger+Installation+Guide

Please note, Apache Ranger  0.4 release is compatible with Hive 0.14. So make 
sure you using the corresponding versions. The install and setup script assumes 
the folders as per Apache Hive install.

Also note, you should follow the instructions as per documentation. Updating 
the install.properties and running enable-hive-plugin.sh helps in updating the 
appropriate configuration file. It seems you have been trying to update it 
manually, which is not advisable, unless your installation folders for Hive or 
other Hadoop components are customized.

Another note, the error you are getting is if you are trying to configure Hive 
using the XASecure 3.5 version documentation. Note, XA Secure 3.5 is the 
erstwhile version of Apache Ranger 0.4 and it is compatible up to Hive 0.13 
only.

Thanks

Bosco


From: Hadoop Solutions <[email protected]<mailto:[email protected]>>
Reply-To: 
"[email protected]<mailto:[email protected]>" 
<[email protected]<mailto:[email protected]>>
Date: Sunday, March 8, 2015 at 7:03 PM
To: "[email protected]<mailto:[email protected]>" 
<[email protected]<mailto:[email protected]>>
Subject: Ranger Hive Plugin issue

Hi,


I have configured hive ranger plugin with the help of HDP doc. After 
configuring Ranger-Hive plugin, tries to run query from HUE. But I am getting 
following exception, unable to run any query.

Error occurred executing hive query: Error while compiling statement: FAILED: 
ClassNotFoundException 
com.xasecure.authorization.hive.hooks.XaSecureSemanticAnalyzerHook

Please help me to resolve the issue.


Thanks,
Shaik





Reply via email to