Hi,

please let me know if you have any other solution for HDFS plugin issue.

Regards,
Shaik

On 10 March 2015 at 10:51, Shaik M <[email protected]> wrote:

> Hi Madhan,
>
> I have reinstalled HDFS plugin, but still i am facing same issue.
> please let me know if you have any other solution.
>
> Thanks,
> Shaik
>
> On 10 March 2015 at 00:01, Madhan Neethiraj <[email protected]>
> wrote:
>
>>  Shaik,
>>
>>  HDFS plugin built from ‘master’ branch uses different configuration
>> files than the one built from 0.4 version. From the error details, it looks
>> like HDFS plugin from the ‘master’ branch was installed in an environment
>> that had earlier version installed.
>>
>>  To fix this issue, please try the following:
>>
>>    - from your Hadoop installation lib directory (for example:
>>    /usr/local/hadoop/lib), remove Ranger libraries or symbolic links –
>>    filename of pattern ranger-*.jar
>>    - reinstall the plugin
>>
>> Hope this helps.
>>
>>  Madhan
>>
>>    From: Shaik M <[email protected]>
>> Reply-To: "[email protected]" <
>> [email protected]>
>> Date: Monday, March 9, 2015 at 3:45 AM
>>
>> To: "[email protected]" <[email protected]>
>> Subject: Re: Ranger Hive Plugin issue
>>
>>   *HDFS Plugin Issue*
>>
>>  Hi Bosco,
>>
>>  I am doing the HDFS Plugin-0.4 installation as per the Apache Ranger
>> documentation. in that installation procedure it is pointing to
>> *xasecure-audit.xml* file. But, after enabling the plugin it is
>> generated following conf files in /etc/hadoop/conf *ranger-hdfs-audit.xml, 
>> ranger-hdfs-security.xml
>> & ranger-policymgr-ssl.xml.*
>>
>>  After restating NameNode, agents are not registered with Ranger-Admin.
>> I have gone through with NameNode logs and i found following error log:
>>
>>  015-03-09 10:32:19,746 FATAL conf.Configuration
>> (Configuration.java:loadResource(2512)) - error parsing conf
>> file:/etc/hadoop/conf.empty/xasecure-audit.xml
>> java.io.FileNotFoundException: /etc/hadoop/conf.empty/xasecure-audit.xml
>> (No such file or directory)
>>         at java.io.FileInputStream.open(Native Method)
>>         at java.io.FileInputStream.<init>(FileInputStream.java:146)
>>         at java.io.FileInputStream.<init>(FileInputStream.java:101)
>>         at
>> sun.net.www.protocol.file.FileURLConnection.connect(FileURLConnection.java:90)
>>         at
>> sun.net.www.protocol.file.FileURLConnection.getInputStream(FileURLConnection.java:188)
>>         at java.net.URL.openStream(URL.java:1037)
>>         at
>> org.apache.hadoop.conf.Configuration.parse(Configuration.java:2342)
>>         at
>> org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:2410)
>>         at
>> org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:2376)
>>         at
>> org.apache.hadoop.conf.Configuration.getProps(Configuration.java:2283)
>>         at
>> org.apache.hadoop.conf.Configuration.get(Configuration.java:1110)
>>         at
>> org.apache.hadoop.hdfs.server.namenode.XaSecureFSPermissionChecker.<clinit>(XaSecureFSPermissionChecker.java:57)
>>         at
>> org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java)
>>         at
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:6515)
>>         at
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getFileInfo(FSNamesystem.java:4143)
>>         at
>> org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getFileInfo(NameNodeRpcServer.java:838)
>>         at
>> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getFileInfo(ClientNamenodeProtocolServerSideTranslatorPB.java:821)
>>         at
>> org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
>>         at
>> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:619)
>>         at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:962)
>>         at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2039)
>>         at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2035)
>>         at java.security.AccessController.doPrivileged(Native Method)
>>         at javax.security.auth.Subject.doAs(Subject.java:415)
>>         at
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628)
>>         at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2033)
>> 2015-03-09 10:32:19,748 WARN  ipc.Server (Server.java:run(2058)) - IPC
>> Server handler 3 on 8020, call
>> org.apache.hadoop.hdfs.protocol.ClientProtocol.getFileInfo from
>> 10.193.153.223:56168 Call#2579 Retry#10
>> java.lang.ExceptionInInitializerError
>>
>>
>>  Please let me know the solution for this.
>>
>>  Regards,
>> Shaik
>>
>> On 9 March 2015 at 18:01, Shaik M <[email protected]> wrote:
>>
>>> Sorry, Bosco...
>>>
>>>  Some DB configuration issue...after fixing that issue has been
>>> resolved. Now all access audits are showing in Audit>Access tab.
>>>
>>>  Thank you for your help!
>>>
>>>  -Shaik
>>>
>>> On 9 March 2015 at 16:51, Shaik M <[email protected]> wrote:
>>>
>>>> Hi Bosco,
>>>>
>>>>  Thank you again..:)
>>>>
>>>>  I have successfully deployed Hive plugin as per Ranger documentation.
>>>> All policies are working fine!!!
>>>>
>>>>  But, schema access audits are not showing in Audit>Access tab.
>>>>
>>>>  Please let me know if any configuration changes are required for
>>>> this...
>>>>
>>>>  Regards,
>>>> Shaik
>>>>
>>>> On 9 March 2015 at 14:18, Don Bosco Durai <[email protected]> wrote:
>>>>
>>>>>  Hi Shaik
>>>>>
>>>>>  For HDFS and Hive plugins, you shouldn’t be needing to do any
>>>>> changes on the Ambari version 1.7 side.  For HDFS, you might want to check
>>>>> the hdf.perm.enable property in Ambari and it is turned off, then you have
>>>>> to enable it. For Hbase you have to do couple of config changes on the
>>>>> Apache Ambari side.
>>>>>
>>>>>  Thanks
>>>>>
>>>>>  Bosco
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>   From: Shaik M <[email protected]>
>>>>> Reply-To: "[email protected]" <
>>>>> [email protected]>
>>>>> Date: Sunday, March 8, 2015 at 8:42 PM
>>>>> To: "[email protected]" <
>>>>> [email protected]>
>>>>> Subject: Re: Ranger Hive Plugin issue
>>>>>
>>>>>   Hi Bosco,
>>>>>
>>>>>  Thank you so much for your suggestions.
>>>>>
>>>>>  I am using HDP2.2 and using "Hive 0.14.0.2.2.0.0". I have reverted
>>>>> all changes based on XASecure 3.5 documentation.
>>>>>
>>>>>  I have a question about: Is it required any changes for Apache
>>>>> Ranger 0.4 on HDP2.2+Ambari  other than mentioned in Ranger 0.4
>>>>> installation procedure?
>>>>>
>>>>>  If any changes are required kindly let me know.
>>>>>
>>>>>  Regards,
>>>>> Shaik M
>>>>>
>>>>>
>>>>> On 9 March 2015 at 13:20, Don Bosco Durai <[email protected]> wrote:
>>>>>
>>>>>>  Hi Shaik
>>>>>>
>>>>>>  There seems to be multiple issues here. The official released
>>>>>> version of Apache Ranger is 0.4 and the instruction to build and install 
>>>>>> is
>>>>>> available at:
>>>>>>
>>>>>>
>>>>>> https://cwiki.apache.org/confluence/display/RANGER/Ranger+Installation+Guide
>>>>>>
>>>>>>  Please note, Apache Ranger  0.4 release is compatible with Hive
>>>>>> 0.14. So make sure you using the corresponding versions. The install and
>>>>>> setup script assumes the folders as per Apache Hive install.
>>>>>>
>>>>>>  Also note, you should follow the instructions as per documentation.
>>>>>> Updating the install.properties and running enable-hive-plugin.sh helps 
>>>>>> in
>>>>>> updating the appropriate configuration file. It seems you have been 
>>>>>> trying
>>>>>> to update it manually, which is not advisable, unless your installation
>>>>>> folders for Hive or other Hadoop components are customized.
>>>>>>
>>>>>>  Another note, the error you are getting is if you are trying to
>>>>>> configure Hive using the XASecure 3.5 version documentation. Note, XA
>>>>>> Secure 3.5 is the erstwhile version of Apache Ranger 0.4 and it is
>>>>>> compatible up to Hive 0.13 only.
>>>>>>
>>>>>>  Thanks
>>>>>>
>>>>>>  Bosco
>>>>>>
>>>>>>
>>>>>>   From: Hadoop Solutions <[email protected]>
>>>>>> Reply-To: "[email protected]" <
>>>>>> [email protected]>
>>>>>> Date: Sunday, March 8, 2015 at 7:03 PM
>>>>>> To: "[email protected]" <
>>>>>> [email protected]>
>>>>>> Subject: Ranger Hive Plugin issue
>>>>>>
>>>>>>   Hi,
>>>>>>
>>>>>>
>>>>>>  I have configured hive ranger plugin with the help of HDP doc.
>>>>>> After configuring Ranger-Hive plugin, tries to run query from HUE. But I 
>>>>>> am
>>>>>> getting following exception, unable to run any query.
>>>>>>
>>>>>>  *Error occurred executing hive query: Error while compiling
>>>>>> statement: FAILED: ClassNotFoundException
>>>>>> com.xasecure.authorization.hive.hooks.XaSecureSemanticAnalyzerHook*
>>>>>>
>>>>>>  Please help me to resolve the issue.
>>>>>>
>>>>>>
>>>>>>  Thanks,
>>>>>> Shaik
>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>

Reply via email to