maybe relevant (maybe not)

https://github.com/apache/zeppelin/pull/1323

Had some issues hitting a cloud HDFS instance a while back- you may be able
to hack a solution out relevant to your problem.

tg

Trevor Grant
Data Scientist
https://github.com/rawkintrevo
http://stackexchange.com/users/3002022/rawkintrevo
http://trevorgrant.org

*"Fortunate is he, who is able to know the causes of things."  -Virgil*


On Wed, May 31, 2017 at 12:52 PM, BigData Consultant <
bigdata.consultant2...@gmail.com> wrote:

> How to set the access for zeppelin user to the HDFS?
>
> On Tue, May 30, 2017 at 2:16 AM, Felix Cheung <felixcheun...@hotmail.com>
> wrote:
>
>> Seems to be with hdfs ACL - does the service user Zeppelin have access to
>> your storage?
>>
>> ------------------------------
>> *From:* BigData Consultant <bigdata.consultant2...@gmail.com>
>> *Sent:* Friday, May 26, 2017 10:56:31 PM
>> *To:* dev@zeppelin.apache.org; us...@zeppelin.apache.org
>> *Subject:* Permission denied: user=zeppelin while using %spark.pyspark
>> interpreter in AWS EMR cluster
>>
>> Hi Team,
>>
>> I have created pyspark structure streaming program and trying to execute
>> in
>> the Zeppelin notebook, I am getting the following error:
>>
>> Py4JJavaError: An error occurred while calling o191.start.
>> : org.apache.hadoop.security.AccessControlException: Permission denied:
>> user=zeppelin, access=WRITE,
>> inode="/mnt/tmp/temporary-e0cf0f09-a6f4-44d6-9a72-3246600856
>> 08/metadata":hdfs:hadoop:drwxr-xr-x
>>
>>
>> I am using Zeppelin Notebook Version 0.7.1 in AWS EMR cluster.
>>
>> Help would be much appreciated.
>>
>> *Full stacktrace:*
>>
>>
>> Traceback (most recent call last):
>> File "/tmp/zeppelin_pyspark-8165971491474576109.py", line 349, in
>> <module>
>> raise Exception(traceback.format_exc())
>> Exception: Traceback (most recent call last):
>> File "/tmp/zeppelin_pyspark-8165971491474576109.py", line 342, in
>> <module>
>> exec(code)
>> File "<stdin>", line 5, in <module>
>> File "/usr/lib/spark/python/pyspark/sql/streaming.py", line 816, in start
>> return self._sq(self._jwrite.start())
>> File "/usr/lib/spark/python/lib/py4j-0.10.4-src.zip/py4j/java_gat
>> eway.py",
>> line 1133, in __call__
>> answer, self.gateway_client, self.target_id, self.name)
>> File "/usr/lib/spark/python/pyspark/sql/utils.py", line 63, in deco
>> return f(*a, **kw)
>> File "/usr/lib/spark/python/lib/py4j-0.10.4-src.zip/py4j/protocol.py",
>> line
>> 319, in get_return_value
>> format(target_id, ".", name), value)
>> Py4JJavaError: An error occurred while calling o191.start.
>> : org.apache.hadoop.security.AccessControlException: Permission denied:
>> user=zeppelin, access=WRITE,
>> inode="/mnt/tmp/temporary-e0cf0f09-a6f4-44d6-9a72-3246600856
>> 08/metadata":hdfs:hadoop:drwxr-xr-x
>> at
>> org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.c
>> heck(FSPermissionChecker.java:320)
>> at
>> org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.c
>> heck(FSPermissionChecker.java:292)
>> at
>> org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.c
>> heckPermission(FSPermissionChecker.java:213)
>> at
>> org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.c
>> heckPermission(FSPermissionChecker.java:190)
>> at
>> org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkPerm
>> ission(FSDirectory.java:1728)
>> at
>> org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkPerm
>> ission(FSDirectory.java:1712)
>> at
>> org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkAnce
>> storAccess(FSDirectory.java:1695)
>> at
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFil
>> eInternal(FSNamesystem.java:2515)
>> at
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFil
>> eInt(FSNamesystem.java:2450)
>> at
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFil
>> e(FSNamesystem.java:2334)
>> at
>> org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.cre
>> ate(NameNodeRpcServer.java:624)
>> at
>> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServ
>> erSideTranslatorPB.create(ClientNamenodeProtocolServerSideTr
>> anslatorPB.java:397)
>> at
>> org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocol
>> Protos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNam
>> enodeProtocolProtos.java)
>> at
>> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcIn
>> voker.call(ProtobufRpcEngine.java:616)
>> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:982)
>> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2049)
>> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2045)
>> at java.security.AccessController.doPrivileged(Native Method)
>> at javax.security.auth.Subject.doAs(Subject.java:422)
>> at
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGro
>> upInformation.java:1698)
>> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2045)
>> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
>> at
>> sun.reflect.NativeConstructorAccessorImpl.newInstance(Native
>> ConstructorAccessorImpl.java:62)
>> at
>> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(De
>> legatingConstructorAccessorImpl.java:45)
>> at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
>> at
>> org.apache.hadoop.ipc.RemoteException.instantiateException(R
>> emoteException.java:106)
>> at
>> org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(
>> RemoteException.java:73)
>> at
>> org.apache.hadoop.hdfs.DFSOutputStream.newStreamForCreate(DF
>> SOutputStream.java:1653)
>> at org.apache.hadoop.hdfs.DFSClient.create(DFSClient.java:1689)
>> at org.apache.hadoop.hdfs.DFSClient.create(DFSClient.java:1624)
>> at
>> org.apache.hadoop.hdfs.DistributedFileSystem$7.doCall(
>> DistributedFileSystem.java:448)
>> at
>> org.apache.hadoop.hdfs.DistributedFileSystem$7.doCall(
>> DistributedFileSystem.java:444)
>> at
>> org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSyst
>> emLinkResolver.java:81)
>> at
>> org.apache.hadoop.hdfs.DistributedFileSystem.create(Distribu
>> tedFileSystem.java:459)
>> at
>> org.apache.hadoop.hdfs.DistributedFileSystem.create(Distribu
>> tedFileSystem.java:387)
>> at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:915)
>> at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:896)
>> at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:793)
>> at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:782)
>> at
>> org.apache.spark.sql.execution.streaming.StreamMetadata$.
>> write(StreamMetadata.scala:76)
>> at
>> org.apache.spark.sql.execution.streaming.StreamExecution$$
>> anonfun$5.apply(StreamExecution.scala:102)
>> at
>> org.apache.spark.sql.execution.streaming.StreamExecution$$
>> anonfun$5.apply(StreamExecution.scala:100)
>> at scala.Option.getOrElse(Option.scala:121)
>> at
>> org.apache.spark.sql.execution.streaming.StreamExecution.<
>> init>(StreamExecution.scala:100)
>> at
>> org.apache.spark.sql.streaming.StreamingQueryManager.createQ
>> uery(StreamingQueryManager.scala:232)
>> at
>> org.apache.spark.sql.streaming.StreamingQueryManager.startQu
>> ery(StreamingQueryManager.scala:269)
>> at
>> org.apache.spark.sql.streaming.DataStreamWriter.start(
>> DataStreamWriter.scala:227)
>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>> at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAcce
>> ssorImpl.java:62)
>> at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMe
>> thodAccessorImpl.java:43)
>> at java.lang.reflect.Method.invoke(Method.java:498)
>> at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244)
>> at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)
>> at py4j.Gateway.invoke(Gateway.java:280)
>> at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
>> at py4j.commands.CallCommand.execute(CallCommand.java:79)
>> at py4j.GatewayConnection.run(GatewayConnection.java:214)
>> at java.lang.Thread.run(Thread.java:745)
>> Caused by:
>> org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.
>> security.AccessControlException):
>> Permission denied: user=zeppelin, access=WRITE,
>> inode="/mnt/tmp/temporary-e0cf0f09-a6f4-44d6-9a72-3246600856
>> 08/metadata":hdfs:hadoop:drwxr-xr-x
>> at
>> org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.c
>> heck(FSPermissionChecker.java:320)
>> at
>> org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.c
>> heck(FSPermissionChecker.java:292)
>> at
>> org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.c
>> heckPermission(FSPermissionChecker.java:213)
>> at
>> org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.c
>> heckPermission(FSPermissionChecker.java:190)
>> at
>> org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkPerm
>> ission(FSDirectory.java:1728)
>> at
>> org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkPerm
>> ission(FSDirectory.java:1712)
>> at
>> org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkAnce
>> storAccess(FSDirectory.java:1695)
>> at
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFil
>> eInternal(FSNamesystem.java:2515)
>> at
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFil
>> eInt(FSNamesystem.java:2450)
>> at
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFil
>> e(FSNamesystem.java:2334)
>> at
>> org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.cre
>> ate(NameNodeRpcServer.java:624)
>> at
>> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServ
>> erSideTranslatorPB.create(ClientNamenodeProtocolServerSideTr
>> anslatorPB.java:397)
>> at
>> org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocol
>> Protos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNam
>> enodeProtocolProtos.java)
>> at
>> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcIn
>> voker.call(ProtobufRpcEngine.java:616)
>> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:982)
>> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2049)
>> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2045)
>> at java.security.AccessController.doPrivileged(Native Method)
>> at javax.security.auth.Subject.doAs(Subject.java:422)
>> at
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGro
>> upInformation.java:1698)
>> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2045)
>> at org.apache.hadoop.ipc.Client.call(Client.java:1475)
>> at org.apache.hadoop.ipc.Client.call(Client.java:1412)
>> at
>> org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(
>> ProtobufRpcEngine.java:229)
>> at com.sun.proxy.$Proxy12.create(Unknown Source)
>> at
>> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTran
>> slatorPB.create(ClientNamenodeProtocolTranslatorPB.java:296)
>> at sun.reflect.GeneratedMethodAccessor14.invoke(Unknown Source)
>> at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMe
>> thodAccessorImpl.java:43)
>> at java.lang.reflect.Method.invoke(Method.java:498)
>> at
>> org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMeth
>> od(RetryInvocationHandler.java:191)
>> at
>> org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(Ret
>> ryInvocationHandler.java:102)
>> at com.sun.proxy.$Proxy13.create(Unknown Source)
>> at
>> org.apache.hadoop.hdfs.DFSOutputStream.newStreamForCreate(DF
>> SOutputStream.java:1648)
>> ... 30 more
>>
>
>

Reply via email to