Hi Moon,

Yes it did set master property and exported Hadoop config

Master= yarn-client
export HADOOP_CONF_DIR =/etc/Hadoop/conf      (in zeppelin-env.sh file)

It is throwing below error.

org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.AccessControlException):
 Permission denied: user=root, access=WRITE, 
inode="/user":hdfs:supergroup:drwxr-xr-x
       at 
org.apache.hadoop.hdfs.server.namenode.DefaultAuthorizationProvider.checkFsPermission(DefaultAuthorizationProvider.java:257)
       at 
org.apache.hadoop.hdfs.server.namenode.DefaultAuthorizationProvider.check(DefaultAuthorizationProvider.java:238)
       at 
org.apache.hadoop.hdfs.server.namenode.DefaultAuthorizationProvider.check(DefaultAuthorizationProvider.java:216)
       at 
org.apache.hadoop.hdfs.server.namenode.DefaultAuthorizationProvider.checkPermission(DefaultAuthorizationProvider.java:145)
       at 
org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:138)
       at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:6596)
       at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:6578)
       at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkAncestorAccess(FSNamesystem.java:6530)
       at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirsInternal(FSNamesystem.java:4334)
       at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirsInt(FSNamesystem.java:4304)
       at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirs(FSNamesystem.java:4277)
       at 
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.mkdirs(NameNodeRpcServer.java:852)
       at 
org.apache.hadoop.hdfs.server.namenode.AuthorizationProviderProxyClientProtocol.mkdirs(AuthorizationProviderProxyClientProtocol.java:321)
       at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.mkdirs(ClientNamenodeProtocolServerSideTranslatorPB.java:601)
       at 
org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
       at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:619)
       at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1060)
       at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2044)
       at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2040)
       at java.security.AccessController.doPrivileged(Native Method)
       at javax.security.auth.Subject.doAs(Subject.java:415)
       at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1671)
       at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2038)

       at org.apache.hadoop.ipc.Client.call(Client.java:1468)
       at org.apache.hadoop.ipc.Client.call(Client.java:1399)
       at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:232)
       at com.sun.proxy.$Proxy14.mkdirs(Unknown Source)
       at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.mkdirs(ClientNamenodeProtocolTranslatorPB.java:539)
       at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
       at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
       at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
       at java.lang.reflect.Method.invoke(Method.java:483)
       at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:187)
       at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)
       at com.sun.proxy.$Proxy15.mkdirs(Unknown Source)
       at org.apache.hadoop.hdfs.DFSClient.primitiveMkdir(DFSClient.java:2760)
       at org.apache.hadoop.hdfs.DFSClient.mkdirs(DFSClient.java:2731)
       at 
org.apache.hadoop.hdfs.DistributedFileSystem$17.doCall(DistributedFileSystem.java:870)
       at 
org.apache.hadoop.hdfs.DistributedFileSystem$17.doCall(DistributedFileSystem.java:866)
       at 
org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
       at 
org.apache.hadoop.hdfs.DistributedFileSystem.mkdirsInternal(DistributedFileSystem.java:866)
       at 
org.apache.hadoop.hdfs.DistributedFileSystem.mkdirs(DistributedFileSystem.java:859)
       at org.apache.hadoop.fs.FileSystem.mkdirs(FileSystem.java:1817)
       at org.apache.hadoop.fs.FileSystem.mkdirs(FileSystem.java:597)
       at 
org.apache.spark.deploy.yarn.Client.prepareLocalResources(Client.scala:224)
       at 
org.apache.spark.deploy.yarn.Client.createContainerLaunchContext(Client.scala:384)
       at 
org.apache.spark.deploy.yarn.Client.submitApplication(Client.scala:102)
       at 
org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.start(YarnClientSchedulerBackend.scala:58)
       at 
org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:141)
       at org.apache.spark.SparkContext.<init>(SparkContext.scala:381)
       at 
org.apache.zeppelin.spark.SparkInterpreter.createSparkContext(SparkInterpreter.java:301)
       at 
org.apache.zeppelin.spark.SparkInterpreter.getSparkContext(SparkInterpreter.java:146)
       at 
org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:423)
       at 
org.apache.zeppelin.interpreter.ClassloaderInterpreter.open(ClassloaderInterpreter.java:74)
       at 
org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:68)
       at 
org.apache.zeppelin.spark.PySparkInterpreter.getSparkInterpreter(PySparkInterpreter.java:353)
       at 
org.apache.zeppelin.spark.PySparkInterpreter.getJavaSparkContext(PySparkInterpreter.java:374)
       at 
org.apache.zeppelin.spark.PySparkInterpreter.open(PySparkInterpreter.java:140)
       at 
org.apache.zeppelin.interpreter.ClassloaderInterpreter.open(ClassloaderInterpreter.java:74)
       at 
org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:68)
       at 
org.apache.zeppelin.interpreter.LazyOpenInterpreter.interpret(LazyOpenInterpreter.java:92)
       at 
org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:275)
       at org.apache.zeppelin.scheduler.Job.run(Job.java:170)
       at 
org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:118)
       at 
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
       at java.util.concurrent.FutureTask.run(FutureTask.java:266)
       at 
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
       at 
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
       at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
       at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
       at java.lang.Thread.run(Thread.java:745)



Thanks
Karthik
From: moon soo Lee [mailto:m...@apache.org]
Sent: Sunday, July 12, 2015 9:05 AM
To: users@zeppelin.incubator.apache.org
Subject: Re: Yarn configuration on Zeppelin

Hi,

Did you set 'master' property to 'yarn-client' in 'Interpreter' menu?
You'll also need export HADOOP_CONF_DIR in bin/zeppelin-env.sh file.

Hope this helps.

Thanks,
moon

On Fri, Jul 10, 2015 at 1:26 PM Vadla, Karthik 
<karthik.va...@intel.com<mailto:karthik.va...@intel.com>> wrote:
Hi All,

I have built my zeppelin binaries with yarn profile. With below command in maven
mvn clean package -Pspark-1.3 -Ppyspark -Dhadoop.version=2.6.0-cdh5.4.2 
-Phadoop-2.6 -Pyarn –DskipTests

I have enabled hive-on-spark option in Cloudera manager and copied 
hive-site.xml to my zeppelin conf/ folder.
But still I’m not able to see any queries ran on spark with zeppelin notebook 
in my Yarn  ResourceManager Web UI 
(master)<http://master.trinity2.cluster.gao-nova:8088/> .

Do I need to do any specific configuration. ?

Reading some previous post I got some idea that zeppelin is using hive server-2 
. Can anyone help me where I can find configuration folder and what files I 
need to copy.

Appreciate your help

Thanks
Karthik Vadla

Reply via email to