All is what’s by default Properties name value args master yarn-client spark.app.name Zeppelin spark.cores.max spark.executor.memory 512m spark.home /usr/hdp/current/spark-client/ spark.yarn.jar hdfs:///apps/zeppelin/zeppelin-spark-0.5.5-SNAPSHOT.jar zeppelin.dep.localrepo local-repo zeppelin.pyspark.python python zeppelin.spark.concurrentSQL false zeppelin.spark.maxResult 1000 zeppelin.spark.useHiveContext true
> On Dec 3, 2015, at 11:51 AM, Felix Cheung <felixcheun...@hotmail.com> wrote: > > Could you send us your configuration of the Spark Interpreter in Zeppelin? > > It can see how both jobs can be long lived in Spark & hive/Tez but they > should not block one or another. > > > _____________________________ > From: Will Du <will...@gmail.com <mailto:will...@gmail.com>> > Sent: Thursday, December 3, 2015 2:15 AM > Subject: Re: zeppelin job is running all the time > To: <users@zeppelin.incubator.apache.org > <mailto:users@zeppelin.incubator.apache.org>> > > > When I run spark-shell and hive cml, everything is good. > One more thing I find is that when I run the hive CML, there is single hive > job running (seen from hue job browser) all the time until I exit from hive > cml. > No matter how many HQL I submit, it is still single job there. I think it is > because of using Tez. Is Tez having conflict with zeppelin? > > On Dec 3, 2015, at 2:15 AM, Felix Cheung < felixcheun...@hotmail.com > <mailto:felixcheun...@hotmail.com>> wrote: > > I don't know enough about HDP, but there should be a way to check user queue > in YARN? > Spark job shouldn't affect Hive job though. Have you tried running > spark-shell (--master yarn-client) and a Hive job at the same time? > > From: will...@gmail.com <mailto:will...@gmail.com> > Subject: Re: zeppelin job is running all the time > Date: Tue, 1 Dec 2015 22:56:55 -0500 > To: users@zeppelin.incubator.apache.org > <mailto:users@zeppelin.incubator.apache.org> > > Do I have to stop zeppelin to make hive job running? > I am thinking to have one notebook have both spark and hive coding running. > > In addition, I am quite sure the first job in the picture is hive job. The > 2nd one is zeppelin running spark. I wonder how to unblock hive job. > > On Dec 1, 2015, at 10:45 PM, Jeff Zhang < zjf...@gmail.com > <mailto:zjf...@gmail.com>> wrote: > > >>> the status of hive query becomes from ACCEPT to RUNNING > The status you see on the hue is the yarn app status which is used for spark. > It should be running forever until you shutdown it in zeppelin > > On Wed, Dec 2, 2015 at 11:42 AM, Will Du <will...@gmail.com > <mailto:will...@gmail.com>> wrote: > The latest HDP sandbox 2.3.2 > > On Dec 1, 2015, at 10:38 PM, Jeff Zhang < zjf...@gmail.com > <mailto:zjf...@gmail.com>> wrote: > > Which version of HDP do you use ? > > On Wed, Dec 2, 2015 at 11:23 AM, Will Du <will...@gmail.com > <mailto:will...@gmail.com>> wrote: > I have assigned an dedicate yarn queue to spark, the status of hive query > becomes from ACCEPT to RUNNING. However, it sees running forever still. > Everything is else is in default config of HDP sandbox. Do I need to set > something else? The browser I saw status is from hue. > > Thanks, > wd > > On Nov 30, 2015, at 12:40 AM, Rick Moritz <rah...@gmail.com > <mailto:rah...@gmail.com>> wrote: > > To explain the previous reply: the SparkContext created by Zeppelin is > persistent and independent of whether it's currently processing a paragraph > or not. Therefore the Zeppelin job will claim all ressources assigned to it > until Zeppelin is stopped. > > On Mon, Nov 30, 2015 at 4:20 AM, Jeff Zhang <zjf...@gmail.com > <mailto:zjf...@gmail.com>> wrote: > I assume this is yarn app Job Browser. How many executors do you specify for > your zeppelin yarn app ? It seems your zeppelin yarn app consume all the > resources so that it block other applications. > > > > On Mon, Nov 30, 2015 at 11:10 AM, Will Du <will...@gmail.com > <mailto:will...@gmail.com>> wrote: > Hi folks, > I am running a simple scala word count from zeppelin in HDP sandbox. The job > is successful with expected result. But the job of zeppelin is shown in job > status of hue forever. It seems block the my hive job. Does anyone know why? > thanks, > wd > <PastedGraphic-1.tiff> > > > > -- > Best Regards > > Jeff Zhang > > > > > > -- > Best Regards > > Jeff Zhang > > > > > -- > Best Regards > > Jeff Zhang > > > >