Re: Zeppelin on EMR doesn't start?

2016-04-14 Thread Chris Miller
Indeed, one of the notebooks JSON files was corrupt... not sure how it
happened, but it wasn't an important one so I just deleted it.

Thanks!!


--
Chris Miller

On Mon, Apr 11, 2016 at 9:06 PM, Hyung Sung Shim <hss...@nflabs.com> wrote:

> Hello.
> It seems like one of your notebook json file has problem.
> Could you check your notebook files are valid? and What zeppelin and EMR
> version are you using?
>
> 2016-04-11 20:41 GMT+09:00 Chris Miller <cmiller11...@gmail.com>:
>
>> Hi. I'm running Zeppelin on the latest EMR distribution. Recently the
>> Zeppelin service will not start. I'm not sure what's going on... here is
>> what I see in the logs:
>>
>> 
>> WARN [2016-04-11 11:36:29,512] ({main}
>> ZeppelinConfiguration.java[create]:95) - Failed to load configuration,
>> proceeding with a default
>>  INFO [2016-04-11 11:36:29,865] ({main} ZeppelinServer.java[main]:101) -
>> Start zeppelin server
>>  INFO [2016-04-11 11:36:29,870] ({main} Server.java[doStart]:272) -
>> jetty-8.1.14.v20131031
>>  INFO [2016-04-11 11:36:30,073] ({main}
>> InterpreterFactory.java[init]:113) - Reading
>> /usr/lib/zeppelin/interpreter/spark
>>  INFO [2016-04-11 11:36:30,139] ({main}
>> InterpreterFactory.java[init]:130) - Interpreter spark.spark found.
>> class=org.apache.zeppelin.spark.SparkInterpreter
>>  INFO [2016-04-11 11:36:30,144] ({main}
>> InterpreterFactory.java[init]:130) - Interpreter spark.pyspark found.
>> class=org.apache.zeppelin.spark.PySparkInterpreter
>>  INFO [2016-04-11 11:36:30,146] ({main}
>> InterpreterFactory.java[init]:130) - Interpreter spark.sql found.
>> class=org.apache.zeppelin.spark.SparkSqlInterpreter
>>  INFO [2016-04-11 11:36:30,149] ({main}
>> InterpreterFactory.java[init]:130) - Interpreter spark.dep found.
>> class=org.apache.zeppelin.spark.DepInterpreter
>>  INFO [2016-04-11 11:36:30,171] ({main}
>> InterpreterFactory.java[init]:113) - Reading
>> /usr/lib/zeppelin/interpreter/sh
>>  INFO [2016-04-11 11:36:30,177] ({main}
>> InterpreterFactory.java[init]:130) - Interpreter sh.sh found.
>> class=org.apache.zeppelin.shell.ShellInterpreter
>>  INFO [2016-04-11 11:36:30,182] ({main}
>> InterpreterFactory.java[init]:113) - Reading
>> /usr/lib/zeppelin/interpreter/lens
>>  INFO [2016-04-11 11:36:30,253] ({main}
>> InterpreterFactory.java[init]:130) - Interpreter lens.lens found.
>> class=org.apache.zeppelin.lens.LensInterpreter
>>  INFO [2016-04-11 11:36:30,256] ({main}
>> InterpreterFactory.java[init]:113) - Reading
>> /usr/lib/zeppelin/interpreter/psql
>>  INFO [2016-04-11 11:36:30,270] ({main}
>> InterpreterFactory.java[init]:130) - Interpreter psql.sql found.
>> class=org.apache.zeppelin.postgresql.PostgreSqlInterpreter
>>  INFO [2016-04-11 11:36:30,270] ({main}
>> InterpreterFactory.java[init]:113) - Reading
>> /usr/lib/zeppelin/interpreter/hive
>>  INFO [2016-04-11 11:36:30,322] ({main}
>> InterpreterFactory.java[init]:130) - Interpreter hive.hql found.
>> class=org.apache.zeppelin.hive.HiveInterpreter
>>  INFO [2016-04-11 11:36:30,329] ({main}
>> InterpreterFactory.java[init]:113) - Reading
>> /usr/lib/zeppelin/interpreter/kylin
>>  INFO [2016-04-11 11:36:30,342] ({main}
>> InterpreterFactory.java[init]:130) - Interpreter kylin.kylin found.
>> class=org.apache.zeppelin.kylin.KylinInterpreter
>>  INFO [2016-04-11 11:36:30,343] ({main}
>> InterpreterFactory.java[init]:113) - Reading
>> /usr/lib/zeppelin/interpreter/cassandra
>>  INFO [2016-04-11 11:36:30,361] ({main}
>> CassandraInterpreter.java[]:154) - Bootstrapping Cassandra
>> Interpreter
>>  INFO [2016-04-11 11:36:30,362] ({main}
>> InterpreterFactory.java[init]:130) - Interpreter cassandra.cassandra found.
>> class=org.apache.zeppelin.cassandra.CassandraInterpreter
>>  INFO [2016-04-11 11:36:30,363] ({main}
>> InterpreterFactory.java[init]:113) - Reading
>> /usr/lib/zeppelin/interpreter/tajo
>>  INFO [2016-04-11 11:36:30,376] ({main}
>> InterpreterFactory.java[init]:130) - Interpreter tajo.tql found.
>> class=org.apache.zeppelin.tajo.TajoInterpreter
>>  INFO [2016-04-11 11:36:30,380] ({main}
>> InterpreterFactory.java[init]:113) - Reading
>> /usr/lib/zeppelin/interpreter/md
>>  INFO [2016-04-11 11:36:30,385] ({main}
>> InterpreterFactory.java[init]:130) - Interpreter md.md found.
>> class=org.apache.zeppelin.markdown.Markdown
>>  INFO [2016-04-11 11:36:30,392] ({main}
>> InterpreterFactory.java[init]:113) - Reading
>> /usr/lib/zeppelin/interpreter/flink
>>  INFO [2016-04-11 11:36:30,443] ({main}
&g

Re: Multi-User Zeppelin Deployment?

2016-03-25 Thread Chris Miller
Thank you for your detailed reply!


--
Chris Miller

On Thu, Mar 10, 2016 at 1:40 AM, moon soo Lee <m...@apache.org> wrote:

> Hi Chris Miller,
>
>
>- If one user is running a job with an interpreter, can another user
>simultaneously run a job (such as, in another notebook) with the same
>interpreter?
>
> Short answer is yes, but it depends.
> Long answer is, first, it depends on which scheduler interpreter
> implementation uses among FIFO and Parallel. Interpreter uses Parallel
> scheduler (e.g. spark sql interpreter, shell interpreter, etc) will able to
> run simultaneously. Interpreter uses FIFO scheduler will not (e.g. spark
> interpreter, etc)
>
> Recently, http://issues.apache.org/jira/browse/ZEPPELIN-513 is resolved.
> Which allows Interpreter use FIFO scheduler run simultaneously, by creating
> interpreter instance per notebook.
>
>
>- Does Zeppelin have any kind of user authentication capabilities?
>
>
> Master branch has authentication capabilities based on Apache Shiro.
> https://issues.apache.org/jira/browse/ZEPPELIN-548
>
>
>
>- Can I give users either read-only or no access to particular
>notebooks but also let users create their own notebooks that only they can
>see?
>
> Recently https://github.com/apache/incubator-zeppelin/pull/681 has been
> merged to master branch. It lets users adjust read/write/execute permission
> for each notebook.
>
>
>
>- Can I have jobs run as the logged-in user rather than a generic
>"zeppelin" user so things like HDFS permissions will apply?
>
>
> User impersonation is on the roadmap
> https://cwiki.apache.org/confluence/display/ZEPPELIN/Zeppelin+Roadmap,
> but not working in progress, yet.
>
> Hope this helps.
>
> Thanks,
> moon
>
>
> On Wed, Mar 9, 2016 at 1:08 AM Chris Miller <cmiller11...@gmail.com>
> wrote:
>
>> Hi,
>>
>> I want to deploy Zeppelin so that multiple users in our organization can
>> use it concurrently. I have a few questions:
>>
>>- If one user is running a job with an interpreter, can another user
>>simultaneously run a job (such as, in another notebook) with the same
>>interpreter?
>>- Does Zeppelin have any kind of user authentication capabilities?
>>- Can I give users either read-only or no access to particular
>>notebooks but also let users create their own notebooks that only they can
>>see?
>>- Can I have jobs run as the logged-in user rather than a generic
>>"zeppelin" user so things like HDFS permissions will apply?
>>
>> For those of you using Zeppelin in production, any other deployment or
>> configuration tips?
>> --
>> Chris Miller
>>
>


Re: Spark + Zeppelin on EC2

2016-03-25 Thread Chris Miller
Curious about this too... I'll be moving Zeppelin off to its own box in the
near future. If you figure this out, post your resolution here.

--
Chris Miller

On Sat, Mar 26, 2016 at 12:54 AM, Marcin Pilarczyk <
marcin.pilarc...@interia.pl> wrote:

> Guys,
>
> I'm trying to switch my zeppelin instance (0.6 snapshot) from the spark
> instance installed on very same machine onto ec2 created spark. Both
> versions of spark are 1.5.2.
>
> I've just created a test instance in EC2, I can submit jobs or use spark
> shell. I have revised logs each and every worker up and running, master
> alive. So far so good.
>
> The next step is to switch the zeppelin on the newly created spark. I'm
> changing two places: zeppelin-env.sh and URL in the interpreter settings.
> I'm SURE these settings are pointing the new instance.
>
> Next step. I'm stopping the spark instance installed together with
> zeppelin.
>
> Final step, zeppelin is restarted, settings are checked. Somehow any
> paragraph that requires computations can't be completed. Master logs are
> ok, in the slave log however I can find a following error:
>
> 16/03/25 12:42:25 INFO Remoting: Starting remoting
> 16/03/25 12:42:25 INFO Remoting: Remoting started; listening on addresses
> :[akka.tcp://driverPropsFetcher@172.31.40.27:36098]
> 16/03/25 12:42:25 INFO util.Utils: Successfully started service
> 'driverPropsFetcher' on port 36098.
> 16/03/25 12:43:28 WARN Remoting: Tried to associate with unreachable
> remote address [akka.tcp://sparkDriver@172.31.41.186:46358]. Address is
> now gated for 5000 ms, all messages to this address will be delivered to
> dead letters.$
> Exception in thread "main" akka.actor.ActorNotFound: Actor not found for:
> ActorSelection[Anchor(akka.tcp://sparkDriver@172.31.41.186:46358/),
> Path(/user/CoarseGrainedScheduler)]
> at
> akka.actor.ActorSelection$$anonfun$resolveOne$1.apply(ActorSelection.scala:65)
> at
> akka.actor.ActorSelection$$anonfun$resolveOne$1.apply(ActorSelection.scala:63)
> at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:32)
> at
> akka.dispatch.BatchingExecutor$Batch$$anonfun$run$1.processBatch$1(BatchingExecutor.scala:67)
> at
> akka.dispatch.BatchingExecutor$Batch$$anonfun$run$1.apply$mcV$sp(BatchingExecutor.scala:82)
> at
> akka.dispatch.BatchingExecutor$Batch$$anonfun$run$1.apply(BatchingExecutor.scala:59)
> at
> akka.dispatch.BatchingExecutor$Batch$$anonfun$run$1.apply(BatchingExecutor.scala:59)
> at
> scala.concurrent.BlockContext$.withBlockContext(BlockContext.scala:72)
> at
> akka.dispatch.BatchingExecutor$Batch.run(BatchingExecutor.scala:58)
> at
> akka.dispatch.ExecutionContexts$sameThreadExecutionContext$.unbatchedExecute(Future.scala:74)
> at
> akka.dispatch.BatchingExecutor$class.execute(BatchingExecutor.scala:110)
> at
> akka.dispatch.ExecutionContexts$sameThreadExecutionContext$.execute(Future.scala:73)
> at
> scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
> at
> scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
> at akka.pattern.PromiseActorRef.$bang(AskSupport.scala:267)
> at akka.actor.EmptyLocalActorRef.specialHandle(ActorRef.scala:508)
> at akka.actor.DeadLetterActorRef.specialHandle(ActorRef.scala:541)
> at akka.actor.DeadLetterActorRef.$bang(ActorRef.scala:531)
> at
> akka.remote.RemoteActorRefProvider$RemoteDeadLetterActorRef.$bang(RemoteActorRefProvider.scala:87)
> at akka.remote.EndpointWriter.postStop(Endpoint.scala:561)
> at akka.actor.Actor$class.aroundPostStop(Actor.scala:475)
> at akka.remote.EndpointActor.aroundPostStop(Endpoint.scala:415)
> at
> akka.actor.dungeon.FaultHandling$class.akka$actor$dungeon$FaultHandling$$finishTerminate(FaultHandling.scala:210)
> at
> akka.actor.dungeon.FaultHandling$class.terminate(FaultHandling.scala:172)
> at akka.actor.ActorCell.terminate(ActorCell.scala:369)
> at akka.actor.ActorCell.invokeAll$1(ActorCell.scala:462)
> at akka.actor.ActorCell.systemInvoke(ActorCell.scala:478)
> at
> akka.dispatch.Mailbox.processAllSystemMessages(Mailbox.scala:263)
> at akka.dispatch.Mailbox.run(Mailbox.scala:219)
> at
> akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:393)
> at
> scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
> at
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
> at
> scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
> at
> sc

Re: Interpreter dependency not loading?

2016-03-09 Thread Chris Miller
Yeah, that isn't very clear. I'll improve it.


--
Chris Miller

On Wed, Mar 9, 2016 at 5:12 PM, mina lee <mina...@apache.org> wrote:

> Glad to hear that it works!
> Actually, there is document
> https://zeppelin.incubator.apache.org/docs/0.5.5-incubating/interpreter/spark.html
>  in
> dependency management section, but obviously it seems hard to find for new
> users. So feel free to improve it.
>
>
> On Wed, Mar 9, 2016 at 6:05 PM Chris Miller <cmiller11...@gmail.com>
> wrote:
>
>> Oh, I see. Yeah, that's not documented... no wonder it's confusing. I'll
>> open a PR with some improvements to the documentation for this case when I
>> have a moment.
>>
>> Changing spark-default.conf as you suggested indeed worked. Thanks!
>>
>>
>> --
>> Chris Miller
>>
>> On Wed, Mar 9, 2016 at 10:04 AM, mina lee <mina...@apache.org> wrote:
>>
>>> Hi Chris,
>>>
>>> there are several ways to load dependencies to Zeppelin 0.5.5.
>>> Using %dep is one of them.
>>> If you want do it by setting spark.jars.packages property, proper way of
>>> doing it is editing your SPARK_HOME/conf/spark-default.conf
>>> and adding below line.(I assume that you set SPARK_HOME in
>>> ZEPPELIN_HOME/conf/zeppelin-env.sh)
>>>
>>> spark.jars.packages   org.apache.avro:avro:1.8.0,org.
>>> joda:joda-convert:1.8.1
>>>
>>> The reason you can import avro dependency is that spark assembly already
>>> includes avro dependencies, not because you added it in Zeppelin
>>> interpreter setting.
>>>
>>> You can add dependencies via GUI with the latest master
>>> branch(0.6.0-incubating-SNAPSHOT) which is experimental at the moment.
>>> Please let me know it answers your question.
>>>
>>> Regards,
>>> Mina
>>>
>>> On Wed, Mar 9, 2016 at 1:41 AM Chris Miller <cmiller11...@gmail.com>
>>> wrote:
>>>
>>>> Hi,
>>>>
>>>> I have a strange situation going on. I'm running Zeppelin 0.5.5 and
>>>> Spark 1.6.0 (on Amazon EMR). I added this property to the interpreter
>>>> settings (and restarted it):
>>>>
>>>>
>>>> spark.jars.packages: org.apache.avro:avro:1.8.0,org.joda:joda-convert:1.8.1
>>>>
>>>> The avro dependency loads fine and I'm able to import and use it.
>>>> However, if I try to import something in the joda-convert package (such as,
>>>> org.joda.convert.FromString), I get an error that "error: object convert is
>>>> not a member of package org.joda".
>>>>
>>>> If I run the spark-shell from the CLI and include the same string above
>>>> in the --package parameter, I'm able to import joda-convert just fine.
>>>> Also, if I restart the interpreter and manually import the dependency with
>>>> z.load(), it also works fine:
>>>>
>>>> %dep
>>>> z.load("org.joda:joda-convert:1.8.1")
>>>>
>>>> So, what's going on here?
>>>>
>>>> --
>>>> Chris Miller
>>>>
>>>
>>


Multi-User Zeppelin Deployment?

2016-03-09 Thread Chris Miller
Hi,

I want to deploy Zeppelin so that multiple users in our organization can
use it concurrently. I have a few questions:

   - If one user is running a job with an interpreter, can another user
   simultaneously run a job (such as, in another notebook) with the same
   interpreter?
   - Does Zeppelin have any kind of user authentication capabilities?
   - Can I give users either read-only or no access to particular notebooks
   but also let users create their own notebooks that only they can see?
   - Can I have jobs run as the logged-in user rather than a generic
   "zeppelin" user so things like HDFS permissions will apply?

For those of you using Zeppelin in production, any other deployment or
configuration tips?
--
Chris Miller


Re: Interpreter dependency not loading?

2016-03-09 Thread Chris Miller
Oh, I see. Yeah, that's not documented... no wonder it's confusing. I'll
open a PR with some improvements to the documentation for this case when I
have a moment.

Changing spark-default.conf as you suggested indeed worked. Thanks!


--
Chris Miller

On Wed, Mar 9, 2016 at 10:04 AM, mina lee <mina...@apache.org> wrote:

> Hi Chris,
>
> there are several ways to load dependencies to Zeppelin 0.5.5.
> Using %dep is one of them.
> If you want do it by setting spark.jars.packages property, proper way of
> doing it is editing your SPARK_HOME/conf/spark-default.conf
> and adding below line.(I assume that you set SPARK_HOME in
> ZEPPELIN_HOME/conf/zeppelin-env.sh)
>
> spark.jars.packages   org.apache.avro:avro:1.8.0,org.
> joda:joda-convert:1.8.1
>
> The reason you can import avro dependency is that spark assembly already
> includes avro dependencies, not because you added it in Zeppelin
> interpreter setting.
>
> You can add dependencies via GUI with the latest master
> branch(0.6.0-incubating-SNAPSHOT) which is experimental at the moment.
> Please let me know it answers your question.
>
> Regards,
> Mina
>
> On Wed, Mar 9, 2016 at 1:41 AM Chris Miller <cmiller11...@gmail.com>
> wrote:
>
>> Hi,
>>
>> I have a strange situation going on. I'm running Zeppelin 0.5.5 and Spark
>> 1.6.0 (on Amazon EMR). I added this property to the interpreter settings
>> (and restarted it):
>>
>>
>> spark.jars.packages: org.apache.avro:avro:1.8.0,org.joda:joda-convert:1.8.1
>>
>> The avro dependency loads fine and I'm able to import and use it.
>> However, if I try to import something in the joda-convert package (such as,
>> org.joda.convert.FromString), I get an error that "error: object convert is
>> not a member of package org.joda".
>>
>> If I run the spark-shell from the CLI and include the same string above
>> in the --package parameter, I'm able to import joda-convert just fine.
>> Also, if I restart the interpreter and manually import the dependency with
>> z.load(), it also works fine:
>>
>> %dep
>> z.load("org.joda:joda-convert:1.8.1")
>>
>> So, what's going on here?
>>
>> --
>> Chris Miller
>>
>


Interpreter dependency not loading?

2016-03-08 Thread Chris Miller
Hi,

I have a strange situation going on. I'm running Zeppelin 0.5.5 and Spark
1.6.0 (on Amazon EMR). I added this property to the interpreter settings
(and restarted it):

spark.jars.packages: org.apache.avro:avro:1.8.0,org.joda:joda-convert:1.8.1

The avro dependency loads fine and I'm able to import and use it. However,
if I try to import something in the joda-convert package (such as,
org.joda.convert.FromString), I get an error that "error: object convert is
not a member of package org.joda".

If I run the spark-shell from the CLI and include the same string above in
the --package parameter, I'm able to import joda-convert just fine. Also,
if I restart the interpreter and manually import the dependency with
z.load(), it also works fine:

%dep
z.load("org.joda:joda-convert:1.8.1")

So, what's going on here?

--
Chris Miller


Notebook Editor 2-space Indentation?

2016-03-07 Thread Chris Miller
Hi,

Is there a way to configure the notebook text editor to use 2-space
indentation when writing Scala (i.e., %spark)?

--
Chris Miller