I forgot to mention that there is also Zeppelin and jove-notebook but I
haven't got any experience with those yet.


*Irfan Ahmad*
CTO | Co-Founder | *CloudPhysics* <http://www.cloudphysics.com>
Best of VMworld Finalist
Best Cloud Management Award
NetworkWorld 10 Startups to Watch
EMA Most Notable Vendor

On Wed, Mar 18, 2015 at 8:01 PM, Irfan Ahmad <ir...@cloudphysics.com> wrote:

> Hi David,
>
> W00t indeed and great questions. On the notebook front, there are two
> options depending on what you are looking for. You can either go with
> iPython 3 with Spark-kernel as a backend or you can use spark-notebook.
> Both have interesting tradeoffs.
>
> If you have looking for a single notebook platform for your data
> scientists that has R, Python as well as a Spark Shell, you'll likely want
> to go with iPython + Spark-kernel. Downsides with the spark-kernel project
> are that data visualization isn't quite there yet, early days for
> documentation and blogs/etc. Upside is that R and Python work beautifully
> and that the ipython committers are super-helpful.
>
> If you are OK with a primarily spark/scala experience, then I suggest you
> with spark-notebook. Upsides are that the project is a little further
> along, visualization support is better than spark-kernel (though not as
> good as iPython with Python) and the committer is awesome with help.
> Downside is that you won't get R and Python.
>
> FWIW: I'm using both at the moment!
>
> Hope that helps.
>
>
> *Irfan Ahmad*
> CTO | Co-Founder | *CloudPhysics* <http://www.cloudphysics.com>
> Best of VMworld Finalist
> Best Cloud Management Award
> NetworkWorld 10 Startups to Watch
> EMA Most Notable Vendor
>
> On Wed, Mar 18, 2015 at 5:45 PM, davidh <dav...@annaisystems.com> wrote:
>
>> hi all, I've been DDGing, Stack Overflowing, Twittering, RTFMing, and
>> scanning through this archive with only moderate success. in other words
>> --
>> my way of saying sorry if this is answered somewhere obvious and I missed
>> it
>> :-)
>>
>> i've been tasked with figuring out how to connect Notebook, Spark, and
>> Accumulo together. The end user will do her work via notebook. thus far,
>> I've successfully setup a Vagrant image containing Spark, Accumulo, and
>> Hadoop. I was able to use some of the Accumulo example code to create a
>> table populated with data, create a simple program in scala that, when
>> fired
>> off to Spark via spark-submit, connects to accumulo and prints the first
>> ten
>> rows of data in the table. so w00t on that - but now I'm left with more
>> questions:
>>
>> 1) I'm still stuck on what's considered 'best practice' in terms of
>> hooking
>> all this together. Let's say Sally, a  user, wants to do some analytic
>> work
>> on her data. She pecks the appropriate commands into notebook and fires
>> them
>> off. how does this get wired together on the back end? Do I, from
>> notebook,
>> use spark-submit to send a job to spark and let spark worry about hooking
>> into accumulo or is it preferable to create some kind of open stream
>> between
>> the two?
>>
>> 2) if I want to extend spark's api, do I need to first submit an endless
>> job
>> via spark-submit that does something like what this gentleman describes
>> <http://blog.madhukaraphatak.com/extending-spark-api>  ? is there an
>> alternative (other than refactoring spark's source) that doesn't involve
>> extending the api via a job submission?
>>
>> ultimately what I'm looking for help locating docs, blogs, etc that may
>> shed
>> some light on this.
>>
>> t/y in advance!
>>
>> d
>>
>>
>>
>> --
>> View this message in context:
>> http://apache-spark-user-list.1001560.n3.nabble.com/iPython-Notebook-Spark-Accumulo-best-practice-tp22137.html
>> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>>
>> ---------------------------------------------------------------------
>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
>> For additional commands, e-mail: user-h...@spark.apache.org
>>
>>
>

Reply via email to