On 09/03/16 06:05, Amos B. Elberg wrote:
Jeff you're correct that when Zeppelin is being professionally administered,
the administrator can take care of all of this.
But why create an additional system administration task? And what about users
without professional systems admins?
The only "benefit" to doing it that way, is we bundle a binary instead of source, when
the source is likely to never need updating. That doesn't seem like a "benefit" at all.
And in exchange for that, the cost is things like having to lock down R or
prevent package updates? That doesn't make much sense to me.
The question of which method has more overhead has been answered empirically:
this issue already broke 702, and there have been a whole series of revisions
to it to address various issues, with no end in sight.
Amos, You mention a few times that 'the issue broke 702...'. I don't see
when and why that particular approach broke anything.
I certainly experimented and reported that the rscala version alignment
is important, like any version of linux, jdk, scala, spark... dependency.
What I changed, after Moon comment, is the download of the jar at build
time instead of shipping the jar in the source tree. This approach has
the advantage of being able to define at build time which version you want.
Meanwhile, this part of 208 has been stable for six months, without a single
user issue.
On Mar 8, 2016, at 8:34 PM, Jeff Steinmetz <jeffrey.steinm...@gmail.com> wrote:
During my tests, I found the rScala installation its dependencies to be
manageable - even though it may not be ideal (i.e. the source is not included).
The Zeppelin build already needs to target the correct version of Spark +
Hadoop so for this exercise, I treat rScala similarly, as an additional build
and install consideration.
I’m looking at this through a specific lens:
“As a technology decision maker in a company, could I and would I deploy this
in our environment? Could I work with our Data Scientist team to implement its
usage? Would the Data Engineering and Data Science team that commonly use R
find it useful?”
Inadvertent rScala updates could be minimized with education, letting the users
know that R package management within a notebook should be avoided. Which
generally seems like a good idea regardless of how R-Zeppelin is implemented
since it’s a shared environment (you don’t want to break other users graphs
with an rGraph update or uninstall ggplot2, etc, etc.)
Even better - what if there was a zeppelin config that disabled
`install.packages()`, `remove.packages()` and `update.packages()`. This would
allow package installation to be carried out only by administrators or devops
outside of Zeppelin.
Although its not clear on the effort vs. benefit, I’m sure somebody crafty
could come up with a way around this with a convoluted Eval or running
something through the shell in Zeppelin.
R, Python and Scala all have pretty wide open door to parts of the underlying
operating system.
A 100% bullet proof way to locking “everything" down is a tough challenge.
----
Jeff Steinmetz
Principal Architect
Akili Interactive
www.akiliinteractive.com <http://www.akiliinteractive.com/>
On 3/8/16, 12:16 PM, "Amos B. Elberg" <amos.elb...@gmail.com> wrote:
Jeff - one of the problems with the rscala approach in 702 is it doesn't take
into account the R library. If rscala gets updated, the user will likely
download and update it automatically when they call update.packages(). The
result will be that the version of the rscala R package doesn't match the
version of the rscala jar, and the interpreter will fail. Or, if the jar is
also updated, it will simply break 702.
This has happened to 702 already-702 changed its library management because an
update to rscala broke the prior version. Actually though, every rscala update
is going to break 702.
Regarding return values from the interpreter, the norm in R is that the return
value of the last expression is shown in a native format. So, if the result is
a data frame, a dataframe should be visualized. If the result is an html
object, it should be interpreted as html by the browser. Do you disagree?
On Mar 8, 2016, at 2:52 PM, Jeff Steinmetz <jeffrey.steinm...@gmail.com> wrote:
RE 702:
I wanted to respond to Eric’s discussion from December 30.
I finally had some time to put aside a good chunk of dedicated, uninterrupted
time.
This means I had a chance to “really” dig into this with a Data Science R
developer hat on.
I also thought about this from a DevOps point of view (deploying in an EC2
cluster, standalone, locally, VM).
I tested it with a spark installation outside of the zeppelin build - as if it
was running on a cluster or standalone install.
I also had a chance to dig under the hood a bit, and explore what the
Java/Scala code in PR 702 is doing.
I like the simplicity of this PR (the source code and approach).
Works as expected, all graphic works, interactive charts works.
I also see your point about Rendering the text result vs TABLE plot when the R
interpreter result is a data frame.
To confirm - the approach is to use %sql to display it in a native Zeppelin
visualization.
Your approach makes sense, since this in line with how this works in other
Zeppelin work flows.
I suppose you could add an R interpreter function, such as:
z.R.showDFAsTable(fooDF) if we wanted to force the data frame into a %table
without having to jump to %sql (perhaps a nice addition in this or a future PR).
It’s GREAT that %r print('%html') works with the Zeppelin display system! (as
well as the other display system methods)
Regarding rscala jar. You have a profile that will allow us to sync up the
version rscala, so that makes sense as well.
This too worked as expected. I specifically installed rscala (as you describe
in your docs) in the VM with:
curl https://cran.r-project.org/src/contrib/Archive/rscala/rscala_1.0.6.tar.gz
-o /tmp/rscala_1.0.6.tar.gz
R CMD INSTALL /tmp/rscala_1.0.6.tar.gz
Installing rscala outside of the Zeppelin dependencies does seem to keep this
PR simpler, and reduces the licensing overhead required to get this PR through
(based on comments I see from others)
I would need to add the two rscala install lines above to PR#751 (I will add
this today)
https://github.com/apache/incubator-zeppelin/pull/751
Regarding the Interpreters. Just having %r as the our first interpreter
keyword makes sense. Loading knitr within the interpreter to enable rendering
(versus having a %knitr interpreter specifically) seems to keep things simple.
In summary - Looks good since everything in your sample R notebook (as well as
a few other tests I tried) worked for me using the VM script in PR#751.
The documentation also facilitated a smooth installation and allowed me to
create a repeatable script, that when paired with the VM worked as expected.
----
Jeff Steinmetz
Principal Architect
Akili Interactive
www.akiliinteractive.com <http://www.akiliinteractive.com/>
From
Eric Charles <e...@apache.org>
Subject
[DISCUSS] PR #208 - R Interpreter for Zeppelin
Date
Wed, 30 Dec 2015 14:04:33 GMT
Hi,
I had a look at https://github.com/apache/incubator-zeppelin/pull/208
(and related Github repo https://github.com/elbamos/Zeppelin-With-R [1])
Here are a few topics for discussion based on my experience developing
https://github.com/datalayer/zeppelin-R [2].
1. rscala jar not in Maven Repository
[1] copies the source (scala and R) code from rscala repo and
changes/extends/repackages it a bit. [2] declares the jar as system
scoped library. I recently had incompatibly issues between the 1.0.8
(the one you get since 2015-12-10 when you install rscala on your R
environment) and the 1.0.6 jar I am using part of the zeppelin-R build.
To avoid such issues, why not the user choosing the version via a
property at build time to fit the version he runs on its host? This will
also allow to benefit from the next rscala releases which fix bugs,
bring not features... This also means we don't have to copy the rscala
code in Zeppelin tree.
2. Interpreters
[1] proposes 2 interpreters %sparkr.r and %sparkr.knitr which are
implemented in their own module apart from the Spark one. To be aligned
the existing pyspark implementation, why not integrating the R code into
the Spark one? Any reason to keep 2 versions which does basically the
same? The unique magic keyword would then be %spark.r
3. Rendering TABLE plot when interpreter result is a dataframe
This may be confusing. What if I display a plot and simply want to print
the first 10 rows at the end of my code? To keep the same behavior as
the other interpreters, we could make this feature optional (disabled by
default, enabled via property).
Thx, Eric