Thanks! In your demo video, were you using RStudio to hit a separate EC2 Spark cluster? I noticed that it appeared your browser that you were using EC2 at that time, so I was just curious. It appears that might be one of the possible workarounds - fire up a separate EC2 instance with RStudio Server that initializes the spark context against a separate Spark cluster.
> On Jun 26, 2015, at 11:46 AM, Shivaram Venkataraman > <shiva...@eecs.berkeley.edu> wrote: > > We don't have a documented way to use RStudio on EC2 right now. We have a > ticket open at https://issues.apache.org/jira/browse/SPARK-8596 > <https://issues.apache.org/jira/browse/SPARK-8596> to discuss work-arounds > and potential solutions for this. > > Thanks > Shivaram > > On Fri, Jun 26, 2015 at 6:27 AM, RedOakMark <m...@redoakstrategic.com > <mailto:m...@redoakstrategic.com>> wrote: > Good morning, > > I am having a bit of trouble finalizing the installation and usage of the > newest Spark version 1.4.0, deploying to an Amazon EC2 instance and using > RStudio to run on top of it. > > Using these instructions ( > http://spark.apache.org/docs/latest/ec2-scripts.html > <http://spark.apache.org/docs/latest/ec2-scripts.html> > <http://spark.apache.org/docs/latest/ec2-scripts.html > <http://spark.apache.org/docs/latest/ec2-scripts.html>> ) we can fire up an > EC2 instance (which we have been successful doing - we have gotten the > cluster to launch from the command line without an issue). Then, I > installed RStudio Server on the same EC2 instance (the master) and > successfully logged into it (using the test/test user) through the web > browser. > > This is where I get stuck - within RStudio, when I try to reference/find the > folder that SparkR was installed, to load the SparkR library and initialize > a SparkContext, I get permissions errors on the folders, or the library > cannot be found because I cannot find the folder in which the library is > sitting. > > Has anyone successfully launched and utilized SparkR 1.4.0 in this way, with > RStudio Server running on top of the master instance? Are we on the right > track, or should we manually launch a cluster and attempt to connect to it > from another instance running R? > > Thank you in advance! > > Mark > > > > -- > View this message in context: > http://apache-spark-user-list.1001560.n3.nabble.com/Spark-1-4-0-Using-SparkR-on-EC2-Instance-tp23506.html > > <http://apache-spark-user-list.1001560.n3.nabble.com/Spark-1-4-0-Using-SparkR-on-EC2-Instance-tp23506.html> > Sent from the Apache Spark User List mailing list archive at Nabble.com. > > --------------------------------------------------------------------- > To unsubscribe, e-mail: user-unsubscr...@spark.apache.org > <mailto:user-unsubscr...@spark.apache.org> > For additional commands, e-mail: user-h...@spark.apache.org > <mailto:user-h...@spark.apache.org> > >