Thanks Shivaram. You got me wondering about the path so I put it in full
and it worked. R does not, of course, expand a "~".

On Thu, Aug 20, 2015 at 4:35 PM, Shivaram Venkataraman <
shiva...@eecs.berkeley.edu> wrote:

> Can you check if the file
> `~/software/spark-1.4.1-bin-hadoop2.4/bin/spark-submit` exists ? The
> error message seems to indicate it is trying to pick up Spark from
> that location and can't seem to find Spark installed there.
>
> Thanks
> Shivaram
>
> On Thu, Aug 20, 2015 at 3:30 PM, Deborah Siegel
> <deborah.sie...@gmail.com> wrote:
> > Hello,
> >
> > I have previously successfully run SparkR in RStudio, with:
> >
> >>Sys.setenv(SPARK_HOME="~/software/spark-1.4.1-bin-hadoop2.4")
> >>.libPaths(c(file.path(Sys.getenv("SPARK_HOME"), "R", "lib"),
> .libPaths()))
> >>library(SparkR)
> >>sc <- sparkR.init(master="local[2]",appName="SparkR-example")
> >
> >
> > Then I tried putting some of it into an .Rprofile. It seemed to work to
> load
> > the paths and SparkR, but I got an error when trying to create the sc. I
> > then removed my .Rprofile, as well as .rstudio-desktop. However, I still
> > cannot create the sc. Here is the error
> >
> >> sc <- sparkR.init(master="local[2]",appName="SparkR-example")
> > Launching java with spark-submit command
> > ~/software/spark-1.4.1-bin-hadoop2.4/bin/spark-submit   sparkr-shell
> >
> /var/folders/p7/k1bpgmx93yd6pjq7dzf35gk80000gn/T//RtmpOitA28/backend_port23377046db
> > sh: ~/software/spark-1.4.1-bin-hadoop2.4/bin/spark-submit: No such file
> or
> > directory
> > Error in sparkR.init(master = "local[2]", appName = "SparkR-example") :
> > JVM is not ready after 10 seconds
> >
> > I suspected there was an incomplete process or something. I checked for
> any
> > running R or Java processes and there were none. Has someone seen this
> type
> > of error? I have the same error in both RStudio and in R shell (but not
> > sparkR wrapper).
> >
> > Thanks,
> > Deb
> >
> >
>

Reply via email to