Hello, I have previously successfully run SparkR in RStudio, with:
>Sys.setenv(SPARK_HOME="~/software/spark-1.4.1-bin-hadoop2.4") >.libPaths(c(file.path(Sys.getenv("SPARK_HOME"), "R", "lib"), .libPaths())) >library(SparkR) >sc <- sparkR.init(master="local[2]",appName="SparkR-example") Then I tried putting some of it into an .Rprofile. It seemed to work to load the paths and SparkR, but I got an error when trying to create the sc. I then removed my .Rprofile, as well as .rstudio-desktop. However, I still cannot create the sc. Here is the error > sc <- sparkR.init(master="local[2]",appName="SparkR-example") Launching java with spark-submit command ~/software/spark-1.4.1-bin-hadoop2.4/bin/spark-submit sparkr-shell /var/folders/p7/k1bpgmx93yd6pjq7dzf35gk80000gn/T//RtmpOitA28/backend_port23377046db sh: ~/software/spark-1.4.1-bin-hadoop2.4/bin/spark-submit: No such file or directory Error in sparkR.init(master = "local[2]", appName = "SparkR-example") : JVM is not ready after 10 seconds I suspected there was an incomplete process or something. I checked for any running R or Java processes and there were none. Has someone seen this type of error? I have the same error in both RStudio and in R shell (but not sparkR wrapper). Thanks, Deb