Many thanks for your response.
Regards,
Ashish
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/Is-it-possible-to-change-the-default-port-number-7077-for-spark-tp23774p23797.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.
Hi,
Try this
Sys.setenv(SPARK_HOME=C:\\spark-1.4.0) # The path to your spark
installation
.libPaths(c(file.path(Sys.getenv(SPARK_HOME), R, lib), .libPaths()))
library(SparkR, lib.loc=C:\\spark-1.4.0\\lib) # The path to the lib
folder in the spark location
library(SparkR)
Please can you explain how did you set this second step in windows
environment?
.libPaths(c(file.path(Sys.getenv(SPARK_HOME), R, lib), .libPaths()))
I mean to ask where do I type this command, at R prompt or in command
prompt?
Thanks for your time.
Regards,
Ashish
--
View this message in
I had been facing this problem for a long time and this practically forced me
to move to pyspark.
This is what I tried after reading the posts here
Sys.setenv(SPARK_HOME=C:\\spark-1.4.0)
.libPaths(c(file.path(Sys.getenv(SPARK_HOME), R, lib), .libPaths()))
library(SparkR,
Hello all,
The configuration of my cluster is as follows;
# 4 noded cluster running on Centos OS 6.4
# spark-1.3.0 installed on all
I would like to use SparkR shipped with spark-1.4.0. I checked Cloudera and
find that the latest release CDH5.4 still does not have the spark-1.4.0.
Forums like
Hello all,
In my lab a colleague installed and configured spark 1.3.0 on a 4 noded
cluster on CDH5.4 environment. The default port number for our spark
configuration is 7456. I have been trying to SSH to spark-master from using
this port number but it fails every time giving error JVM is timed
Not really a clean solution but I solved the problem by reinstalling Anaconda
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/DLL-load-failed-1-is-not-a-valid-win32-application-on-invoking-pyspark-tp23733p23743.html
Sent from the Apache Spark User List
Hi,
I get the error, DLL load failed: %1 is not a valid win32 application
whenever I invoke pyspark. Attached is the screenshot of the same.
Is there any way I can get rid of it. Still being new to PySpark and have
had, a not so pleasant experience so far most probably because I am on a
windows
Hi,
I am trying to connect a worker to the master. The spark master is on
cloudera manager and I know the master IP address and port number.
I downloaded the spark binary for CDH4 on the worker machine and then when I
try to invoke the command
sc = sparkR.init(master=ip address:port number) I