Re: Is it possible to change the default port number 7077 for spark?

2015-07-13 Thread ashishdutt
Many thanks for your response. Regards, Ashish -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Is-it-possible-to-change-the-default-port-number-7077-for-spark-tp23774p23797.html Sent from the Apache Spark User List mailing list archive at Nabble.com.

Re: SparkR Error in sparkR.init(master=“local”) in RStudio

2015-07-13 Thread ashishdutt
Hi, Try this Sys.setenv(SPARK_HOME=C:\\spark-1.4.0) # The path to your spark installation .libPaths(c(file.path(Sys.getenv(SPARK_HOME), R, lib), .libPaths())) library(SparkR, lib.loc=C:\\spark-1.4.0\\lib) # The path to the lib folder in the spark location library(SparkR)

Re: sparkR

2015-07-13 Thread ashishdutt
Please can you explain how did you set this second step in windows environment? .libPaths(c(file.path(Sys.getenv(SPARK_HOME), R, lib), .libPaths())) I mean to ask where do I type this command, at R prompt or in command prompt? Thanks for your time. Regards, Ashish -- View this message in

Re: sparkR

2015-07-13 Thread ashishdutt
I had been facing this problem for a long time and this practically forced me to move to pyspark. This is what I tried after reading the posts here Sys.setenv(SPARK_HOME=C:\\spark-1.4.0) .libPaths(c(file.path(Sys.getenv(SPARK_HOME), R, lib), .libPaths())) library(SparkR,

Upgrade Spark-1.3.0 to Spark-1.4.0 in CDH5.4

2015-07-13 Thread ashishdutt
Hello all, The configuration of my cluster is as follows; # 4 noded cluster running on Centos OS 6.4 # spark-1.3.0 installed on all I would like to use SparkR shipped with spark-1.4.0. I checked Cloudera and find that the latest release CDH5.4 still does not have the spark-1.4.0. Forums like

Is it possible to change the default port number 7077 for spark?

2015-07-10 Thread ashishdutt
Hello all, In my lab a colleague installed and configured spark 1.3.0 on a 4 noded cluster on CDH5.4 environment. The default port number for our spark configuration is 7456. I have been trying to SSH to spark-master from using this port number but it fails every time giving error JVM is timed

Re: DLL load failed: %1 is not a valid win32 application on invoking pyspark

2015-07-09 Thread ashishdutt
Not really a clean solution but I solved the problem by reinstalling Anaconda -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/DLL-load-failed-1-is-not-a-valid-win32-application-on-invoking-pyspark-tp23733p23743.html Sent from the Apache Spark User List

DLL load failed: %1 is not a valid win32 application on invoking pyspark

2015-07-08 Thread ashishdutt
Hi, I get the error, DLL load failed: %1 is not a valid win32 application whenever I invoke pyspark. Attached is the screenshot of the same. Is there any way I can get rid of it. Still being new to PySpark and have had, a not so pleasant experience so far most probably because I am on a windows

JVM is not ready after 10 seconds

2015-07-06 Thread ashishdutt
Hi, I am trying to connect a worker to the master. The spark master is on cloudera manager and I know the master IP address and port number. I downloaded the spark binary for CDH4 on the worker machine and then when I try to invoke the command sc = sparkR.init(master=ip address:port number) I