Re: Is it possible to change the default port number 7077 for spark?

2015-07-13 Thread ashishdutt
Many thanks for your response.
Regards,
Ashish



--
View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/Is-it-possible-to-change-the-default-port-number-7077-for-spark-tp23774p23797.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.

-
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org



Re: SparkR Error in sparkR.init(master=“local”) in RStudio

2015-07-13 Thread ashishdutt
Hi,
Try this
 Sys.setenv(SPARK_HOME=C:\\spark-1.4.0)  # The path to your spark
 installation
 .libPaths(c(file.path(Sys.getenv(SPARK_HOME), R, lib), .libPaths()))
 library(SparkR, lib.loc=C:\\spark-1.4.0\\lib) # The path to the lib
 folder in the spark location
 library(SparkR)
 sc=sparkR.init(master=local)
Launching java with spark-submit command C:\spark-1.4.0/bin/spark-submit.cmd 
sparkr-shell
C:\Users\Ashish\AppData\Local\Temp\RtmpWqFsOB\backend_portbdc329477c6 
 

Cheers,
Ashish



--
View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/SparkR-Error-in-sparkR-init-master-local-in-RStudio-tp23768p23800.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.

-
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org



Re: sparkR

2015-07-13 Thread ashishdutt
Please can you explain how did you set this second step in windows
environment?
.libPaths(c(file.path(Sys.getenv(SPARK_HOME), R, lib), .libPaths())) 
I mean to ask where do I type this command, at R prompt or in command
prompt?

Thanks for your time.

Regards,
Ashish



--
View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/sparkR-tp23412p23798.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.

-
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org



Re: sparkR

2015-07-13 Thread ashishdutt
I had been facing this problem for a long time and this practically forced me
to move to pyspark.
This is what I tried after reading the posts here
 Sys.setenv(SPARK_HOME=C:\\spark-1.4.0)
 .libPaths(c(file.path(Sys.getenv(SPARK_HOME), R, lib), .libPaths()))
 library(SparkR, lib.loc=C:\\spark-1.4.0\\lib)
 library(SparkR)
 sc=sparkR.init(master=local)

OMFG It works, Bravo... 
Thanks a ton.

Cheers
Ashish



--
View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/sparkR-tp23412p23799.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.

-
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org



Upgrade Spark-1.3.0 to Spark-1.4.0 in CDH5.4

2015-07-13 Thread ashishdutt
Hello all,
The configuration of my cluster is as follows;
# 4 noded cluster running on Centos OS 6.4
# spark-1.3.0 installed on all

I would like to use SparkR shipped with spark-1.4.0. I checked Cloudera and
find that the latest release CDH5.4 still does not have the spark-1.4.0.
Forums like StackExchange and Google groups suggest to do the process
manually by building the binaries and also caution that if spark-1.4.0 will
be used in CDH5.4 then it will not be supported by Cloudera as of now  see
here
http://community.cloudera.com/t5/Advanced-Analytics-Apache-Spark/can-i-upgrade-spark-from-1-3-0-to-1-3-1-in-CDH-5-4-0/td-p/27137
 
. (this link refers to an older version though)
I dont mind if it will be supported by Cloudera or not because eventually it
will but dont know when.
Meanwhile, production is at stake and I am not sure how will i go about
building the binaries (coming from Windows environment)
Therefore, can you suggest how can I upgrade spark-1.3.0 to spark-1.4.0 on
Centos 6.4 Final  or do i wait for Cloudera to roll out the upgrade?

Thanks for your help.

Cheers,
Ashish



--
View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/Upgrade-Spark-1-3-0-to-Spark-1-4-0-in-CDH5-4-tp23814.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.

-
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org



Is it possible to change the default port number 7077 for spark?

2015-07-10 Thread ashishdutt
Hello all, 
In my lab a colleague installed and configured spark 1.3.0 on a 4 noded
cluster on CDH5.4 environment. The default port number for our spark
configuration is 7456. I have been trying to SSH to spark-master from using
this port number but it fails every time giving error JVM is timed out. 
After reading the documentation , given by  Cloudera
http://www.cloudera.com/content/cloudera/en/documentation/core/latest/topics/cdh_ig_ports_cdh5.html
 
it says that the default port number for spark configuration should be 7077
and that is what i see in all the posts here and elsewhere on search results
in Google. So now I have three question please if anyone can help me answer
all or any of them Q1) Will the spark configuration work only with port
number 7077? If yes, then how can I change the port number? Q2) Do i need to
install spark on all the machines in the cluster? Q3) Do run any spark job
do I always have to SSH into the spark-master machine ? or is it possible to
connect my laptop to the spark-master and invoke commands from my laptop to
spark-master and worker machines?  
Thank you for your time. 
Ashish



--
View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/Is-it-possible-to-change-the-default-port-number-7077-for-spark-tp23774.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.

-
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org



Re: DLL load failed: %1 is not a valid win32 application on invoking pyspark

2015-07-09 Thread ashishdutt
Not really a clean solution but I solved the problem by reinstalling Anaconda 



--
View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/DLL-load-failed-1-is-not-a-valid-win32-application-on-invoking-pyspark-tp23733p23743.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.

-
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org



DLL load failed: %1 is not a valid win32 application on invoking pyspark

2015-07-08 Thread ashishdutt
Hi,

I get the error, DLL load failed: %1 is not a valid win32 application
whenever I invoke pyspark. Attached is the screenshot of the same.
Is there any way I can get rid of it. Still being new to PySpark and have
had,  a not so pleasant experience so far most probably because I am on a
windows environment. 
Therefore, I am afraid that this error might cause me trouble as I continue
my journey exploring pyspark.
I have already check SO and this user list but there are no posts for the
same.

My environment is python version 2.7, OS- Windows 7, Spark- ver 1.3.0
Appreciate your help.

 
Sincerely,
Ashish Dutt error.PNG
http://apache-spark-user-list.1001560.n3.nabble.com/file/n23733/error.PNG  



--
View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/DLL-load-failed-1-is-not-a-valid-win32-application-on-invoking-pyspark-tp23733.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.

-
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org



JVM is not ready after 10 seconds

2015-07-06 Thread ashishdutt
Hi,

I am trying to connect a worker to the master. The spark master is on
cloudera manager and I know the master IP address and port number. 
I downloaded the spark binary for CDH4 on the worker machine and then when I
try to invoke the command 
 sc = sparkR.init(master=ip address:port number) I get the following
 error. 

 sc=sparkR.init(master=spark://10.229.200.250:7377)
Launching java with spark-submit command
C:\spark-1.4.0\bin/bin/spark-submit.cmd  sparkr-shell
C:\Users\ASHISH~1\AppData\Local\Temp\Rtmp82kCxH\backend_port4281739d85 
Error in sparkR.init(master = spark://10.229.200.250:7377) : 
  JVM is not ready after 10 seconds
In addition: Warning message:
running command 'C:\spark-1.4.0\bin/bin/spark-submit.cmd  sparkr-shell
C:\Users\ASHISH~1\AppData\Local\Temp\Rtmp82kCxH\backend_port4281739d85' had
status 127 

 
I am using windows 7 as the OS on the worker machine and I am invoking the
sparkR.init() from RStudio

Any help in this reference will be appreciated

Thank you,
Ashish Dutt



--
View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/JVM-is-not-ready-after-10-seconds-tp23658.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.

-
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org