Hello,
at first you will need to make sure that JAVA is installed, or
install it otherwise. Then install scala and a build tool (sbt or
maven). In my point of view, IntelliJ IDEA is a good option to create
your Spark applications. At the end you have to install a distributed
file system e.g HDFS.
I think there is no an all-in-one configuration. But there are
examples about how to configure you Spark cluster (e.g
https://github.com/jaceklaskowski/mastering-apache-spark-book/blob/master/spark-standalone-example-2-workers-on-1-node-cluster.adoc).
Best,
--Iacovos
On 5/6/19 5:50 π.μ., ya wrote:
Dear list,
I am very new to spark, and I am having trouble installing it on my
mac. I have following questions, please give me some guidance. Thank
you very much.
1. How many and what software should I install before installing
spark? I have been searching online, people discussing their
experiences on this topic with different opinions, some says there is
no need to install hadoop before install spark, some says hadoop has
to be installed before spark. Some other people say scala has to be
installed, whereas others say scala is included in spark, and it is
installed automatically once spark in installed. So I am confused what
to install for a start.
2. Is there an simple way to configure these software? for instance,
an all-in-one configuration file? It takes forever for me to configure
things before I can really use it for data analysis.
I hope my questions make sense. Thank you very much.
Best regards,
YA