Hi,
I was running a WordCount application on Spark, and the machine I used has
4 physical cores. However, in spark-env.sh file, I set SPARK_WORKER_CORES
= 32. The web UI says it launched one executor with 32 cores and the
executor could execute 32 tasks simultaneously. Does spark create 32
I am a beginner to Spark, having some simple questions regarding the use of
RDD in python.
Suppose I have a matrix called data_matrix, I pass it to RDD using
RDD_matrix = sc.parallelize(data_matrix)
but I will have a problem if I want to know the dimension of the matrix in
Spark, because Sparkk