What I tried to say is, I didn't start spark master/worker at all, for a
standalone deployment.
But I still can login into pyspark to run the job. I don't know why.
$ ps -efw|grep spark
$ netstat -ntlp
both the output above have no spark related info.
And this machine is managed by myself, I know how to start spark
correctly. But I didn't start them yet, and I still can login to pyspark
to run the jobs. for exmaple:
df = sc.parallelize([("t1",1),("t2",2)]).toDF(["name","number"])
df.show()
+----+------+
|name|number|
+----+------+
| t1| 1|
| t2| 2|
+----+------+
do you know why?
Thank you.
frakass.
---------------------------------------------------------------------
To unsubscribe e-mail: user-unsubscr...@spark.apache.org