Yes, scheduling is centralized in the driver.

For debugging, I think you'd want to set the executor JVM, not the worker
JVM flags.


On Thu, Jun 30, 2016 at 11:36 AM, cbruegg <m...@cbruegg.com> wrote:

> Hello everyone,
>
> I'm a student assistant in research at the University of Paderborn, working
> on integrating Spark (v1.6.2) with a new network resource management
> system.
> I have already taken a deep dive into the source code of spark-core w.r.t.
> its scheduling systems.
>
> We are running a cluster in standalone mode consisting of a master node and
> three slave nodes. Am I right to assume that tasks are scheduled within the
> TaskSchedulerImpl using the DAGScheduler in this mode? I need to find a
> place where the execution plan (and each stage) for a job is computed and
> can be analyzed, so I placed some breakpoints in these two classes.
>
> The remote debugging session within IntelliJ IDEA has been established by
> running the following commands on the master node before:
>
>   export SPARK_WORKER_OPTS="-Xdebug
> -Xrunjdwp:server=y,transport=dt_socket,address=4000,suspend=n"
>   export SPARK_MASTER_OPTS="-Xdebug
> -Xrunjdwp:server=y,transport=dt_socket,address=4000,suspend=n"
>
> Port 4000 has been forwarded to my local machine. Unfortunately, none of my
> breakpoints through the class get hit when I invoke a task like
> sc.parallelize(1 to 1000).count() in spark-shell on the master node (using
> --master spark://...), though when I pause all threads I can see that the
> process I am debugging runs some kind of event queue, which means that the
> debugger is connected to /something/.
>
> Do I rely on false assumptions or should these breakpoints in fact get hit?
> I am not too familiar with Spark, so please bear with me if I got something
> wrong. Many thanks in advance for your help.
>
> Best regards,
> Christian Brüggemann
>
>
>
> --
> View this message in context:
> http://apache-spark-developers-list.1001551.n3.nabble.com/Debugging-Spark-itself-in-standalone-cluster-mode-tp18139.html
> Sent from the Apache Spark Developers List mailing list archive at
> Nabble.com.
>
> ---------------------------------------------------------------------
> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>
>

Reply via email to