[GitHub] spark pull request #22167: [SPARK-25170][DOC] Add list and short description...

2018-09-11 Thread LucaCanali
Github user LucaCanali closed the pull request at:

https://github.com/apache/spark/pull/22167


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #22167: [SPARK-25170][DOC] Add list and short description...

2018-09-05 Thread kiszk
Github user kiszk commented on a diff in the pull request:

https://github.com/apache/spark/pull/22167#discussion_r215315965
  
--- Diff: docs/monitoring.md ---
@@ -388,6 +388,163 @@ value triggering garbage collection on jobs, and 
`spark.ui.retainedStages` that
 Note that the garbage collection takes place on playback: it is possible 
to retrieve
 more entries by increasing these values and restarting the history server.
 
+### Executor Task Metrics
+
+The REST API exposes the values of the Task Metrics collected by Spark 
executors at the
+task execution level. The metrics can be used for performance 
troubleshooting.
+A list of the available metrics with a short description:
+
+
+  Spark Executor Task Metric name
+  Short description
+  
+  
+executorRunTime
+Time the executor spent running this task. This includes time 
fetching shuffle data.
+The value is expressed in milliseconds.
+  
+  
+executorCpuTime
+CPU Time the executor spent running this task. This includes time 
fetching shuffle data.
--- End diff --

nit: `CPU time`?


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #22167: [SPARK-25170][DOC] Add list and short description...

2018-09-05 Thread kiszk
Github user kiszk commented on a diff in the pull request:

https://github.com/apache/spark/pull/22167#discussion_r215315657
  
--- Diff: docs/monitoring.md ---
@@ -388,6 +388,163 @@ value triggering garbage collection on jobs, and 
`spark.ui.retainedStages` that
 Note that the garbage collection takes place on playback: it is possible 
to retrieve
 more entries by increasing these values and restarting the history server.
 
+### Executor Task Metrics
+
+The REST API exposes the values of the Task Metrics collected by Spark 
executors at the
+task execution level. The metrics can be used for performance 
troubleshooting.
+A list of the available metrics with a short description:
+
+
+  Spark Executor Task Metric name
+  Short description
+  
+  
+executorRunTime
+Time the executor spent running this task. This includes time 
fetching shuffle data.
+The value is expressed in milliseconds.
+  
+  
+executorCpuTime
+CPU Time the executor spent running this task. This includes time 
fetching shuffle data.
+The value is expressed in nanoseconds.
+  
+  
+executorDeserializeTime
+Time taken on the executor to deserialize this task.
+The value is expressed in milliseconds.
+  
+  
+executorDeserializeCpuTime
+CPU Time taken on the executor to deserialize this task.
+ The value is expressed in nanoseconds.
+  
+  
+resultSize
+The number of bytes this task transmitted back to the driver as 
the TaskResult.
+  
+  
+jvmGCTime
+Amount of time the JVM spent in garbage collection while executing 
this task.
--- End diff --

Why do we start with `amount of` while the above parameters start `Time` or 
`CPU Time`?


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #22167: [SPARK-25170][DOC] Add list and short description...

2018-09-05 Thread kiszk
Github user kiszk commented on a diff in the pull request:

https://github.com/apache/spark/pull/22167#discussion_r215315711
  
--- Diff: docs/monitoring.md ---
@@ -388,6 +388,163 @@ value triggering garbage collection on jobs, and 
`spark.ui.retainedStages` that
 Note that the garbage collection takes place on playback: it is possible 
to retrieve
 more entries by increasing these values and restarting the history server.
 
+### Executor Task Metrics
+
+The REST API exposes the values of the Task Metrics collected by Spark 
executors at the
+task execution level. The metrics can be used for performance 
troubleshooting.
+A list of the available metrics with a short description:
+
+
+  Spark Executor Task Metric name
+  Short description
+  
+  
+executorRunTime
+Time the executor spent running this task. This includes time 
fetching shuffle data.
+The value is expressed in milliseconds.
+  
+  
+executorCpuTime
+CPU Time the executor spent running this task. This includes time 
fetching shuffle data.
+The value is expressed in nanoseconds.
+  
+  
+executorDeserializeTime
+Time taken on the executor to deserialize this task.
+The value is expressed in milliseconds.
+  
+  
+executorDeserializeCpuTime
+CPU Time taken on the executor to deserialize this task.
+ The value is expressed in nanoseconds.
+  
+  
+resultSize
+The number of bytes this task transmitted back to the driver as 
the TaskResult.
+  
+  
+jvmGCTime
+Amount of time the JVM spent in garbage collection while executing 
this task.
+The value is expressed in milliseconds.
+  
+  
+resultSerializationTime
+Amount of time spent serializing the task result.
--- End diff --

ditto


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #22167: [SPARK-25170][DOC] Add list and short description...

2018-09-05 Thread kiszk
Github user kiszk commented on a diff in the pull request:

https://github.com/apache/spark/pull/22167#discussion_r215314710
  
--- Diff: docs/monitoring.md ---
@@ -388,6 +388,163 @@ value triggering garbage collection on jobs, and 
`spark.ui.retainedStages` that
 Note that the garbage collection takes place on playback: it is possible 
to retrieve
 more entries by increasing these values and restarting the history server.
 
+### Executor Task Metrics
+
+The REST API exposes the values of the Task Metrics collected by Spark 
executors at the
+task execution level. The metrics can be used for performance 
troubleshooting.
+A list of the available metrics with a short description:
+
+
+  Spark Executor Task Metric name
+  Short description
+  
+  
+executorRunTime
+Time the executor spent running this task. This includes time 
fetching shuffle data.
+The value is expressed in milliseconds.
+  
+  
+executorCpuTime
--- End diff --

Do we miss `` in these two lines?


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #22167: [SPARK-25170][DOC] Add list and short description...

2018-09-05 Thread kiszk
Github user kiszk commented on a diff in the pull request:

https://github.com/apache/spark/pull/22167#discussion_r215314527
  
--- Diff: docs/monitoring.md ---
@@ -388,6 +388,163 @@ value triggering garbage collection on jobs, and 
`spark.ui.retainedStages` that
 Note that the garbage collection takes place on playback: it is possible 
to retrieve
 more entries by increasing these values and restarting the history server.
 
+### Executor Task Metrics
+
+The REST API exposes the values of the Task Metrics collected by Spark 
executors at the
+task execution level. The metrics can be used for performance 
troubleshooting.
+A list of the available metrics with a short description:
+
+
+  Spark Executor Task Metric name
+  Short description
+  
+  
+executorRunTime
+Time the executor spent running this task. This includes time 
fetching shuffle data.
--- End diff --

Does `Time` mean `elapsed time` or other `time`?


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #22167: [SPARK-25170][DOC] Add list and short description...

2018-08-21 Thread LucaCanali
GitHub user LucaCanali opened a pull request:

https://github.com/apache/spark/pull/22167

[SPARK-25170][DOC] Add list and short description of Spark Executor Task 
Metrics to the documentation

## What changes were proposed in this pull request?

Add description of Task Metrics to the documentation.

## How was this patch tested?

None.


You can merge this pull request into a Git repository by running:

$ git pull https://github.com/LucaCanali/spark docMonitoringTaskMetrics

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/spark/pull/22167.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #22167


commit a8db1605adbc271c785fda24b4945bf87149a4cd
Author: LucaCanali 
Date:   2018-08-20T14:12:52Z

Document Spark Executor Task Metrics




---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org