Github user kiszk commented on a diff in the pull request: https://github.com/apache/spark/pull/22167#discussion_r215315711 --- Diff: docs/monitoring.md --- @@ -388,6 +388,163 @@ value triggering garbage collection on jobs, and `spark.ui.retainedStages` that Note that the garbage collection takes place on playback: it is possible to retrieve more entries by increasing these values and restarting the history server. +### Executor Task Metrics + +The REST API exposes the values of the Task Metrics collected by Spark executors at the +task execution level. The metrics can be used for performance troubleshooting. +A list of the available metrics with a short description: + +<table class="table"> + <tr><th>Spark Executor Task Metric name</th> + <th>Short description</th> + </tr> + <tr> + <td>executorRunTime</td> + <td>Time the executor spent running this task. This includes time fetching shuffle data. + The value is expressed in milliseconds.</td> + </tr> + <tr> + <td>executorCpuTime + <td>CPU Time the executor spent running this task. This includes time fetching shuffle data. + The value is expressed in nanoseconds. + </tr> + <tr> + <td>executorDeserializeTime</td> + <td>Time taken on the executor to deserialize this task. + The value is expressed in milliseconds.</td> + </tr> + <tr> + <td>executorDeserializeCpuTime</td> + <td>CPU Time taken on the executor to deserialize this task. + The value is expressed in nanoseconds.</td> + </tr> + <tr> + <td>resultSize</td> + <td>The number of bytes this task transmitted back to the driver as the TaskResult.</td> + </tr> + <tr> + <td>jvmGCTime</td> + <td>Amount of time the JVM spent in garbage collection while executing this task. + The value is expressed in milliseconds.</td> + </tr> + <tr> + <td>resultSerializationTime</td> + <td>Amount of time spent serializing the task result. --- End diff -- ditto
--- --------------------------------------------------------------------- To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For additional commands, e-mail: reviews-h...@spark.apache.org