liuzqt opened a new pull request, #47192:
URL: https://github.com/apache/spark/pull/47192

   
   
   ### What changes were proposed in this pull request?
   
   Add task on/off heap execution memory in `TaskMetrics`, tracked in 
`TaskMemoryManager`.
   
   ### Why are the changes needed?
   
   Currently there is no task on/off heap execution memory metrics. 
   
   There is a 
[peakExecutionMemory](https://github.com/apache/spark/blob/3cd35f8cb6462051c621cf49de54b9c5692aae1d/core/src/main/scala/org/apache/spark/executor/TaskMetrics.scala#L114)
  metrics, however, the semantic is a confusing: it only cover the execution 
memory used by shuffle/join/aggregate/sort, which is accumulated in specific 
operators and thus not really reflect the real execution memory.
   
   Therefore it's necessary to add these two metrics. 
   
   How to: we can easily maintain the whole task-level peak memory in 
TaskMemoryManager, assuming acquireExecutionMemory is the only one narrow waist 
for acquiring execution memory.
   
   Also I created two followup sub tickets:
   
   https://issues.apache.org/jira/browse/SPARK-48788 :accumulate task metrics 
in stage, and display in Spark UI
   https://issues.apache.org/jira/browse/SPARK-48789 : deprecate 
`peakExecutionMemory` once we have replacement for it.
   
   
   ### Does this PR introduce _any_ user-facing change?
   
   Supposedly no. But two followup sub ticket will have user-facing change: new 
metrics exposed to Spark UI, and old metrics deprecation.
   
   
   ### How was this patch tested?
   new test
   
   ### Was this patch authored or co-authored using generative AI tooling?
   NO
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org

Reply via email to