Github user andrewor14 commented on a diff in the pull request:

    https://github.com/apache/spark/pull/10835#discussion_r50805566
  
    --- Diff: 
core/src/main/scala/org/apache/spark/scheduler/TaskResultGetter.scala ---
    @@ -82,7 +84,19 @@ private[spark] class TaskResultGetter(sparkEnv: 
SparkEnv, scheduler: TaskSchedul
                   (deserializedResult, size)
               }
     
    -          result.metrics.setResultSize(size)
    +          // Set the task result size in the accumulator updates received 
from the executors.
    +          // We need to do this here on the driver because if we did this 
on the executors then
    +          // we would have to serialize the result again after updating 
the size.
    +          result.accumUpdates = result.accumUpdates.map { a =>
    +            if (a.name == InternalAccumulator.RESULT_SIZE) {
    +              assert(a.update.getOrElse(0L) == 0L,
    --- End diff --
    
    actually it should be equal to Some(0L). Maybe I should just explicitly 
write that.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org

Reply via email to