Github user maropu commented on a diff in the pull request:

    https://github.com/apache/spark/pull/21968#discussion_r212959634
  
    --- Diff: 
sql/core/src/main/scala/org/apache/spark/sql/execution/aggregate/RowBasedHashMapGenerator.scala
 ---
    @@ -141,9 +141,6 @@ class RowBasedHashMapGenerator(
            |    if (buckets[idx] == -1) {
            |      if (numRows < capacity && !isBatchFull) {
            |        // creating the unsafe for new entry
    -       |        
org.apache.spark.sql.catalyst.expressions.codegen.UnsafeRowWriter agg_rowWriter
    -       |          = new 
org.apache.spark.sql.catalyst.expressions.codegen.UnsafeRowWriter(
    -       |              ${groupingKeySchema.length}, ${numVarLenFields * 
32});
            |        agg_rowWriter.reset(); //TODO: investigate if reset or 
zeroout are actually needed
            |        agg_rowWriter.zeroOutNullBytes();
    --- End diff --
    
    btw, if `groupingKeySchema` has no nullable field, can we drop 
`agg_rowWriter.zeroOutNullBytes()`?


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org

Reply via email to