Github user maropu commented on a diff in the pull request:
https://github.com/apache/spark/pull/21968#discussion_r212959634
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/execution/aggregate/RowBasedHashMapGenerator.scala
---
@@ -141,9 +141,6 @@ class RowBasedHashMapGenerator(
| if (buckets[idx] == -1) {
| if (numRows < capacity && !isBatchFull) {
| // creating the unsafe for new entry
- |
org.apache.spark.sql.catalyst.expressions.codegen.UnsafeRowWriter agg_rowWriter
- | = new
org.apache.spark.sql.catalyst.expressions.codegen.UnsafeRowWriter(
- | ${groupingKeySchema.length}, ${numVarLenFields *
32});
| agg_rowWriter.reset(); //TODO: investigate if reset or
zeroout are actually needed
| agg_rowWriter.zeroOutNullBytes();
--- End diff --
btw, if `groupingKeySchema` has no nullable field, can we drop
`agg_rowWriter.zeroOutNullBytes()`?
---
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]