Github user cxzl25 commented on the issue:

    https://github.com/apache/spark/pull/21311
  
    @cloud-fan 
    LongToUnsafeRowMap#append(key: Long, row: UnsafeRow)
    when row.getSizeInBytes > newPageSize( oldPage.length * 8L * 2),still use 
newPageSize value.
    When the new page size is insufficient to hold the entire row of data, 
Platform.copyMemory is still called.No error.
    At this time, the last remaining data was discarded.
    When reading data, read this buffer according to offset and length,the last 
data is unpredictable.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to