dongjoon-hyun commented on pull request #34846:
URL: https://github.com/apache/spark/pull/34846#issuecomment-1002336872


   The new proposed code focused only on fitting to JVM G1HeapRegionSize and it 
claims that it will reduce the Spark's pageSize (humongous) to smaller ones. In 
this case, there is no guarantee that new defaultPageSizeBytes is better than 
the previous calculated ones (whatever they were in the production) in terms of 
the executor time (if no GC occurs because there is enough memory in the 
extreme cases), isn't it, @srowen ?
   ```
     private lazy val defaultPageSizeBytes = Utils.maybeG1HeapRegionSize match {
        case Some(heapRegionSize) if tungstenMemoryMode == MemoryMode.ON_HEAP =>
          heapRegionSize - Platform.LONG_ARRAY_OFFSET
   ```
   
   If the smaller page size always wins, we may want to have a default value, 
`1MB` (the minimum value of G1HeapRegionSize) - Platform.LONG_ARRAY_OFFSET, for 
`spark.buffer.pageSize` configuration, instead of this PR.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]



---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to