goiri commented on a change in pull request #3423:
URL: https://github.com/apache/hadoop/pull/3423#discussion_r708538792
##########
File path:
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/Text.java
##########
@@ -301,9 +305,18 @@ public void clear() {
*/
private boolean ensureCapacity(final int capacity) {
if (bytes.length < capacity) {
+ // use long to allow overflow
+ long tmpLength = bytes.length;
+ long tmpCapacity = capacity;
+
// Try to expand the backing array by the factor of 1.5x
- // (by taking the current size + diving it by half)
- int targetSize = Math.max(capacity, bytes.length + (bytes.length >> 1));
+ // (by taking the current size + diving it by half).
+ //
+ // If the calculated value is beyond the size
+ // limit, we cap it to ARRAY_MAX_SIZE
+ int targetSize = (int)Math.min(ARRAY_MAX_SIZE,
Review comment:
I would prefer to cast inside so it uses Math.max(int, int) and keeps
everything as ints.
The splitting for the min and the max is more a matter of readability.
##########
File path:
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/Text.java
##########
@@ -73,6 +73,10 @@ protected CharsetDecoder initialValue() {
}
};
+ // max size of the byte array, seems to be a safe choice for multiple VMs
Review comment:
JVMs then
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]