David Mollitor created HIVE-21466:
-------------------------------------

             Summary: Increase Default Size of SPLIT_MAXSIZE
                 Key: HIVE-21466
                 URL: https://issues.apache.org/jira/browse/HIVE-21466
             Project: Hive
          Issue Type: Improvement
          Components: Configuration
    Affects Versions: 4.0.0, 3.2.0
            Reporter: David Mollitor
            Assignee: David Mollitor
         Attachments: HIVE-21466.1.patch

{code:java}
 MAPREDMAXSPLITSIZE(FileInputFormat.SPLIT_MAXSIZE, 256000000L, "", true),
{code}
[https://github.com/apache/hive/blob/8d4300a02691777fc96f33861ed27e64fed72f2c/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java#L682]

This field specifies a maximum size for each MR (maybe other?) splits.

This number should be a multiple of the HDFS Block size. The way that this 
maximum is implemented, is that each block is added to the split, and if the 
split grows to be larger than the maximum allowed, the split is submitted to 
the cluster and a new split is opened.

So, imagine the following scenario:
 * HDFS block size of 16 bytes
 * Maximum size of 40 bytes

This will produce a split with 3 blocks. (2x16) = 32; another block will be 
inserted, (3x16) = 48 bytes in the split. So, while many operators would assume 
a split of 2 blocks, the actual is 3 blocks. Setting the maximum split size to 
a multiple of the HDFS block size will make this behavior less confusing.

The current setting is ~256MB and when this was introduced, the default HDFS 
block size was 64MB. That is a factor of 4x. However, now HDFS block sizes are 
128MB by default, so I propose setting this to 4x128MB.  The larger splits 
(fewer tasks) should give a nice performance boost for modern hardware.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

Reply via email to