[ 
https://issues.apache.org/jira/browse/HIVE-17276?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16152168#comment-16152168
 ] 

Lefty Leverenz commented on HIVE-17276:
---------------------------------------

Doc note:  This adds the configuration parameter 
*hive.auto.convert.join.shuffle.max.size* and fixes a typo in the description 
of *hive.auto.convert.join.hashtable.max.entries* (which was introduced in 
2.2.0 by HIVE-12492) so they need to be documented in the wiki.

* [Configuration Properties -- Query and DDL Execution | 
https://cwiki.apache.org/confluence/display/Hive/Configuration+Properties#ConfigurationProperties-QueryandDDLExecution]

Added a TODOC3.0 label.

> Check max shuffle size when converting to dynamically partitioned hash join
> ---------------------------------------------------------------------------
>
>                 Key: HIVE-17276
>                 URL: https://issues.apache.org/jira/browse/HIVE-17276
>             Project: Hive
>          Issue Type: Bug
>          Components: Physical Optimizer
>    Affects Versions: 3.0.0
>            Reporter: Jesus Camacho Rodriguez
>            Assignee: Jesus Camacho Rodriguez
>              Labels: TODOC3.0
>             Fix For: 3.0.0
>
>         Attachments: HIVE-17276.01.patch, HIVE-17276.02.patch, 
> HIVE-17276.03.patch, HIVE-17276.patch
>
>
> Currently we only check that the max number of entries in the hashmap for a 
> MapJoin surpasses a certain threshold to decide whether to execute a 
> dynamically partitioned hash join.
> We would like to factor the size of the large input that we will shuffle for 
> the dynamically partitioned hash join into the cost model too.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

Reply via email to