[
https://issues.apache.org/jira/browse/HIVE-17114?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16092780#comment-16092780
]
liyunzhang_intel commented on HIVE-17114:
-----------------------------------------
[~lirui]: several questions:
1.{quote}
Spark decides the reducer task for each record by computing
hash(key)%numReducers
{quote}
this is in hive on spark code or in spark code? can you point out detail code
place?
2. when i view HIVE-7121, the problem mentioned in the jira description only
relates table with bucket ?
{code}
CREATE TABLE bucket1_1(key int, value string) CLUSTERED BY (key) INTO 100
BUCKETS;
{code}
> HoS: Possible skew in shuffling when data is not really skewed
> --------------------------------------------------------------
>
> Key: HIVE-17114
> URL: https://issues.apache.org/jira/browse/HIVE-17114
> Project: Hive
> Issue Type: Bug
> Reporter: Rui Li
> Assignee: Rui Li
> Priority: Minor
> Attachments: HIVE-17114.1.patch
>
>
> Observed in HoS and may apply to other engines as well.
> When we join 2 tables on a single int key, we use the key itself as hash code
> in {{ObjectInspectorUtils.hashCode}}:
> {code}
> case INT:
> return ((IntObjectInspector) poi).get(o);
> {code}
> Suppose the keys are different but are all some multiples of 10. And if we
> choose 10 as #reducers, the shuffle will be skewed.
--
This message was sent by Atlassian JIRA
(v6.4.14#64029)