[ 
https://issues.apache.org/jira/browse/HIVE-7526?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14084450#comment-14084450
 ] 

Rui Li commented on HIVE-7526:
------------------------------

Hi [~xuefuz] [~csun], it seems in SparkShuffler, we lost the # of partitions 
when applying the shuffle transformations. It may be useful if user can specify 
it (e.g. HIVE-7540). Should we add that to the "shuffle" method?

> Research to use groupby transformation to replace Hive existing 
> partitionByKey and SparkCollector combination
> -------------------------------------------------------------------------------------------------------------
>
>                 Key: HIVE-7526
>                 URL: https://issues.apache.org/jira/browse/HIVE-7526
>             Project: Hive
>          Issue Type: Task
>          Components: Spark
>            Reporter: Xuefu Zhang
>            Assignee: Chao
>             Fix For: spark-branch
>
>         Attachments: HIVE-7526.2.patch, HIVE-7526.3.patch, 
> HIVE-7526.4-spark.patch, HIVE-7526.5-spark.patch, HIVE-7526.patch
>
>
> Currently SparkClient shuffles data by calling paritionByKey(). This 
> transformation outputs <key, value> tuples. However, Hive's ExecMapper 
> expects <key, iterator<value>> tuples, and Spark's groupByKey() seems 
> outputing this directly. Thus, using groupByKey, we may be able to avoid its 
> own key clustering mechanism (in HiveReduceFunction). This research is to 
> have a try.



--
This message was sent by Atlassian JIRA
(v6.2#6252)

Reply via email to