[
https://issues.apache.org/jira/browse/HIVE-16391?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16501408#comment-16501408
]
ASF GitHub Bot commented on HIVE-16391:
---------------------------------------
GitHub user jerryshao opened a pull request:
https://github.com/apache/hive/pull/364
HIVE-16391: Add a new classifier for hive-exec to be used by Spark
This fix adding a new classifier for hive-exec artifact (`core-spark`),
which is specifically used for Spark. Details in
[SPARK-20202](https://issues.apache.org/jira/browse/SPARK-20202).
This is because original hive-exec packages many transitive dependencies
into shaded jar without relocation, this makes conflicts in Spark. Spark only
needs to relocate protobuf and kryo jar. So here propose to add a new
classifier to generate a new artifact only for Spark.
You can merge this pull request into a Git repository by running:
$ git pull https://github.com/jerryshao/hive 1.2-spark-fix
Alternatively you can review and apply these changes as the patch at:
https://github.com/apache/hive/pull/364.patch
To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:
This closes #364
----
commit bb27b260d82fa0a77d9fea3c123f2af8f1ea88aa
Author: jerryshao <sshao@...>
Date: 2018-06-05T06:59:37Z
HIVE-16391: Add a new classifier for hive-exec to be used by Spark
----
> Publish proper Hive 1.2 jars (without including all dependencies in uber jar)
> -----------------------------------------------------------------------------
>
> Key: HIVE-16391
> URL: https://issues.apache.org/jira/browse/HIVE-16391
> Project: Hive
> Issue Type: Task
> Components: Build Infrastructure
> Reporter: Reynold Xin
> Priority: Major
> Labels: pull-request-available
>
> Apache Spark currently depends on a forked version of Apache Hive. AFAIK, the
> only change in the fork is to work around the issue that Hive publishes only
> two sets of jars: one set with no dependency declared, and another with all
> the dependencies included in the published uber jar. That is to say, Hive
> doesn't publish a set of jars with the proper dependencies declared.
> There is general consensus on both sides that we should remove the forked
> Hive.
> The change in the forked version is recorded here
> https://github.com/JoshRosen/hive/tree/release-1.2.1-spark2
> Note that the fork in the past included other fixes but those have all become
> unnecessary.
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)