[
https://issues.apache.org/jira/browse/SPARK-5134?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14353047#comment-14353047
]
Sean Owen commented on SPARK-5134:
----------------------------------
Yep, I confirmed that ...
{code}
[INFO] \- org.apache.spark:spark-core_2.10:jar:1.2.1:compile
...
[INFO] +- org.apache.hadoop:hadoop-client:jar:2.2.0:compile
[INFO] | +- org.apache.hadoop:hadoop-common:jar:2.2.0:compile
[INFO] | | +- commons-cli:commons-cli:jar:1.2:compile
...
{code}
Well, FWIW, although unintentional I do think there are upsides to this change.
It would be good to codify that in the build, I suppose, by updating the
default version number. How about updating to 2.2.0 to match what has actually
happened? This would not entail activating the Hadoop build profiles by default
or anything.
[~rdub] would you care to do the honors?
> Bump default Hadoop version to 2+
> ---------------------------------
>
> Key: SPARK-5134
> URL: https://issues.apache.org/jira/browse/SPARK-5134
> Project: Spark
> Issue Type: Improvement
> Components: Build
> Affects Versions: 1.2.0
> Reporter: Ryan Williams
> Priority: Minor
>
> [~srowen] and I discussed bumping [the default hadoop version in the parent
> POM|https://github.com/apache/spark/blob/bb38ebb1abd26b57525d7d29703fd449e40cd6de/pom.xml#L122]
> from {{1.0.4}} to something more recent.
> There doesn't seem to be a good reason that it was set/kept at {{1.0.4}}.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]