Github user srowen commented on the pull request:
https://github.com/apache/spark/pull/5786#issuecomment-101438765
`spark-shell` works locally for me. You're right, this may not work on
Hadoop cluster X, but haven't we always generally needed to build Spark for
Hadoop X to avoid this? I get it though, maybe the inconsistent Hadoop client
libs don't work whereas a consistent Hadoop 1.x client lib set did, even
against a mismatched cluster version.
Fair point and all that but this isn't the right way to build Spark anyway,
and I'm afraid this change was effectively already released. I'm narrowly
arguing against undoing the `hadoop.version=2.2.0` change. I'm also asserting
that the 1.4 release artifacts will be fine.
And then saying we should fix-forward the rest of this for 1.5, if not 1.4.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]