Github user srowen commented on a diff in the pull request:
https://github.com/apache/spark/pull/3070#discussion_r19720976
--- Diff: mllib/pom.xml ---
@@ -46,6 +46,11 @@
<version>${project.version}</version>
</dependency>
<dependency>
+ <groupId>org.apache.spark</groupId>
+ <artifactId>spark-sql_${scala.binary.version}</artifactId>
--- End diff --
This still feels weird to me, MLlib depending on SQL. It seems like they
are both wanting to depend on a `SchemaRDD` that is specific to neither. I'm
afraid of making the jar hell in Spark worse by attaching more subprojects
together. That said, the SQL module itself doesn't, for instance, bring in
Hive. Is this going to add much to the MLlib deps? or can the commonality not
be factored out into Core?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]