Github user mengxr commented on a diff in the pull request:

    https://github.com/apache/spark/pull/3070#discussion_r19754487
  
    --- Diff: mllib/pom.xml ---
    @@ -46,6 +46,11 @@
           <version>${project.version}</version>
         </dependency>
         <dependency>
    +      <groupId>org.apache.spark</groupId>
    +      <artifactId>spark-sql_${scala.binary.version}</artifactId>
    --- End diff --
    
    @srowen Yes, it feels weird if we say ML depends on SQL, the "query 
language". Spark SQL provides RDD with schema support and execution plan 
optimization, both of which are need by MLlib. We need flexible table-like 
datasets and I/O support, and operations that "carry over" additional columns 
during the training phrase. It is natural to say that ML depends on RDD with 
schema support and execution plan optimization.
    
    I agree that we should factor the common part out or make SchemaRDD a 
first-class citizen in Core, but that definitely takes time for both design and 
development. This dependence change has no effect on the content we deliver to 
users, and UDTs are internal to Spark.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to