I'll note that the DSL is pretty experimental.  That said you should be
able to do something like "user.id".attr

On Mon, Sep 29, 2014 at 3:39 PM, Benyi Wang <bewang.t...@gmail.com> wrote:

> scala> user
> res19: org.apache.spark.sql.SchemaRDD =
> SchemaRDD[0] at RDD at SchemaRDD.scala:98
> == Query Plan ==
> ParquetTableScan [id#0,name#1], (ParquetRelation
> /user/hive/warehouse/user), None
>
> scala> order
> res20: org.apache.spark.sql.SchemaRDD =
> SchemaRDD[72] at RDD at SchemaRDD.scala:98
> == Query Plan ==
> ParquetTableScan [id#8,userid#9,unit#10], (ParquetRelation
> /user/hive/warehouse/orders), None
>
> For joining SchemaRDD user and order, This will generate Ambiguous issue
> because both of tables have 'id.
>
> user.join(order, on=Some('id === 'userid))
>
> How can I specify an expression which can use SchemaRDD name and column
> together? Something might be like 'user.'id. This expression currently
> doesn't work in Spark 1.0.0 CDH 5.1.0.
>

Reply via email to