Spark doesn't support subqueries in WHERE clause, IIRC. It supports subqueries only in the FROM clause as of now. See this ticket <https://issues.apache.org/jira/browse/SPARK-4226> for more on this.
[image: http://] Tariq, Mohammad about.me/mti [image: http://] <http://about.me/mti> On Fri, Feb 26, 2016 at 7:01 AM, ayan guha <guha.a...@gmail.com> wrote: > Why is this not working for you? Are you trying on dataframe? What error > are you getting? > > On Thu, Feb 25, 2016 at 10:23 PM, Ashok Kumar < > ashok34...@yahoo.com.invalid> wrote: > >> Hi, >> >> What is the equivalent of this in Spark please >> >> select * from mytable where column1 in (select max(column1) from mytable) >> >> Thanks >> > > > > -- > Best Regards, > Ayan Guha >