Spark doesn't support subqueries in WHERE clause, IIRC. It supports
subqueries only in the FROM clause as of now. See this ticket
<https://issues.apache.org/jira/browse/SPARK-4226> for more on this.



[image: http://]

Tariq, Mohammad
about.me/mti
[image: http://]
<http://about.me/mti>


On Fri, Feb 26, 2016 at 7:01 AM, ayan guha <guha.a...@gmail.com> wrote:

> Why is this not working for you? Are you trying on dataframe? What error
> are you getting?
>
> On Thu, Feb 25, 2016 at 10:23 PM, Ashok Kumar <
> ashok34...@yahoo.com.invalid> wrote:
>
>> Hi,
>>
>> What is the equivalent of this in Spark please
>>
>> select * from mytable where column1 in (select max(column1) from mytable)
>>
>> Thanks
>>
>
>
>
> --
> Best Regards,
> Ayan Guha
>

Reply via email to