Hi all,

I have a dataframe like following,

+---------+----------+
|client_id|Date      |
+-------- +----------+
| a       |2016-11-23|
| b       |2016-11-18|
| a       |2016-11-23|
| a       |2016-11-23|
| a       |2016-11-24|
+---------+----------+

I want to find unique dates of each client_id using spark dataframe.

expected output

a  (2016-11-23, 2016-11-24)
b   2016-11-18

I tried with df.groupBy("client_id").But I don't know how to find distinct
values after groupBy().
How to do this?
Is any other efficient methods are available for doing this ?
I am using scala 2.11.8 & spark 2.0


Thanks

Reply via email to