Hi Ayan, yup that works very well however I believe Kant's other mail
"Queries with streaming sources must be executed with writeStream.start()"
is adding more context.
I think he is trying leverage on structured streaming and applying the rdd
conversion to a streaming dataset is breaking the streaming lineage.
Best,

On Mon, Sep 11, 2017 at 3:51 AM, ayan guha <guha.a...@gmail.com> wrote:

> Sorry for side-line question, but for Python, isn't following the easiest:
>
> >>> import json
> >>>* df1 = df.rdd.map(lambda r: json.dumps(r.asDict()))*
> >>> df1.take(10)
> ['{"id": 1}', '{"id": 2}', '{"id": 3}', '{"id": 4}', '{"id": 5}']
>
>
>
>
> On Mon, Sep 11, 2017 at 4:22 AM, Riccardo Ferrari <ferra...@gmail.com>
> wrote:
>
>> Hi Kant,
>>
>> You can check the getValuesMap
>> <https://spark.apache.org/docs/2.2.0/api/java/org/apache/spark/sql/Row.html#getValuesMap-scala.collection.Seq->.
>> I found this post <https://stackoverflow.com/a/41602178/8356352> useful,
>> it is in Scala but should be a good starting point.
>> An alternative approach is combine the 'struct' and 'to_json' functions
>> <https://spark.apache.org/docs/2.2.0/api/java/org/apache/spark/sql/functions.html>.
>> I have not tested this in Java but I am using it in Python.
>>
>> Best,
>>
>> On Sun, Sep 10, 2017 at 1:45 AM, kant kodali <kanth...@gmail.com> wrote:
>>
>>> toJSON on Row object.
>>>
>>> On Sat, Sep 9, 2017 at 4:18 PM, Felix Cheung <felixcheun...@hotmail.com>
>>> wrote:
>>>
>>>> toJSON on Dataset/DataFrame?
>>>>
>>>> ------------------------------
>>>> *From:* kant kodali <kanth...@gmail.com>
>>>> *Sent:* Saturday, September 9, 2017 4:15:49 PM
>>>> *To:* user @spark
>>>> *Subject:* How to convert Row to JSON in Java?
>>>>
>>>> Hi All,
>>>>
>>>> How to convert Row to JSON in Java? It would be nice to have .toJson()
>>>> method in the Row class.
>>>>
>>>> Thanks,
>>>> kant
>>>>
>>>
>>>
>>
>
>
> --
> Best Regards,
> Ayan Guha
>

Reply via email to