Something like this?

val json = myRDD.map(*map_obj* => new JSONObject(*map_obj*))

​Here map_obj will be a map containing values (eg: *Map("name" -> "Akhil",
"mail" -> "xyz@xyz")*)​

Performance wasn't so good with this one though.

Thanks
Best Regards

On Wed, Nov 5, 2014 at 3:02 AM, Yin Huai <huaiyin....@gmail.com> wrote:

> Hello Andrejs,
>
> For now, you need to use a JSON lib to serialize records of your datasets
> as JSON strings. In future, we will add a method to SchemaRDD to let you
> write a SchemaRDD in JSON format (I have created
> https://issues.apache.org/jira/browse/SPARK-4228 to track it).
>
> Thanks,
>
> Yin
>
> On Tue, Nov 4, 2014 at 3:48 AM, Andrejs Abele <
> andrejs.ab...@insight-centre.org> wrote:
>
>> Hi,
>> Can some one pleas sugest me, what is the best way to output spark data
>> as JSON file. (File where each line is a JSON object)
>> Cheers,
>> Andrejs
>>
>
>

Reply via email to