Well, I could try to do that,
but *partitionBy *method is anyway only supported for Parquet format even
in Spark 1.5.1
Narek
Narek Galstyan
Նարեկ Գալստյան
On 27 December 2015 at 21:50, Ted Yu wrote:
> Is upgrading to 1.5.x a possibility for you ?
>
> Cheers
>
> On Sun,
I don't think thats true (though if the docs are wrong we should fix
that). In Spark 1.5 we converted JSON to go through the same code path as
parquet.
On Mon, Dec 28, 2015 at 12:20 AM, Նարեկ Գալստեան
wrote:
> Well, I could try to do that,
> but *partitionBy *method is
have you tried to specify format of your output, might be parquet is
default format?
df.write().format("json").mode(SaveMode.Overwrite).save("/tmp/path");
On 27 December 2015 at 15:18, Նարեկ Գալստեան wrote:
> Hey all!
> I am willing to partition *json *data by a column
Is upgrading to 1.5.x a possibility for you ?
Cheers
On Sun, Dec 27, 2015 at 9:28 AM, Նարեկ Գալստեան
wrote:
>
> http://spark.apache.org/docs/1.4.1/api/scala/index.html#org.apache.spark.sql.DataFrameWriter
> I did try but it all was in vain.
> It is also explicitly
http://spark.apache.org/docs/1.4.1/api/scala/index.html#org.apache.spark.sql.DataFrameWriter
I did try but it all was in vain.
It is also explicitly written in api docs that it only supports Parquet.
Narek Galstyan
Նարեկ Գալստյան
On 27 December 2015 at 17:52, Igor Berman
Hey all!
I am willing to partition *json *data by a column name and store the result
as a collection of json files to be loaded to another database.
I could use spark's built in *partitonBy *function but it only outputs in
parquet format which is not desirable for me.
Could you suggest me a way