Have a look at MultipleOutputs in the hadoop API. Spark can read and write to 
arbitrary hadoop formats. 

> On Apr 4, 2014, at 6:01 AM, dmpour23 <dmpou...@gmail.com> wrote:
> 
> Hi all,
> Say I have an input file which I would like to partition using
> HashPartitioner k times.
> 
> Calling  rdd.saveAsTextFile(""hdfs://"); will save k files as part-00000
> part-0000k    
> Is there a way to save each partition in specific folders?
> 
> i.e. src
>      part0/part-00000 
>      part1/part-00001
>      part1/part-0000k
> 
> thanks
> Dimitri
> 
> 
> 
> 
> 
> --
> View this message in context: 
> http://apache-spark-user-list.1001560.n3.nabble.com/how-to-save-RDD-partitions-in-different-folders-tp3754.html
> Sent from the Apache Spark User List mailing list archive at Nabble.com.

Reply via email to