Hi, Does the output hadoop block size depend on spark tasks number?
In my application, when the number of tasks increases, hadoop block size decreases. For a big number of tasks, I get a very high number of 1 MB files generated by saveAsSequenceFile(). How is it possible to control the block size by spark?
