Hi,

Does the output hadoop block size depend on spark tasks number?

In my application, when the number of tasks increases, hadoop block size
decreases. For a big number of tasks, I get a very high number of 1 MB
files generated by saveAsSequenceFile().

How is it possible to control the block size by spark?

Reply via email to