Hi

One question about the power of spark.shuffle.spill -
(I know this has been asked several times :-)

Basically, in handling a (cached) dataset that doesn't fit in memory, Spark can 
spill it to disk.

However, can I say that, when this is enabled, Spark can handle the situation 
faultlessly, no matter -

(1)    How big the data set is (as compared to the available memory)

(2)    How complex the detailed calculation is being carried out
Can spark.shuffle.spill handle this perfectly?

Here we assume that (1) the disk space has no limitations and (2) the code is 
correctly written according to the functional requirements.

The reason to ask this is, under such situations, I kept receiving warnings 
like "FetchFailed", if memory usage reaches the limit.

Thanks
YC

Reply via email to