Re: java.lang.OutOfMemoryError at simple local test

2014-11-26 Thread rzykov
We made some changes in code (it generates 1000 * 1000 elements) and memory limits up to 100M: def generate = { for{ j <- 1 to 10 i <- 1 to 1000 } yield(j, i) } ~/soft/spark-1.1.0-bin-hadoop2.3/bin/spark-submit --master local --executor-memory 100M --driver-memory 100M --class Spill -

Re: java.lang.OutOfMemoryError at simple local test

2014-11-22 Thread Sean Owen
10M is tiny compared to all of the overhead of running a lot complex Scala based app in a JVM. I think you may be bumping up against practical minimum sizes and that you may find it is not really the data size? I don't think it really scales down this far. On Nov 22, 2014 2:24 PM, "rzykov" wrote: