Re: java.lang.OutOfMemoryError at simple local test

2014-11-26 Thread rzykov
We made some changes in code (it generates 1000 * 1000 elements) and memory limits up to 100M: def generate = { for{ j - 1 to 10 i - 1 to 1000 } yield(j, i) } ~/soft/spark-1.1.0-bin-hadoop2.3/bin/spark-submit --master local --executor-memory 100M --driver-memory 100M --class Spill

java.lang.OutOfMemoryError at simple local test

2014-11-22 Thread rzykov
Dear all, Unfortunately I've not got ant respond in users forum. That's why I decided to publish this question here. We encountered problems of failed jobs with huge amount of data. For example, an application works perfectly with relative small sized data, but when it grows in 2 times this