I got a 40 node cdh 5.1 cluster and attempting to run a simple spark app that
processes about 10-15GB raw data but I keep running into this error:
 
  java.lang.OutOfMemoryError: GC overhead limit exceeded
 
Each node has 8 cores and 2GB memory. I notice the heap size on the
executors is set to 512MB with total heap size on each executor is set to
2GB. Wanted to know whats the heap size needs to be set to for such data
sizes and if anyone had input on other config changes that will help as
well.
 
Thanks for the input!



--
View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/Spark-app-throwing-java-lang-OutOfMemoryError-GC-overhead-limit-exceeded-tp11350.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.

---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org

Reply via email to