I ran into some issues with it a while ago, and submitted a couple PRs to fix it:
https://github.com/apache/spark/pull/2401 https://github.com/apache/spark/pull/3024 Do these look relevant? What version of Spark are you running? On Sat, Apr 11, 2015 at 9:33 AM, Tom Arnfeld <[email protected]> wrote: > Hey, > > Not sure whether it's best to ask this on the spark mailing list or the > mesos one, so I'll try here first :-) > > I'm having a bit of trouble with out of memory errors in my spark jobs... > it seems fairly odd to me that memory resources can only be set at the > executor level, and not also at the task level. For example, as far as I > can tell there's only a *spark.executor.memory* config option. > > Surely the memory requirements of a single executor are quite dramatically > influenced by the number of concurrent tasks running? Given a shared > cluster, I have no idea what % of an individual slave my executor is going > to get, so I basically have to set the executor memory to a value that's > correct when the whole machine is in use... > > Has anyone else running Spark on Mesos come across this, or maybe someone > could correct my understanding of the config options? > > Thanks! > > Tom. >

