thanks.
On Thu, Sep 25, 2014 at 10:25 PM, Marcelo Vanzin [via Apache Spark User List] <ml-node+s1001560n15177...@n3.nabble.com> wrote: > From spark-submit --help: > > YARN-only: > --executor-cores NUM Number of cores per executor (Default: 1). > --queue QUEUE_NAME The YARN queue to submit to (Default: > "default"). > --num-executors NUM Number of executors to launch (Default: 2). > --archives ARCHIVES Comma separated list of archives to be > extracted into the > working directory of each executor. > > On Thu, Sep 25, 2014 at 2:20 PM, Tamas Jambor <[hidden email]> wrote: > >> Thank you. >> >> Where is the number of containers set? >> >> On Thu, Sep 25, 2014 at 7:17 PM, Marcelo Vanzin <[hidden email]> wrote: >>> On Thu, Sep 25, 2014 at 8:55 AM, jamborta <[hidden email]> wrote: >>>> I am running spark with the default settings in yarn client mode. For >>>> some >>>> reason yarn always allocates three containers to the application >>>> (wondering >>>> where it is set?), and only uses two of them. >>> >>> The default number of executors in Yarn mode is 2; so you have 2 >>> executors + the application master, so 3 containers. >>> >>>> Also the cpus on the cluster never go over 50%, I turned off the fair >>>> scheduler and set high spark.cores.max. Is there some additional >>>> settings I >>>> am missing? >>> >>> You probably need to request more cores (--executor-cores). Don't >>> remember if that is respected in Yarn, but should be. >>> >>> -- >>> Marcelo > > > > -- > Marcelo > > --------------------------------------------------------------------- > To unsubscribe, e-mail: [hidden email] > For additional commands, e-mail: [hidden email] > > > > ________________________________ > If you reply to this email, your message will be added to the discussion > below: > http://apache-spark-user-list.1001560.n3.nabble.com/Yarn-number-of-containers-tp15148p15177.html > To unsubscribe from Yarn number of containers, click here. > NAML -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Yarn-number-of-containers-tp15148p15191.html Sent from the Apache Spark User List mailing list archive at Nabble.com.