Any more thoughts on this? I'm not sure how to do this yet.
On Fri, Aug 29, 2014 at 12:10 PM, Victor Tso-Guillen <v...@paxata.com> wrote: > Standalone. I'd love to tell it that my one executor can simultaneously > serve, say, 16 tasks at once for an arbitrary number of distinct jobs. > > > On Fri, Aug 29, 2014 at 11:29 AM, Matei Zaharia <matei.zaha...@gmail.com> > wrote: > >> Yes, executors run one task per core of your machine by default. You can >> also manually launch them with more worker threads than you have cores. >> What cluster manager are you on? >> >> Matei >> >> On August 29, 2014 at 11:24:33 AM, Victor Tso-Guillen (v...@paxata.com) >> wrote: >> >> I'm thinking of local mode where multiple virtual executors occupy the >> same vm. Can we have the same configuration in spark standalone cluster >> mode? >> >> >