perhaps creating Fair Scheduler Pools might help?  there's no way to pin
certain nodes to a pool, but you can specify minShares (cpu's).  not sure
if that would help, but worth looking in to.


On Tue, Jul 8, 2014 at 7:37 PM, haopu <hw...@qilinsoft.com> wrote:

> In a standalone cluster, is there way to specify the stage to be running
> on a
> faster worker?
>
> That stage is reading HDFS file and then doing some filter operations.  The
> tasks are assigned to the slower worker also, but the slower worker delays
> to launch because it's running some tasks from other stages.
>
> So I think it may be better to assign stage to a worker. Any suggestions?
>
> And will the cluster on Yarn help?
>
>
>
> --
> View this message in context:
> http://apache-spark-user-list.1001560.n3.nabble.com/slower-worker-node-in-the-cluster-tp9125.html
> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>

Reply via email to