One nice feature is to attach different configuration for different job. 
You can do it either by command line "-conf" or programatically.  We are 
using that extensively while experimenting different setting. 

Zhu, Guojun
Modeling Sr Graduate
571-3824370
guojun_...@freddiemac.com
Financial Engineering
Freddie Mac



   Radim Kolar <h...@filez.com> 
   05/10/2012 07:56 AM
   Please respond to
mapreduce-user@hadoop.apache.org


To
mapreduce-user@hadoop.apache.org
cc

Subject
Re: max 1 mapper per node







> We've been against these 'features' since it leads to very bad 
> behaviour across the cluster with multiple apps/users etc.
Its not new feature, its extension of existing resource scheduling which 
works good enough only for RAM. There are 2 other resources - CPU cores 
and network IO which needs to be considered.

We have job which is doing lot of network IO in mapper and its desirable 
to run mappers on different nodes even if reading blocks from HDFS will 
not be local.

Our second job is burning all CPU cores on machine while doing 
computations, its important for mappers not to land on same node.

Reply via email to