Hi,

This may be cause, depending on your scheduler, only one Reducer may
be allocated per TT heartbeat. A reasoning of why this is the case is
explained here: http://search-hadoop.com/m/KYv8JhkOHc1

You may have better results in 1.0.3 using an alternative scheduler
such as FairScheduler with multiple-assignments-per-heartbeat turned
on (See http://hadoop.apache.org/common/docs/current/fair_scheduler.html
and boolean property "mapred.fairscheduler.assignmultiple" to enable)
or via CapacityScheduler (See
http://hadoop.apache.org/common/docs/current/capacity_scheduler.html)
which does it as well (OOB).

On Tue, May 22, 2012 at 5:36 PM, Andrés Durán <du...@tadium.es> wrote:
> Hello,
>
>        I'm working with a Hadoop, version is 1.0.3 and configured in 
> pseudo-distributed mode.
>
>        I have 128 reducers tasks and it's running in a local machine with 32 
> cores. The job is working fine and fast it  takes 1 hour and 30 minutes to 
> fininsh. But when the Job starts, the reducers are comming to the running 
> phase from the tasks queue very slow, it takes 7 minutes to allocate 32 tasks 
> in the running phase. Why is too slow to allocate task in running mode? It's 
> possible to adjust any variable in the jobs tracker setup to reduce this 
> allocation time?
>
>  Thanks to all!
>
>  Best regards,
>        Andrés Durán



-- 
Harsh J

Reply via email to