A minor correction: CapacityScheduler doesn't seem to do multi-reducer assignments (or at least not in 1.x), but does do multi-map assignments. This is for the same reason as http://search-hadoop.com/m/KYv8JhkOHc1. FairScheduler in 1.x supports multi-map and multi-reducer assignments over single heartbeats, which should do good on your single 32-task machine.
Do give it a try and let us know! On Tue, May 22, 2012 at 5:51 PM, Harsh J <ha...@cloudera.com> wrote: > Hi, > > This may be cause, depending on your scheduler, only one Reducer may > be allocated per TT heartbeat. A reasoning of why this is the case is > explained here: http://search-hadoop.com/m/KYv8JhkOHc1 > > You may have better results in 1.0.3 using an alternative scheduler > such as FairScheduler with multiple-assignments-per-heartbeat turned > on (See http://hadoop.apache.org/common/docs/current/fair_scheduler.html > and boolean property "mapred.fairscheduler.assignmultiple" to enable) > or via CapacityScheduler (See > http://hadoop.apache.org/common/docs/current/capacity_scheduler.html) > which does it as well (OOB). > > On Tue, May 22, 2012 at 5:36 PM, Andrés Durán <du...@tadium.es> wrote: >> Hello, >> >> I'm working with a Hadoop, version is 1.0.3 and configured in >> pseudo-distributed mode. >> >> I have 128 reducers tasks and it's running in a local machine with 32 >> cores. The job is working fine and fast it takes 1 hour and 30 minutes to >> fininsh. But when the Job starts, the reducers are comming to the running >> phase from the tasks queue very slow, it takes 7 minutes to allocate 32 >> tasks in the running phase. Why is too slow to allocate task in running >> mode? It's possible to adjust any variable in the jobs tracker setup to >> reduce this allocation time? >> >> Thanks to all! >> >> Best regards, >> Andrés Durán > > > > -- > Harsh J -- Harsh J