I see mapred.tasktracker.reduce.tasks.maximum and 
mapred.tasktracker.map.tasks.maximum, but I'm wondering if there isn't another 
tuning parameter I need to look at.

I can tune the task tracker so that when I have many jobs running, with many 
simultaneous maps and reduces I utilize 95% of cpu and memory. 

Inevitably though I end up with a huge final reduce task that only uses half of 
of my cluster because I have reserved the other half for Mapping. 

Is there a way around this problem?  

Seems like there should also be a maximum number of reducers conditional on no 
Map tasks running. 

-JD

Reply via email to