how many reducer do you have? You should make this value larger then 1 to
make mapper and reducer run concurrently. You can set this value from
JobConf.*setNumReduceTasks*().


2009/3/2 Rasit OZDAS <rasitoz...@gmail.com>

> Hi!
>
> Whatever code I run on hadoop, reduce starts a few seconds after map
> finishes.
> And worse, when I run 10 jobs parallely (using threads and sending one
> after
> another)
> all maps finish sequentially, then after 8-10 seconds reduces start.
> I use reducer also as combiner, my cluster has 6 machines, namenode and
> jobtracker run also as slaves.
> There were 44 maps and 6 reduces in the last example, I never tried a
> bigger
> job.
>
> What can the problem be? I've read somewhere that this is not the normal
> behaviour.
> Replication factor is 3.
> Thank you in advance for any pointers.
>
> Rasit
>



-- 
http://daily.appspot.com/food/

Reply via email to