The only difference would be that with 4 reducers your keys would get partitioned based on their hashCode() implementation (if you use the default hash partitioner) (I'd check the key impl. here, first thing, if its a custom key impl.), and each be sent to one reducer.
Check the input record counters on your reducers, and the total map output record counters - they should add up and be equal to the latter. Also make sure you aren't skipping out on the reducer iterator under any condition, when you are doing the reducer op. I'm guessing its mostly your logic that's somehow causing this but I do not have your source bits to say that for sure. On 21-Jan-2012, at 11:47 PM, Thamizhannal Paramasivam wrote: > Hi All, > I am experimenting MapReduce program on Hadoop-0.19. This program has single > input file with 7 records(later it can have many records on multiple files) > and each input suppose to produce 11 output records. When it runs with > no_of_reducer=4. It produces only 33 records. But, when I ran with > no_of_reducer=1 then it produces 77 records as expected. > > What could be the reason for this ? I am missing any configuration parameter. > > Thanks > Tamil > -- Harsh J Customer Ops. Engineer, Cloudera