You are looking for total order partitioning of your map-emitted data. Have a look at the TotalOrderPartitioner class/usage.
On Tue, Mar 22, 2011 at 7:54 PM, JunYoung Kim <[email protected]> wrote: > hi, > > I run almost 60 ruduce tasks for a single job. > > if the outputs of a job are from part00 to part 59. > > is there way to write rows sequentially by sorted keys? > > curretly my outputs are like this. > > part00) > 1 > 10 > 12 > 14 > > part 01) > 2 > 4 > 6 > 11 > 13 > > part 02) > 3 > 5 > 7 > 8 > 9 > > but, my aim is to get the following results. > > part00) > 1 > 2 > 3 > 4 > 5 > > part01) > 6 > 7 > 8 > 9 > 10 > > part02) > 11 > 12 > 13 > 14 > 15 > > the hadoop is able to support this kind of one? > > thanks > -- Harsh J http://harshj.com
