Set mapred.min.split.size D
On Mon, Apr 23, 2012 at 4:30 PM, Sam William <[email protected]> wrote: > I have a file on HDFS with a reduced block size. I created this overriding > the dfs.block.size param on the hadoop fs -put command . hadoop fsck shows > that this file has 15 blocks (as opposed to the normal 1 block) I did it so > as to force Pig to use more maps than normal . On my pig command line I > specify ' pig -Dpig.splitCombination=false ' to turn off the defaulr split > combination logic. The jobs still ends running just one mapper. How can I > achieve multiple maps ? Splitting the original file into multiple files > would be my last resort. > > > > Sam William > [email protected] > > >
