I have a file on HDFS with a reduced block size. I created this overriding the dfs.block.size param on the hadoop fs -put command . hadoop fsck shows that this file has 15 blocks (as opposed to the normal 1 block) I did it so as to force Pig to use more maps than normal . On my pig command line I specify ' pig -Dpig.splitCombination=false ' to turn off the defaulr split combination logic. The jobs still ends running just one mapper. How can I achieve multiple maps ? Splitting the original file into multiple files would be my last resort.
Sam William [email protected]
