Re: Mapper basic question

2012-07-11 Thread Manoj Babu
Regards > Bejoy KS > > Sent from handheld, please excuse typos. > -- > *From: * Manoj Babu > *Date: *Wed, 11 Jul 2012 18:17:41 +0530 > *To: * > *ReplyTo: * mapreduce-user@hadoop.apache.org > *Subject: *Re: Mapper basic question > > Hi

Re: Mapper basic question

2012-07-11 Thread Bejoy KS
2012 18:17:41 To: Reply-To: mapreduce-user@hadoop.apache.org Subject: Re: Mapper basic question Hi Tariq \Arun, The no of blocks(splits) = *total no of file size/hdfs block size * replicate value* The no of splits is again nothing but the blocks here. Other than increasing the block size(input

Re: Mapper basic question

2012-07-11 Thread Manoj Babu
Hi Tariq \Arun, The no of blocks(splits) = *total no of file size/hdfs block size * replicate value* The no of splits is again nothing but the blocks here. Other than increasing the block size(input splits) is it possible to limit that no of mappers? Cheers! Manoj. On Wed, Jul 11, 2012 at 6

Re: Mapper basic question

2012-07-11 Thread Arun C Murthy
Take a look at CombineFileInputFormat - this will create 'meta splits' which include multiple small spilts, thus reducing #maps which are run. Arun On Jul 11, 2012, at 5:29 AM, Manoj Babu wrote: > Hi, > > The no of mappers is depends on the no of blocks. Is it possible to limit the > no of ma

Re: Mapper basic question

2012-07-11 Thread Mohammad Tariq
Hello Manoj, It is not the block that determines the no of mappers. It is rather based on the no of input splits. No of mappers = no of input splits. And I did not get what do you mean by 'no of mapper size'. It is possible to configure the input splits though. Hope it helps. Regards, Mo