Regards
> Bejoy KS
>
> Sent from handheld, please excuse typos.
> --
> *From: * Manoj Babu
> *Date: *Wed, 11 Jul 2012 18:17:41 +0530
> *To: *
> *ReplyTo: * mapreduce-user@hadoop.apache.org
> *Subject: *Re: Mapper basic question
>
> Hi
2012 18:17:41
To:
Reply-To: mapreduce-user@hadoop.apache.org
Subject: Re: Mapper basic question
Hi Tariq \Arun,
The no of blocks(splits) = *total no of file size/hdfs block size *
replicate value*
The no of splits is again nothing but the blocks here.
Other than increasing the block size(input
Hi Tariq \Arun,
The no of blocks(splits) = *total no of file size/hdfs block size *
replicate value*
The no of splits is again nothing but the blocks here.
Other than increasing the block size(input splits) is it possible to limit
that no of mappers?
Cheers!
Manoj.
On Wed, Jul 11, 2012 at 6
Take a look at CombineFileInputFormat - this will create 'meta splits' which
include multiple small spilts, thus reducing #maps which are run.
Arun
On Jul 11, 2012, at 5:29 AM, Manoj Babu wrote:
> Hi,
>
> The no of mappers is depends on the no of blocks. Is it possible to limit the
> no of ma
Hello Manoj,
It is not the block that determines the no of mappers. It is
rather based on the no of input splits. No of mappers = no of input
splits.
And I did not get what do you mean by 'no of mapper size'. It is
possible to configure the input splits though. Hope it helps.
Regards,
Mo