The larger question is how many blocks are required to store a 100MB file if the HDFS block size is 64MB.
If it takes 2 blocks then when you run your job, you will have 1 mapper per block, unless the file is not splittable. (But from your example its a simple text file which is splittable.) Again, where are you getting these questions? On Nov 7, 2012, at 10:46 AM, Ramasubramanian Narayanan <ramasubramanian.naraya...@gmail.com> wrote: > Hi, > > Thanks! > > But it is given as 100 Mappers... I think we can also use 'n' number of > Mappers as the same as the number of input files... (not for this > question)... If you know more detail on that please share.. > > Note : I forgot from where this question I taken :) > > regards, > Rams. > > On Wed, Nov 7, 2012 at 10:01 PM, Michael Segel <michael_se...@hotmail.com> > wrote: > 0 Custer didn't run. He got surrounded and then massacred. :-P (See Custer's > last stand at Little Big Horn) > > Ok... plain text files 100 files 2 blocks each would by default attempt to > schedule 200 mappers. > > Is this one of those online Cert questions? > > On Nov 7, 2012, at 10:20 AM, Ramasubramanian Narayanan > <ramasubramanian.naraya...@gmail.com> wrote: > > > Hi, > > > > If a Custer’s HDFS block size is 64MB. You have a directory containing 100 > > plain text files, each of which Is 100MB in size. The InputFormat for your > > job is TextInputFormat. How many Mappers will run? > > > > regards, > > Rams > > > > > > > >