Hi

    Suppose that your input file are 10 with total size 64mb , I think you will 
get the 10 maps.

    By the ways,the uber mode is only for yarn . Suppose you have actually 1 
map ,yarn will at least create two containers , one for app master and the 
other for the map , if uber mode is enabled with the yarn , yarn will only 
create 1 container for both app master and the map. 
    

发自我的 iPhone

在 2013-5-6,22:45,Rahul Bhattacharjee <[email protected]> 写道:

> Hi,
> 
> I was going through the definition of Uber Job of Hadoop.
> 
> A job is considered uber when it has 10 or less maps , one reducer and the 
> complete data is less than one dfs block size.
> 
> I have some doubts here-
> 
> Splits are created as per the dfs block size.Creating 10 mappers are possible 
> from one block of data by some settings change (changing the max split size). 
> But trying to understand , why would some job need to run around 10 maps for 
> 64 MB of data.
> One thing may be that the job is immensely CUP intensive. Will it be a 
> correct assumption? or is there is any other reason for this.
> 
> Thanks,
> Rahul
> 
> 

Reply via email to