Thanks Sean.

On Thu, Feb 19, 2015 at 5:33 PM, Sean Owen <so...@cloudera.com> wrote:

> I think that the newer Hadoop API does not expose this suggested min
> partitions parameter like the old one did. I believe you can try
> setting mapreduce.input.fileinputformat.split.{min,max}size instead on
> the Hadoop Configuration to suggest a max/min split size, and
> therefore bound the number of partitions you get back.
>
> On Thu, Feb 19, 2015 at 11:07 AM, twinkle sachdeva
> <twinkle.sachd...@gmail.com> wrote:
> > Hi,
> >
> > In our job, we need to process the data in small chunks, so  as to avoid
> GC
> > and other stuff. For this, we are using old API of hadoop as that let us
> > specify parameter like minPartitions.
> >
> > Does any one knows, If  there a way to do the same via newHadoopAPI also?
> > How that way will be different from older API?
> >
> > I am little bit aware of split size stuff, but not much aware regarding
> any
> > promise that minimum number of partitions criteria gets satisfied or not.
> >
> > Any pointers will be of help.
> >
> > Thanks,
> > Twinkle
>

Reply via email to