thanks.  I think I will investigate capacity scheduler.

On Fri, May 11, 2012 at 7:26 AM, Michael Segel <michael_se...@hotmail.com>wrote:

> Just a quick note...
>
> If your task is currently occupying a slot,  the only way to release the
> slot is to kill the specific task.
> If you are using FS, you can move the task to another queue and/or you can
> lower the job's priority which will cause new tasks to spawn  slower than
> other jobs so you will eventually free up the cluster.
>
> There isn't a way to 'freeze' or stop a job mid state.
>
> Is the issue that the job has a large number of slots, or is it an issue
> of the individual tasks taking a  long time to complete?
>
> If its the latter, you will probably want to go to a capacity scheduler
> over the fair scheduler.
>
> HTH
>
> -Mike
>
> On May 11, 2012, at 6:08 AM, Harsh J wrote:
>
> > I do not know about the per-host slot control (that is most likely not
> > supported, or not yet anyway - and perhaps feels wrong to do), but the
> > rest of the needs can be doable if you use schedulers and
> > queues/pools.
> >
> > If you use FairScheduler (FS), ensure that this job always goes to a
> > special pool and when you want to freeze the pool simply set the
> > pool's maxMaps and maxReduces to 0. Likewise, control max simultaneous
> > tasks as you wish, to constrict instead of freeze. When you make
> > changes to the FairScheduler configs, you do not need to restart the
> > JT, and you may simply wait a few seconds for FairScheduler to refresh
> > its own configs.
> >
> > More on FS at
> http://hadoop.apache.org/common/docs/current/fair_scheduler.html
> >
> > If you use CapacityScheduler (CS), then I believe you can do this by
> > again making sure the job goes to a specific queue, and when needed to
> > freeze it, simply set the queue's maximum-capacity to 0 (percentage)
> > or to constrict it, choose a lower, positive percentage value as you
> > need. You can also refresh CS to pick up config changes by refreshing
> > queues via mradmin.
> >
> > More on CS at
> http://hadoop.apache.org/common/docs/current/capacity_scheduler.html
> >
> > Either approach will not freeze/constrict the job immediately, but
> > should certainly prevent it from progressing. Meaning, their existing
> > running tasks during the time of changes made to scheduler config will
> > continue to run till completion but further tasks scheduling from
> > those jobs shall begin seeing effect of the changes made.
> >
> > P.s. A better solution would be to make your job not take as many
> > days, somehow? :-)
> >
> > On Fri, May 11, 2012 at 4:13 PM, Rita <rmorgan...@gmail.com> wrote:
> >> I have a rather large map reduce job which takes few days. I was
> wondering
> >> if its possible for me to freeze the job or make the job less
> intensive. Is
> >> it possible to reduce the number of slots per host and then I can
> increase
> >> them overnight?
> >>
> >>
> >> tia
> >>
> >> --
> >> --- Get your facts first, then you can distort them as you please.--
> >
> >
> >
> > --
> > Harsh J
> >
>
>


-- 
--- Get your facts first, then you can distort them as you please.--

Reply via email to