Presumably the limit could be made dynamic. The limit could be max(static_limit, number of cores in cluster / # active jobs)
On 1/10/08 9:56 AM, "Joydeep Sen Sarma" <[EMAIL PROTECTED]> wrote: > this may be simple - but is this the right solution? (and i have the same > concern about hod) > > if the cluster is unused - why restrict parallelism? if someone's willing to > wake up at 4am to beat the crowd - they would just absolutely hate this. > > > -----Original Message----- > From: Doug Cutting [mailto:[EMAIL PROTECTED] > Sent: Thu 1/10/2008 9:50 AM > To: hadoop-user@lucene.apache.org > Subject: Re: Question on running simultaneous jobs > > Aaron Kimball wrote: >> Multiple students should be able to submit jobs and if one student's >> poorly-written task is grinding up a lot of cycles on a shared cluster, >> other students still need to be able to test their code in the meantime; > > I think a simple approach to address this is to limit the number of > tasks from a job that are permitted to execute simultaneously. If, for > example, you have a cluster of 50 dual-core nodes, with 100 map task > slots and 100 reduce task slots, and the configured limit is 25 > simultaneous tasks/job, then four or more jobs will be able to run at a > time. This will permit faster jobs to pass slower jobs. This approach > also avoids some problems we've seen with HOD, where nodes are > underutilized during the tail of jobs, and with input locality. > > The JobTracker already handles simultaneously executing jobs, so the > primary change required is just to task allocation, and thus should not > prove intractable. > > I've added a Jira issue for this: > > https://issues.apache.org/jira/browse/HADOOP-2573 > > Please add further comments there. > > Doug >