Re: process limits for streaming jar

2008-06-27 Thread Rick Cox
On Fri, Jun 27, 2008 at 08:57, Chris Anderson <[EMAIL PROTECTED]> wrote: > The problem is that when there are a large number of map tasks to > complete, Hadoop doesn't seem to obey the map.tasks.maximum. Instead, > it is spawning 8 map tasks per tasktracker (even when I change the > mapred.tasktra

Re: process limits for streaming jar

2008-06-27 Thread Chris Anderson
Having experimented some more, I've found that the simple solution is to limit the resource usage by limiting the # of map tasks and the memory they are allowed to consume. I'm specifying the constraints on the command line like this: -jobconf mapred.tasktracker.map.tasks.maximum=2 mapred.child.u

Re: process limits for streaming jar

2008-06-27 Thread Vinod KV
Allen Wittenauer wrote: This is essentially what we're doing via torque (and therefore hod). When we intend to move away from HOD(and thus torque) to using Hadoop Resource manager(HADOOP-3421) and scheduler(HADOOP-3412) interfaces, we need to move this resource management functionality into

Re: process limits for streaming jar

2008-06-26 Thread Allen Wittenauer
On 6/26/08 10:14 AM, "Joydeep Sen Sarma" <[EMAIL PROTECTED]> wrote: > However - in our environment - we spawn all streaming tasks through a > wrapper program (with the wrapper being defaulted across all users). We > can control resource uses from this wrapper (and also do some level of > job contro

RE: process limits for streaming jar

2008-06-26 Thread Joydeep Sen Sarma
Memory limits were handled in this jira: https://issues.apache.org/jira/browse/HADOOP-2765 However - in our environment - we spawn all streaming tasks through a wrapper program (with the wrapper being defaulted across all users). We can control resource uses from this wrapper (and also do some lev