I noticed it is not possible to run simultaneously two jobs by the same
user, the second job gets stuck at map 0% reduce 0%


On 10/01/2008, Jeff Hammerbacher <[EMAIL PROTECTED]> wrote:
>
> it's a stopgap and doesn't seem to be working well for y!:
> https://issues.apache.org/jira/browse/HADOOP-2510
>
> On Jan 9, 2008 5:30 PM, Ted Dunning <[EMAIL PROTECTED]> wrote:
>
> >
> > What is the status of Hadoop on Demand?  Is it ready for prime time?
> >
> >
> > On 1/9/08 4:58 PM, "Aaron Kimball" <[EMAIL PROTECTED]> wrote:
> >
> > > I will add to the discussion that the ability to have multiple tasks
> of
> > > equal priority all making progress simultaneously is important in
> > > academic environments. There are a number of undergraduate programs
> > > which are starting to use Hadoop in code labs for students.
> > >
> > > Multiple students should be able to submit jobs and if one student's
> > > poorly-written task is grinding up a lot of cycles on a shared
> cluster,
> > > other students still need to be able to test their code in the
> meantime;
> > > ideally, they would not need to enter a lengthy job queue. ... I'd say
> > > that this actually applies to development clusters in general, where
> > > individual task performance is less important than the ability of
> > > multiple developers to test code concurrently.
> > >
> > > - Aaron
> > >
> > >
> > >
> > > Joydeep Sen Sarma wrote:
> > >>> that can run(per job) at any given time.
> > >>
> > >> not possible afaik - but i will be happy to hear otherwise.
> > >>
> > >> priorities are a good substitute though. there's no point needlessly
> > >> restricting concurrency if there's nothing else to run. if there is
> > something
> > >> else more important to run - then in most cases, assigning a higher
> > priority
> > >> to that other thing would make the right thing happen.
> > >>
> > >> except with long running tasks (usually reducers) that cannot be
> > preempted.
> > >> (Hadoop does not seem to use OS process priorities at all. I wonder
> if
> > >> process priorities can be used as a substitute for pre-emption.)
> > >>
> > >> HOD is another solution that you might want to look into - my
> > understanding
> > >> is that with HOD u can restrict the number of machines used by a job.
> > >>
> > >> ________________________________
> > >>
> > >> From: Xavier Stevens [mailto:[EMAIL PROTECTED]
> > >> Sent: Wed 1/9/2008 2:57 PM
> > >> To: hadoop-user@lucene.apache.org
> > >> Subject: RE: Question on running simultaneous jobs
> > >>
> > >>
> > >>
> > >> This doesn't work to solve this issue because it sets the total
> number
> > >> of map/reduce tasks. When setting the total number of map tasks I get
> > an
> > >> ArrayOutOfBoundsException within Hadoop; I believe because of the
> input
> > >> dataset size (around 90 million lines).
> > >>
> > >> I think it is important to make a distinction between setting total
> > >> number of map/reduce tasks and the number that can run(per job) at
> any
> > >> given time.  I would like only to restrict the later, while allowing
> > >> Hadoop to divide the data into chunks as it sees fit.
> > >>
> > >>
> > >> -----Original Message-----
> > >> From: Ted Dunning [mailto:[EMAIL PROTECTED]
> > >> Sent: Wednesday, January 09, 2008 1:50 PM
> > >> To: hadoop-user@lucene.apache.org
> > >> Subject: Re: Question on running simultaneous jobs
> > >>
> > >>
> > >> You may need to upgrade, but 15.1 does just fine with multiple jobs
> in
> > >> the cluster.  Use conf.setNumMapTasks(int) and
> > >> conf.setNumReduceTasks(int).
> > >>
> > >>
> > >> On 1/9/08 11:25 AM, "Xavier Stevens" <[EMAIL PROTECTED]> wrote:
> > >>
> > >>> Does Hadoop support running simultaneous jobs?  If so, what
> parameters
> > >>
> > >>> do I need to set in my job configuration?  We basically want to give
> a
> > >>
> > >>> job that takes a really long time, half of the total resources of
> the
> > >>> cluster so other jobs don't queue up behind it.
> > >>>
> > >>> I am using Hadoop 0.14.2 currently.  I tried setting
> > >>> mapred.tasktracker.tasks.maximum to be half of the maximum specified
> > >>> in mapred-default.xml.  This shows the change in the web
> > >>> administration page for the job, but it has no effect on the actual
> > >>> numbers of tasks running.
> > >>>
> > >>> Thanks,
> > >>>
> > >>> Xavier
> > >>>
> > >>
> > >>
> > >>
> > >>
> > >>
> > >>
> >
> >
>



-- 
---------------------------------------------------------
شهر مبارك كريم
كل عام و أنتم بخير
---------------------------------------------------------
Honsali Khalil − 本査理 カリル
Academic>Japan>NIT>Grad. Sc. Eng.>Dept. CS>Matsuo&Tsumura Lab.
http://www.matlab.nitech.ac.jp/~k-hon/
+81 (zero-)eight-zero 5134 8119
[EMAIL PROTECTED] (instant reply mail)

Reply via email to