Hi Kenneth, There currently isn't a great way to do this, especially on Hadoop 1. You can give your spawned job a unique name, and have your Java action check for any running jobs with that name before running a new instance of the job. In later versions of Hadoop 2, we can actually set "tags" on jobs, so we've added a similar check to most actions, but using a tag instead of the job name.
- Robert On Thu, Oct 23, 2014 at 10:53 AM, Kenneth Kavaliauskas < [email protected]> wrote: > I’m using an oozie java action step to start a java main. This java > application does some calculations and then runs another map-reduce job > based on that data. > > Since the oozie java action runs as a map-only job through oozie it is > also seen in job tracker. > > > > One of our nodes was low on memory so the task tracker killed the oozie > map-only job and restarted it on another node. > > However before killing it, the java application had already spawned its > own map reduce job. > > When the oozie map-only job was restarted on the other node, it again > spawned yet another map-reduce job with the same data as the former one. > > Looking in job tracker now has duplicate map-reduce jobs running against > the same data. > > > > How do you prevent/manage/alter settings such that the java program that > oozie initiates in the map-only process only get run once. > > > > Any help would be appreciated, > > Ken > > [image: Description: Description: > http://insite.telcordia.com/corporate/pr/brand/downloads/templates/EricssonLine.gif] > > Ken Kavaliauskas > > Software Engineer > > > > Ericsson > > [email protected] > > >
