Hello,

Adding to Aleksandr's suggestion, you could also lower the timeout if
a throw condition can't be determined.

On Thu, Aug 4, 2011 at 5:10 AM, Aleksandr Elbakyan <[email protected]> wrote:
> Hello,
>
> You can just throw run time exception. In that case it will fail :)
>
> Regards,
> Aleksandr
>
> --- On Wed, 8/3/11, Adam Shook <[email protected]> wrote:
>
> From: Adam Shook <[email protected]>
> Subject: Kill Task Programmatically
> To: "[email protected]" <[email protected]>
> Date: Wednesday, August 3, 2011, 3:33 PM
>
> Is there any way I can programmatically kill or fail a task, preferably from 
> inside a Mapper or Reducer?
>
> At any time during a map or reduce task, I have a use case where I know it 
> won't succeed based solely on the machine it is running on.  It is rare, but 
> I would prefer to kill the task and have Hadoop start it up on a different 
> machine as usual instead of waiting for the 10 minute default timeout.
>
> I suppose the speculative execution could take care of it, but I would rather 
> not rely on it if I am able to kill it myself.
>
> Thanks,
> Adam
>



-- 
Harsh J

Reply via email to