Hey David,

The only time I've seen orphaned containers is when the NM dies. If the NM
isn't running, the RM has no means to kill the containers on a node. Can
you verify that the NM was healthy at the time of the shut down?

If it wasn't healthy and/or it was restarted, one option that may help is
NM Recovery:
https://hadoop.apache.org/docs/r2.7.2/hadoop-yarn/hadoop-yarn-site/NodeManagerRestart.html

With NM Recovery, the NM will resume control over containers that were
running when the NM shut down. This option has virtually eliminated
orphaned containers in our clusters.

-Jake

On Tue, May 17, 2016 at 11:54 PM, David Yu <david...@optimizely.com> wrote:

> Samza version = 0.10.0
> YARN version = Hadoop 2.6.0-cdh5.4.9
>
> We are experience issues when killing a Samza job:
>
> $ yarn application -kill application_1463512986427_0007
>
> Killing application application_1463512986427_0007
>
> 16/05/18 06:29:05 INFO impl.YarnClientImpl: Killed application
> application_1463512986427_0007
>
> RM shows that the job is killed. However, the samza containers are still
> left running.
>
> Any idea why this is happening?
>
> Thanks,
> David
>

Reply via email to