Probably the cleanup work like clean shuffle files, tmp files cost too much
of CPUs, since if we run Spark Streaming for a long time, lots of files
will be generated, so cleanup this files before app is exited could be
time-consuming.

Thanks
Jerry

2015-03-11 10:43 GMT+08:00 Tathagata Das <t...@databricks.com>:

> Do you have event logging enabled?
> That could be the problem. The Master tries to aggressively recreate the
> web ui of the completed job with the event logs (when it is enabled)
> causing the Master to stall.
> I created a JIRA for this.
> https://issues.apache.org/jira/browse/SPARK-6270
>
> On Tue, Mar 10, 2015 at 7:10 PM, Xuelin Cao <xuelincao2...@gmail.com>
> wrote:
>
>>
>> Hey,
>>
>>      Recently, we found in our cluster, that when we kill a spark
>> streaming app, the whole cluster cannot response for 10 minutes.
>>
>>      And, we investigate the master node, and found the master process
>> consumes 100% CPU when we kill the spark streaming app.
>>
>>      How could it happen? Did anyone had the similar problem before?
>>
>>
>>
>>
>>
>

Reply via email to