I see the same thing. 

A workaround is to put a Thread.sleep(5000) statement before sc.stop()

Let us know how it goes. 



> On Sep 17, 2014, at 3:43 AM, "wyphao.2007" <wyphao.2...@163.com> wrote:
> 
> Hi,  When I run spark job on yarn,and the job finished success,but I found 
> there are some error logs in the logfile as follow(the red color text):
> 
> 
> 14/09/17 18:25:03 INFO ui.SparkUI: Stopped Spark web UI at 
> http://sparkserver2.cn:63937
> 14/09/17 18:25:03 INFO scheduler.DAGScheduler: Stopping DAGScheduler
> 14/09/17 18:25:03 INFO cluster.YarnClusterSchedulerBackend: Shutting down all 
> executors
> 14/09/17 18:25:03 INFO cluster.YarnClusterSchedulerBackend: Asking each 
> executor to shut down
> 14/09/17 18:25:03 INFO network.ConnectionManager: Removing SendingConnection 
> to ConnectionManagerId(sparkserver2.cn,9072)
> 14/09/17 18:25:03 INFO network.ConnectionManager: Removing 
> ReceivingConnection to ConnectionManagerId(sparkserver2.cn,9072)
> 14/09/17 18:25:03 ERROR network.ConnectionManager: Corresponding 
> SendingConnection to ConnectionManagerId(sparkserver2.cn,9072) not found
> 14/09/17 18:25:03 INFO network.ConnectionManager: Removing 
> ReceivingConnection to ConnectionManagerId(sparkserver2.cn,14474)
> 14/09/17 18:25:03 INFO network.ConnectionManager: Removing SendingConnection 
> to ConnectionManagerId(sparkserver2.cn,14474)
> 14/09/17 18:25:03 INFO network.ConnectionManager: Removing SendingConnection 
> to ConnectionManagerId(sparkserver2.cn,14474)
> 14/09/17 18:25:04 INFO spark.MapOutputTrackerMasterActor: 
> MapOutputTrackerActor stopped!
> 14/09/17 18:25:04 INFO network.ConnectionManager: Selector thread was 
> interrupted!
> 14/09/17 18:25:04 INFO network.ConnectionManager: Removing SendingConnection 
> to ConnectionManagerId(sparkserver2.cn,9072)
> 14/09/17 18:25:04 INFO network.ConnectionManager: Removing SendingConnection 
> to ConnectionManagerId(sparkserver2.cn,14474)
> 14/09/17 18:25:04 INFO network.ConnectionManager: Removing 
> ReceivingConnection to ConnectionManagerId(sparkserver2.cn,9072)
> 14/09/17 18:25:04 ERROR network.ConnectionManager: Corresponding 
> SendingConnection to ConnectionManagerId(sparkserver2.cn,9072) not found
> 14/09/17 18:25:04 INFO network.ConnectionManager: Removing 
> ReceivingConnection to ConnectionManagerId(sparkserver2.cn,14474)
> 14/09/17 18:25:04 ERROR network.ConnectionManager: Corresponding 
> SendingConnection to ConnectionManagerId(sparkserver2.cn,14474) not found
> 14/09/17 18:25:04 WARN network.ConnectionManager: All connections not cleaned 
> up
> 14/09/17 18:25:04 INFO network.ConnectionManager: ConnectionManager stopped
> 14/09/17 18:25:04 INFO storage.MemoryStore: MemoryStore cleared
> 14/09/17 18:25:04 INFO storage.BlockManager: BlockManager stopped
> 14/09/17 18:25:04 INFO storage.BlockManagerMaster: BlockManagerMaster stopped
> 14/09/17 18:25:04 INFO spark.SparkContext: Successfully stopped SparkContext
> 14/09/17 18:25:04 INFO yarn.ApplicationMaster: Unregistering 
> ApplicationMaster with SUCCEEDED
> 14/09/17 18:25:04 INFO remote.RemoteActorRefProvider$RemotingTerminator: 
> Shutting down remote daemon.
> 14/09/17 18:25:04 INFO remote.RemoteActorRefProvider$RemotingTerminator: 
> Remote daemon shut down; proceeding with flushing remote transports.
> 14/09/17 18:25:04 INFO impl.AMRMClientImpl: Waiting for application to be 
> successfully unregistered.
> 14/09/17 18:25:04 INFO Remoting: Remoting shut down
> 14/09/17 18:25:04 INFO remote.RemoteActorRefProvider$RemotingTerminator: 
> Remoting shut down.
> 
> 
> What is the cause of this error? My spark version is 1.1.0 &  hadoop version 
> is 2.2.0.
> Thank you.

---------------------------------------------------------------------
To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org
For additional commands, e-mail: dev-h...@spark.apache.org

Reply via email to