I found the error. Due to some refactoring, a wrong message was sent to the
JobManager in the JobManagerInfoServlet.java. I pushed a fix. Could you try
it out again?

On Wed, Feb 11, 2015 at 11:34 AM, Till Rohrmann <till.rohrm...@gmail.com>
wrote:

> Could you check the rebasing because it seems as if the web server is now
> sending RequestArchivedJobs messages to the JobManager which should not
> happen. These messages should go directly to the MemoryArchivist. The
> corresponding file is JobManagerInfoServlet.java, I think.
>
> On Wed, Feb 11, 2015 at 11:11 AM, Chesnay Schepler <
> chesnay.schep...@fu-berlin.de> wrote:
>
>> I just tried Till's fix, rebased to the latest master and got a whole lot
>> of these exceptions right away:
>>
>> java.lang.Exception: The slot in which the task was scheduled has been
>> killed (probably loss of TaskManager).
>>     at org.apache.flink.runtime.instance.SimpleSlot.cancel(
>> SimpleSlot.java:98)
>>     at org.apache.flink.runtime.jobmanager.scheduler.
>> SlotSharingGroupAssignment.releaseSimpleSlot(SlotSharingGroupAssignment.
>> java:320)
>>     at org.apache.flink.runtime.jobmanager.scheduler.
>> SlotSharingGroupAssignment.releaseSharedSlot(SlotSharingGroupAssignment.
>> java:304)
>>     at org.apache.flink.runtime.instance.SharedSlot.
>> releaseSlot(SharedSlot.java:106)
>>     at org.apache.flink.runtime.instance.Instance.markDead(
>> Instance.java:148)
>>     at org.apache.flink.runtime.instance.InstanceManager.
>> shutdown(InstanceManager.java:111)
>>     at org.apache.flink.runtime.jobmanager.JobManager.
>> postStop(JobManager.scala:132)
>>     at org.apache.flink.runtime.jobmanager.JobManager$$
>> anonfun$main$1$$anon$1.org$apache$flink$runtime$
>> jobmanager$WithWebServer$$super$postStop(JobManager.scala:559)
>>     at org.apache.flink.runtime.jobmanager.WithWebServer$
>> class.postStop(WithWebServer.scala:38)
>>     at org.apache.flink.runtime.jobmanager.JobManager$$
>> anonfun$main$1$$anon$1.postStop(JobManager.scala:559)
>>     at akka.actor.Actor$class.preRestart(Actor.scala:533)
>>     at org.apache.flink.runtime.jobmanager.JobManager.
>> preRestart(JobManager.scala:80)
>>     at akka.actor.Actor$class.aroundPreRestart(Actor.scala:480)
>>     at org.apache.flink.runtime.jobmanager.JobManager.
>> aroundPreRestart(JobManager.scala:80)
>>     at akka.actor.dungeon.FaultHandling$class.
>> faultRecreate(FaultHandling.scala:67)
>>     at akka.actor.ActorCell.faultRecreate(ActorCell.scala:369)
>>     at akka.actor.ActorCell.invokeAll$1(ActorCell.scala:459)
>>     at akka.actor.ActorCell.systemInvoke(ActorCell.scala:478)
>>     at akka.dispatch.Mailbox.processAllSystemMessages(Mailbox.scala:279)
>>     at akka.dispatch.Mailbox.run(Mailbox.scala:220)
>>     at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
>>     at scala.concurrent.forkjoin.ForkJoinTask.doExec(
>> ForkJoinTask.java:260)
>>     at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.
>> runTask(ForkJoinPool.java:1339)
>>     at scala.concurrent.forkjoin.ForkJoinPool.runWorker(
>> ForkJoinPool.java:1979)
>>     at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(
>> ForkJoinWorkerThread.java:107)
>>
>> the following is an exempt from the jobmanager log:
>>
>> 10:47:13,567 ERROR akka.actor.OneForOneStrategy
>>         - Received unknown message RequestArchivedJobs
>> java.lang.RuntimeException: Received unknown message RequestArchivedJobs
>>     at org.apache.flink.runtime.jobmanager.JobManager.
>> unhandled(JobManager.scala:510)
>>     at akka.actor.Actor$$anonfun$aroundReceive$1.apply(Actor.scala:465)
>>     at akka.actor.Actor$$anonfun$aroundReceive$1.apply(Actor.scala:465)
>>     at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
>>     at org.apache.flink.runtime.ActorLogMessages$$anon$1.
>> applyOrElse(ActorLogMessages.scala:30)
>>     at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
>>     at org.apache.flink.runtime.jobmanager.JobManager.
>> aroundReceive(JobManager.scala:80)
>>     at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
>>     at akka.actor.ActorCell.invoke(ActorCell.scala:487)
>>     at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
>>     at akka.dispatch.Mailbox.run(Mailbox.scala:221)
>>     at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
>>     at scala.concurrent.forkjoin.ForkJoinTask.doExec(
>> ForkJoinTask.java:260)
>>     at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.
>> runTask(ForkJoinPool.java:1339)
>>     at scala.concurrent.forkjoin.ForkJoinPool.runWorker(
>> ForkJoinPool.java:1979)
>>     at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(
>> ForkJoinWorkerThread.java:107)
>> 10:47:13,569 INFO 
>> org.apache.flink.runtime.jobmanager.JobManager$$anonfun$main$1$$anon$1
>> - Stopping webserver.
>> 10:47:13,620 WARN org.eclipse.jetty.util.log
>>       - /jobsInfo
>> org.eclipse.jetty.io.RuntimeIOException: org.eclipse.jetty.io.
>> EofException
>>     at org.eclipse.jetty.io.UncheckedPrintWriter.setError(
>> UncheckedPrintWriter.java:107)
>>     at org.eclipse.jetty.io.UncheckedPrintWriter.write(
>> UncheckedPrintWriter.java:280)
>>     at org.eclipse.jetty.io.UncheckedPrintWriter.write(
>> UncheckedPrintWriter.java:295)
>>     at org.eclipse.jetty.io.UncheckedPrintWriter.print(
>> UncheckedPrintWriter.java:460)
>>     at org.apache.flink.runtime.jobmanager.web.
>> JobManagerInfoServlet.doGet(JobManagerInfoServlet.java:158)
>>     at javax.servlet.http.HttpServlet.service(HttpServlet.java:734)
>>     at javax.servlet.http.HttpServlet.service(HttpServlet.java:847)
>>     at org.eclipse.jetty.servlet.ServletHolder.handle(
>> ServletHolder.java:532)
>>     at org.eclipse.jetty.servlet.ServletHandler.doHandle(
>> ServletHandler.java:453)
>>     at org.eclipse.jetty.server.session.SessionHandler.
>> doHandle(SessionHandler.java:227)
>>     at org.eclipse.jetty.server.handler.ContextHandler.
>> doHandle(ContextHandler.java:965)
>>     at org.eclipse.jetty.servlet.ServletHandler.doScope(
>> ServletHandler.java:388)
>>     at org.eclipse.jetty.server.session.SessionHandler.
>> doScope(SessionHandler.java:187)
>>     at org.eclipse.jetty.server.handler.ContextHandler.
>> doScope(ContextHandler.java:901)
>>     at org.eclipse.jetty.server.handler.ScopedHandler.handle(
>> ScopedHandler.java:117)
>>     at org.eclipse.jetty.server.handler.HandlerList.handle(
>> HandlerList.java:47)
>>     at org.eclipse.jetty.server.handler.HandlerWrapper.handle(
>> HandlerWrapper.java:113)
>>     at org.eclipse.jetty.server.Server.handle(Server.java:352)
>>     at org.eclipse.jetty.server.HttpConnection.handleRequest(
>> HttpConnection.java:596)
>>     at org.eclipse.jetty.server.HttpConnection$RequestHandler.
>> headerComplete(HttpConnection.java:1048)
>>     at org.eclipse.jetty.http.HttpParser.parseNext(HttpParser.java:549)
>>     at org.eclipse.jetty.http.HttpParser.parseAvailable(
>> HttpParser.java:211)
>>     at org.eclipse.jetty.server.HttpConnection.handle(
>> HttpConnection.java:425)
>>     at org.eclipse.jetty.io.nio.SelectChannelEndPoint.run(
>> SelectChannelEndPoint.java:489)
>>     at org.eclipse.jetty.util.thread.QueuedThreadPool$2.run(
>> QueuedThreadPool.java:436)
>>     at java.lang.Thread.run(Thread.java:745)
>> Caused by: org.eclipse.jetty.io.EofException
>>     at org.eclipse.jetty.server.HttpOutput.write(HttpOutput.java:142)
>>     at org.eclipse.jetty.server.HttpOutput.write(HttpOutput.java:86)
>>     at java.io.ByteArrayOutputStream.writeTo(ByteArrayOutputStream.
>> java:167)
>>     at org.eclipse.jetty.server.HttpWriter.write(HttpWriter.java:258)
>>     at org.eclipse.jetty.server.HttpWriter.write(HttpWriter.java:107)
>>     at org.eclipse.jetty.io.UncheckedPrintWriter.write(
>> UncheckedPrintWriter.java:271)
>>     ... 24 more
>> 10:47:13,623 INFO 
>> org.apache.flink.runtime.jobmanager.JobManager$$anonfun$main$1$$anon$1
>> - Stopped webserver.
>> 10:47:13,624 INFO 
>> org.apache.flink.runtime.jobmanager.JobManager$$anonfun$main$1$$anon$1
>> - Stopping job manager akka://flink/user/jobmanager.
>> 10:47:13,624 INFO 
>> org.apache.flink.runtime.jobmanager.JobManager$$anonfun$main$1$$anon$1
>> - Starting job manager at akka://flink/user/jobmanager.
>> 10:47:13,625 INFO org.apache.flink.runtime.blob.BlobServer
>>         - Started BLOB server on port 34038
>> 10:47:13,625 INFO org.apache.flink.runtime.blob.BlobServer
>>         - Created BLOB server storage directory
>> /tmp/blobStore-88f5ebb0-15e2-47a6-ad56-fb2970d83ee2
>> 10:47:13,626 INFO org.apache.flink.runtime.jobmanager.web.WebInfoServer
>>        - Setting up web info server, using web-root directoryjar:file: ...
>> 10:47:13,626 INFO 
>> org.apache.flink.runtime.jobmanager.JobManager$$anonfun$main$1$$anon$1
>> - Started job manager. Waiting for incoming messages.
>> 10:47:13,626 INFO org.apache.flink.runtime.jobmanager.web.WebInfoServer
>>        - Web info server will display information about flink job-manager
>> on ...
>> 10:47:13,627 INFO org.apache.flink.runtime.jobmanager.web.WebInfoServer
>>        - Starting web info server for JobManager on port ...
>> 10:47:13,627 INFO org.eclipse.jetty.util.log
>>       - jetty-0.9-SNAPSHOT
>> 10:47:13,738 INFO org.eclipse.jetty.util.log
>>       - Started SelectChannelConnector@0.0.0.0:8082
>> 10:47:14,032 ERROR 
>> org.apache.flink.runtime.jobmanager.JobManager$$anonfun$main$1$$anon$1
>> - Cannot find execution graph for job ID 1e0c0e5a1a7a741a1182791ea597ba
>> 7e.
>> 10:47:14,068 ERROR 
>> org.apache.flink.runtime.jobmanager.JobManager$$anonfun$main$1$$anon$1
>> - Cannot find execution graph for job ID 1e0c0e5a1a7a741a1182791ea597ba
>> 7e.
>> 10:47:14,069 ERROR 
>> org.apache.flink.runtime.jobmanager.JobManager$$anonfun$main$1$$anon$1
>> - Cannot find execution graph for job ID 1e0c0e5a1a7a741a1182791ea597ba
>> 7e.
>> 10:47:14,107 ERROR 
>> org.apache.flink.runtime.jobmanager.JobManager$$anonfun$main$1$$anon$1
>> - Cannot find execution graph for job ID 1e0c0e5a1a7a741a1182791ea597ba
>> 7e.
>>
>>
>>
>> On 05.02.2015 11:09, Till Rohrmann wrote:
>>
>>> I checked and indeed the scheduleOrUpdateConsumers method can throw an
>>> IllegalStateException without properly handling such an exception on the
>>> JobManager level.
>>>
>>> It is a design decision of Scala not to complain about unhandled
>>> exceptions
>>> which are otherwise properly annotated in Java code. We should definitely
>>> pay attention in Scala to properly handle thrown exceptions of Java code.
>>>
>>> On Thu, Feb 5, 2015 at 10:42 AM, Stephan Ewen <se...@apache.org> wrote:
>>>
>>>  I suspect that this is one of the cases where an exception in an actor
>>>> causes the actor to die (here the job manager)
>>>>
>>>> On Thu, Feb 5, 2015 at 10:40 AM, Till Rohrmann <trohrm...@apache.org>
>>>> wrote:
>>>>
>>>>  It looks to me that the TaskManager does not receive a
>>>>> ConsumerNotificationResult after having send the
>>>>>
>>>> ScheduleOrUpdateConsumers
>>>>
>>>>> message. This can either mean that something went wrong in
>>>>> ExecutionGraph.scheduleOrUpdateConsumers method or the connection was
>>>>> disassociated for some reasons. The logs would indeed be very helpful
>>>>> to
>>>>> understand what happened.
>>>>>
>>>>> On Thu, Feb 5, 2015 at 10:36 AM, Ufuk Celebi <u...@apache.org> wrote:
>>>>>
>>>>>  Hey Chesnay,
>>>>>>
>>>>>> I will look into it. Can you share the complete LOGs?
>>>>>>
>>>>>> – Ufuk
>>>>>>
>>>>>> On 04 Feb 2015, at 14:49, Chesnay Schepler <
>>>>>>
>>>>> chesnay.schep...@fu-berlin.de>
>>>>>
>>>>>> wrote:
>>>>>>
>>>>>>  Hello,
>>>>>>>
>>>>>>> I'm trying to run python jobs with the latest master on a cluster and
>>>>>>>
>>>>>> get the following exception:
>>>>>>
>>>>>>> Error: The program execution failed: JobManager not reachable
>>>>>>>
>>>>>> anymore.
>>>>
>>>>> Terminate waiting for job answer.
>>>>>>
>>>>>>> org.apache.flink.client.program.ProgramInvocationException: The
>>>>>>>
>>>>>> program
>>>>
>>>>> execution failed: JobManager not reachable anymore. Terminate waiting
>>>>>>
>>>>> for
>>>>
>>>>> job answer.
>>>>>>
>>>>>>>     at org.apache.flink.client.program.Client.run(Client.java:345)
>>>>>>>     at org.apache.flink.client.program.Client.run(Client.java:304)
>>>>>>>     at org.apache.flink.client.program.Client.run(Client.java:298)
>>>>>>>     at
>>>>>>>
>>>>>> org.apache.flink.client.program.ContextEnvironment.
>>>> execute(ContextEnvironment.java:55)
>>>>
>>>>>     at
>>>>>>>
>>>>>> org.apache.flink.api.java.ExecutionEnvironment.execute(
>>>> ExecutionEnvironment.java:677)
>>>>
>>>>>     at
>>>>>>>
>>>>>> org.apache.flink.languagebinding.api.java.python.PythonPlanBinder.
>>>> runPlan(PythonPlanBinder.java:106)
>>>>
>>>>>     at
>>>>>>>
>>>>>> org.apache.flink.languagebinding.api.java.
>>>> python.PythonPlanBinder.main(PythonPlanBinder.java:79)
>>>>
>>>>>     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>>>>>     at
>>>>>>>
>>>>>> sun.reflect.NativeMethodAccessorImpl.invoke(
>>>> NativeMethodAccessorImpl.java:57)
>>>>
>>>>>     at
>>>>>>>
>>>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(
>>>> DelegatingMethodAccessorImpl.java:43)
>>>>
>>>>>     at java.lang.reflect.Method.invoke(Method.java:606)
>>>>>>>     at
>>>>>>>
>>>>>> org.apache.flink.client.program.PackagedProgram.callMainMethod(
>>>> PackagedProgram.java:437)
>>>>
>>>>>     at
>>>>>>>
>>>>>> org.apache.flink.client.program.PackagedProgram.
>>>> invokeInteractiveModeForExecution(PackagedProgram.java:353)
>>>>
>>>>>     at org.apache.flink.client.program.Client.run(Client.java:250)
>>>>>>>     at
>>>>>>>
>>>>>> org.apache.flink.client.CliFrontend.executeProgram(
>>>> CliFrontend.java:387)
>>>>
>>>>>     at org.apache.flink.client.CliFrontend.run(CliFrontend.java:356)
>>>>>>>     at
>>>>>>>
>>>>>> org.apache.flink.client.CliFrontend.parseParameters(
>>>> CliFrontend.java:1066)
>>>>
>>>>>     at org.apache.flink.client.CliFrontend.main(CliFrontend.java:1090)
>>>>>>>
>>>>>>> In the jobmanager log file i find this exception:
>>>>>>>
>>>>>>> java.lang.IllegalStateException: Buffer has already been recycled.
>>>>>>>         at
>>>>>>>
>>>>>> org.apache.flink.shaded.com.google.common.base.
>>>> Preconditions.checkState(Preconditions.java:176)
>>>>
>>>>>         at
>>>>>>>
>>>>>> org.apache.flink.runtime.io.network.buffer.Buffer.
>>>> ensureNotRecycled(Buffer.java:131)
>>>>
>>>>>         at
>>>>>>>
>>>>>> org.apache.flink.runtime.io.network.buffer.Buffer.setSize(
>>>> Buffer.java:95)
>>>>
>>>>>         at
>>>>>>>
>>>>>> org.apache.flink.runtime.io.network.api.serialization.
>>>> SpanningRecordSerializer.getCurrentBuffer(
>>>> SpanningRecordSerializer.java:151)
>>>>
>>>>>         at
>>>>>>>
>>>>>> org.apache.flink.runtime.io.network.api.writer.
>>>> RecordWriter.clearBuffers(RecordWriter.java:158)
>>>>
>>>>>         at
>>>>>>>
>>>>>> org.apache.flink.runtime.operators.RegularPactTask.
>>>> clearWriters(RegularPactTask.java:1533)
>>>>
>>>>>         at
>>>>>>>
>>>>>> org.apache.flink.runtime.operators.RegularPactTask.
>>>> invoke(RegularPactTask.java:367)
>>>>
>>>>>         at
>>>>>>>
>>>>>> org.apache.flink.runtime.execution.RuntimeEnvironment.
>>>> run(RuntimeEnvironment.java:204)
>>>>
>>>>>         at java.lang.Thread.run(Thread.java:745)
>>>>>>>
>>>>>>> the same exception is in the task manager logs, along with the
>>>>>>>
>>>>>> following
>>>>>
>>>>>> one:
>>>>>>
>>>>>>> java.util.concurrent.TimeoutException: Futures timed out after [100
>>>>>>>
>>>>>> seconds]
>>>>>>
>>>>>>>         at
>>>>>>>
>>>>>> scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
>>>>>>
>>>>>>>         at
>>>>>>>
>>>>>> scala.concurrent.impl.Promise$DefaultPromise.result(Promise.
>>>>>> scala:223)
>>>>>>
>>>>>>>         at
>>>>>>>
>>>>>> scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
>>>>>>
>>>>>>>         at
>>>>>>>
>>>>>> scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(
>>>> BlockContext.scala:53)
>>>>
>>>>>         at scala.concurrent.Await$.result(package.scala:107)
>>>>>>>         at
>>>>>>>
>>>>>> org.apache.flink.runtime.akka.AkkaUtils$.ask(AkkaUtils.scala:265)
>>>>>>
>>>>>>>         at
>>>>>>>
>>>>>> org.apache.flink.runtime.akka.AkkaUtils.ask(AkkaUtils.scala)
>>>>
>>>>>         at
>>>>>>>
>>>>>> org.apache.flink.runtime.io.network.partition.
>>>> IntermediateResultPartition.scheduleOrUpdateConsumers(
>>>> IntermediateResultPartition.java:247)
>>>>
>>>>>         at
>>>>>>>
>>>>>> org.apache.flink.runtime.io.network.partition.
>>>> IntermediateResultPartition.maybeNotifyConsumers(
>>>> IntermediateResultPartition.java:240)
>>>>
>>>>>         at
>>>>>>>
>>>>>> org.apache.flink.runtime.io.network.partition.
>>>> IntermediateResultPartition.add(IntermediateResultPartition.java:144)
>>>>
>>>>>         at
>>>>>>>
>>>>>> org.apache.flink.runtime.io.network.api.writer.
>>>> BufferWriter.writeBuffer(BufferWriter.java:74)
>>>>
>>>>>         at
>>>>>>>
>>>>>> org.apache.flink.runtime.io.network.api.writer.RecordWriter.emit(
>>>> RecordWriter.java:91)
>>>>
>>>>>         at
>>>>>>>
>>>>>> org.apache.flink.runtime.operators.shipping.OutputCollector.collect(
>>>> OutputCollector.java:88)
>>>>
>>>>>         at
>>>>>>>
>>>>>> org.apache.flink.languagebinding.api.java.common.streaming.Receiver.
>>>> collectBuffer(Receiver.java:253)
>>>>
>>>>>         at
>>>>>>>
>>>>>> org.apache.flink.languagebinding.api.java.common.streaming.Streamer.
>>>> streamBufferWithoutGroups(Streamer.java:193)
>>>>
>>>>>         at
>>>>>>>
>>>>>> org.apache.flink.languagebinding.api.java.python.functions.
>>>> PythonMapPartition.mapPartition(PythonMapPartition.java:54)
>>>>
>>>>>         at
>>>>>>>
>>>>>> org.apache.flink.runtime.operators.MapPartitionDriver.
>>>> run(MapPartitionDriver.java:98)
>>>>
>>>>>         at
>>>>>>>
>>>>>> org.apache.flink.runtime.operators.RegularPactTask.run(
>>>> RegularPactTask.java:496)
>>>>
>>>>>         at
>>>>>>>
>>>>>> org.apache.flink.runtime.operators.RegularPactTask.
>>>> invoke(RegularPactTask.java:360)
>>>>
>>>>>         at
>>>>>>>
>>>>>> org.apache.flink.runtime.execution.RuntimeEnvironment.
>>>> run(RuntimeEnvironment.java:204)
>>>>
>>>>>         at java.lang.Thread.run(Thread.java:745)
>>>>>>>
>>>>>>>
>>>>>>
>>
>

Reply via email to