Hi Philipp, it seems that Stephan was right and that your JobGraph is somehow corrupted. You can see it in the exception JobSubmissionException that the JobGraph contains a vertex whose InvokableClassName is null. Furthermore, even the ID and the vertex name are null. This is a strong indicator, that the JobGraph is not correct.
Can you also post the log of the JobManager? Do you have the code of your job online? Cheers, Till On Thu, Jul 16, 2015 at 1:20 PM, Philipp Goetze < philipp.goe...@tu-ilmenau.de> wrote: > Hey Tim, > > here the console output now with log4j: > > 0 [pool-7-thread-1-ScalaTest-running-FlinkCompileIt] INFO > org.apache.flink.client.program.Client - Starting program in interactive > mode > 121 [pool-7-thread-1-ScalaTest-running-FlinkCompileIt] DEBUG > org.apache.flink.api.scala.ClosureCleaner$ - accessedFields: Map() > 137 [pool-7-thread-1-ScalaTest-running-FlinkCompileIt] DEBUG > org.apache.flink.api.scala.ClosureCleaner$ - accessedFields: Map() > 183 [pool-7-thread-1-ScalaTest-running-FlinkCompileIt] INFO > org.apache.flink.api.java.ExecutionEnvironment - The job has 0 registered > types and 0 default Kryo serializers > 188 [pool-7-thread-1-ScalaTest-running-FlinkCompileIt] DEBUG > org.apache.flink.api.java.ExecutionEnvironment - Registered Kryo types: > 188 [pool-7-thread-1-ScalaTest-running-FlinkCompileIt] DEBUG > org.apache.flink.api.java.ExecutionEnvironment - Registered Kryo with > Serializers types: > 188 [pool-7-thread-1-ScalaTest-running-FlinkCompileIt] DEBUG > org.apache.flink.api.java.ExecutionEnvironment - Registered Kryo with > Serializer Classes types: > 188 [pool-7-thread-1-ScalaTest-running-FlinkCompileIt] DEBUG > org.apache.flink.api.java.ExecutionEnvironment - Registered Kryo default > Serializers: > 188 [pool-7-thread-1-ScalaTest-running-FlinkCompileIt] DEBUG > org.apache.flink.api.java.ExecutionEnvironment - Registered Kryo default > Serializers Classes > 188 [pool-7-thread-1-ScalaTest-running-FlinkCompileIt] DEBUG > org.apache.flink.api.java.ExecutionEnvironment - Registered POJO types: > 188 [pool-7-thread-1-ScalaTest-running-FlinkCompileIt] DEBUG > org.apache.flink.api.java.ExecutionEnvironment - Static code analysis > mode: DISABLE > 188 [pool-7-thread-1-ScalaTest-running-FlinkCompileIt] DEBUG > org.apache.flink.client.program.Client - Set parallelism 1, plan default > parallelism 1 > 198 [pool-7-thread-1-ScalaTest-running-FlinkCompileIt] DEBUG > org.apache.flink.optimizer.Optimizer - Beginning compilation of program > 'Starting Query' > 198 [pool-7-thread-1-ScalaTest-running-FlinkCompileIt] DEBUG > org.apache.flink.optimizer.Optimizer - Using a default parallelism of 1 > 198 [pool-7-thread-1-ScalaTest-running-FlinkCompileIt] DEBUG > org.apache.flink.optimizer.Optimizer - Using default data exchange mode > PIPELINED > 266 [pool-7-thread-1-ScalaTest-running-FlinkCompileIt] DEBUG > org.apache.flink.api.common.io.FileInputFormat - Opening input split > file:/home/blaze/Documents/TU_Ilmenau/Masterthesis/projects/pigspark/src/it/resources/file.txt > [0,32] > 269 [pool-7-thread-1-ScalaTest-running-FlinkCompileIt] DEBUG > org.apache.flink.api.common.io.FileInputFormat - Opening input split > file:/home/blaze/Documents/TU_Ilmenau/Masterthesis/projects/pigspark/src/it/resources/file.txt > [16,16] > 412 [pool-7-thread-1-ScalaTest-running-FlinkCompileIt] INFO > org.apache.flink.client.program.Client - JobManager actor system address > is localhost/127.0.0.1:6123 > 412 [pool-7-thread-1-ScalaTest-running-FlinkCompileIt] INFO > org.apache.flink.client.program.Client - Starting client actor system > 415 [pool-7-thread-1-ScalaTest-running-FlinkCompileIt] INFO > org.apache.flink.runtime.client.JobClient - Starting JobClient actor system > 922 [flink-akka.actor.default-dispatcher-2] INFO > akka.event.slf4j.Slf4jLogger - Slf4jLogger started > 953 [flink-akka.actor.default-dispatcher-2] DEBUG akka.event.EventStream > - logger log1-Slf4jLogger started > 954 [flink-akka.actor.default-dispatcher-2] DEBUG akka.event.EventStream > - Default Loggers started > 1044 [flink-akka.actor.default-dispatcher-4] INFO Remoting - Starting > remoting > 1117 [flink-akka.remote.default-remote-dispatcher-6] DEBUG > org.jboss.netty.channel.socket.nio.SelectorUtil - Using select timeout of > 500 > 1118 [flink-akka.remote.default-remote-dispatcher-6] DEBUG > org.jboss.netty.channel.socket.nio.SelectorUtil - Epoll-bug workaround > enabled = false > 1325 [flink-akka.actor.default-dispatcher-2] INFO Remoting - Remoting > started; listening on addresses :[akka.tcp://flink@127.0.0.1:58455] > 1343 [pool-7-thread-1-ScalaTest-running-FlinkCompileIt] INFO > org.apache.flink.runtime.client.JobClient - Started JobClient actor system > at 127.0.0.1:58455 > 1343 [pool-7-thread-1-ScalaTest-running-FlinkCompileIt] INFO > org.apache.flink.client.program.Client - Looking up JobManager > 1542 [flink-akka.actor.default-dispatcher-2] DEBUG > akka.serialization.Serialization(akka://flink) - Using > serializer[akka.serialization.JavaSerializer] for message > [akka.actor.Identify] > 1567 [flink-akka.actor.default-dispatcher-2] DEBUG > akka.remote.EndpointWriter - Drained buffer with maxWriteCount: 50, > fullBackoffCount: 1, smallBackoffCount: 0, noBackoffCount: 0 , > adaptiveBackoff: 1000 > 1599 [pool-7-thread-1-ScalaTest-running-FlinkCompileIt] INFO > org.apache.flink.client.program.Client - JobManager runs at > akka.tcp://flink@127.0.0.1:6123/user/jobmanager > 1600 [pool-7-thread-1-ScalaTest-running-FlinkCompileIt] INFO > org.apache.flink.client.program.Client - Communication between client and > JobManager will have a timeout of 100000 milliseconds > 1600 [pool-7-thread-1-ScalaTest-running-FlinkCompileIt] INFO > org.apache.flink.client.program.Client - Checking and uploading JAR files > 1606 [flink-akka.actor.default-dispatcher-2] DEBUG > akka.serialization.Serialization(akka://flink) - Using > serializer[akka.serialization.JavaSerializer] for message > [org.apache.flink.runtime.messages.JobManagerMessages$RequestBlobManagerPort$] > 1638 [pool-7-thread-1-ScalaTest-running-FlinkCompileIt] DEBUG > org.apache.flink.runtime.blob.BlobClient - PUT content addressable BLOB > stream to /127.0.0.1:42947 > 1660 [flink-akka.actor.default-dispatcher-4] INFO > org.apache.flink.runtime.client.JobClient - Sending message to JobManager > akka.tcp://flink@127.0.0.1:6123/user/jobmanager to submit job Starting > Query (227a3733c283899991ba8a5237a0f2a8) and wait for progress > 1667 [flink-akka.actor.default-dispatcher-2] DEBUG > akka.serialization.Serialization(akka://flink) - Using > serializer[akka.serialization.JavaSerializer] for message > [org.apache.flink.runtime.messages.JobManagerMessages$SubmitJob] > 1712 [flink-akka.actor.default-dispatcher-4] DEBUG > akka.remote.RemoteWatcher - Watching: [akka://flink/user/$a -> > akka.tcp://flink@127.0.0.1:6123/user/jobmanager] > 1781 [flink-akka.actor.default-dispatcher-4] DEBUG > akka.serialization.Serialization(akka://flink) - Using > serializer[akka.serialization.JavaSerializer] for message > [akka.dispatch.sysmsg.Watch] > 1819 [flink-akka.actor.default-dispatcher-4] DEBUG > org.apache.flink.runtime.client.JobClient - Received failure from > JobManager > org.apache.flink.runtime.client.JobSubmissionException: The vertex null > (null) has no invokable class. > at > org.apache.flink.runtime.jobmanager.JobManager$$anonfun$org$apache$flink$runtime$jobmanager$JobManager$$submitJob$4.apply(JobManager.scala:511) > at > org.apache.flink.runtime.jobmanager.JobManager$$anonfun$org$apache$flink$runtime$jobmanager$JobManager$$submitJob$4.apply(JobManager.scala:507) > at scala.collection.Iterator$class.foreach(Iterator.scala:743) > at scala.collection.AbstractIterator.foreach(Iterator.scala:1195) > at scala.collection.IterableLike$class.foreach(IterableLike.scala:72) > at scala.collection.AbstractIterable.foreach(Iterable.scala:54) > at org.apache.flink.runtime.jobmanager.JobManager.org > $apache$flink$runtime$jobmanager$JobManager$$submitJob(JobManager.scala:507) > at > org.apache.flink.runtime.jobmanager.JobManager$$anonfun$receiveWithLogMessages$1.applyOrElse(JobManager.scala:190) > at > scala.runtime.AbstractPartialFunction.apply(AbstractPartialFunction.scala:36) > at > org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:43) > at > org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:29) > at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:123) > at > org.apache.flink.runtime.ActorLogMessages$$anon$1.applyOrElse(ActorLogMessages.scala:29) > at akka.actor.Actor$class.aroundReceive(Actor.scala:465) > at > org.apache.flink.runtime.jobmanager.JobManager.aroundReceive(JobManager.scala:92) > at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516) > at akka.actor.ActorCell.invoke(ActorCell.scala:487) > at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254) > at akka.dispatch.Mailbox.run(Mailbox.scala:221) > at akka.dispatch.Mailbox.exec(Mailbox.scala:231) > at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) > at > scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) > at > scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) > at > scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) > 1858 [flink-akka.actor.default-dispatcher-3] DEBUG > akka.serialization.Serialization(akka://flink) - Using > serializer[akka.serialization.JavaSerializer] for message > [akka.dispatch.sysmsg.Unwatch] > 1862 [flink-akka.actor.default-dispatcher-3] DEBUG > akka.remote.RemoteWatcher - Unwatching: [akka://flink/user/$a -> > akka.tcp://flink@127.0.0.1:6123/user/jobmanager] > 1863 [flink-akka.actor.default-dispatcher-3] DEBUG > akka.remote.RemoteWatcher - Cleanup self watch of [ > akka.tcp://flink@127.0.0.1:6123/user/jobmanager] > 1879 [flink-akka.actor.default-dispatcher-3] DEBUG > akka.remote.RemoteWatcher - Unwatched last watchee of node: [ > akka.tcp://flink@127.0.0.1:6123] > 1932 [flink-akka.actor.default-dispatcher-2] INFO > akka.remote.RemoteActorRefProvider$RemotingTerminator - Shutting down > remote daemon. > 1935 [flink-akka.actor.default-dispatcher-2] INFO > akka.remote.RemoteActorRefProvider$RemotingTerminator - Remote daemon shut > down; proceeding with flushing remote transports. > 2037 [flink-akka.actor.default-dispatcher-4] INFO > akka.remote.RemoteActorRefProvider$RemotingTerminator - Remoting shut down. > org.apache.flink.client.program.ProgramInvocationException: The program > execution failed: The vertex null (null) has no invokable class. > at org.apache.flink.client.program.Client.run(Client.java:412) > at org.apache.flink.client.program.Client.run(Client.java:355) > at org.apache.flink.client.program.Client.run(Client.java:348) > at > org.apache.flink.client.program.ContextEnvironment.execute(ContextEnvironment.java:63) > at > org.apache.flink.api.scala.ExecutionEnvironment.execute(ExecutionEnvironment.scala:590) > at > load$.main(/home/blaze/Documents/TU_Ilmenau/Masterthesis/projects/pigspark/load/load.scala:20) > at > load.main(/home/blaze/Documents/TU_Ilmenau/Masterthesis/projects/pigspark/load/load.scala) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:497) > at > org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:437) > at > org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:353) > at org.apache.flink.client.program.Client.run(Client.java:315) > at dbis.pig.tools.FlinkRun.submitJar(FlinkRun.scala:62) > at dbis.pig.tools.FlinkRun.execute(FlinkRun.scala:37) > at dbis.pig.PigCompiler$.run(PigCompiler.scala:106) > at dbis.pig.PigCompiler$.main(PigCompiler.scala:69) > at > dbis.test.flink.FlinkCompileIt$$anonfun$1$$anonfun$apply$mcV$sp$1.apply(FlinkCompileIt.scala:62) > at > dbis.test.flink.FlinkCompileIt$$anonfun$1$$anonfun$apply$mcV$sp$1.apply(FlinkCompileIt.scala:53) > at > org.scalatest.prop.TableFor4$$anonfun$apply$17.apply(TableFor1.scala:797) > at > org.scalatest.prop.TableFor4$$anonfun$apply$17.apply(TableFor1.scala:795) > at > scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:778) > at > scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) > at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48) > at > scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:777) > at org.scalatest.prop.TableFor4.apply(TableFor1.scala:795) > at > org.scalatest.prop.TableDrivenPropertyChecks$class.forAll(TableDrivenPropertyChecks.scala:418) > at > org.scalatest.prop.TableDrivenPropertyChecks$.forAll(TableDrivenPropertyChecks.scala:665) > at > dbis.test.flink.FlinkCompileIt$$anonfun$1.apply$mcV$sp(FlinkCompileIt.scala:53) > at > dbis.test.flink.FlinkCompileIt$$anonfun$1.apply(FlinkCompileIt.scala:53) > at > dbis.test.flink.FlinkCompileIt$$anonfun$1.apply(FlinkCompileIt.scala:53) > at > org.scalatest.Transformer$$anonfun$apply$1.apply$mcV$sp(Transformer.scala:22) > at org.scalatest.OutcomeOf$class.outcomeOf(OutcomeOf.scala:85) > at org.scalatest.OutcomeOf$.outcomeOf(OutcomeOf.scala:104) > at org.scalatest.Transformer.apply(Transformer.scala:22) > at org.scalatest.Transformer.apply(Transformer.scala:20) > at org.scalatest.FlatSpecLike$$anon$1.apply(FlatSpecLike.scala:1647) > at org.scalatest.Suite$class.withFixture(Suite.scala:1122) > at org.scalatest.FlatSpec.withFixture(FlatSpec.scala:1683) > at > org.scalatest.FlatSpecLike$class.invokeWithFixture$1(FlatSpecLike.scala:1644) > at > org.scalatest.FlatSpecLike$$anonfun$runTest$1.apply(FlatSpecLike.scala:1656) > at > org.scalatest.FlatSpecLike$$anonfun$runTest$1.apply(FlatSpecLike.scala:1656) > at org.scalatest.SuperEngine.runTestImpl(Engine.scala:306) > at org.scalatest.FlatSpecLike$class.runTest(FlatSpecLike.scala:1656) > at org.scalatest.FlatSpec.runTest(FlatSpec.scala:1683) > at > org.scalatest.FlatSpecLike$$anonfun$runTests$1.apply(FlatSpecLike.scala:1714) > at > org.scalatest.FlatSpecLike$$anonfun$runTests$1.apply(FlatSpecLike.scala:1714) > at > org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:413) > at > org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:401) > at scala.collection.immutable.List.foreach(List.scala:381) > at org.scalatest.SuperEngine.traverseSubNodes$1(Engine.scala:401) > at org.scalatest.SuperEngine.org > $scalatest$SuperEngine$$runTestsInBranch(Engine.scala:390) > at > org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:427) > at > org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:401) > at scala.collection.immutable.List.foreach(List.scala:381) > at org.scalatest.SuperEngine.traverseSubNodes$1(Engine.scala:401) > at org.scalatest.SuperEngine.org > $scalatest$SuperEngine$$runTestsInBranch(Engine.scala:396) > at org.scalatest.SuperEngine.runTestsImpl(Engine.scala:483) > at org.scalatest.FlatSpecLike$class.runTests(FlatSpecLike.scala:1714) > at org.scalatest.FlatSpec.runTests(FlatSpec.scala:1683) > at org.scalatest.Suite$class.run(Suite.scala:1424) > at org.scalatest.FlatSpec.org > $scalatest$FlatSpecLike$$super$run(FlatSpec.scala:1683) > at > org.scalatest.FlatSpecLike$$anonfun$run$1.apply(FlatSpecLike.scala:1760) > at > org.scalatest.FlatSpecLike$$anonfun$run$1.apply(FlatSpecLike.scala:1760) > at org.scalatest.SuperEngine.runImpl(Engine.scala:545) > at org.scalatest.FlatSpecLike$class.run(FlatSpecLike.scala:1760) > at org.scalatest.FlatSpec.run(FlatSpec.scala:1683) > at org.scalatest.tools.Framework.org > $scalatest$tools$Framework$$runSuite(Framework.scala:462) > at > org.scalatest.tools.Framework$ScalaTestTask.execute(Framework.scala:671) > at sbt.TestRunner.runTest$1(TestFramework.scala:76) > at sbt.TestRunner.run(TestFramework.scala:85) > at > sbt.TestFramework$$anon$2$$anonfun$$init$$1$$anonfun$apply$8.apply(TestFramework.scala:202) > at > sbt.TestFramework$$anon$2$$anonfun$$init$$1$$anonfun$apply$8.apply(TestFramework.scala:202) > at > sbt.TestFramework$.sbt$TestFramework$$withContextLoader(TestFramework.scala:185) > at > sbt.TestFramework$$anon$2$$anonfun$$init$$1.apply(TestFramework.scala:202) > at > sbt.TestFramework$$anon$2$$anonfun$$init$$1.apply(TestFramework.scala:202) > at sbt.TestFunction.apply(TestFramework.scala:207) > at sbt.Tests$.sbt$Tests$$processRunnable$1(Tests.scala:239) > at sbt.Tests$$anonfun$makeSerial$1.apply(Tests.scala:245) > at sbt.Tests$$anonfun$makeSerial$1.apply(Tests.scala:245) > at sbt.std.Transform$$anon$3$$anonfun$apply$2.apply(System.scala:44) > at sbt.std.Transform$$anon$3$$anonfun$apply$2.apply(System.scala:44) > at sbt.std.Transform$$anon$4.work(System.scala:63) > at > sbt.Execute$$anonfun$submit$1$$anonfun$apply$1.apply(Execute.scala:226) > at > sbt.Execute$$anonfun$submit$1$$anonfun$apply$1.apply(Execute.scala:226) > at sbt.ErrorHandling$.wideConvert(ErrorHandling.scala:17) > at sbt.Execute.work(Execute.scala:235) > at sbt.Execute$$anonfun$submit$1.apply(Execute.scala:226) > at sbt.Execute$$anonfun$submit$1.apply(Execute.scala:226) > at > sbt.ConcurrentRestrictions$$anon$4$$anonfun$1.apply(ConcurrentRestrictions.scala:159) > at sbt.CompletionService$$anon$2.call(CompletionService.scala:28) > at java.util.concurrent.FutureTask.run(FutureTask.java:266) > at > java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) > at java.util.concurrent.FutureTask.run(FutureTask.java:266) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) > at java.lang.Thread.run(Thread.java:745) > Caused by: org.apache.flink.runtime.client.JobSubmissionException: The > vertex null (null) has no invokable class. > at > org.apache.flink.runtime.jobmanager.JobManager$$anonfun$org$apache$flink$runtime$jobmanager$JobManager$$submitJob$4.apply(JobManager.scala:511) > at > org.apache.flink.runtime.jobmanager.JobManager$$anonfun$org$apache$flink$runtime$jobmanager$JobManager$$submitJob$4.apply(JobManager.scala:507) > at scala.collection.Iterator$class.foreach(Iterator.scala:743) > at scala.collection.AbstractIterator.foreach(Iterator.scala:1195) > at scala.collection.IterableLike$class.foreach(IterableLike.scala:72) > at scala.collection.AbstractIterable.foreach(Iterable.scala:54) > at org.apache.flink.runtime.jobmanager.JobManager.org > $apache$flink$runtime$jobmanager$JobManager$$submitJob(JobManager.scala:507) > at > org.apache.flink.runtime.jobmanager.JobManager$$anonfun$receiveWithLogMessages$1.applyOrElse(JobManager.scala:190) > at > scala.runtime.AbstractPartialFunction.apply(AbstractPartialFunction.scala:36) > at > org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:43) > at > org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:29) > at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:123) > at > org.apache.flink.runtime.ActorLogMessages$$anon$1.applyOrElse(ActorLogMessages.scala:29) > at akka.actor.Actor$class.aroundReceive(Actor.scala:465) > at > org.apache.flink.runtime.jobmanager.JobManager.aroundReceive(JobManager.scala:92) > at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516) > at akka.actor.ActorCell.invoke(ActorCell.scala:487) > at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254) > at akka.dispatch.Mailbox.run(Mailbox.scala:221) > at akka.dispatch.Mailbox.exec(Mailbox.scala:231) > at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) > at > scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) > at > scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) > at > scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) > > Best Regards, > Philipp > > > > On 16.07.2015 11:45, Till Rohrmann wrote: > > When you run your program from the IDE, then you can specify a > log4j.properties file. There you can configure where and what to log. It > should be enough to place the log4j.properties file in the resource > folder of your project. An example properties file could look like: > > log4j.rootLogger=INFO, testlogger > > log4j.appender.testlogger=org.apache.log4j.ConsoleAppender > log4j.appender.testlogger.target = System.err > log4j.appender.testlogger.layout=org.apache.log4j.PatternLayout > log4j.appender.testlogger.layout.ConversionPattern=%-4r [%t] %-5p %c %x - %m%n > > Alternatively, you can specify it via a JVM option: > -Dlog4j.configuration=<path > to properties file> > > Cheers, > Till > > > On Thu, Jul 16, 2015 at 11:23 AM, Philipp Goetze < > philipp.goe...@tu-ilmenau.de> wrote: > >> Hi Till, >> >> the problem is that this is the only output :( Or is it possible to get a >> more verbose log output? >> >> Maybe it is important to note, that both Flink and our project is built >> with Scala 2.11. >> >> Best Regards, >> Philipp >> >> >> On 16.07.2015 11:12, Till Rohrmann wrote: >> >> Hi Philipp, >> >> could you post the complete log output. This might help to get to the >> bottom of the problem. >> >> Cheers, >> Till >> >> On Thu, Jul 16, 2015 at 11:01 AM, Philipp Goetze < >> philipp.goe...@tu-ilmenau.de> wrote: >> >>> Hi community, >>> >>> in our project we try to submit built Flink programs to the jobmanager >>> from within Scala code. The test program is executed correctly when >>> submitted via the wrapper script "bin/flink run ..." and also with the >>> webclient. But when executed from within the Scala code nothing seems to >>> happen, but the following warning is found in the log: >>> >>> 10:47:18,153 WARN akka.remote.ReliableDeliverySupervisor >>> - Association with remote system [akka.tcp://flink@127.0.0.1:34074] >>> has failed, address is now gated for [5000] ms. Reason is: >>> [org.apache.flink.runtime.jobgraph.AbstractJobVertex] >>> >>> >>> Our submit method looks like that: >>> >>> def submitJar(master: String, path: String, className: String, args: >>> String*) = { >>> >>> val file = new File(path) >>> >>> val parallelism = 1 >>> >>> val wait = true >>> >>> try { >>> >>> val program = new PackagedProgram(file, className, args:_*) >>> >>> val jobManagerAddress = getInetFromHostport(master) >>> >>> val client = new Client(jobManagerAddress, new Configuration(), >>> program.getUserCodeClassLoader(), 1) >>> >>> println("Executing " + path); >>> >>> client.run(program, parallelism, wait); >>> >>> } catch { >>> >>> case e: ProgramInvocationException => e.printStackTrace() >>> >>> } >>> >>> } >>> >>> >>> I took this as a reference: >>> http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/How-to-submit-flink-jars-from-plain-Java-programs-td656.html >>> >>> I hope you can help. >>> >>> Best Regards, >>> Philipp Götze >>> >> >> >> > >