I figured out that for my stream job the best was just to use the default MemoryStateBackend. I load a table from a file of 725MB in a UDF. I am also not using Flink ListState since I don't have to change the values of this table. i only do a lookup.
The only thing that I need was more memory for the TM and a bit larger timeout. Currently, my configurations are these. I am not sure if there are a better configuration heartbeat.timeout: 100000 taskmanager.memory.flink.size: 12g taskmanager.memory.jvm-overhead.max: 4g taskmanager.memory.jvm-metaspace.size: 2048m # default: 1024m Another thing that is not working is this parameter that when I set it I get an JVM argument error and the TM does not start. taskmanager.memory.task.heap.size: 2048m # default: 1024m # Flink error Best, Felipe -- -- Felipe Gutierrez -- skype: felipe.o.gutierrez -- https://felipeogutierrez.blogspot.com On Tue, Jul 7, 2020 at 2:17 PM Yun Tang <myas...@live.com> wrote: > > Hi Felipe > > flink_taskmanager_Status_JVM_Memory_Direct_MemoryUsed cannot tell you how > much memory is used by RocksDB as it mallocate memory from os directly > instead from JVM. > > Moreover, I cannot totally understand why you ask how to increase the memory > of the JM and TM when using the PredefinedOptions.SPINNING_DISK_OPTIMIZED for > RocksDB. > Did you mean how to increase the total process memory? If so, as Flink uses > managed memory to control RocksDB [1] by default, you could increase total > memory by increasing managed memory [2][3] > > [1] > https://ci.apache.org/projects/flink/flink-docs-stable/ops/config.html#state-backend-rocksdb-memory-managed > [2] > https://ci.apache.org/projects/flink/flink-docs-stable/ops/config.html#taskmanager-memory-managed-fraction > [3] > https://ci.apache.org/projects/flink/flink-docs-stable/ops/config.html#taskmanager-memory-managed-size > > Best, > Yun Tang > > > > ________________________________ > From: Felipe Gutierrez <felipe.o.gutier...@gmail.com> > Sent: Monday, July 6, 2020 19:17 > To: Yun Tang <myas...@live.com> > Cc: Ori Popowski <ori....@gmail.com>; user <user@flink.apache.org> > Subject: Re: Timeout when using RockDB to handle large state in a stream app > > Hi all, > > I tested the two TPC-H query 03 [1] and 10 [2] using Datastream API on > the cluster with RocksDB state backend. One thing that I did that > improved a lot was to replace the List<LineItem> POJO to a > List<Tuple2<>>. Then I could load a table of 200MB in memory as my > state. However, the original table is 725MB, and turned out that I > need another configuration. I am not sure what I can do more to reduce > the size of my state. If one of you have an idea I am thankful to > hear. > > Now, speaking about the flink-conf.yaml file and the RocksDB > configuration. When I use these configurations on the flink-conf.yaml > the stream job still runs out of memory. > jobmanager.heap.size: 4g # default: 2048m > heartbeat.timeout: 100000 > taskmanager.memory.process.size: 2g # default: 1728m > > Then I changed for this configuration which I can set > programmatically. The stream job seems to behave better. It starts to > process something, then the metrics disappear for some time and appear > again. The available and used memory on the TM > (flink_taskmanager_Status_JVM_Memory_Direct_MemoryUsed) is 167MB. And > the available and used memory on the JM > (flink_jobmanager_Status_JVM_Memory_Direct_MemoryUsed) is 610KB. I > guess the PredefinedOptions.SPINNING_DISK_OPTIMIZED configuration is > overwriting the configuration on the flink-conf.yaml file. > > RocksDBStateBackend stateBackend = new RocksDBStateBackend(stateDir, true); > stateBackend.setPredefinedOptions(PredefinedOptions.SPINNING_DISK_OPTIMIZED); > env.setStateBackend(stateBackend); > > How can I increase the memory of the JM and TM when I am still using > the PredefinedOptions.SPINNING_DISK_OPTIMIZED for RocksDB? > > [1] > https://github.com/felipegutierrez/explore-flink/blob/master/src/main/java/org/sense/flink/examples/stream/tpch/TPCHQuery03.java > [2] > https://github.com/felipegutierrez/explore-flink/blob/master/src/main/java/org/sense/flink/examples/stream/tpch/TPCHQuery10.java > > -- > -- Felipe Gutierrez > -- skype: felipe.o.gutierrez > -- https://felipeogutierrez.blogspot.com > > On Fri, Jul 3, 2020 at 9:01 AM Felipe Gutierrez > <felipe.o.gutier...@gmail.com> wrote: > > > > yes. I agree. because RocsDB will spill data to disk if there is not > > enough space in memory. > > Thanks > > -- > > -- Felipe Gutierrez > > -- skype: felipe.o.gutierrez > > -- https://felipeogutierrez.blogspot.com > > > > On Fri, Jul 3, 2020 at 8:27 AM Yun Tang <myas...@live.com> wrote: > > > > > > Hi Felipe, > > > > > > I noticed my previous mail has a typo: RocksDB is executed in task main > > > thread which does not take the role to respond to heart beat. Sorry for > > > previous typo, and the key point I want to clarify is that RocksDB should > > > not have business for heartbeat problem. > > > > > > Best > > > Yun Tang > > > ________________________________ > > > From: Felipe Gutierrez <felipe.o.gutier...@gmail.com> > > > Sent: Tuesday, June 30, 2020 17:46 > > > To: Yun Tang <myas...@live.com> > > > Cc: Ori Popowski <ori....@gmail.com>; user <user@flink.apache.org> > > > Subject: Re: Timeout when using RockDB to handle large state in a stream > > > app > > > > > > Hi, > > > > > > I reduced the size of the tables that I am loading on a ListState and > > > the query worked. One of them was about 700MB [1] [2]. > > > > > > Now I am gonna deploy it on the cluster and check if it works. I will > > > probably need to increase the heartbeat timeout. > > > > > > Thanks, > > > Felipe > > > [1] > > > https://github.com/apache/flink/blob/master/flink-examples/flink-examples-batch/src/main/java/org/apache/flink/examples/java/relational/TPCHQuery3.java > > > [2] > > > https://github.com/apache/flink/blob/master/flink-examples/flink-examples-batch/src/main/java/org/apache/flink/examples/java/relational/TPCHQuery10.java > > > -- > > > -- Felipe Gutierrez > > > -- skype: felipe.o.gutierrez > > > -- https://felipeogutierrez.blogspot.com > > > > > > On Tue, Jun 30, 2020 at 10:51 AM Yun Tang <myas...@live.com> wrote: > > > > > > > > Hi Felipe > > > > > > > > RocksDB is executed in task main thread which does take the role to > > > > respond to heart beat and RocksDB mainly use native memory which is > > > > decoupled from JVM heap to not bring any GC pressure. Thus, timeout > > > > should have no relationship with RocksDB in general if your task > > > > manager is really heartbeat timeout instead of crash to exit. > > > > > > > > Try to increase the heartbeat timeout [1] and watch the GC detail logs > > > > to see anything weird. > > > > > > > > [1] > > > > https://ci.apache.org/projects/flink/flink-docs-stable/ops/config.html#heartbeat-timeout > > > > > > > > Best > > > > Yun Tang > > > > > > > > ________________________________ > > > > From: Ori Popowski <ori....@gmail.com> > > > > Sent: Monday, June 29, 2020 17:44 > > > > Cc: user <user@flink.apache.org> > > > > Subject: Re: Timeout when using RockDB to handle large state in a > > > > stream app > > > > > > > > Hi there, > > > > > > > > I'm currently experiencing the exact same issue. > > > > > > > > http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/Heartbeat-of-TaskManager-timed-out-td36228.html > > > > > > > > I've found out that GC is causing the problem, but I still haven't > > > > managed to solve this. > > > > > > > > > > > > > > > > On Mon, Jun 29, 2020 at 12:39 PM Felipe Gutierrez > > > > <felipe.o.gutier...@gmail.com> wrote: > > > > > > > > Hi community, > > > > > > > > I am trying to run a stream application with large state in a > > > > standalone flink cluster [3]. I configured the RocksDB state backend > > > > and I increased the memory of the Job Manager and Task Manager. > > > > However, I am still getting the timeout message > > > > "java.util.concurrent.TimeoutException: Heartbeat of TaskManager with > > > > id cb1091d792f52ca4743f345790d87dd5 timed out.". I am using Flink > > > > 1.10.1 and here are the configurations that I changed on the > > > > flink-conf.yaml. For the "state.checkpoints.dir" I am still using the > > > > filesystem. I am not sure if I need to use HDFS here since I am > > > > testing only in one machine. > > > > > > > > jobmanager.heap.size: 12g > > > > taskmanager.memory.process.size: 8g > > > > state.backend: rocksdb > > > > state.checkpoints.dir: file:///tmp/flink/state > > > > > > > > In the stream application I am using RocksDB as well (full code [3]): > > > > StreamExecutionEnvironment env = > > > > StreamExecutionEnvironment.getExecutionEnvironment(); > > > > env.setStateBackend(new RocksDBStateBackend("file:///tmp/flink/state", > > > > true)); > > > > > > > > I have some operators that hold a large state when the load a static > > > > table on their state. I use them in two aggregate operations [1] and > > > > [2]. > > > > > > > > [1] > > > > https://github.com/felipegutierrez/explore-flink/blob/acb4d4675f60c59f5c3de70c9e0ba82031205744/src/main/java/org/sense/flink/examples/stream/tpch/TPCHQuery03.java#L128 > > > > [2] > > > > https://github.com/felipegutierrez/explore-flink/blob/acb4d4675f60c59f5c3de70c9e0ba82031205744/src/main/java/org/sense/flink/examples/stream/tpch/TPCHQuery03.java#L199 > > > > [3] > > > > https://github.com/felipegutierrez/explore-flink/blob/acb4d4675f60c59f5c3de70c9e0ba82031205744/src/main/java/org/sense/flink/examples/stream/tpch/TPCHQuery03.java > > > > > > > > Here is my stack trace error: > > > > > > > > org.apache.flink.runtime.JobException: Recovery is suppressed by > > > > NoRestartBackoffTimeStrategy > > > > at > > > > org.apache.flink.runtime.executiongraph.failover.flip1.ExecutionFailureHandler.handleFailure(ExecutionFailureHandler.java:110) > > > > at > > > > org.apache.flink.runtime.executiongraph.failover.flip1.ExecutionFailureHandler.getFailureHandlingResult(ExecutionFailureHandler.java:76) > > > > at > > > > org.apache.flink.runtime.scheduler.DefaultScheduler.handleTaskFailure(DefaultScheduler.java:192) > > > > at > > > > org.apache.flink.runtime.scheduler.DefaultScheduler.maybeHandleTaskFailure(DefaultScheduler.java:186) > > > > at > > > > org.apache.flink.runtime.scheduler.DefaultScheduler.updateTaskExecutionStateInternal(DefaultScheduler.java:180) > > > > at > > > > org.apache.flink.runtime.scheduler.SchedulerBase.updateTaskExecutionState(SchedulerBase.java:496) > > > > at > > > > org.apache.flink.runtime.scheduler.UpdateSchedulerNgOnInternalFailuresListener.notifyTaskFailure(UpdateSchedulerNgOnInternalFailuresListener.java:49) > > > > at > > > > org.apache.flink.runtime.executiongraph.ExecutionGraph.notifySchedulerNgAboutInternalTaskFailure(ExecutionGraph.java:1703) > > > > at > > > > org.apache.flink.runtime.executiongraph.Execution.processFail(Execution.java:1252) > > > > at > > > > org.apache.flink.runtime.executiongraph.Execution.processFail(Execution.java:1220) > > > > at > > > > org.apache.flink.runtime.executiongraph.Execution.fail(Execution.java:955) > > > > at > > > > org.apache.flink.runtime.jobmaster.slotpool.SingleLogicalSlot.signalPayloadRelease(SingleLogicalSlot.java:173) > > > > at > > > > org.apache.flink.runtime.jobmaster.slotpool.SingleLogicalSlot.release(SingleLogicalSlot.java:165) > > > > at > > > > org.apache.flink.runtime.jobmaster.slotpool.SlotSharingManager$SingleTaskSlot.release(SlotSharingManager.java:732) > > > > at > > > > org.apache.flink.runtime.jobmaster.slotpool.SlotSharingManager$MultiTaskSlot.release(SlotSharingManager.java:537) > > > > at > > > > org.apache.flink.runtime.jobmaster.slotpool.AllocatedSlot.releasePayload(AllocatedSlot.java:149) > > > > at > > > > org.apache.flink.runtime.jobmaster.slotpool.SlotPoolImpl.releaseTaskManagerInternal(SlotPoolImpl.java:818) > > > > at > > > > org.apache.flink.runtime.jobmaster.slotpool.SlotPoolImpl.releaseTaskManager(SlotPoolImpl.java:777) > > > > at > > > > org.apache.flink.runtime.jobmaster.JobMaster.disconnectTaskManager(JobMaster.java:429) > > > > at > > > > org.apache.flink.runtime.jobmaster.JobMaster$TaskManagerHeartbeatListener.notifyHeartbeatTimeout(JobMaster.java:1147) > > > > at > > > > org.apache.flink.runtime.heartbeat.HeartbeatMonitorImpl.run(HeartbeatMonitorImpl.java:109) > > > > at > > > > java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) > > > > at java.util.concurrent.FutureTask.run(FutureTask.java:266) > > > > at > > > > org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRunAsync(AkkaRpcActor.java:402) > > > > at > > > > org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcMessage(AkkaRpcActor.java:195) > > > > at > > > > org.apache.flink.runtime.rpc.akka.FencedAkkaRpcActor.handleRpcMessage(FencedAkkaRpcActor.java:74) > > > > at > > > > org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleMessage(AkkaRpcActor.java:152) > > > > at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:26) > > > > at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:21) > > > > at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:123) > > > > at akka.japi.pf.UnitCaseStatement.applyOrElse(CaseStatements.scala:21) > > > > at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:170) > > > > at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171) > > > > at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171) > > > > at akka.actor.Actor$class.aroundReceive(Actor.scala:517) > > > > at akka.actor.AbstractActor.aroundReceive(AbstractActor.scala:225) > > > > at akka.actor.ActorCell.receiveMessage(ActorCell.scala:592) > > > > at akka.actor.ActorCell.invoke(ActorCell.scala:561) > > > > at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:258) > > > > at akka.dispatch.Mailbox.run(Mailbox.scala:225) > > > > at akka.dispatch.Mailbox.exec(Mailbox.scala:235) > > > > at akka.dispatch.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) > > > > at > > > > akka.dispatch.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) > > > > at akka.dispatch.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) > > > > at > > > > akka.dispatch.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) > > > > Caused by: java.util.concurrent.TimeoutException: Heartbeat of > > > > TaskManager with id cb1091d792f52ca4743f345790d87dd5 timed out. > > > > ... 26 more > > > > > > > > Thanks, > > > > Felipe > > > > -- > > > > -- Felipe Gutierrez > > > > -- skype: felipe.o.gutierrez > > > > -- https://felipeogutierrez.blogspot.com