Are you using the RocksDB backend in native mode? If so then the off-heap
memory may be there.

On Tue, Nov 29, 2016 at 9:54 AM, <rimin...@sina.cn> wrote:

> i have the same problem,but i put the flink job into yarn.
> but i put the job into yarn on the computer 22,and the job can success
> run,and the jobmanager is 79 and taskmanager is 69,they three different
> compu345ter,
> however,on computer 22,the pid=3463,which is the job that put into yarn,is
> have 2.3g memory,15% of total,
> the commend is : ./flink run -m yarn-cluster -yn 1 -ys 1 -yjm 1024 -ytm
> 1024 ....
> why in conputer 22,has occupy so much momory?the job is running computer
> 79 and computer 69.
> What would be the possible causes of such behavior ?
> Best Regards,
> ----- 原始邮件 -----
> 发件人:Daniel Santos <dsan...@cryptolab.net>
> 收件人:user@flink.apache.org
> 主题:JVM Non Heap Memory
> 日期:2016年11月29日 22点26分
>
>
> Hello,
> Is it common to have high usage of Non-Heap in JVM ?
> I am running flink in stand-alone cluster and in docker, with each
> docker bieng capped at 6G of memory.
> I have been struggling to keep memory usage in check.
> The non-heap increases to no end. It start with just 100MB of usage and
> after a day it reaches to 1,3GB.
> Then evetually reaches to 2GB and then eventually the docker is killed
> because it has reached the memory limit.
> My configuration for each flink task manager is the following :
> ----------- flink-conf.yaml --------------
> taskmanager.heap.mb: 3072
> taskmanager.numberOfTaskSlots: 8
> taskmanager.memory.preallocate: false
> taskmanager.network.numberOfBuffers: 12500
> taskmanager.memory.off-heap: false
> ---------------------------------------------
> What would be the possible causes of such behavior ?
> Best Regards,
> Daniel Santos
>

Reply via email to