[ 
https://issues.apache.org/jira/browse/HDDS-11234?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Wei-Chiu Chuang updated HDDS-11234:
-----------------------------------
    Attachment: ozone_dn_leak.pdf

> Manage Netty native memory consumption
> --------------------------------------
>
>                 Key: HDDS-11234
>                 URL: https://issues.apache.org/jira/browse/HDDS-11234
>             Project: Apache Ozone
>          Issue Type: Sub-task
>            Reporter: Wei-Chiu Chuang
>            Priority: Major
>         Attachments: ozone_dn_leak.pdf
>
>
> We observed an Ozone DataNode that used way too much memory with HBase 
> LoadTest workload.
> Most of its memory allocation was non-heap, so we thought there might be a 
> native memory leak.
> Used jemalloc and jeprof tools, I managed to produce the native memory 
> allocation map, and it shows most of it comes from Unsafe_AllocateMemory. 
> After some digging, it led me to this post 
> [https://github.com/netty/netty/issues/11835] where it suggests Netty's 
> internal memory management is to blame. A workaround is to disable it  (Java 
> property -Dio.netty.allocator.type=unpooled) and another is to reduce the 
> native memory size using -Dio.netty.maxDirectMemory=<size>
>  
> By default, the size is the same as JDK's maximum direct memory size 
> (-XX:MaxDirectMemorySize), which is usally the same as max heap size (-Xmx). 
> We should provide a best practice for users. In addition, we have Ratis 
> shaded Netty and gRPC. They use different Netty properties to configure 
> memory size (-Dio.netty.maxDirectMemory, 
> -Dorg.apache.ratis.thirdparty.io.netty.maxDirectMemory) So in theory the 
> memory consumption can go up to 3x of maximum heap size.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to