Kihwal Lee commented on HDFS-13234:

bq. Configuration occupies over 600MB
How big is a single instance in your use case? Bloated conf in dfs client is 
obviously a serious issue, but it can create bigger issues in apps/jobs. 
Sometimes a conf can get embedded in another conf.  Avoiding unnecessarily 
duplicated confs is a good thing, but looking into what is causing the bloat 
and fixing that will also be important.

> Remove renew configuration instance in ConfiguredFailoverProxyProvider and 
> reduce memory footprint for client
> -------------------------------------------------------------------------------------------------------------
>                 Key: HDFS-13234
>                 URL: https://issues.apache.org/jira/browse/HDFS-13234
>             Project: Hadoop HDFS
>          Issue Type: Improvement
>          Components: fs, ha, hdfs-client
>            Reporter: He Xiaoqiao
>            Priority: Major
>         Attachments: HDFS-13234.001.patch
> The memory footprint of #DFSClient is very considerable in some special 
> scenario since there are many #Configuration instances and occupy much memory 
> resource (In an extreme case, org.apache.hadoop.conf.Configuration occupies 
> over 600MB we meet under HDFS Federation an HA with QJM and there are dozens 
> of NameNodes). I think some new Configuration instance is not necessary. Such 
> as  #ConfiguredFailoverProxyProvider initialization.
> {code:java}
>   public ConfiguredFailoverProxyProvider(Configuration conf, URI uri,
>       Class<T> xface, HAProxyFactory<T> factory) {
>     this.xface = xface;
>     this.conf = new Configuration(conf);
>     ......
>   }
> {code}

This message was sent by Atlassian JIRA

To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org

Reply via email to