How much memory does your Namenode machine have?

You should look at the number of files, directories and blocks on your installation. All these numbers are available via NamenodeFsck.Result

HADOOP-1687 (http://issues.apache.org/jira/browse/HADOOP-1687) has a detailed discussion of the amount of memory used by Namenode data structures.

Sameer

Taeho Kang wrote:
Dear All,

Hi, my name is Taeho and I am trying to figure out the maximum number of
files a namenode can hold.
The main reason for doing this is that I want to have some estimates on how
many files I can put into the HDFS without overflowing the Namenode
machine's memory.

I know the number depends on the size of memory and how much is allocated
for the running JVM.
For the memory usage by the namenode, I can simply use Runtime object of
JDK.
For the total number of files residing in the DFS, I am thinking of using
getTotailfiles() funcion of NamenodeFsck class in
org.apache.hadoop.dfspacakge. Am I correct here in using NamenodeFsck?

Or, has anybody done similar experiments?

Any comments/suggestions will be appreciated.
Thanks in advance.
Best Regards,

Reply via email to