[ https://issues.apache.org/jira/browse/HADOOP-1139?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12495210 ]
dhruba borthakur commented on HADOOP-1139: ------------------------------------------ I ran an experiment on a cluster that has about 15M blocks. I switched on logging when a block is allocated, when the namenode receives a block-confirmation from the datanode for a block. The log size at the end of a 8 hour random-writer run was about 6.5GB. This shows that increasing the debug-levels of block transitions might flood the logs. An alternative is to print log messages at the following events only: 1. When a file is created and then closed, log all blocks that belong to that file. 2. When a file gets deleted, log all the blocks that belong to that file. 3. When the namenode replication engine triggers re-replication of a block, log it. 4. When a replica is detected to be corrupt, it is deleted. Log it. > All block trasitions should be logged at log level INFO > ------------------------------------------------------- > > Key: HADOOP-1139 > URL: https://issues.apache.org/jira/browse/HADOOP-1139 > Project: Hadoop > Issue Type: Bug > Components: dfs > Reporter: dhruba borthakur > > The namenode records block trasitions in its log file. It is seen that some > of the block transition messages were being logged at debug level. These > should be done at INFO level. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.