Namenode unstable when replicating 500k blocks at once ------------------------------------------------------
Key: HADOOP-2606 URL: https://issues.apache.org/jira/browse/HADOOP-2606 Project: Hadoop Issue Type: Bug Components: dfs Affects Versions: 0.14.3 Reporter: Koji Noguchi Fix For: 0.17.0 We tried to decommission about 40 nodes at once, each containing 12k blocks. (about 500k total) (This also happened when we first tried to decommission 2 million blocks) Clients started experiencing "java.lang.RuntimeException: java.net.SocketTimeoutException: timed out waiting for rpc response" and namenode was in 100% cpu state. It was spending most of its time on one thread, "[EMAIL PROTECTED]" daemon prio=10 tid=0x0000002e10702800 nid=0x6718 runnable [0x0000000041a42000..0x0000000041a42a30] java.lang.Thread.State: RUNNABLE at org.apache.hadoop.dfs.FSNamesystem.containingNodeList(FSNamesystem.java:2766) at org.apache.hadoop.dfs.FSNamesystem.pendingTransfers(FSNamesystem.java:2870) - locked <0x0000002aa3cef720> (a org.apache.hadoop.dfs.UnderReplicatedBlocks) - locked <0x0000002aa3c42e28> (a org.apache.hadoop.dfs.FSNamesystem) at org.apache.hadoop.dfs.FSNamesystem.computeDatanodeWork(FSNamesystem.java:1928) at org.apache.hadoop.dfs.FSNamesystem$ReplicationMonitor.run(FSNamesystem.java:1868) at java.lang.Thread.run(Thread.java:619) We confirmed that Namenode was not in the fullGC states when these problem happened. Also, dfsadmin -metasave was showing "Blocks waiting for replication" was decreasing very slowly. I believe this is not specific to decommission and same problem would happen if we lose one rack. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.