Author: eli
Date: Fri May 11 00:05:15 2012
New Revision: 1336973
URL: http://svn.apache.org/viewvc?rev=1336973&view=rev
Log:
HDFS-3401. svn merge -c 1336972 from trunk
Modified:
hadoop/common/branches/branch-2/hadoop-hdfs-project/ (props changed)
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/ (props
changed)
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/main/java/
(props changed)
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/main/native/
(props changed)
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/main/webapps/datanode/
(props changed)
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/main/webapps/hdfs/
(props changed)
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/main/webapps/secondary/
(props changed)
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/hdfs/
(props changed)
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/DFSTestUtil.java
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDFSClientRetries.java
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestGetBlocks.java
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/protocolPB/TestPBHelper.java
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestBlockManager.java
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestHost2NodesMap.java
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestReplicationPolicy.java
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestBPOfferService.java
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestBlockRecovery.java
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/TestInterDatanodeProtocol.java
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/net/TestNetworkTopology.java
Propchange: hadoop/common/branches/branch-2/hadoop-hdfs-project/
------------------------------------------------------------------------------
Merged /hadoop/common/trunk/hadoop-hdfs-project:r1336972
Propchange: hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/
------------------------------------------------------------------------------
Merged /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs:r1336972
Modified:
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
URL:
http://svn.apache.org/viewvc/hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt?rev=1336973&r1=1336972&r2=1336973&view=diff
==============================================================================
--- hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
(original)
+++ hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
Fri May 11 00:05:15 2012
@@ -308,6 +308,10 @@ Release 2.0.0 - UNRELEASED
HDFS-3134. harden edit log loader against malformed or malicious input.
(Colin Patrick McCabe via eli)
+ HDFS-3230. Cleanup DatanodeID creation in the tests. (eli)
+
+ HDFS-3401. Cleanup DatanodeDescriptor creation in the tests. (eli)
+
OPTIMIZATIONS
HDFS-2477. Optimize computing the diff between a block report and the
@@ -521,8 +525,6 @@ Release 2.0.0 - UNRELEASED
HDFS-3395. NN doesn't start with HA+security enabled and HTTP address set
to 0.0.0.0. (atm)
- HDFS-3230. Cleanup DatanodeID creation in the tests. (eli)
-
BREAKDOWN OF HDFS-1623 SUBTASKS
HDFS-2179. Add fencing framework and mechanisms for NameNode HA. (todd)
Propchange:
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/main/java/
------------------------------------------------------------------------------
Merged
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java:r1336972
Propchange:
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/main/native/
------------------------------------------------------------------------------
Merged
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/native:r1336972
Propchange:
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/main/webapps/datanode/
------------------------------------------------------------------------------
Merged
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/webapps/datanode:r1336972
Propchange:
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/main/webapps/hdfs/
------------------------------------------------------------------------------
Merged
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/webapps/hdfs:r1336972
Propchange:
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/main/webapps/secondary/
------------------------------------------------------------------------------
Merged
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/webapps/secondary:r1336972
Propchange:
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/hdfs/
------------------------------------------------------------------------------
Merged
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/hdfs:r1336972
Modified:
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/DFSTestUtil.java
URL:
http://svn.apache.org/viewvc/hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/DFSTestUtil.java?rev=1336973&r1=1336972&r2=1336973&view=diff
==============================================================================
---
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/DFSTestUtil.java
(original)
+++
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/DFSTestUtil.java
Fri May 11 00:05:15 2012
@@ -59,6 +59,7 @@ import org.apache.hadoop.hdfs.MiniDFSClu
import org.apache.hadoop.hdfs.client.HdfsDataInputStream;
import org.apache.hadoop.hdfs.protocol.DatanodeID;
import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
+import org.apache.hadoop.hdfs.protocol.DatanodeInfo.AdminStates;
import org.apache.hadoop.hdfs.protocol.ExtendedBlock;
import org.apache.hadoop.hdfs.protocol.HdfsConstants;
import org.apache.hadoop.hdfs.protocol.LocatedBlock;
@@ -706,12 +707,19 @@ public class DFSTestUtil {
.join(nameservices));
}
+ private static DatanodeID getDatanodeID(String ipAddr) {
+ return new DatanodeID(ipAddr, "localhost",
+ DFSConfigKeys.DFS_DATANODE_DEFAULT_PORT);
+ }
+
public static DatanodeID getLocalDatanodeID() {
- return new DatanodeID("127.0.0.1", "localhost",
DFSConfigKeys.DFS_DATANODE_DEFAULT_PORT);
+ return new DatanodeID("127.0.0.1", "localhost",
+ DFSConfigKeys.DFS_DATANODE_DEFAULT_PORT);
}
public static DatanodeID getLocalDatanodeID(int port) {
- return new DatanodeID("127.0.0.1", "localhost", "", port, port, port);
+ return new DatanodeID("127.0.0.1", "localhost", "",
+ port, port, port);
}
public static DatanodeDescriptor getLocalDatanodeDescriptor() {
@@ -722,7 +730,36 @@ public class DFSTestUtil {
return new DatanodeInfo(getLocalDatanodeID());
}
+ public static DatanodeInfo getDatanodeInfo(String ipAddr) {
+ return new DatanodeInfo(getDatanodeID(ipAddr));
+ }
+
public static DatanodeInfo getLocalDatanodeInfo(int port) {
return new DatanodeInfo(getLocalDatanodeID(port));
}
+
+ public static DatanodeInfo getDatanodeInfo(String ipAddr,
+ String host, int port) {
+ return new DatanodeInfo(new DatanodeID(ipAddr, host, port));
+ }
+
+ public static DatanodeInfo getLocalDatanodeInfo(String ipAddr,
+ String hostname, AdminStates adminState) {
+ return new DatanodeInfo(ipAddr, hostname, "storage",
+ DFSConfigKeys.DFS_DATANODE_DEFAULT_PORT,
+ DFSConfigKeys.DFS_DATANODE_HTTP_DEFAULT_PORT,
+ DFSConfigKeys.DFS_DATANODE_IPC_DEFAULT_PORT,
+ 1, 2, 3, 4, 5, 6, "local", adminState);
+ }
+
+ public static DatanodeDescriptor getDatanodeDescriptor(String ipAddr,
+ String rackLocation) {
+ return getDatanodeDescriptor(ipAddr,
DFSConfigKeys.DFS_DATANODE_DEFAULT_PORT,
+ rackLocation);
+ }
+
+ public static DatanodeDescriptor getDatanodeDescriptor(String ipAddr,
+ int port, String rackLocation) {
+ return new DatanodeDescriptor(new DatanodeID(ipAddr, port), rackLocation);
+ }
}
Modified:
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDFSClientRetries.java
URL:
http://svn.apache.org/viewvc/hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDFSClientRetries.java?rev=1336973&r1=1336972&r2=1336973&view=diff
==============================================================================
---
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDFSClientRetries.java
(original)
+++
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDFSClientRetries.java
Fri May 11 00:05:15 2012
@@ -332,7 +332,7 @@ public class TestDFSClientRetries extend
LocatedBlock badLocatedBlock = new LocatedBlock(
goodLocatedBlock.getBlock(),
new DatanodeInfo[] {
- new DatanodeInfo(new DatanodeID("255.255.255.255", 234))
+ DFSTestUtil.getDatanodeInfo("1.2.3.4", "bogus", 1234)
},
goodLocatedBlock.getStartOffset(),
false);
Modified:
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestGetBlocks.java
URL:
http://svn.apache.org/viewvc/hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestGetBlocks.java?rev=1336973&r1=1336972&r2=1336973&view=diff
==============================================================================
---
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestGetBlocks.java
(original)
+++
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestGetBlocks.java
Fri May 11 00:05:15 2012
@@ -121,8 +121,7 @@ public class TestGetBlocks extends TestC
getBlocksWithException(namenode, dataNodes[0], -1);
// get blocks of size BlockSize from a non-existent datanode
- DatanodeInfo info = DFSTestUtil.getLocalDatanodeInfo();
- info.setIpAddr("1.2.3.4");
+ DatanodeInfo info = DFSTestUtil.getDatanodeInfo("1.2.3.4");
getBlocksWithException(namenode, info, 2);
} finally {
cluster.shutdown();
Modified:
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/protocolPB/TestPBHelper.java
URL:
http://svn.apache.org/viewvc/hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/protocolPB/TestPBHelper.java?rev=1336973&r1=1336972&r2=1336973&view=diff
==============================================================================
---
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/protocolPB/TestPBHelper.java
(original)
+++
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/protocolPB/TestPBHelper.java
Fri May 11 00:05:15 2012
@@ -400,16 +400,11 @@ public class TestPBHelper {
@Test
public void testConvertLocatedBlock() {
- DatanodeInfo [] dnInfos = new DatanodeInfo[3];
- dnInfos[0] = new DatanodeInfo("127.0.0.1", "host1", "0",
- 5000, 5001, 5002, 20000, 10001, 9999,
- 59, 69, 32, "local", AdminStates.DECOMMISSION_INPROGRESS);
- dnInfos[1] = new DatanodeInfo("127.0.0.1", "host2", "1",
- 5000, 5001, 5002, 20000, 10001, 9999,
- 59, 69, 32, "local", AdminStates.DECOMMISSIONED);
- dnInfos[2] = new DatanodeInfo("127.0.0.1", "host3", "2",
- 5000, 5001, 5002, 20000, 10001, 9999,
- 59, 69, 32, "local", AdminStates.NORMAL);
+ DatanodeInfo [] dnInfos = {
+ DFSTestUtil.getLocalDatanodeInfo("1.1.1.1", "h1",
AdminStates.DECOMMISSION_INPROGRESS),
+ DFSTestUtil.getLocalDatanodeInfo("2.2.2.2", "h2",
AdminStates.DECOMMISSIONED),
+ DFSTestUtil.getLocalDatanodeInfo("3.3.3.3", "h3", AdminStates.NORMAL)
+ };
LocatedBlock lb = new LocatedBlock(
new ExtendedBlock("bp12", 12345, 10, 53), dnInfos, 5, false);
LocatedBlockProto lbProto = PBHelper.convert(lb);
Modified:
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestBlockManager.java
URL:
http://svn.apache.org/viewvc/hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestBlockManager.java?rev=1336973&r1=1336972&r2=1336973&view=diff
==============================================================================
---
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestBlockManager.java
(original)
+++
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestBlockManager.java
Fri May 11 00:05:15 2012
@@ -26,6 +26,7 @@ import java.util.Map.Entry;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.hdfs.DFSConfigKeys;
+import org.apache.hadoop.hdfs.DFSTestUtil;
import org.apache.hadoop.hdfs.HdfsConfiguration;
import org.apache.hadoop.hdfs.protocol.Block;
import org.apache.hadoop.hdfs.protocol.DatanodeID;
@@ -76,12 +77,12 @@ public class TestBlockManager {
Mockito.doReturn(true).when(fsn).hasWriteLock();
bm = new BlockManager(fsn, fsn, conf);
nodes = ImmutableList.of(
- new DatanodeDescriptor(new DatanodeID("1.1.1.1", 5020), "/rackA"),
- new DatanodeDescriptor(new DatanodeID("2.2.2.2", 5020), "/rackA"),
- new DatanodeDescriptor(new DatanodeID("3.3.3.3", 5020), "/rackA"),
- new DatanodeDescriptor(new DatanodeID("4.4.4.4", 5020), "/rackB"),
- new DatanodeDescriptor(new DatanodeID("5.5.5.5", 5020), "/rackB"),
- new DatanodeDescriptor(new DatanodeID("6.6.6.6", 5020), "/rackB")
+ DFSTestUtil.getDatanodeDescriptor("1.1.1.1", "/rackA"),
+ DFSTestUtil.getDatanodeDescriptor("2.2.2.2", "/rackA"),
+ DFSTestUtil.getDatanodeDescriptor("3.3.3.3", "/rackA"),
+ DFSTestUtil.getDatanodeDescriptor("4.4.4.4", "/rackB"),
+ DFSTestUtil.getDatanodeDescriptor("5.5.5.5", "/rackB"),
+ DFSTestUtil.getDatanodeDescriptor("6.6.6.6", "/rackB")
);
rackA = nodes.subList(0, 3);
rackB = nodes.subList(3, 6);
@@ -277,7 +278,7 @@ public class TestBlockManager {
// the block is still under-replicated. Add a new node. This should allow
// the third off-rack replica.
DatanodeDescriptor rackCNode =
- new DatanodeDescriptor(new DatanodeID("7.7.7.7", 100), "/rackC");
+ DFSTestUtil.getDatanodeDescriptor("7.7.7.7", "/rackC");
addNodes(ImmutableList.of(rackCNode));
try {
DatanodeDescriptor[] pipeline2 = scheduleSingleReplication(blockInfo);
@@ -317,13 +318,13 @@ public class TestBlockManager {
@Test
public void testBlocksAreNotUnderreplicatedInSingleRack() throws Exception {
- List<DatanodeDescriptor> nodes = ImmutableList.of(
- new DatanodeDescriptor(new DatanodeID("h1", 5020), "/rackA"),
- new DatanodeDescriptor(new DatanodeID("h2", 5020), "/rackA"),
- new DatanodeDescriptor(new DatanodeID("h3", 5020), "/rackA"),
- new DatanodeDescriptor(new DatanodeID("h4", 5020), "/rackA"),
- new DatanodeDescriptor(new DatanodeID("h5", 5020), "/rackA"),
- new DatanodeDescriptor(new DatanodeID("h6", 5020), "/rackA")
+ List<DatanodeDescriptor> nodes = ImmutableList.of(
+ DFSTestUtil.getDatanodeDescriptor("1.1.1.1", "/rackA"),
+ DFSTestUtil.getDatanodeDescriptor("2.2.2.2", "/rackA"),
+ DFSTestUtil.getDatanodeDescriptor("3.3.3.3", "/rackA"),
+ DFSTestUtil.getDatanodeDescriptor("4.4.4.4", "/rackA"),
+ DFSTestUtil.getDatanodeDescriptor("5.5.5.5", "/rackA"),
+ DFSTestUtil.getDatanodeDescriptor("6.6.6.6", "/rackA")
);
addNodes(nodes);
List<DatanodeDescriptor> origNodes = nodes.subList(0, 3);;
Modified:
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestHost2NodesMap.java
URL:
http://svn.apache.org/viewvc/hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestHost2NodesMap.java?rev=1336973&r1=1336972&r2=1336973&view=diff
==============================================================================
---
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestHost2NodesMap.java
(original)
+++
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestHost2NodesMap.java
Fri May 11 00:05:15 2012
@@ -18,7 +18,7 @@
package org.apache.hadoop.hdfs.server.blockmanagement;
-import org.apache.hadoop.hdfs.protocol.DatanodeID;
+import org.apache.hadoop.hdfs.DFSTestUtil;
import org.junit.Before;
import org.junit.Test;
@@ -32,10 +32,10 @@ public class TestHost2NodesMap {
@Before
public void setup() {
dataNodes = new DatanodeDescriptor[] {
- new DatanodeDescriptor(new DatanodeID("1.1.1.1", 5020), "/d1/r1"),
- new DatanodeDescriptor(new DatanodeID("2.2.2.2", 5020), "/d1/r1"),
- new DatanodeDescriptor(new DatanodeID("3.3.3.3", 5020), "/d1/r2"),
- new DatanodeDescriptor(new DatanodeID("3.3.3.3", 5030), "/d1/r2"),
+ DFSTestUtil.getDatanodeDescriptor("1.1.1.1", "/d1/r1"),
+ DFSTestUtil.getDatanodeDescriptor("2.2.2.2", "/d1/r1"),
+ DFSTestUtil.getDatanodeDescriptor("3.3.3.3", "/d1/r2"),
+ DFSTestUtil.getDatanodeDescriptor("3.3.3.3", 5021, "/d1/r2"),
};
for (DatanodeDescriptor node : dataNodes) {
map.add(node);
@@ -46,7 +46,7 @@ public class TestHost2NodesMap {
@Test
public void testContains() throws Exception {
DatanodeDescriptor nodeNotInMap =
- new DatanodeDescriptor(new DatanodeID("3.3.3.3", 5040), "/d1/r4");
+ DFSTestUtil.getDatanodeDescriptor("3.3.3.3", "/d1/r4");
for (int i = 0; i < dataNodes.length; i++) {
assertTrue(map.contains(dataNodes[i]));
}
@@ -66,7 +66,7 @@ public class TestHost2NodesMap {
@Test
public void testRemove() throws Exception {
DatanodeDescriptor nodeNotInMap =
- new DatanodeDescriptor(new DatanodeID("3.3.3.3", 5040), "/d1/r4");
+ DFSTestUtil.getDatanodeDescriptor("3.3.3.3", "/d1/r4");
assertFalse(map.remove(nodeNotInMap));
assertTrue(map.remove(dataNodes[0]));
Modified:
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestReplicationPolicy.java
URL:
http://svn.apache.org/viewvc/hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestReplicationPolicy.java?rev=1336973&r1=1336972&r2=1336973&view=diff
==============================================================================
---
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestReplicationPolicy.java
(original)
+++
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestReplicationPolicy.java
Fri May 11 00:05:15 2012
@@ -56,12 +56,12 @@ public class TestReplicationPolicy {
public static void setupCluster() throws Exception {
Configuration conf = new HdfsConfiguration();
dataNodes = new DatanodeDescriptor[] {
- new DatanodeDescriptor(new DatanodeID("1.1.1.1", 5020), "/d1/r1"),
- new DatanodeDescriptor(new DatanodeID("2.2.2.2", 5020), "/d1/r1"),
- new DatanodeDescriptor(new DatanodeID("3.3.3.3", 5020), "/d1/r2"),
- new DatanodeDescriptor(new DatanodeID("4.4.4.4", 5020), "/d1/r2"),
- new DatanodeDescriptor(new DatanodeID("5.5.5.5", 5020), "/d2/r3"),
- new DatanodeDescriptor(new DatanodeID("6.6.6.6", 5020), "/d2/r3")
+ DFSTestUtil.getDatanodeDescriptor("1.1.1.1", "/d1/r1"),
+ DFSTestUtil.getDatanodeDescriptor("2.2.2.2", "/d1/r1"),
+ DFSTestUtil.getDatanodeDescriptor("3.3.3.3", "/d1/r2"),
+ DFSTestUtil.getDatanodeDescriptor("4.4.4.4", "/d1/r2"),
+ DFSTestUtil.getDatanodeDescriptor("5.5.5.5", "/d2/r3"),
+ DFSTestUtil.getDatanodeDescriptor("6.6.6.6", "/d2/r3")
};
FileSystem.setDefaultUri(conf, "hdfs://localhost:0");
@@ -329,6 +329,7 @@ public class TestReplicationPolicy {
HdfsConstants.MIN_BLOCKS_FOR_WRITE*BLOCK_SIZE, 0L, 0, 0);
}
}
+
/**
* In this testcase, client is is a node outside of file system.
* So the 1st replica can be placed on any node.
@@ -338,8 +339,8 @@ public class TestReplicationPolicy {
*/
@Test
public void testChooseTarget5() throws Exception {
- DatanodeDescriptor writerDesc =
- new DatanodeDescriptor(new DatanodeID("7.7.7.7", 5020), "/d2/r4");
+ DatanodeDescriptor writerDesc =
+ DFSTestUtil.getDatanodeDescriptor("7.7.7.7", "/d2/r4");
DatanodeDescriptor[] targets;
targets = replicator.chooseTarget(filename,
Modified:
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestBPOfferService.java
URL:
http://svn.apache.org/viewvc/hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestBPOfferService.java?rev=1336973&r1=1336972&r2=1336973&view=diff
==============================================================================
---
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestBPOfferService.java
(original)
+++
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestBPOfferService.java
Fri May 11 00:05:15 2012
@@ -115,7 +115,7 @@ public class TestBPOfferService {
0, HdfsConstants.LAYOUT_VERSION))
.when(mock).versionRequest();
- Mockito.doReturn(new DatanodeRegistration("fake-node", 100))
+ Mockito.doReturn(new DatanodeRegistration("1.2.3.4", 100))
.when(mock).registerDatanode(Mockito.any(DatanodeRegistration.class));
Mockito.doAnswer(new HeartbeatAnswer(nnIdx))
Modified:
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestBlockRecovery.java
URL:
http://svn.apache.org/viewvc/hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestBlockRecovery.java?rev=1336973&r1=1336972&r2=1336973&view=diff
==============================================================================
---
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestBlockRecovery.java
(original)
+++
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestBlockRecovery.java
Fri May 11 00:05:15 2012
@@ -47,6 +47,7 @@ import org.apache.hadoop.fs.CommonConfig
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.FileUtil;
import org.apache.hadoop.hdfs.DFSConfigKeys;
+import org.apache.hadoop.hdfs.DFSTestUtil;
import org.apache.hadoop.hdfs.HdfsConfiguration;
import org.apache.hadoop.hdfs.MiniDFSCluster;
import org.apache.hadoop.hdfs.protocol.DatanodeID;
@@ -197,9 +198,9 @@ public class TestBlockRecovery {
locs, RECOVERY_ID);
ArrayList<BlockRecord> syncList = new ArrayList<BlockRecord>(2);
BlockRecord record1 = new BlockRecord(
- new DatanodeID("xx", "yy", "zz", 1, 2, 3), dn1, replica1);
+ DFSTestUtil.getDatanodeInfo("1.2.3.4", "bogus", 1234), dn1, replica1);
BlockRecord record2 = new BlockRecord(
- new DatanodeID("aa", "bb", "cc", 1, 2, 3), dn2, replica2);
+ DFSTestUtil.getDatanodeInfo("1.2.3.4", "bogus", 1234), dn2, replica2);
syncList.add(record1);
syncList.add(record2);
@@ -401,8 +402,7 @@ public class TestBlockRecovery {
private Collection<RecoveringBlock> initRecoveringBlocks() throws
IOException {
Collection<RecoveringBlock> blocks = new ArrayList<RecoveringBlock>(1);
- DatanodeInfo mockOtherDN = new DatanodeInfo(
- new DatanodeID("127.0.0.1", "localhost", "storage-1234", 0, 0, 0));
+ DatanodeInfo mockOtherDN = DFSTestUtil.getLocalDatanodeInfo();
DatanodeInfo[] locs = new DatanodeInfo[] {
new DatanodeInfo(dn.getDNRegistrationForBP(block.getBlockPoolId())),
mockOtherDN };
Modified:
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/TestInterDatanodeProtocol.java
URL:
http://svn.apache.org/viewvc/hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/TestInterDatanodeProtocol.java?rev=1336973&r1=1336972&r2=1336973&view=diff
==============================================================================
---
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/TestInterDatanodeProtocol.java
(original)
+++
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/TestInterDatanodeProtocol.java
Fri May 11 00:05:15 2012
@@ -357,8 +357,7 @@ public class TestInterDatanodeProtocol {
server.start();
final InetSocketAddress addr = NetUtils.getConnectAddress(server);
- DatanodeID fakeDnId = new DatanodeID(
- "localhost", "localhost", "fake-storage", addr.getPort(), 0,
addr.getPort());
+ DatanodeID fakeDnId = DFSTestUtil.getLocalDatanodeID(addr.getPort());
DatanodeInfo dInfo = new DatanodeInfo(fakeDnId);
InterDatanodeProtocol proxy = null;
Modified:
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/net/TestNetworkTopology.java
URL:
http://svn.apache.org/viewvc/hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/net/TestNetworkTopology.java?rev=1336973&r1=1336972&r2=1336973&view=diff
==============================================================================
---
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/net/TestNetworkTopology.java
(original)
+++
hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/net/TestNetworkTopology.java
Fri May 11 00:05:15 2012
@@ -21,6 +21,7 @@ package org.apache.hadoop.net;
import java.util.HashMap;
import java.util.Map;
+import org.apache.hadoop.hdfs.DFSTestUtil;
import org.apache.hadoop.hdfs.protocol.DatanodeID;
import org.apache.hadoop.hdfs.server.blockmanagement.DatanodeDescriptor;
@@ -36,13 +37,13 @@ public class TestNetworkTopology {
@Before
public void setupDatanodes() {
dataNodes = new DatanodeDescriptor[] {
- new DatanodeDescriptor(new DatanodeID("1.1.1.1", 5020), "/d1/r1"),
- new DatanodeDescriptor(new DatanodeID("2.2.2.2", 5020), "/d1/r1"),
- new DatanodeDescriptor(new DatanodeID("3.3.3.3", 5020), "/d1/r2"),
- new DatanodeDescriptor(new DatanodeID("4.4.4.4", 5020), "/d1/r2"),
- new DatanodeDescriptor(new DatanodeID("5.5.5.5", 5020), "/d1/r2"),
- new DatanodeDescriptor(new DatanodeID("6.6.6.6", 5020), "/d2/r3"),
- new DatanodeDescriptor(new DatanodeID("7.7.7.7", 5020), "/d2/r3")
+ DFSTestUtil.getDatanodeDescriptor("1.1.1.1", "/d1/r1"),
+ DFSTestUtil.getDatanodeDescriptor("2.2.2.2", "/d1/r1"),
+ DFSTestUtil.getDatanodeDescriptor("3.3.3.3", "/d1/r2"),
+ DFSTestUtil.getDatanodeDescriptor("4.4.4.4", "/d1/r2"),
+ DFSTestUtil.getDatanodeDescriptor("5.5.5.5", "/d1/r2"),
+ DFSTestUtil.getDatanodeDescriptor("6.6.6.6", "/d2/r3"),
+ DFSTestUtil.getDatanodeDescriptor("7.7.7.7", "/d2/r3")
};
for (int i = 0; i < dataNodes.length; i++) {
cluster.add(dataNodes[i]);
@@ -52,7 +53,7 @@ public class TestNetworkTopology {
@Test
public void testContains() throws Exception {
DatanodeDescriptor nodeNotInMap =
- new DatanodeDescriptor(new DatanodeID("8.8.8.8", 5020), "/d2/r4");
+ DFSTestUtil.getDatanodeDescriptor("8.8.8.8", "/d2/r4");
for (int i=0; i < dataNodes.length; i++) {
assertTrue(cluster.contains(dataNodes[i]));
}
@@ -68,9 +69,9 @@ public class TestNetworkTopology {
public void testCreateInvalidTopology() throws Exception {
NetworkTopology invalCluster = new NetworkTopology();
DatanodeDescriptor invalDataNodes[] = new DatanodeDescriptor[] {
- new DatanodeDescriptor(new DatanodeID("1.1.1.1", 5020), "/d1/r1"),
- new DatanodeDescriptor(new DatanodeID("2.2.2.2", 5020), "/d1/r1"),
- new DatanodeDescriptor(new DatanodeID("3.3.3.3", 5020), "/d1")
+ DFSTestUtil.getDatanodeDescriptor("1.1.1.1", "/d1/r1"),
+ DFSTestUtil.getDatanodeDescriptor("2.2.2.2", "/d1/r1"),
+ DFSTestUtil.getDatanodeDescriptor("3.3.3.3", "/d1")
};
invalCluster.add(invalDataNodes[0]);
invalCluster.add(invalDataNodes[1]);