This is an automated email from the ASF dual-hosted git repository. sammichen pushed a commit to branch HDDS-2034 in repository https://gitbox.apache.org/repos/asf/hadoop-ozone.git
commit d13f9600c58e82c5e7ed434b97fbb31392a341f0 Author: Sammi Chen <sammic...@apache.org> AuthorDate: Tue Oct 15 14:52:02 2019 +0800 Fix checkstyle and code improvement --- .../org/apache/hadoop/hdds/scm/pipeline/Pipeline.java | 2 -- .../common/report/PipelineReportPublisher.java | 1 - .../common/statemachine/DatanodeStateMachine.java | 8 -------- .../hdds/scm/container/ContainerStateManager.java | 18 +++++++++--------- .../hdds/scm/pipeline/RatisPipelineProvider.java | 1 - .../scm/container/TestCloseContainerEventHandler.java | 9 +-------- .../hdds/scm/pipeline/TestSCMPipelineManager.java | 1 - .../client/rpc/TestContainerStateMachineFailures.java | 7 ++++++- 8 files changed, 16 insertions(+), 31 deletions(-) diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/pipeline/Pipeline.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/pipeline/Pipeline.java index f2ca2fa..4414368 100644 --- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/pipeline/Pipeline.java +++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/pipeline/Pipeline.java @@ -50,7 +50,6 @@ public final class Pipeline { private Map<DatanodeDetails, Long> nodeStatus; // nodes with ordered distance to client private ThreadLocal<List<DatanodeDetails>> nodesInOrder = new ThreadLocal<>(); - private final long creationTime; /** * The immutable properties of pipeline object is used in @@ -65,7 +64,6 @@ public final class Pipeline { this.factor = factor; this.state = state; this.nodeStatus = nodeStatus; - this.creationTime = System.currentTimeMillis(); } /** diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/report/PipelineReportPublisher.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/report/PipelineReportPublisher.java index e1dd098..e7f4347 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/report/PipelineReportPublisher.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/report/PipelineReportPublisher.java @@ -68,7 +68,6 @@ public class PipelineReportPublisher extends @Override protected PipelineReportsProto getReport() { - System.out.println("Pipeline Report Generate"); return getContext().getParent().getContainer().getPipelineReport(); } } diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/statemachine/DatanodeStateMachine.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/statemachine/DatanodeStateMachine.java index 926f19c..6832153 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/statemachine/DatanodeStateMachine.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/statemachine/DatanodeStateMachine.java @@ -492,12 +492,4 @@ public class DatanodeStateMachine implements Closeable { public ReplicationSupervisor getSupervisor() { return supervisor; } - - public Configuration getConf() { - return conf; - } - - public CertificateClient getCertificateClient() { - return dnCertClient; - } } diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/ContainerStateManager.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/ContainerStateManager.java index 78a944f..cefc185 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/ContainerStateManager.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/ContainerStateManager.java @@ -249,19 +249,19 @@ public class ContainerStateManager { throws IOException { Pipeline pipeline; - final List<Pipeline> pipelines = pipelineManager - .getPipelines(type, replicationFactor, Pipeline.PipelineState.OPEN); - if (pipelines.isEmpty()) { - try { - pipeline = pipelineManager.createPipeline(type, replicationFactor); - pipelineManager.waitPipelineReady(pipeline.getId(), 0); - } catch (IOException e) { - LOG.error("Fail to create pipeline for " + e.getMessage()); + try { + // TODO: #CLUTIL remove creation logic when all replication types and + // factors are handled by pipeline creator job. + pipeline = pipelineManager.createPipeline(type, replicationFactor); + pipelineManager.waitPipelineReady(pipeline.getId(), 0); + } catch (IOException e) { + final List<Pipeline> pipelines = pipelineManager + .getPipelines(type, replicationFactor, Pipeline.PipelineState.OPEN); + if (pipelines.isEmpty()) { throw new IOException("Could not allocate container. Cannot get any" + " matching pipeline for Type:" + type + ", Factor:" + replicationFactor + ", State:PipelineState.OPEN"); } - } else { pipeline = pipelines.get((int) containerCount.get() % pipelines.size()); } synchronized (pipeline) { diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/pipeline/RatisPipelineProvider.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/pipeline/RatisPipelineProvider.java index 126b0c7..8663471 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/pipeline/RatisPipelineProvider.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/pipeline/RatisPipelineProvider.java @@ -200,7 +200,6 @@ public class RatisPipelineProvider implements PipelineProvider { * @throws IOException */ public void close(Pipeline pipeline) { - LOG.debug("Destroy pipeline:{}", pipeline.getId()); final ClosePipelineCommand closeCommand = new ClosePipelineCommand(pipeline.getId()); pipeline.getNodes().stream().forEach(node -> { diff --git a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/TestCloseContainerEventHandler.java b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/TestCloseContainerEventHandler.java index b27f8a3..dca49cd 100644 --- a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/TestCloseContainerEventHandler.java +++ b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/TestCloseContainerEventHandler.java @@ -26,7 +26,6 @@ import org.apache.hadoop.hdds.protocol.DatanodeDetails; import org.apache.hadoop.hdds.protocol.proto.HddsProtos; import org.apache.hadoop.hdds.scm.TestUtils; import org.apache.hadoop.hdds.scm.pipeline.MockRatisPipelineProvider; -import org.apache.hadoop.hdds.scm.pipeline.Pipeline; import org.apache.hadoop.hdds.scm.pipeline.PipelineProvider; import org.apache.hadoop.hdds.scm.pipeline.SCMPipelineManager; import org.apache.hadoop.hdds.server.events.EventQueue; @@ -68,6 +67,7 @@ public class TestCloseContainerEventHandler { configuration .set(HddsConfigKeys.OZONE_METADATA_DIRS, testDir.getAbsolutePath()); nodeManager = new MockNodeManager(true, 10); + eventQueue = new EventQueue(); pipelineManager = new SCMPipelineManager(configuration, nodeManager, eventQueue); PipelineProvider mockRatisProvider = @@ -78,7 +78,6 @@ public class TestCloseContainerEventHandler { containerManager = new SCMContainerManager(configuration, nodeManager, pipelineManager, new EventQueue()); - eventQueue = new EventQueue(); eventQueue.addHandler(CLOSE_CONTAINER, new CloseContainerEventHandler(pipelineManager, containerManager)); eventQueue.addHandler(DATANODE_COMMAND, nodeManager); @@ -119,9 +118,6 @@ public class TestCloseContainerEventHandler { @Test public void testCloseContainerEventWithValidContainers() throws IOException { - Pipeline pipeline = pipelineManager.createPipeline( - HddsProtos.ReplicationType.RATIS, HddsProtos.ReplicationFactor.ONE); - pipelineManager.openPipeline(pipeline.getId()); ContainerInfo container = containerManager .allocateContainer(HddsProtos.ReplicationType.RATIS, HddsProtos.ReplicationFactor.ONE, "ozone"); @@ -139,9 +135,6 @@ public class TestCloseContainerEventHandler { @Test public void testCloseContainerEventWithRatis() throws IOException { - Pipeline pipeline = pipelineManager.createPipeline( - HddsProtos.ReplicationType.RATIS, HddsProtos.ReplicationFactor.THREE); - pipelineManager.openPipeline(pipeline.getId()); GenericTestUtils.LogCapturer logCapturer = GenericTestUtils.LogCapturer .captureLogs(CloseContainerEventHandler.LOG); ContainerInfo container = containerManager diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/pipeline/TestSCMPipelineManager.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/pipeline/TestSCMPipelineManager.java index 5d934eb..59bccc2 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/pipeline/TestSCMPipelineManager.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/pipeline/TestSCMPipelineManager.java @@ -31,7 +31,6 @@ import org.apache.hadoop.hdds.scm.TestUtils; import org.apache.hadoop.hdds.scm.container.ContainerID; import org.apache.hadoop.hdds.scm.container.MockNodeManager; import org.apache.hadoop.hdds.scm.safemode.SCMSafeModeManager; -import org.apache.hadoop.hdds.scm.server.SCMDatanodeHeartbeatDispatcher; import org.apache.hadoop.hdds.scm.server.SCMDatanodeHeartbeatDispatcher .PipelineReportFromDatanode; import org.apache.hadoop.hdds.server.events.EventQueue; diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestContainerStateMachineFailures.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestContainerStateMachineFailures.java index 9ac45b8..a45d258 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestContainerStateMachineFailures.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestContainerStateMachineFailures.java @@ -68,6 +68,8 @@ import static org.apache.hadoop.hdds.HddsConfigKeys. HDDS_COMMAND_STATUS_REPORT_INTERVAL; import static org.apache.hadoop.hdds.HddsConfigKeys. HDDS_CONTAINER_REPORT_INTERVAL; +import static org.apache.hadoop.hdds.HddsConfigKeys + .HDDS_PIPELINE_REPORT_INTERVAL; import static org.apache.hadoop.hdds.protocol.datanode.proto.ContainerProtos. ContainerDataProto.State.UNHEALTHY; import static org.apache.hadoop.hdds.scm.ScmConfigKeys. @@ -112,6 +114,8 @@ public class TestContainerStateMachineFailures { TimeUnit.MILLISECONDS); conf.setTimeDuration(HDDS_COMMAND_STATUS_REPORT_INTERVAL, 200, TimeUnit.MILLISECONDS); + conf.setTimeDuration(HDDS_PIPELINE_REPORT_INTERVAL, 200, + TimeUnit.MILLISECONDS); conf.setTimeDuration(HDDS_SCM_WATCHER_TIMEOUT, 1000, TimeUnit.MILLISECONDS); conf.setTimeDuration(OZONE_SCM_STALENODE_INTERVAL, 3, TimeUnit.SECONDS); conf.setTimeDuration(OZONE_SCM_PIPELINE_DESTROY_TIMEOUT, 10, @@ -458,6 +462,7 @@ public class TestContainerStateMachineFailures { Assert.assertTrue(!dispatcher.getMissingContainerSet().isEmpty()); Assert .assertTrue(dispatcher.getMissingContainerSet().contains(containerID)); + // write a new key key = objectStore.getVolume(volumeName).getBucket(bucketName) .createKey("ratis", 1024, ReplicationType.RATIS, ReplicationFactor.ONE, @@ -481,7 +486,7 @@ public class TestContainerStateMachineFailures { byte[] blockCommitSequenceIdKey = DFSUtil.string2Bytes(OzoneConsts.BLOCK_COMMIT_SEQUENCE_ID_PREFIX); - // modify the bcsid for the container in the ROCKS DB tereby inducing + // modify the bcsid for the container in the ROCKS DB thereby inducing // corruption db.getStore().put(blockCommitSequenceIdKey, Longs.toByteArray(0)); db.decrementReference(); --------------------------------------------------------------------- To unsubscribe, e-mail: hdfs-commits-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-commits-h...@hadoop.apache.org