[
https://issues.apache.org/jira/browse/HDFS-12378?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16167111#comment-16167111
]
Hadoop QA commented on HDFS-12378:
----------------------------------
| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m
38s{color} | {color:blue} Docker mode activated. {color} |
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:red}-1{color} | {color:red} test4tests {color} | {color:red} 0m
0s{color} | {color:red} The patch doesn't appear to include any new or modified
tests. Please justify why no new tests are needed for this patch. Also please
list what manual steps were performed to verify this patch. {color} |
|| || || || {color:brown} trunk Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 18m
41s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m
13s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m
47s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m
17s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 2m
8s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m
55s{color} | {color:green} trunk passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 1m
5s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m
3s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green} 1m
3s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m
42s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m
5s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m
0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 2m
6s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m
51s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:red}-1{color} | {color:red} unit {color} | {color:red}117m 26s{color}
| {color:red} hadoop-hdfs in the patch failed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m
21s{color} | {color:green} The patch does not generate ASF License warnings.
{color} |
| {color:black}{color} | {color:black} {color} | {color:black}152m 0s{color} |
{color:black} {color} |
\\
\\
|| Reason || Tests ||
| Failed junit tests |
hadoop.hdfs.server.datanode.TestDataNodeVolumeFailureReporting |
| | hadoop.hdfs.server.namenode.ha.TestEditLogTailer |
| | hadoop.hdfs.TestDFSUpgrade |
| | hadoop.hdfs.TestLeaseRecoveryStriped |
| | hadoop.hdfs.TestReconstructStripedFile |
| | hadoop.hdfs.server.namenode.TestNameNodeMetadataConsistency |
| | hadoop.hdfs.TestReadWhileWriting |
| | hadoop.hdfs.server.namenode.metrics.TestNameNodeMetrics |
| Timed out junit tests | org.apache.hadoop.hdfs.TestWriteReadStripedFile |
\\
\\
|| Subsystem || Report/Notes ||
| Docker | Image:yetus/hadoop:71bbb86 |
| JIRA Issue | HDFS-12378 |
| JIRA Patch URL |
https://issues.apache.org/jira/secure/attachment/12887182/HDFS-12378.01.patch |
| Optional Tests | asflicense compile javac javadoc mvninstall mvnsite
unit findbugs checkstyle |
| uname | Linux 5f8d457ecde9 3.13.0-116-generic #163-Ubuntu SMP Fri Mar 31
14:13:22 UTC 2017 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh
|
| git revision | trunk / 7ee02d1 |
| Default Java | 1.8.0_144 |
| findbugs | v3.1.0-RC1 |
| unit |
https://builds.apache.org/job/PreCommit-HDFS-Build/21142/artifact/patchprocess/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt
|
| Test Results |
https://builds.apache.org/job/PreCommit-HDFS-Build/21142/testReport/ |
| modules | C: hadoop-hdfs-project/hadoop-hdfs U:
hadoop-hdfs-project/hadoop-hdfs |
| Console output |
https://builds.apache.org/job/PreCommit-HDFS-Build/21142/console |
| Powered by | Apache Yetus 0.6.0-SNAPSHOT http://yetus.apache.org |
This message was automatically generated.
> TestClientProtocolForPipelineRecovery#testZeroByteBlockRecovery fails on trunk
> ------------------------------------------------------------------------------
>
> Key: HDFS-12378
> URL: https://issues.apache.org/jira/browse/HDFS-12378
> Project: Hadoop HDFS
> Issue Type: Bug
> Components: test
> Affects Versions: 3.0.0-alpha4
> Reporter: Xiao Chen
> Assignee: Lei (Eddy) Xu
> Priority: Blocker
> Labels: flaky-test
> Attachments: HDFS-12378.00.patch, HDFS-12378.01.patch
>
>
> Saw on
> https://builds.apache.org/job/PreCommit-HDFS-Build/20928/testReport/org.apache.hadoop.hdfs/TestClientProtocolForPipelineRecovery/testZeroByteBlockRecovery/:
> Error Message
> {noformat}
> Failed to replace a bad datanode on the existing pipeline due to no more good
> datanodes being available to try. (Nodes:
> current=[DatanodeInfoWithStorage[127.0.0.1:51925,DS-274e8cc9-280b-4370-b494-6a4f0d67ccf4,DISK]],
>
> original=[DatanodeInfoWithStorage[127.0.0.1:51925,DS-274e8cc9-280b-4370-b494-6a4f0d67ccf4,DISK]]).
> The current failed datanode replacement policy is ALWAYS, and a client may
> configure this via
> 'dfs.client.block.write.replace-datanode-on-failure.policy' in its
> configuration.
> {noformat}
> Stacktrace
> {noformat}
> java.io.IOException: Failed to replace a bad datanode on the existing
> pipeline due to no more good datanodes being available to try. (Nodes:
> current=[DatanodeInfoWithStorage[127.0.0.1:51925,DS-274e8cc9-280b-4370-b494-6a4f0d67ccf4,DISK]],
>
> original=[DatanodeInfoWithStorage[127.0.0.1:51925,DS-274e8cc9-280b-4370-b494-6a4f0d67ccf4,DISK]]).
> The current failed datanode replacement policy is ALWAYS, and a client may
> configure this via
> 'dfs.client.block.write.replace-datanode-on-failure.policy' in its
> configuration.
> at
> org.apache.hadoop.hdfs.DataStreamer.findNewDatanode(DataStreamer.java:1322)
> at
> org.apache.hadoop.hdfs.DataStreamer.addDatanode2ExistingPipeline(DataStreamer.java:1388)
> at
> org.apache.hadoop.hdfs.DataStreamer.handleDatanodeReplacement(DataStreamer.java:1587)
> at
> org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1488)
> at
> org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1470)
> at
> org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1274)
> at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:684)
> {noformat}
> Standard Output
> {noformat}
> 2017-08-30 18:02:37,714 [main] INFO hdfs.MiniDFSCluster
> (MiniDFSCluster.java:<init>(469)) - starting cluster: numNameNodes=1,
> numDataNodes=3
> Formatting using clusterid: testClusterID
> 2017-08-30 18:02:37,716 [main] INFO namenode.FSEditLog
> (FSEditLog.java:newInstance(224)) - Edit logging is async:false
> 2017-08-30 18:02:37,716 [main] INFO namenode.FSNamesystem
> (FSNamesystem.java:<init>(742)) - KeyProvider: null
> 2017-08-30 18:02:37,716 [main] INFO namenode.FSNamesystem
> (FSNamesystemLock.java:<init>(120)) - fsLock is fair: true
> 2017-08-30 18:02:37,716 [main] INFO namenode.FSNamesystem
> (FSNamesystemLock.java:<init>(136)) - Detailed lock hold time metrics
> enabled: false
> 2017-08-30 18:02:37,717 [main] INFO namenode.FSNamesystem
> (FSNamesystem.java:<init>(763)) - fsOwner = jenkins (auth:SIMPLE)
> 2017-08-30 18:02:37,717 [main] INFO namenode.FSNamesystem
> (FSNamesystem.java:<init>(764)) - supergroup = supergroup
> 2017-08-30 18:02:37,717 [main] INFO namenode.FSNamesystem
> (FSNamesystem.java:<init>(765)) - isPermissionEnabled = true
> 2017-08-30 18:02:37,717 [main] INFO namenode.FSNamesystem
> (FSNamesystem.java:<init>(776)) - HA Enabled: false
> 2017-08-30 18:02:37,718 [main] INFO common.Util
> (Util.java:isDiskStatsEnabled(395)) -
> dfs.datanode.fileio.profiling.sampling.percentage set to 0. Disabling file IO
> profiling
> 2017-08-30 18:02:37,718 [main] INFO blockmanagement.DatanodeManager
> (DatanodeManager.java:<init>(301)) - dfs.block.invalidate.limit:
> configured=1000, counted=60, effected=1000
> 2017-08-30 18:02:37,718 [main] INFO blockmanagement.DatanodeManager
> (DatanodeManager.java:<init>(309)) -
> dfs.namenode.datanode.registration.ip-hostname-check=true
> 2017-08-30 18:02:37,719 [main] INFO blockmanagement.BlockManager
> (InvalidateBlocks.java:printBlockDeletionTime(76)) -
> dfs.namenode.startup.delay.block.deletion.sec is set to 000:00:00:00.000
> 2017-08-30 18:02:37,719 [main] INFO blockmanagement.BlockManager
> (InvalidateBlocks.java:printBlockDeletionTime(82)) - The block deletion will
> start around 2017 Aug 30 18:02:37
> 2017-08-30 18:02:37,719 [main] INFO util.GSet
> (LightWeightGSet.java:computeCapacity(395)) - Computing capacity for map
> BlocksMap
> 2017-08-30 18:02:37,719 [main] INFO util.GSet
> (LightWeightGSet.java:computeCapacity(396)) - VM type = 64-bit
> 2017-08-30 18:02:37,720 [main] INFO util.GSet
> (LightWeightGSet.java:computeCapacity(397)) - 2.0% max memory 1.8 GB = 36.4 MB
> 2017-08-30 18:02:37,720 [main] INFO util.GSet
> (LightWeightGSet.java:computeCapacity(402)) - capacity = 2^22 = 4194304
> entries
> 2017-08-30 18:02:37,726 [main] INFO blockmanagement.BlockManager
> (BlockManager.java:createBlockTokenSecretManager(560)) -
> dfs.block.access.token.enable = false
> 2017-08-30 18:02:37,726 [main] INFO Configuration.deprecation
> (Configuration.java:logDeprecation(1285)) - No unit for
> dfs.namenode.safemode.extension(0) assuming MILLISECONDS
> 2017-08-30 18:02:37,727 [main] INFO blockmanagement.BlockManagerSafeMode
> (BlockManagerSafeMode.java:<init>(161)) - dfs.namenode.safemode.threshold-pct
> = 0.9990000128746033
> 2017-08-30 18:02:37,727 [main] INFO blockmanagement.BlockManagerSafeMode
> (BlockManagerSafeMode.java:<init>(162)) - dfs.namenode.safemode.min.datanodes
> = 0
> 2017-08-30 18:02:37,727 [main] INFO blockmanagement.BlockManagerSafeMode
> (BlockManagerSafeMode.java:<init>(164)) - dfs.namenode.safemode.extension = 0
> 2017-08-30 18:02:37,727 [main] INFO blockmanagement.BlockManager
> (BlockManager.java:<init>(546)) - defaultReplication = 3
> 2017-08-30 18:02:37,727 [main] INFO blockmanagement.BlockManager
> (BlockManager.java:<init>(547)) - maxReplication = 512
> 2017-08-30 18:02:37,728 [main] INFO blockmanagement.BlockManager
> (BlockManager.java:<init>(548)) - minReplication = 1
> 2017-08-30 18:02:37,728 [main] INFO blockmanagement.BlockManager
> (BlockManager.java:<init>(549)) - maxReplicationStreams = 2
> 2017-08-30 18:02:37,728 [main] INFO blockmanagement.BlockManager
> (BlockManager.java:<init>(550)) - redundancyRecheckInterval = 3000ms
> 2017-08-30 18:02:37,728 [main] INFO blockmanagement.BlockManager
> (BlockManager.java:<init>(551)) - encryptDataTransfer = false
> 2017-08-30 18:02:37,728 [main] INFO blockmanagement.BlockManager
> (BlockManager.java:<init>(552)) - maxNumBlocksToLog = 1000
> 2017-08-30 18:02:37,729 [main] INFO util.GSet
> (LightWeightGSet.java:computeCapacity(395)) - Computing capacity for map
> INodeMap
> 2017-08-30 18:02:37,729 [main] INFO util.GSet
> (LightWeightGSet.java:computeCapacity(396)) - VM type = 64-bit
> 2017-08-30 18:02:37,729 [main] INFO util.GSet
> (LightWeightGSet.java:computeCapacity(397)) - 1.0% max memory 1.8 GB = 18.2 MB
> 2017-08-30 18:02:37,730 [main] INFO util.GSet
> (LightWeightGSet.java:computeCapacity(402)) - capacity = 2^21 = 2097152
> entries
> 2017-08-30 18:02:37,734 [main] INFO namenode.FSDirectory
> (FSDirectory.java:<init>(278)) - ACLs enabled? false
> 2017-08-30 18:02:37,734 [main] INFO namenode.FSDirectory
> (FSDirectory.java:<init>(282)) - POSIX ACL inheritance enabled? true
> 2017-08-30 18:02:37,734 [main] INFO namenode.FSDirectory
> (FSDirectory.java:<init>(286)) - XAttrs enabled? true
> 2017-08-30 18:02:37,734 [main] INFO namenode.NameNode
> (FSDirectory.java:<init>(350)) - Caching file names occurring more than 10
> times
> 2017-08-30 18:02:37,734 [main] INFO snapshot.SnapshotManager
> (SnapshotManager.java:<init>(105)) - Loaded config captureOpenFiles:
> falseskipCaptureAccessTimeOnlyChange: false
> 2017-08-30 18:02:37,734 [main] INFO util.GSet
> (LightWeightGSet.java:computeCapacity(395)) - Computing capacity for map
> cachedBlocks
> 2017-08-30 18:02:37,734 [main] INFO util.GSet
> (LightWeightGSet.java:computeCapacity(396)) - VM type = 64-bit
> 2017-08-30 18:02:37,735 [main] INFO util.GSet
> (LightWeightGSet.java:computeCapacity(397)) - 0.25% max memory 1.8 GB = 4.6 MB
> 2017-08-30 18:02:37,735 [main] INFO util.GSet
> (LightWeightGSet.java:computeCapacity(402)) - capacity = 2^19 = 524288
> entries
> 2017-08-30 18:02:37,736 [main] INFO metrics.TopMetrics
> (TopMetrics.java:logConf(76)) - NNTop conf:
> dfs.namenode.top.window.num.buckets = 10
> 2017-08-30 18:02:37,736 [main] INFO metrics.TopMetrics
> (TopMetrics.java:logConf(78)) - NNTop conf: dfs.namenode.top.num.users = 10
> 2017-08-30 18:02:37,736 [main] INFO metrics.TopMetrics
> (TopMetrics.java:logConf(80)) - NNTop conf: dfs.namenode.top.windows.minutes
> = 1,5,25
> 2017-08-30 18:02:37,736 [main] INFO namenode.FSNamesystem
> (FSNamesystem.java:initRetryCache(968)) - Retry cache on namenode is enabled
> 2017-08-30 18:02:37,737 [main] INFO namenode.FSNamesystem
> (FSNamesystem.java:initRetryCache(976)) - Retry cache will use 0.03 of total
> heap and retry cache entry expiry time is 600000 millis
> 2017-08-30 18:02:37,737 [main] INFO util.GSet
> (LightWeightGSet.java:computeCapacity(395)) - Computing capacity for map
> NameNodeRetryCache
> 2017-08-30 18:02:37,737 [main] INFO util.GSet
> (LightWeightGSet.java:computeCapacity(396)) - VM type = 64-bit
> 2017-08-30 18:02:37,737 [main] INFO util.GSet
> (LightWeightGSet.java:computeCapacity(397)) - 0.029999999329447746% max
> memory 1.8 GB = 559.3 KB
> 2017-08-30 18:02:37,737 [main] INFO util.GSet
> (LightWeightGSet.java:computeCapacity(402)) - capacity = 2^16 = 65536
> entries
> 2017-08-30 18:02:37,738 [main] INFO namenode.FSImage
> (FSImage.java:format(164)) - Allocated new BlockPoolId:
> BP-1775344991-172.17.0.2-1504116157738
> 2017-08-30 18:02:37,740 [main] INFO common.Storage
> (NNStorage.java:format(583)) - Storage directory
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/name-0-1
> has been successfully formatted.
> 2017-08-30 18:02:37,741 [main] INFO common.Storage
> (NNStorage.java:format(583)) - Storage directory
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/name-0-2
> has been successfully formatted.
> 2017-08-30 18:02:37,742 [FSImageSaver for
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/name-0-1
> of type IMAGE_AND_EDITS] INFO namenode.FSImageFormatProtobuf
> (FSImageFormatProtobuf.java:save(429)) - Saving image file
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/name-0-1/current/fsimage.ckpt_0000000000000000000
> using no compression
> 2017-08-30 18:02:37,742 [FSImageSaver for
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/name-0-2
> of type IMAGE_AND_EDITS] INFO namenode.FSImageFormatProtobuf
> (FSImageFormatProtobuf.java:save(429)) - Saving image file
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/name-0-2/current/fsimage.ckpt_0000000000000000000
> using no compression
> 2017-08-30 18:02:37,748 [FSImageSaver for
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/name-0-2
> of type IMAGE_AND_EDITS] INFO namenode.FSImageFormatProtobuf
> (FSImageFormatProtobuf.java:save(432)) - Image file
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/name-0-2/current/fsimage.ckpt_0000000000000000000
> of size 335 bytes saved in 0 seconds.
> 2017-08-30 18:02:37,748 [FSImageSaver for
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/name-0-1
> of type IMAGE_AND_EDITS] INFO namenode.FSImageFormatProtobuf
> (FSImageFormatProtobuf.java:save(432)) - Image file
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/name-0-1/current/fsimage.ckpt_0000000000000000000
> of size 335 bytes saved in 0 seconds.
> 2017-08-30 18:02:37,750 [main] INFO namenode.NNStorageRetentionManager
> (NNStorageRetentionManager.java:getImageTxIdToRetain(203)) - Going to retain
> 1 images with txid >= 0
> 2017-08-30 18:02:37,751 [main] INFO namenode.NameNode
> (NameNode.java:createNameNode(1577)) - createNameNode []
> 2017-08-30 18:02:37,753 [main] INFO impl.MetricsConfig
> (MetricsConfig.java:loadFirst(121)) - loaded properties from
> hadoop-metrics2-namenode.properties
> 2017-08-30 18:02:37,754 [main] INFO impl.MetricsSystemImpl
> (MetricsSystemImpl.java:startTimer(374)) - Scheduled Metric snapshot period
> at 1 second(s).
> 2017-08-30 18:02:37,754 [main] INFO impl.MetricsSystemImpl
> (MetricsSystemImpl.java:start(191)) - NameNode metrics system started
> 2017-08-30 18:02:37,755 [main] INFO namenode.NameNode
> (NameNode.java:setClientNamenodeAddress(464)) - fs.defaultFS is
> hdfs://127.0.0.1:0
> 2017-08-30 18:02:37,757
> [org.apache.hadoop.util.JvmPauseMonitor$Monitor@65f2f9b0] INFO
> util.JvmPauseMonitor (JvmPauseMonitor.java:run(188)) - Starting JVM pause
> monitor
> 2017-08-30 18:02:37,758 [main] INFO hdfs.DFSUtil
> (DFSUtil.java:httpServerTemplateForNNAndJN(1512)) - Starting Web-server for
> hdfs at: http://localhost:0
> 2017-08-30 18:02:37,759 [main] INFO server.AuthenticationFilter
> (AuthenticationFilter.java:constructSecretProvider(240)) - Unable to
> initialize FileSignerSecretProvider, falling back to use random secrets.
> 2017-08-30 18:02:37,760 [main] INFO http.HttpRequestLog
> (HttpRequestLog.java:getRequestLog(81)) - Http request log for
> http.requests.namenode is not defined
> 2017-08-30 18:02:37,760 [main] INFO http.HttpServer2
> (HttpServer2.java:addGlobalFilter(921)) - Added global filter 'safety'
> (class=org.apache.hadoop.http.HttpServer2$QuotingInputFilter)
> 2017-08-30 18:02:37,761 [main] INFO http.HttpServer2
> (HttpServer2.java:addFilter(894)) - Added filter static_user_filter
> (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to
> context hdfs
> 2017-08-30 18:02:37,761 [main] INFO http.HttpServer2
> (HttpServer2.java:addFilter(904)) - Added filter static_user_filter
> (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to
> context static
> 2017-08-30 18:02:37,761 [main] INFO http.HttpServer2
> (HttpServer2.java:addFilter(904)) - Added filter static_user_filter
> (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to
> context logs
> 2017-08-30 18:02:37,762 [main] INFO http.HttpServer2
> (NameNodeHttpServer.java:initWebHdfs(98)) - Added filter
> 'org.apache.hadoop.hdfs.web.AuthFilter'
> (class=org.apache.hadoop.hdfs.web.AuthFilter)
> 2017-08-30 18:02:37,762 [main] INFO http.HttpServer2
> (HttpServer2.java:addJerseyResourcePackage(779)) - addJerseyResourcePackage:
> packageName=org.apache.hadoop.hdfs.server.namenode.web.resources;org.apache.hadoop.hdfs.web.resources,
> pathSpec=/webhdfs/v1/*
> 2017-08-30 18:02:37,762 [main] INFO http.HttpServer2
> (HttpServer2.java:bindListener(1121)) - Jetty bound to port 35619
> 2017-08-30 18:02:37,763 [main] INFO server.Server (Server.java:doStart(346))
> - jetty-9.3.19.v20170502
> 2017-08-30 18:02:37,764 [main] INFO handler.ContextHandler
> (ContextHandler.java:doStart(781)) - Started
> o.e.j.s.ServletContextHandler@46c3a14d{/logs,file:///testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/log/,AVAILABLE}
> 2017-08-30 18:02:37,765 [main] INFO handler.ContextHandler
> (ContextHandler.java:doStart(781)) - Started
> o.e.j.s.ServletContextHandler@5e8cda75{/static,file:///testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test-classes/webapps/static/,AVAILABLE}
> 2017-08-30 18:02:37,771 [main] INFO handler.ContextHandler
> (ContextHandler.java:doStart(781)) - Started
> o.e.j.w.WebAppContext@5633dafd{/,file:///testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test-classes/webapps/hdfs/,AVAILABLE}{/hdfs}
> 2017-08-30 18:02:37,772 [main] INFO server.AbstractConnector
> (AbstractConnector.java:doStart(278)) - Started
> ServerConnector@5d5160e6{HTTP/1.1,[http/1.1]}{localhost:35619}
> 2017-08-30 18:02:37,772 [main] INFO server.Server (Server.java:doStart(414))
> - Started @37288ms
> 2017-08-30 18:02:37,775 [main] INFO namenode.FSEditLog
> (FSEditLog.java:newInstance(224)) - Edit logging is async:false
> 2017-08-30 18:02:37,775 [main] INFO namenode.FSNamesystem
> (FSNamesystem.java:<init>(742)) - KeyProvider: null
> 2017-08-30 18:02:37,775 [main] INFO namenode.FSNamesystem
> (FSNamesystemLock.java:<init>(120)) - fsLock is fair: true
> 2017-08-30 18:02:37,775 [main] INFO namenode.FSNamesystem
> (FSNamesystemLock.java:<init>(136)) - Detailed lock hold time metrics
> enabled: false
> 2017-08-30 18:02:37,776 [main] INFO namenode.FSNamesystem
> (FSNamesystem.java:<init>(763)) - fsOwner = jenkins (auth:SIMPLE)
> 2017-08-30 18:02:37,776 [main] INFO namenode.FSNamesystem
> (FSNamesystem.java:<init>(764)) - supergroup = supergroup
> 2017-08-30 18:02:37,776 [main] INFO namenode.FSNamesystem
> (FSNamesystem.java:<init>(765)) - isPermissionEnabled = true
> 2017-08-30 18:02:37,776 [main] INFO namenode.FSNamesystem
> (FSNamesystem.java:<init>(776)) - HA Enabled: false
> 2017-08-30 18:02:37,777 [main] INFO common.Util
> (Util.java:isDiskStatsEnabled(395)) -
> dfs.datanode.fileio.profiling.sampling.percentage set to 0. Disabling file IO
> profiling
> 2017-08-30 18:02:37,777 [main] INFO blockmanagement.DatanodeManager
> (DatanodeManager.java:<init>(301)) - dfs.block.invalidate.limit:
> configured=1000, counted=60, effected=1000
> 2017-08-30 18:02:37,777 [main] INFO blockmanagement.DatanodeManager
> (DatanodeManager.java:<init>(309)) -
> dfs.namenode.datanode.registration.ip-hostname-check=true
> 2017-08-30 18:02:37,777 [main] INFO blockmanagement.BlockManager
> (InvalidateBlocks.java:printBlockDeletionTime(76)) -
> dfs.namenode.startup.delay.block.deletion.sec is set to 000:00:00:00.000
> 2017-08-30 18:02:37,778 [main] INFO blockmanagement.BlockManager
> (InvalidateBlocks.java:printBlockDeletionTime(82)) - The block deletion will
> start around 2017 Aug 30 18:02:37
> 2017-08-30 18:02:37,778 [main] INFO util.GSet
> (LightWeightGSet.java:computeCapacity(395)) - Computing capacity for map
> BlocksMap
> 2017-08-30 18:02:37,778 [main] INFO util.GSet
> (LightWeightGSet.java:computeCapacity(396)) - VM type = 64-bit
> 2017-08-30 18:02:37,778 [main] INFO util.GSet
> (LightWeightGSet.java:computeCapacity(397)) - 2.0% max memory 1.8 GB = 36.4 MB
> 2017-08-30 18:02:37,779 [main] INFO util.GSet
> (LightWeightGSet.java:computeCapacity(402)) - capacity = 2^22 = 4194304
> entries
> 2017-08-30 18:02:37,784 [main] INFO blockmanagement.BlockManager
> (BlockManager.java:createBlockTokenSecretManager(560)) -
> dfs.block.access.token.enable = false
> 2017-08-30 18:02:37,785 [main] INFO Configuration.deprecation
> (Configuration.java:logDeprecation(1285)) - No unit for
> dfs.namenode.safemode.extension(0) assuming MILLISECONDS
> 2017-08-30 18:02:37,785 [main] INFO blockmanagement.BlockManagerSafeMode
> (BlockManagerSafeMode.java:<init>(161)) - dfs.namenode.safemode.threshold-pct
> = 0.9990000128746033
> 2017-08-30 18:02:37,785 [main] INFO blockmanagement.BlockManagerSafeMode
> (BlockManagerSafeMode.java:<init>(162)) - dfs.namenode.safemode.min.datanodes
> = 0
> 2017-08-30 18:02:37,785 [main] INFO blockmanagement.BlockManagerSafeMode
> (BlockManagerSafeMode.java:<init>(164)) - dfs.namenode.safemode.extension = 0
> 2017-08-30 18:02:37,785 [main] INFO blockmanagement.BlockManager
> (BlockManager.java:<init>(546)) - defaultReplication = 3
> 2017-08-30 18:02:37,785 [main] INFO blockmanagement.BlockManager
> (BlockManager.java:<init>(547)) - maxReplication = 512
> 2017-08-30 18:02:37,786 [main] INFO blockmanagement.BlockManager
> (BlockManager.java:<init>(548)) - minReplication = 1
> 2017-08-30 18:02:37,786 [main] INFO blockmanagement.BlockManager
> (BlockManager.java:<init>(549)) - maxReplicationStreams = 2
> 2017-08-30 18:02:37,786 [main] INFO blockmanagement.BlockManager
> (BlockManager.java:<init>(550)) - redundancyRecheckInterval = 3000ms
> 2017-08-30 18:02:37,786 [main] INFO blockmanagement.BlockManager
> (BlockManager.java:<init>(551)) - encryptDataTransfer = false
> 2017-08-30 18:02:37,786 [main] INFO blockmanagement.BlockManager
> (BlockManager.java:<init>(552)) - maxNumBlocksToLog = 1000
> 2017-08-30 18:02:37,786 [main] INFO util.GSet
> (LightWeightGSet.java:computeCapacity(395)) - Computing capacity for map
> INodeMap
> 2017-08-30 18:02:37,787 [main] INFO util.GSet
> (LightWeightGSet.java:computeCapacity(396)) - VM type = 64-bit
> 2017-08-30 18:02:37,787 [main] INFO util.GSet
> (LightWeightGSet.java:computeCapacity(397)) - 1.0% max memory 1.8 GB = 18.2 MB
> 2017-08-30 18:02:37,787 [main] INFO util.GSet
> (LightWeightGSet.java:computeCapacity(402)) - capacity = 2^21 = 2097152
> entries
> 2017-08-30 18:02:37,790 [main] INFO namenode.FSDirectory
> (FSDirectory.java:<init>(278)) - ACLs enabled? false
> 2017-08-30 18:02:37,791 [main] INFO namenode.FSDirectory
> (FSDirectory.java:<init>(282)) - POSIX ACL inheritance enabled? true
> 2017-08-30 18:02:37,791 [main] INFO namenode.FSDirectory
> (FSDirectory.java:<init>(286)) - XAttrs enabled? true
> 2017-08-30 18:02:37,791 [main] INFO namenode.NameNode
> (FSDirectory.java:<init>(350)) - Caching file names occurring more than 10
> times
> 2017-08-30 18:02:37,791 [main] INFO snapshot.SnapshotManager
> (SnapshotManager.java:<init>(105)) - Loaded config captureOpenFiles:
> falseskipCaptureAccessTimeOnlyChange: false
> 2017-08-30 18:02:37,791 [main] INFO util.GSet
> (LightWeightGSet.java:computeCapacity(395)) - Computing capacity for map
> cachedBlocks
> 2017-08-30 18:02:37,791 [main] INFO util.GSet
> (LightWeightGSet.java:computeCapacity(396)) - VM type = 64-bit
> 2017-08-30 18:02:37,792 [main] INFO util.GSet
> (LightWeightGSet.java:computeCapacity(397)) - 0.25% max memory 1.8 GB = 4.6 MB
> 2017-08-30 18:02:37,792 [main] INFO util.GSet
> (LightWeightGSet.java:computeCapacity(402)) - capacity = 2^19 = 524288
> entries
> 2017-08-30 18:02:37,793 [main] INFO metrics.TopMetrics
> (TopMetrics.java:logConf(76)) - NNTop conf:
> dfs.namenode.top.window.num.buckets = 10
> 2017-08-30 18:02:37,793 [main] INFO metrics.TopMetrics
> (TopMetrics.java:logConf(78)) - NNTop conf: dfs.namenode.top.num.users = 10
> 2017-08-30 18:02:37,793 [main] INFO metrics.TopMetrics
> (TopMetrics.java:logConf(80)) - NNTop conf: dfs.namenode.top.windows.minutes
> = 1,5,25
> 2017-08-30 18:02:37,793 [main] INFO namenode.FSNamesystem
> (FSNamesystem.java:initRetryCache(968)) - Retry cache on namenode is enabled
> 2017-08-30 18:02:37,793 [main] INFO namenode.FSNamesystem
> (FSNamesystem.java:initRetryCache(976)) - Retry cache will use 0.03 of total
> heap and retry cache entry expiry time is 600000 millis
> 2017-08-30 18:02:37,793 [main] INFO util.GSet
> (LightWeightGSet.java:computeCapacity(395)) - Computing capacity for map
> NameNodeRetryCache
> 2017-08-30 18:02:37,794 [main] INFO util.GSet
> (LightWeightGSet.java:computeCapacity(396)) - VM type = 64-bit
> 2017-08-30 18:02:37,794 [main] INFO util.GSet
> (LightWeightGSet.java:computeCapacity(397)) - 0.029999999329447746% max
> memory 1.8 GB = 559.3 KB
> 2017-08-30 18:02:37,794 [main] INFO util.GSet
> (LightWeightGSet.java:computeCapacity(402)) - capacity = 2^16 = 65536
> entries
> 2017-08-30 18:02:37,831 [main] INFO common.Storage
> (Storage.java:tryLock(847)) - Lock on
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/name-0-1/in_use.lock
> acquired by nodename 26462@0c74973bd755
> 2017-08-30 18:02:37,833 [main] INFO common.Storage
> (Storage.java:tryLock(847)) - Lock on
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/name-0-2/in_use.lock
> acquired by nodename 26462@0c74973bd755
> 2017-08-30 18:02:37,834 [main] INFO namenode.FileJournalManager
> (FileJournalManager.java:recoverUnfinalizedSegments(396)) - Recovering
> unfinalized segments in
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/name-0-1/current
> 2017-08-30 18:02:37,834 [main] INFO namenode.FileJournalManager
> (FileJournalManager.java:recoverUnfinalizedSegments(396)) - Recovering
> unfinalized segments in
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/name-0-2/current
> 2017-08-30 18:02:37,834 [main] INFO namenode.FSImage
> (FSImage.java:loadFSImage(712)) - No edit log streams selected.
> 2017-08-30 18:02:37,835 [main] INFO namenode.FSImage
> (FSImage.java:loadFSImageFile(775)) - Planning to load image:
> FSImageFile(file=/testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/name-0-1/current/fsimage_0000000000000000000,
> cpktTxId=0000000000000000000)
> 2017-08-30 18:02:37,835 [main] INFO namenode.FSImageFormatPBINode
> (FSImageFormatPBINode.java:loadINodeSection(266)) - Loading 1 INodes.
> 2017-08-30 18:02:37,836 [main] INFO namenode.FSImageFormatProtobuf
> (FSImageFormatProtobuf.java:load(185)) - Loaded FSImage in 0 seconds.
> 2017-08-30 18:02:37,836 [main] INFO namenode.FSImage
> (FSImage.java:loadFSImage(941)) - Loaded image for txid 0 from
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/name-0-1/current/fsimage_0000000000000000000
> 2017-08-30 18:02:37,836 [main] INFO namenode.FSNamesystem
> (FSNamesystem.java:loadFSImage(1080)) - Need to save fs image? false
> (staleImage=false, haEnabled=false, isRollingUpgrade=false)
> 2017-08-30 18:02:37,836 [main] INFO namenode.FSEditLog
> (FSEditLog.java:startLogSegment(1324)) - Starting log segment at 1
> 2017-08-30 18:02:37,841 [main] INFO namenode.NameCache
> (NameCache.java:initialized(143)) - initialized with 0 entries 0 lookups
> 2017-08-30 18:02:37,841 [main] INFO namenode.FSNamesystem
> (FSNamesystem.java:loadFromDisk(714)) - Finished loading FSImage in 46 msecs
> 2017-08-30 18:02:37,841 [main] INFO namenode.NameNode
> (NameNodeRpcServer.java:<init>(439)) - RPC server is binding to localhost:0
> 2017-08-30 18:02:37,842 [main] INFO ipc.CallQueueManager
> (CallQueueManager.java:<init>(84)) - Using callQueue: class
> java.util.concurrent.LinkedBlockingQueue queueCapacity: 1000 scheduler: class
> org.apache.hadoop.ipc.DefaultRpcScheduler
> 2017-08-30 18:02:37,842 [Socket Reader #1 for port 47551] INFO ipc.Server
> (Server.java:run(1067)) - Starting Socket Reader #1 for port 47551
> 2017-08-30 18:02:37,847 [main] INFO namenode.NameNode
> (NameNode.java:initialize(737)) - Clients are to use localhost:47551 to
> access this namenode/service.
> 2017-08-30 18:02:37,847 [main] INFO namenode.FSNamesystem
> (FSNamesystem.java:registerMBean(4811)) - Registered FSNamesystemState,
> ReplicatedBlocksState and ECBlockGroupsState MBeans.
> 2017-08-30 18:02:37,858 [main] INFO namenode.LeaseManager
> (LeaseManager.java:getNumUnderConstructionBlocks(158)) - Number of blocks
> under construction: 0
> 2017-08-30 18:02:37,859 [main] INFO blockmanagement.BlockManager
> (BlockManager.java:initializeReplQueues(4650)) - initializing replication
> queues
> 2017-08-30 18:02:37,860 [main] INFO hdfs.StateChange
> (BlockManagerSafeMode.java:leaveSafeMode(396)) - STATE* Leaving safe mode
> after 0 secs
> 2017-08-30 18:02:37,860 [main] INFO hdfs.StateChange
> (BlockManagerSafeMode.java:leaveSafeMode(402)) - STATE* Network topology has
> 0 racks and 0 datanodes
> 2017-08-30 18:02:37,860 [main] INFO hdfs.StateChange
> (BlockManagerSafeMode.java:leaveSafeMode(404)) - STATE* UnderReplicatedBlocks
> has 0 blocks
> 2017-08-30 18:02:37,866 [Reconstruction Queue Initializer] INFO
> blockmanagement.BlockManager
> (BlockManager.java:processMisReplicatesAsync(3355)) - Total number of blocks
> = 0
> 2017-08-30 18:02:37,867 [Reconstruction Queue Initializer] INFO
> blockmanagement.BlockManager
> (BlockManager.java:processMisReplicatesAsync(3356)) - Number of invalid
> blocks = 0
> 2017-08-30 18:02:37,867 [Reconstruction Queue Initializer] INFO
> blockmanagement.BlockManager
> (BlockManager.java:processMisReplicatesAsync(3357)) - Number of
> under-replicated blocks = 0
> 2017-08-30 18:02:37,867 [Reconstruction Queue Initializer] INFO
> blockmanagement.BlockManager
> (BlockManager.java:processMisReplicatesAsync(3358)) - Number of
> over-replicated blocks = 0
> 2017-08-30 18:02:37,867 [Reconstruction Queue Initializer] INFO
> blockmanagement.BlockManager
> (BlockManager.java:processMisReplicatesAsync(3360)) - Number of blocks being
> written = 0
> 2017-08-30 18:02:37,867 [Reconstruction Queue Initializer] INFO
> hdfs.StateChange (BlockManager.java:processMisReplicatesAsync(3363)) - STATE*
> Replication Queue initialization scan for invalid, over- and under-replicated
> blocks completed in 8 msec
> 2017-08-30 18:02:37,869 [IPC Server Responder] INFO ipc.Server
> (Server.java:run(1307)) - IPC Server Responder: starting
> 2017-08-30 18:02:37,869 [IPC Server listener on 47551] INFO ipc.Server
> (Server.java:run(1146)) - IPC Server listener on 47551: starting
> 2017-08-30 18:02:37,872 [main] INFO namenode.NameNode
> (NameNode.java:startCommonServices(830)) - NameNode RPC up at:
> localhost/127.0.0.1:47551
> 2017-08-30 18:02:37,873 [main] INFO namenode.FSNamesystem
> (FSNamesystem.java:startActiveServices(1192)) - Starting services required
> for active state
> 2017-08-30 18:02:37,873 [main] INFO namenode.FSDirectory
> (FSDirectory.java:updateCountForQuota(706)) - Initializing quota with 4
> thread(s)
> 2017-08-30 18:02:37,873 [main] INFO namenode.FSDirectory
> (FSDirectory.java:updateCountForQuota(715)) - Quota initialization completed
> in 0 milliseconds
> name space=1
> storage space=0
> storage types=RAM_DISK=0, SSD=0, DISK=0, ARCHIVE=0
> 2017-08-30 18:02:37,875 [CacheReplicationMonitor(1864849826)] INFO
> blockmanagement.CacheReplicationMonitor
> (CacheReplicationMonitor.java:run(160)) - Starting CacheReplicationMonitor
> with interval 30000 milliseconds
> 2017-08-30 18:02:37,879 [main] INFO hdfs.MiniDFSCluster
> (MiniDFSCluster.java:startDataNodes(1591)) - Starting DataNode 0 with
> dfs.datanode.data.dir:
> [DISK]file:/testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data1,[DISK]file:/testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data2
> 2017-08-30 18:02:37,879 [main] INFO checker.ThrottledAsyncChecker
> (ThrottledAsyncChecker.java:schedule(122)) - Scheduling a check for
> [DISK]file:/testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data1
> 2017-08-30 18:02:37,880 [main] INFO checker.ThrottledAsyncChecker
> (ThrottledAsyncChecker.java:schedule(122)) - Scheduling a check for
> [DISK]file:/testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data2
> 2017-08-30 18:02:37,887 [main] INFO impl.MetricsSystemImpl
> (MetricsSystemImpl.java:init(158)) - DataNode metrics system started (again)
> 2017-08-30 18:02:37,887 [main] INFO common.Util
> (Util.java:isDiskStatsEnabled(395)) -
> dfs.datanode.fileio.profiling.sampling.percentage set to 0. Disabling file IO
> profiling
> 2017-08-30 18:02:37,887 [main] INFO datanode.BlockScanner
> (BlockScanner.java:<init>(184)) - Initialized block scanner with
> targetBytesPerSec 1048576
> 2017-08-30 18:02:37,888 [main] INFO datanode.DataNode
> (DataNode.java:<init>(496)) - Configured hostname is 127.0.0.1
> 2017-08-30 18:02:37,888 [main] INFO common.Util
> (Util.java:isDiskStatsEnabled(395)) -
> dfs.datanode.fileio.profiling.sampling.percentage set to 0. Disabling file IO
> profiling
> 2017-08-30 18:02:37,888 [main] INFO datanode.DataNode
> (DataNode.java:startDataNode(1385)) - Starting DataNode with maxLockedMemory
> = 0
> 2017-08-30 18:02:37,888 [main] INFO datanode.DataNode
> (DataNode.java:initDataXceiver(1142)) - Opened streaming server at
> /127.0.0.1:46794
> 2017-08-30 18:02:37,889 [main] INFO datanode.DataNode
> (DataXceiverServer.java:<init>(78)) - Balancing bandwidth is 10485760 bytes/s
> 2017-08-30 18:02:37,889 [main] INFO datanode.DataNode
> (DataXceiverServer.java:<init>(79)) - Number threads for balancing is 50
> 2017-08-30 18:02:37,892 [main] INFO server.AuthenticationFilter
> (AuthenticationFilter.java:constructSecretProvider(240)) - Unable to
> initialize FileSignerSecretProvider, falling back to use random secrets.
> 2017-08-30 18:02:37,892 [main] INFO http.HttpRequestLog
> (HttpRequestLog.java:getRequestLog(81)) - Http request log for
> http.requests.datanode is not defined
> 2017-08-30 18:02:37,893 [main] INFO http.HttpServer2
> (HttpServer2.java:addGlobalFilter(921)) - Added global filter 'safety'
> (class=org.apache.hadoop.http.HttpServer2$QuotingInputFilter)
> 2017-08-30 18:02:37,893 [main] INFO http.HttpServer2
> (HttpServer2.java:addFilter(894)) - Added filter static_user_filter
> (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to
> context datanode
> 2017-08-30 18:02:37,894 [main] INFO http.HttpServer2
> (HttpServer2.java:addFilter(904)) - Added filter static_user_filter
> (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to
> context static
> 2017-08-30 18:02:37,894 [main] INFO http.HttpServer2
> (HttpServer2.java:addFilter(904)) - Added filter static_user_filter
> (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to
> context logs
> 2017-08-30 18:02:37,894 [main] INFO http.HttpServer2
> (HttpServer2.java:bindListener(1121)) - Jetty bound to port 46189
> 2017-08-30 18:02:37,895 [main] INFO server.Server (Server.java:doStart(346))
> - jetty-9.3.19.v20170502
> 2017-08-30 18:02:37,897 [main] INFO handler.ContextHandler
> (ContextHandler.java:doStart(781)) - Started
> o.e.j.s.ServletContextHandler@4e1ce44{/logs,file:///testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/log/,AVAILABLE}
> 2017-08-30 18:02:37,898 [main] INFO handler.ContextHandler
> (ContextHandler.java:doStart(781)) - Started
> o.e.j.s.ServletContextHandler@7a7cc52c{/static,file:///testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test-classes/webapps/static/,AVAILABLE}
> 2017-08-30 18:02:37,903 [main] INFO handler.ContextHandler
> (ContextHandler.java:doStart(781)) - Started
> o.e.j.w.WebAppContext@29eda4f8{/,file:///testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test-classes/webapps/datanode/,AVAILABLE}{/datanode}
> 2017-08-30 18:02:37,904 [main] INFO server.AbstractConnector
> (AbstractConnector.java:doStart(278)) - Started
> ServerConnector@1b9776f5{HTTP/1.1,[http/1.1]}{localhost:46189}
> 2017-08-30 18:02:37,905 [main] INFO server.Server (Server.java:doStart(414))
> - Started @37420ms
> 2017-08-30 18:02:37,911 [main] INFO web.DatanodeHttpServer
> (DatanodeHttpServer.java:start(239)) - Listening HTTP traffic on
> /127.0.0.1:52625
> 2017-08-30 18:02:37,911
> [org.apache.hadoop.util.JvmPauseMonitor$Monitor@79d9214d] INFO
> util.JvmPauseMonitor (JvmPauseMonitor.java:run(188)) - Starting JVM pause
> monitor
> 2017-08-30 18:02:37,911 [main] INFO datanode.DataNode
> (DataNode.java:startDataNode(1412)) - dnUserName = jenkins
> 2017-08-30 18:02:37,911 [main] INFO datanode.DataNode
> (DataNode.java:startDataNode(1413)) - supergroup = supergroup
> 2017-08-30 18:02:37,912 [main] INFO ipc.CallQueueManager
> (CallQueueManager.java:<init>(84)) - Using callQueue: class
> java.util.concurrent.LinkedBlockingQueue queueCapacity: 1000 scheduler: class
> org.apache.hadoop.ipc.DefaultRpcScheduler
> 2017-08-30 18:02:37,912 [Socket Reader #1 for port 58248] INFO ipc.Server
> (Server.java:run(1067)) - Starting Socket Reader #1 for port 58248
> 2017-08-30 18:02:37,915 [main] INFO datanode.DataNode
> (DataNode.java:initIpcServer(1029)) - Opened IPC server at /127.0.0.1:58248
> 2017-08-30 18:02:37,916 [main] INFO datanode.DataNode
> (BlockPoolManager.java:refreshNamenodes(149)) - Refresh request received for
> nameservices: null
> 2017-08-30 18:02:37,917 [main] INFO datanode.DataNode
> (BlockPoolManager.java:doRefreshNamenodes(198)) - Starting BPOfferServices
> for nameservices: <default>
> 2017-08-30 18:02:37,917 [Thread-997] INFO datanode.DataNode
> (BPServiceActor.java:run(809)) - Block pool <registering> (Datanode Uuid
> unassigned) service to localhost/127.0.0.1:47551 starting to offer service
> 2017-08-30 18:02:37,918 [IPC Server Responder] INFO ipc.Server
> (Server.java:run(1307)) - IPC Server Responder: starting
> 2017-08-30 18:02:37,919 [IPC Server listener on 58248] INFO ipc.Server
> (Server.java:run(1146)) - IPC Server listener on 58248: starting
> 2017-08-30 18:02:37,922 [main] INFO hdfs.MiniDFSCluster
> (MiniDFSCluster.java:startDataNodes(1591)) - Starting DataNode 1 with
> dfs.datanode.data.dir:
> [DISK]file:/testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data3,[DISK]file:/testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data4
> 2017-08-30 18:02:37,923 [Thread-997] INFO datanode.DataNode
> (BPOfferService.java:verifyAndSetNamespaceInfo(364)) - Acknowledging ACTIVE
> Namenode during handshakeBlock pool <registering> (Datanode Uuid unassigned)
> service to localhost/127.0.0.1:47551
> 2017-08-30 18:02:37,924 [Thread-997] INFO common.Storage
> (DataStorage.java:getParallelVolumeLoadThreadsNum(346)) - Using 2 threads to
> upgrade data directories (dfs.datanode.parallel.volumes.load.threads.num=2,
> dataDirs=2)
> 2017-08-30 18:02:37,925 [main] INFO checker.ThrottledAsyncChecker
> (ThrottledAsyncChecker.java:schedule(122)) - Scheduling a check for
> [DISK]file:/testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data3
> 2017-08-30 18:02:37,925 [main] INFO checker.ThrottledAsyncChecker
> (ThrottledAsyncChecker.java:schedule(122)) - Scheduling a check for
> [DISK]file:/testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data4
> 2017-08-30 18:02:37,925 [Thread-997] INFO common.Storage
> (Storage.java:tryLock(847)) - Lock on
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data1/in_use.lock
> acquired by nodename 26462@0c74973bd755
> 2017-08-30 18:02:37,925 [Thread-997] INFO common.Storage
> (DataStorage.java:loadStorageDirectory(273)) - Storage directory with
> location
> [DISK]file:/testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data1
> is not formatted for namespace 785578744. Formatting...
> 2017-08-30 18:02:37,925 [Thread-997] INFO common.Storage
> (DataStorage.java:createStorageID(150)) - Generated new storageID
> DS-c7f21b9d-2106-4d12-bb20-c414e1397d77 for directory
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data1
> 2017-08-30 18:02:37,927 [Thread-997] INFO common.Storage
> (Storage.java:tryLock(847)) - Lock on
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data2/in_use.lock
> acquired by nodename 26462@0c74973bd755
> 2017-08-30 18:02:37,927 [Thread-997] INFO common.Storage
> (DataStorage.java:loadStorageDirectory(273)) - Storage directory with
> location
> [DISK]file:/testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data2
> is not formatted for namespace 785578744. Formatting...
> 2017-08-30 18:02:37,928 [Thread-997] INFO common.Storage
> (DataStorage.java:createStorageID(150)) - Generated new storageID
> DS-230ffa8a-d07d-4ac0-86e2-9acc7403dc9d for directory
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data2
> 2017-08-30 18:02:37,934 [main] INFO impl.MetricsSystemImpl
> (MetricsSystemImpl.java:init(158)) - DataNode metrics system started (again)
> 2017-08-30 18:02:37,934 [main] INFO common.Util
> (Util.java:isDiskStatsEnabled(395)) -
> dfs.datanode.fileio.profiling.sampling.percentage set to 0. Disabling file IO
> profiling
> 2017-08-30 18:02:37,935 [main] INFO datanode.BlockScanner
> (BlockScanner.java:<init>(184)) - Initialized block scanner with
> targetBytesPerSec 1048576
> 2017-08-30 18:02:37,935 [main] INFO datanode.DataNode
> (DataNode.java:<init>(496)) - Configured hostname is 127.0.0.1
> 2017-08-30 18:02:37,935 [main] INFO common.Util
> (Util.java:isDiskStatsEnabled(395)) -
> dfs.datanode.fileio.profiling.sampling.percentage set to 0. Disabling file IO
> profiling
> 2017-08-30 18:02:37,936 [main] INFO datanode.DataNode
> (DataNode.java:startDataNode(1385)) - Starting DataNode with maxLockedMemory
> = 0
> 2017-08-30 18:02:37,936 [main] INFO datanode.DataNode
> (DataNode.java:initDataXceiver(1142)) - Opened streaming server at
> /127.0.0.1:51925
> 2017-08-30 18:02:37,936 [main] INFO datanode.DataNode
> (DataXceiverServer.java:<init>(78)) - Balancing bandwidth is 10485760 bytes/s
> 2017-08-30 18:02:37,937 [main] INFO datanode.DataNode
> (DataXceiverServer.java:<init>(79)) - Number threads for balancing is 50
> 2017-08-30 18:02:37,939 [main] INFO server.AuthenticationFilter
> (AuthenticationFilter.java:constructSecretProvider(240)) - Unable to
> initialize FileSignerSecretProvider, falling back to use random secrets.
> 2017-08-30 18:02:37,939 [main] INFO http.HttpRequestLog
> (HttpRequestLog.java:getRequestLog(81)) - Http request log for
> http.requests.datanode is not defined
> 2017-08-30 18:02:37,940 [main] INFO http.HttpServer2
> (HttpServer2.java:addGlobalFilter(921)) - Added global filter 'safety'
> (class=org.apache.hadoop.http.HttpServer2$QuotingInputFilter)
> 2017-08-30 18:02:37,941 [Thread-997] INFO common.Storage
> (BlockPoolSliceStorage.java:recoverTransitionRead(252)) - Analyzing storage
> directories for bpid BP-1775344991-172.17.0.2-1504116157738
> 2017-08-30 18:02:37,941 [main] INFO http.HttpServer2
> (HttpServer2.java:addFilter(894)) - Added filter static_user_filter
> (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to
> context datanode
> 2017-08-30 18:02:37,941 [Thread-997] INFO common.Storage
> (Storage.java:lock(806)) - Locking is disabled for
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data1/current/BP-1775344991-172.17.0.2-1504116157738
> 2017-08-30 18:02:37,941 [main] INFO http.HttpServer2
> (HttpServer2.java:addFilter(904)) - Added filter static_user_filter
> (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to
> context logs
> 2017-08-30 18:02:37,941 [Thread-997] INFO common.Storage
> (BlockPoolSliceStorage.java:loadStorageDirectory(168)) - Block pool storage
> directory for location
> [DISK]file:/testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data1
> and block pool id BP-1775344991-172.17.0.2-1504116157738 is not formatted
> for BP-1775344991-172.17.0.2-1504116157738. Formatting ...
> 2017-08-30 18:02:37,941 [Thread-997] INFO common.Storage
> (BlockPoolSliceStorage.java:format(280)) - Formatting block pool
> BP-1775344991-172.17.0.2-1504116157738 directory
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data1/current/BP-1775344991-172.17.0.2-1504116157738/current
> 2017-08-30 18:02:37,941 [main] INFO http.HttpServer2
> (HttpServer2.java:addFilter(904)) - Added filter static_user_filter
> (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to
> context static
> 2017-08-30 18:02:37,942 [main] INFO http.HttpServer2
> (HttpServer2.java:bindListener(1121)) - Jetty bound to port 55765
> 2017-08-30 18:02:37,942 [main] INFO server.Server (Server.java:doStart(346))
> - jetty-9.3.19.v20170502
> 2017-08-30 18:02:37,945 [main] INFO handler.ContextHandler
> (ContextHandler.java:doStart(781)) - Started
> o.e.j.s.ServletContextHandler@6631cb64{/logs,file:///testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/log/,AVAILABLE}
> 2017-08-30 18:02:37,945 [main] INFO handler.ContextHandler
> (ContextHandler.java:doStart(781)) - Started
> o.e.j.s.ServletContextHandler@481e91b6{/static,file:///testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test-classes/webapps/static/,AVAILABLE}
> 2017-08-30 18:02:37,951 [Thread-997] INFO common.Storage
> (BlockPoolSliceStorage.java:recoverTransitionRead(252)) - Analyzing storage
> directories for bpid BP-1775344991-172.17.0.2-1504116157738
> 2017-08-30 18:02:37,952 [Thread-997] INFO common.Storage
> (Storage.java:lock(806)) - Locking is disabled for
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data2/current/BP-1775344991-172.17.0.2-1504116157738
> 2017-08-30 18:02:37,952 [Thread-997] INFO common.Storage
> (BlockPoolSliceStorage.java:loadStorageDirectory(168)) - Block pool storage
> directory for location
> [DISK]file:/testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data2
> and block pool id BP-1775344991-172.17.0.2-1504116157738 is not formatted
> for BP-1775344991-172.17.0.2-1504116157738. Formatting ...
> 2017-08-30 18:02:37,952 [Thread-997] INFO common.Storage
> (BlockPoolSliceStorage.java:format(280)) - Formatting block pool
> BP-1775344991-172.17.0.2-1504116157738 directory
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data2/current/BP-1775344991-172.17.0.2-1504116157738/current
> 2017-08-30 18:02:37,953 [main] INFO handler.ContextHandler
> (ContextHandler.java:doStart(781)) - Started
> o.e.j.w.WebAppContext@42b28ff1{/,file:///testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test-classes/webapps/datanode/,AVAILABLE}{/datanode}
> 2017-08-30 18:02:37,954 [Thread-997] INFO datanode.DataNode
> (DataNode.java:initStorage(1693)) - Setting up storage:
> nsid=785578744;bpid=BP-1775344991-172.17.0.2-1504116157738;lv=-57;nsInfo=lv=-64;cid=testClusterID;nsid=785578744;c=1504116157738;bpid=BP-1775344991-172.17.0.2-1504116157738;dnuuid=null
> 2017-08-30 18:02:37,954 [main] INFO server.AbstractConnector
> (AbstractConnector.java:doStart(278)) - Started
> ServerConnector@36061cf3{HTTP/1.1,[http/1.1]}{localhost:55765}
> 2017-08-30 18:02:37,954 [main] INFO server.Server (Server.java:doStart(414))
> - Started @37470ms
> 2017-08-30 18:02:37,955 [Thread-997] INFO datanode.DataNode
> (DataNode.java:checkDatanodeUuid(1517)) - Generated and persisted new
> Datanode UUID d7ee1847-d046-47ea-9a05-7fd25bcb1e86
> 2017-08-30 18:02:37,959 [Thread-997] INFO impl.FsDatasetImpl
> (FsVolumeList.java:addVolume(306)) - Added new volume:
> DS-c7f21b9d-2106-4d12-bb20-c414e1397d77
> 2017-08-30 18:02:37,962 [Thread-997] INFO impl.FsDatasetImpl
> (FsDatasetImpl.java:addVolume(435)) - Added volume -
> [DISK]file:/testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data1,
> StorageType: DISK
> 2017-08-30 18:02:37,963 [Thread-997] INFO impl.FsDatasetImpl
> (FsVolumeList.java:addVolume(306)) - Added new volume:
> DS-230ffa8a-d07d-4ac0-86e2-9acc7403dc9d
> 2017-08-30 18:02:37,964 [Thread-997] INFO impl.FsDatasetImpl
> (FsDatasetImpl.java:addVolume(435)) - Added volume -
> [DISK]file:/testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data2,
> StorageType: DISK
> 2017-08-30 18:02:37,964 [Thread-997] INFO impl.FsDatasetImpl
> (FsDatasetImpl.java:registerMBean(2178)) - Registered FSDatasetState MBean
> 2017-08-30 18:02:37,965 [Thread-997] INFO checker.ThrottledAsyncChecker
> (ThrottledAsyncChecker.java:schedule(122)) - Scheduling a check for
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data1
> 2017-08-30 18:02:37,967 [Thread-997] INFO checker.DatasetVolumeChecker
> (DatasetVolumeChecker.java:checkAllVolumes(210)) - Scheduled health check for
> volume
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data1
> 2017-08-30 18:02:37,967 [main] INFO web.DatanodeHttpServer
> (DatanodeHttpServer.java:start(239)) - Listening HTTP traffic on
> /127.0.0.1:42070
> 2017-08-30 18:02:37,967 [Thread-997] INFO checker.ThrottledAsyncChecker
> (ThrottledAsyncChecker.java:schedule(122)) - Scheduling a check for
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data2
> 2017-08-30 18:02:37,967 [Thread-997] INFO checker.DatasetVolumeChecker
> (DatasetVolumeChecker.java:checkAllVolumes(210)) - Scheduled health check for
> volume
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data2
> 2017-08-30 18:02:37,968 [main] INFO datanode.DataNode
> (DataNode.java:startDataNode(1412)) - dnUserName = jenkins
> 2017-08-30 18:02:37,968 [main] INFO datanode.DataNode
> (DataNode.java:startDataNode(1413)) - supergroup = supergroup
> 2017-08-30 18:02:37,968
> [org.apache.hadoop.util.JvmPauseMonitor$Monitor@76889e60] INFO
> util.JvmPauseMonitor (JvmPauseMonitor.java:run(188)) - Starting JVM pause
> monitor
> 2017-08-30 18:02:37,968 [Thread-997] INFO impl.FsDatasetImpl
> (FsDatasetImpl.java:addBlockPool(2664)) - Adding block pool
> BP-1775344991-172.17.0.2-1504116157738
> 2017-08-30 18:02:37,969 [main] INFO ipc.CallQueueManager
> (CallQueueManager.java:<init>(84)) - Using callQueue: class
> java.util.concurrent.LinkedBlockingQueue queueCapacity: 1000 scheduler: class
> org.apache.hadoop.ipc.DefaultRpcScheduler
> 2017-08-30 18:02:37,969 [Thread-1025] INFO impl.FsDatasetImpl
> (FsVolumeList.java:run(409)) - Scanning block pool
> BP-1775344991-172.17.0.2-1504116157738 on volume
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data1...
> 2017-08-30 18:02:37,969 [Thread-1026] INFO impl.FsDatasetImpl
> (FsVolumeList.java:run(409)) - Scanning block pool
> BP-1775344991-172.17.0.2-1504116157738 on volume
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data2...
> 2017-08-30 18:02:37,969 [Socket Reader #1 for port 50140] INFO ipc.Server
> (Server.java:run(1067)) - Starting Socket Reader #1 for port 50140
> 2017-08-30 18:02:37,975 [Thread-1026] INFO impl.FsDatasetImpl
> (FsVolumeList.java:run(414)) - Time taken to scan block pool
> BP-1775344991-172.17.0.2-1504116157738 on
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data2:
> 7ms
> 2017-08-30 18:02:37,977 [main] INFO datanode.DataNode
> (DataNode.java:initIpcServer(1029)) - Opened IPC server at /127.0.0.1:50140
> 2017-08-30 18:02:37,979 [Thread-1025] INFO impl.FsDatasetImpl
> (FsVolumeList.java:run(414)) - Time taken to scan block pool
> BP-1775344991-172.17.0.2-1504116157738 on
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data1:
> 10ms
> 2017-08-30 18:02:37,979 [Thread-997] INFO impl.FsDatasetImpl
> (FsVolumeList.java:addBlockPool(440)) - Total time to scan all replicas for
> block pool BP-1775344991-172.17.0.2-1504116157738: 10ms
> 2017-08-30 18:02:37,979 [Thread-1032] INFO impl.FsDatasetImpl
> (FsVolumeList.java:run(198)) - Adding replicas to map for block pool
> BP-1775344991-172.17.0.2-1504116157738 on volume
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data1...
> 2017-08-30 18:02:37,979 [Thread-1033] INFO impl.FsDatasetImpl
> (FsVolumeList.java:run(198)) - Adding replicas to map for block pool
> BP-1775344991-172.17.0.2-1504116157738 on volume
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data2...
> 2017-08-30 18:02:37,980 [Thread-1032] INFO impl.BlockPoolSlice
> (BlockPoolSlice.java:readReplicasFromCache(770)) - Replica Cache file:
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data1/current/BP-1775344991-172.17.0.2-1504116157738/current/replicas
> doesn't exist
> 2017-08-30 18:02:37,980 [Thread-1033] INFO impl.BlockPoolSlice
> (BlockPoolSlice.java:readReplicasFromCache(770)) - Replica Cache file:
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data2/current/BP-1775344991-172.17.0.2-1504116157738/current/replicas
> doesn't exist
> 2017-08-30 18:02:37,980 [Thread-1032] INFO impl.FsDatasetImpl
> (FsVolumeList.java:run(203)) - Time to add replicas to map for block pool
> BP-1775344991-172.17.0.2-1504116157738 on volume
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data1:
> 0ms
> 2017-08-30 18:02:37,980 [Thread-1033] INFO impl.FsDatasetImpl
> (FsVolumeList.java:run(203)) - Time to add replicas to map for block pool
> BP-1775344991-172.17.0.2-1504116157738 on volume
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data2:
> 0ms
> 2017-08-30 18:02:37,980 [Thread-997] INFO impl.FsDatasetImpl
> (FsVolumeList.java:getAllVolumesMap(229)) - Total time to add all replicas to
> map: 1ms
> 2017-08-30 18:02:37,980 [main] INFO datanode.DataNode
> (BlockPoolManager.java:refreshNamenodes(149)) - Refresh request received for
> nameservices: null
> 2017-08-30 18:02:37,980 [main] INFO datanode.DataNode
> (BlockPoolManager.java:doRefreshNamenodes(198)) - Starting BPOfferServices
> for nameservices: <default>
> 2017-08-30 18:02:37,980
> [VolumeScannerThread(/testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data1)]
> INFO datanode.VolumeScanner
> (VolumeScanner.java:findNextUsableBlockIter(382)) - Now scanning bpid
> BP-1775344991-172.17.0.2-1504116157738 on volume
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data1
> 2017-08-30 18:02:37,981 [Thread-997] INFO datanode.DirectoryScanner
> (DirectoryScanner.java:start(281)) - Periodic Directory Tree Verification
> scan starting at 8/30/17 8:12 PM with interval of 21600000ms
> 2017-08-30 18:02:37,980
> [VolumeScannerThread(/testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data2)]
> INFO datanode.VolumeScanner
> (VolumeScanner.java:findNextUsableBlockIter(382)) - Now scanning bpid
> BP-1775344991-172.17.0.2-1504116157738 on volume
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data2
> 2017-08-30 18:02:37,981
> [VolumeScannerThread(/testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data1)]
> INFO datanode.VolumeScanner (VolumeScanner.java:runLoop(538)) -
> VolumeScanner(/testptch/hadoop/h
> ...[truncated 43082 chars]...
> :02:38,122 [IPC Server handler 0 on 47551] INFO net.NetworkTopology
> (NetworkTopology.java:add(145)) - Adding a new node:
> /default-rack/127.0.0.1:46540
> 2017-08-30 18:02:38,122 [IPC Server handler 0 on 47551] INFO
> blockmanagement.BlockReportLeaseManager
> (BlockReportLeaseManager.java:registerNode(204)) - Registered DN
> 303e97cb-9549-4084-8530-e6f511ebcbe1 (127.0.0.1:46540).
> 2017-08-30 18:02:38,123 [BP-1775344991-172.17.0.2-1504116157738 heartbeating
> to localhost/127.0.0.1:47551] INFO datanode.DataNode
> (BPServiceActor.java:register(782)) - Block pool Block pool
> BP-1775344991-172.17.0.2-1504116157738 (Datanode Uuid
> 303e97cb-9549-4084-8530-e6f511ebcbe1) service to localhost/127.0.0.1:47551
> successfully registered with NN
> 2017-08-30 18:02:38,123 [BP-1775344991-172.17.0.2-1504116157738 heartbeating
> to localhost/127.0.0.1:47551] INFO datanode.DataNode
> (BPServiceActor.java:offerService(612)) - For namenode
> localhost/127.0.0.1:47551 using BLOCKREPORT_INTERVAL of 21600000msec
> CACHEREPORT_INTERVAL of 10000msec Initial delay: 0msec; heartBeatInterval=3000
> 2017-08-30 18:02:38,126 [IPC Server handler 1 on 47551] INFO
> blockmanagement.DatanodeDescriptor
> (DatanodeDescriptor.java:updateStorage(937)) - Adding new storage ID
> DS-3f87b7ae-9ac7-4485-885e-fc9d2475df21 for DN 127.0.0.1:46540
> 2017-08-30 18:02:38,126 [IPC Server handler 1 on 47551] INFO
> blockmanagement.DatanodeDescriptor
> (DatanodeDescriptor.java:updateStorage(937)) - Adding new storage ID
> DS-a58905ea-f4b4-4b7f-8c16-587ae6b9033b for DN 127.0.0.1:46540
> 2017-08-30 18:02:38,127 [Block report processor] INFO BlockStateChange
> (BlockManager.java:processReport(2437)) - BLOCK* processReport
> 0x703473b98bf5005f: Processing first storage report for
> DS-a58905ea-f4b4-4b7f-8c16-587ae6b9033b from datanode
> 303e97cb-9549-4084-8530-e6f511ebcbe1
> 2017-08-30 18:02:38,128 [Block report processor] INFO BlockStateChange
> (BlockManager.java:processReport(2463)) - BLOCK* processReport
> 0x703473b98bf5005f: from storage DS-a58905ea-f4b4-4b7f-8c16-587ae6b9033b node
> DatanodeRegistration(127.0.0.1:46540,
> datanodeUuid=303e97cb-9549-4084-8530-e6f511ebcbe1, infoPort=41583,
> infoSecurePort=0, ipcPort=56578,
> storageInfo=lv=-57;cid=testClusterID;nsid=785578744;c=1504116157738), blocks:
> 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0
> 2017-08-30 18:02:38,128 [Block report processor] INFO BlockStateChange
> (BlockManager.java:processReport(2437)) - BLOCK* processReport
> 0x703473b98bf5005f: Processing first storage report for
> DS-3f87b7ae-9ac7-4485-885e-fc9d2475df21 from datanode
> 303e97cb-9549-4084-8530-e6f511ebcbe1
> 2017-08-30 18:02:38,128 [Block report processor] INFO BlockStateChange
> (BlockManager.java:processReport(2463)) - BLOCK* processReport
> 0x703473b98bf5005f: from storage DS-3f87b7ae-9ac7-4485-885e-fc9d2475df21 node
> DatanodeRegistration(127.0.0.1:46540,
> datanodeUuid=303e97cb-9549-4084-8530-e6f511ebcbe1, infoPort=41583,
> infoSecurePort=0, ipcPort=56578,
> storageInfo=lv=-57;cid=testClusterID;nsid=785578744;c=1504116157738), blocks:
> 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0
> 2017-08-30 18:02:38,129 [BP-1775344991-172.17.0.2-1504116157738 heartbeating
> to localhost/127.0.0.1:47551] INFO datanode.DataNode
> (BPServiceActor.java:blockReport(422)) - Successfully sent block report
> 0x703473b98bf5005f, containing 2 storage report(s), of which we sent 2. The
> reports had 0 total blocks and used 1 RPC(s). This took 0 msec to generate
> and 2 msecs for RPC and NN processing. Got back one command:
> FinalizeCommand/5.
> 2017-08-30 18:02:38,129 [BP-1775344991-172.17.0.2-1504116157738 heartbeating
> to localhost/127.0.0.1:47551] INFO datanode.DataNode
> (BPOfferService.java:processCommandFromActive(745)) - Got finalize command
> for block pool BP-1775344991-172.17.0.2-1504116157738
> 2017-08-30 18:02:38,160 [main] INFO hdfs.MiniDFSCluster
> (MiniDFSCluster.java:waitActive(2621)) - Cluster is active
> 2017-08-30 18:02:38,163 [main] INFO hdfs.MiniDFSCluster
> (MiniDFSCluster.java:waitActive(2621)) - Cluster is active
> 2017-08-30 18:02:38,168 [IPC Server handler 5 on 47551] INFO
> FSNamesystem.audit (FSNamesystem.java:logAuditMessage(7598)) - allowed=true
> ugi=jenkins (auth:SIMPLE) ip=/127.0.0.1 cmd=create
> src=/user/jenkins/noheartbeat.dat dst=null
> perm=jenkins:supergroup:rw-r--r-- proto=rpc
> 2017-08-30 18:02:38,174 [IPC Server handler 6 on 47551] INFO
> hdfs.StateChange (FSDirWriteFileOp.java:logAllocatedBlock(789)) - BLOCK*
> allocate blk_1073741825_1001, replicas=127.0.0.1:46794, 127.0.0.1:51925 for
> /user/jenkins/noheartbeat.dat
> 2017-08-30 18:02:38,177 [DataXceiver for client
> DFSClient_NONMAPREDUCE_-172490240_1 at /127.0.0.1:54220 [Receiving block
> BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001]] INFO
> datanode.DataNode (DataXceiver.java:writeBlock(742)) - Receiving
> BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001 src:
> /127.0.0.1:54220 dest: /127.0.0.1:46794
> 2017-08-30 18:02:38,178 [DataXceiver for client
> DFSClient_NONMAPREDUCE_-172490240_1 at /127.0.0.1:56224 [Receiving block
> BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001]] INFO
> datanode.DataNode (DataXceiver.java:writeBlock(742)) - Receiving
> BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001 src:
> /127.0.0.1:56224 dest: /127.0.0.1:51925
> 2017-08-30 18:02:39,181 [DataXceiver for client
> DFSClient_NONMAPREDUCE_-172490240_1 at /127.0.0.1:56224 [Receiving block
> BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001]] INFO
> datanode.DataNode (BlockReceiver.java:receiveBlock(1001)) - Exception for
> BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001
> java.net.SocketTimeoutException: 1000 millis timeout while waiting for
> channel to be ready for read. ch : java.nio.channels.SocketChannel[connected
> local=/127.0.0.1:51925 remote=/127.0.0.1:56224]
> at
> org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:164)
> at
> org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161)
> at
> org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131)
> at java.io.BufferedInputStream.fill(BufferedInputStream.java:246)
> at java.io.BufferedInputStream.read1(BufferedInputStream.java:286)
> at java.io.BufferedInputStream.read(BufferedInputStream.java:345)
> at java.io.DataInputStream.read(DataInputStream.java:149)
> at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:208)
> at
> org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:211)
> at
> org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:134)
> at
> org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:109)
> at
> org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:522)
> at
> org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:962)
> at
> org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:900)
> at
> org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:173)
> at
> org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:107)
> at
> org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:290)
> at java.lang.Thread.run(Thread.java:748)
> 2017-08-30 18:02:39,182 [PacketResponder:
> BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001,
> type=LAST_IN_PIPELINE] INFO datanode.DataNode (BlockReceiver.java:run(1461))
> - PacketResponder:
> BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001,
> type=LAST_IN_PIPELINE: Thread is interrupted.
> 2017-08-30 18:02:39,182 [PacketResponder:
> BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001,
> type=LAST_IN_PIPELINE] INFO datanode.DataNode (BlockReceiver.java:run(1497))
> - PacketResponder:
> BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001,
> type=LAST_IN_PIPELINE terminating
> 2017-08-30 18:02:39,182 [DataXceiver for client
> DFSClient_NONMAPREDUCE_-172490240_1 at /127.0.0.1:56224 [Receiving block
> BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001]] INFO
> datanode.DataNode (DataXceiver.java:writeBlock(933)) - opWriteBlock
> BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001 received exception
> java.net.SocketTimeoutException: 1000 millis timeout while waiting for
> channel to be ready for read. ch : java.nio.channels.SocketChannel[connected
> local=/127.0.0.1:51925 remote=/127.0.0.1:56224]
> 2017-08-30 18:02:39,183 [DataXceiver for client
> DFSClient_NONMAPREDUCE_-172490240_1 at /127.0.0.1:56224 [Receiving block
> BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001]] ERROR
> datanode.DataNode (DataXceiver.java:run(323)) - 127.0.0.1:51925:DataXceiver
> error processing WRITE_BLOCK operation src: /127.0.0.1:56224 dst:
> /127.0.0.1:51925
> java.net.SocketTimeoutException: 1000 millis timeout while waiting for
> channel to be ready for read. ch : java.nio.channels.SocketChannel[connected
> local=/127.0.0.1:51925 remote=/127.0.0.1:56224]
> at
> org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:164)
> at
> org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161)
> at
> org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131)
> at java.io.BufferedInputStream.fill(BufferedInputStream.java:246)
> at java.io.BufferedInputStream.read1(BufferedInputStream.java:286)
> at java.io.BufferedInputStream.read(BufferedInputStream.java:345)
> at java.io.DataInputStream.read(DataInputStream.java:149)
> at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:208)
> at
> org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:211)
> at
> org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:134)
> at
> org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:109)
> at
> org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:522)
> at
> org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:962)
> at
> org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:900)
> at
> org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:173)
> at
> org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:107)
> at
> org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:290)
> at java.lang.Thread.run(Thread.java:748)
> 2017-08-30 18:02:39,183 [PacketResponder:
> BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001,
> type=HAS_DOWNSTREAM_IN_PIPELINE, downstreams=1:[127.0.0.1:51925]] INFO
> datanode.DataNode (BlockReceiver.java:packetSentInTime(393)) - A packet was
> last sent 1006 milliseconds ago.
> 2017-08-30 18:02:39,186 [PacketResponder:
> BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001,
> type=HAS_DOWNSTREAM_IN_PIPELINE, downstreams=1:[127.0.0.1:51925]] WARN
> datanode.DataNode (BlockReceiver.java:run(1436)) - The downstream error might
> be due to congestion in upstream including this node. Propagating the error:
> java.io.EOFException: Unexpected EOF while trying to read response from server
> at
> org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:444)
> at
> org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:213)
> at
> org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.run(BlockReceiver.java:1375)
> at java.lang.Thread.run(Thread.java:748)
> 2017-08-30 18:02:39,186 [PacketResponder:
> BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001,
> type=HAS_DOWNSTREAM_IN_PIPELINE, downstreams=1:[127.0.0.1:51925]] WARN
> datanode.DataNode (BlockReceiver.java:run(1480)) - IOException in
> BlockReceiver.run():
> java.io.EOFException: Unexpected EOF while trying to read response from server
> at
> org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:444)
> at
> org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:213)
> at
> org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.run(BlockReceiver.java:1375)
> at java.lang.Thread.run(Thread.java:748)
> 2017-08-30 18:02:39,186 [PacketResponder:
> BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001,
> type=HAS_DOWNSTREAM_IN_PIPELINE, downstreams=1:[127.0.0.1:51925]] INFO
> datanode.DataNode (BlockReceiver.java:run(1483)) - PacketResponder:
> BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001,
> type=HAS_DOWNSTREAM_IN_PIPELINE, downstreams=1:[127.0.0.1:51925]
> java.io.EOFException: Unexpected EOF while trying to read response from server
> at
> org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:444)
> at
> org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:213)
> at
> org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.run(BlockReceiver.java:1375)
> at java.lang.Thread.run(Thread.java:748)
> 2017-08-30 18:02:39,186 [PacketResponder:
> BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001,
> type=HAS_DOWNSTREAM_IN_PIPELINE, downstreams=1:[127.0.0.1:51925]] INFO
> datanode.DataNode (BlockReceiver.java:run(1497)) - PacketResponder:
> BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001,
> type=HAS_DOWNSTREAM_IN_PIPELINE, downstreams=1:[127.0.0.1:51925] terminating
> 2017-08-30 18:02:39,186 [DataXceiver for client
> DFSClient_NONMAPREDUCE_-172490240_1 at /127.0.0.1:54220 [Receiving block
> BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001]] INFO
> datanode.DataNode (BlockReceiver.java:handleMirrorOutError(447)) -
> DatanodeRegistration(127.0.0.1:46794,
> datanodeUuid=d7ee1847-d046-47ea-9a05-7fd25bcb1e86, infoPort=52625,
> infoSecurePort=0, ipcPort=58248,
> storageInfo=lv=-57;cid=testClusterID;nsid=785578744;c=1504116157738):Exception
> writing BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001 to mirror
> 127.0.0.1:51925
> java.io.IOException: Interrupted while sleeping. Bailing out.
> at
> org.apache.hadoop.hdfs.TestClientProtocolForPipelineRecovery$6.stopSendingPacketDownstream(TestClientProtocolForPipelineRecovery.java:558)
> at
> org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:581)
> at
> org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:962)
> at
> org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:900)
> at
> org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:173)
> at
> org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:107)
> at
> org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:290)
> at java.lang.Thread.run(Thread.java:748)
> 2017-08-30 18:02:49,187 [ResponseProcessor for block
> BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001] WARN
> hdfs.DataStreamer (DataStreamer.java:run(1214)) - Exception for
> BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001
> java.net.SocketTimeoutException: 11000 millis timeout while waiting for
> channel to be ready for read. ch : java.nio.channels.SocketChannel[connected
> local=/127.0.0.1:54220 remote=/127.0.0.1:46794]
> at
> org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:164)
> at
> org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161)
> at
> org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131)
> at
> org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:118)
> at java.io.FilterInputStream.read(FilterInputStream.java:83)
> at java.io.FilterInputStream.read(FilterInputStream.java:83)
> at
> org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:442)
> at
> org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:213)
> at
> org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1104)
> 2017-08-30 18:02:49,188 [DataStreamer for file /user/jenkins/noheartbeat.dat
> block BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001] WARN
> hdfs.DataStreamer (DataStreamer.java:handleBadDatanode(1560)) - Error
> Recovery for BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001 in
> pipeline
> [DatanodeInfoWithStorage[127.0.0.1:46794,DS-230ffa8a-d07d-4ac0-86e2-9acc7403dc9d,DISK],
>
> DatanodeInfoWithStorage[127.0.0.1:51925,DS-274e8cc9-280b-4370-b494-6a4f0d67ccf4,DISK]]:
> datanode
> 0(DatanodeInfoWithStorage[127.0.0.1:46794,DS-230ffa8a-d07d-4ac0-86e2-9acc7403dc9d,DISK])
> is bad.
> 2017-08-30 18:02:49,188 [DataXceiver for client
> DFSClient_NONMAPREDUCE_-172490240_1 at /127.0.0.1:54220 [Receiving block
> BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001]] INFO
> datanode.DataNode (BlockReceiver.java:receiveBlock(1001)) - Exception for
> BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001
> java.io.IOException: Premature EOF from inputStream
> at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:210)
> at
> org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:211)
> at
> org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:134)
> at
> org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:109)
> at
> org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:522)
> at
> org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:962)
> at
> org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:900)
> at
> org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:173)
> at
> org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:107)
> at
> org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:290)
> at java.lang.Thread.run(Thread.java:748)
> 2017-08-30 18:02:49,188 [DataXceiver for client
> DFSClient_NONMAPREDUCE_-172490240_1 at /127.0.0.1:54220 [Receiving block
> BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001]] INFO
> datanode.DataNode (DataXceiver.java:writeBlock(933)) - opWriteBlock
> BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001 received exception
> java.io.IOException: Premature EOF from inputStream
> 2017-08-30 18:02:49,189 [DataXceiver for client
> DFSClient_NONMAPREDUCE_-172490240_1 at /127.0.0.1:54220 [Receiving block
> BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001]] ERROR
> datanode.DataNode (DataXceiver.java:run(323)) - 127.0.0.1:46794:DataXceiver
> error processing WRITE_BLOCK operation src: /127.0.0.1:54220 dst:
> /127.0.0.1:46794
> java.io.IOException: Premature EOF from inputStream
> at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:210)
> at
> org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:211)
> at
> org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:134)
> at
> org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:109)
> at
> org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:522)
> at
> org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:962)
> at
> org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:900)
> at
> org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:173)
> at
> org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:107)
> at
> org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:290)
> at java.lang.Thread.run(Thread.java:748)
> 2017-08-30 18:02:49,214 [DataXceiver for client
> DFSClient_NONMAPREDUCE_-172490240_1 at /127.0.0.1:55316 [Receiving block
> BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001]] INFO
> datanode.DataNode (DataXceiver.java:writeBlock(742)) - Receiving
> BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001 src:
> /127.0.0.1:55316 dest: /127.0.0.1:46540
> 2017-08-30 18:02:49,214 [DataXceiver for client
> DFSClient_NONMAPREDUCE_-172490240_1 at /127.0.0.1:55315 [Receiving block
> BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001]] INFO
> datanode.DataNode (DataXceiver.java:writeBlock(742)) - Receiving
> BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001 src:
> /127.0.0.1:55315 dest: /127.0.0.1:46540
> 2017-08-30 18:02:49,215 [DataXceiver for client
> DFSClient_NONMAPREDUCE_-172490240_1 at /127.0.0.1:55316 [Receiving block
> BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001]] INFO
> impl.FsDatasetImpl (FsDatasetImpl.java:recoverRbw(1376)) - Recover RBW
> replica BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001
> 2017-08-30 18:02:49,215
> [org.apache.hadoop.hdfs.server.datanode.DataNode$DataTransfer@47d9e1e8] INFO
> datanode.DataNode (DataNode.java:run(2520)) - DataTransfer, at
> 127.0.0.1:51925: Transmitted
> BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001 (numBytes=0) to
> /127.0.0.1:46540
> 2017-08-30 18:02:49,215 [DataXceiver for client
> DFSClient_NONMAPREDUCE_-172490240_1 at /127.0.0.1:55316 [Receiving block
> BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001]] INFO
> datanode.DataNode (DataXceiver.java:writeBlock(933)) - opWriteBlock
> BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001 received exception
> org.apache.hadoop.hdfs.server.datanode.ReplicaNotFoundException: Cannot
> recover a non-RBW replica LocalReplicaInPipeline, blk_1073741825_1001,
> TEMPORARY
> getNumBytes() = 0
> getBytesOnDisk() = 0
> getVisibleLength()= -1
> getVolume() =
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data5
> getBlockURI() =
> file:/testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data5/current/BP-1775344991-172.17.0.2-1504116157738/tmp/blk_1073741825
> bytesAcked=0
> bytesOnDisk=0
> 2017-08-30 18:02:49,215 [DataXceiver for client
> DFSClient_NONMAPREDUCE_-172490240_1 at /127.0.0.1:55315 [Receiving block
> BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001]] INFO
> impl.FsDatasetImpl (FsDatasetImpl.java:convertTemporaryToRbw(1470)) - Convert
> BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001 from Temporary to
> RBW, visible length=0
> 2017-08-30 18:02:49,215 [DataXceiver for client
> DFSClient_NONMAPREDUCE_-172490240_1 at /127.0.0.1:55316 [Receiving block
> BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001]] ERROR
> datanode.DataNode (DataXceiver.java:run(323)) - 127.0.0.1:46540:DataXceiver
> error processing WRITE_BLOCK operation src: /127.0.0.1:55316 dst:
> /127.0.0.1:46540
> org.apache.hadoop.hdfs.server.datanode.ReplicaNotFoundException: Cannot
> recover a non-RBW replica LocalReplicaInPipeline, blk_1073741825_1001,
> TEMPORARY
> getNumBytes() = 0
> getBytesOnDisk() = 0
> getVisibleLength()= -1
> getVolume() =
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data5
> getBlockURI() =
> file:/testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data5/current/BP-1775344991-172.17.0.2-1504116157738/tmp/blk_1073741825
> bytesAcked=0
> bytesOnDisk=0
> at
> org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.recoverRbw(FsDatasetImpl.java:1385)
> at
> org.apache.hadoop.hdfs.server.datanode.BlockReceiver.<init>(BlockReceiver.java:221)
> at
> org.apache.hadoop.hdfs.server.datanode.DataXceiver.getBlockReceiver(DataXceiver.java:1291)
> at
> org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:758)
> at
> org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:173)
> at
> org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:107)
> at
> org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:290)
> at java.lang.Thread.run(Thread.java:748)
> 2017-08-30 18:02:49,216 [DataStreamer for file /user/jenkins/noheartbeat.dat
> block BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001] INFO
> hdfs.DataStreamer (DataStreamer.java:createBlockOutputStream(1778)) -
> Exception in createBlockOutputStream blk_1073741825_1001
> java.io.EOFException: Unexpected EOF while trying to read response from server
> at
> org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:444)
> at
> org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1751)
> at
> org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1496)
> at
> org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1470)
> at
> org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1274)
> at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:684)
> 2017-08-30 18:02:49,217 [DataStreamer for file /user/jenkins/noheartbeat.dat
> block BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001] WARN
> hdfs.DataStreamer (DataStreamer.java:handleBadDatanode(1560)) - Error
> Recovery for BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001 in
> pipeline
> [DatanodeInfoWithStorage[127.0.0.1:46540,DS-a58905ea-f4b4-4b7f-8c16-587ae6b9033b,DISK],
>
> DatanodeInfoWithStorage[127.0.0.1:51925,DS-274e8cc9-280b-4370-b494-6a4f0d67ccf4,DISK]]:
> datanode
> 0(DatanodeInfoWithStorage[127.0.0.1:46540,DS-a58905ea-f4b4-4b7f-8c16-587ae6b9033b,DISK])
> is bad.
> 2017-08-30 18:02:49,218 [IPC Server handler 0 on 47551] WARN
> blockmanagement.BlockPlacementPolicy
> (BlockPlacementPolicyDefault.java:chooseTarget(405)) - Failed to place enough
> replicas, still in need of 1 to reach 2 (unavailableStorages=[],
> storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK],
> creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=false) For
> more information, please enable DEBUG log level on
> org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and
> org.apache.hadoop.net.NetworkTopology
> 2017-08-30 18:02:49,219 [IPC Server handler 0 on 47551] WARN
> blockmanagement.BlockPlacementPolicy
> (BlockPlacementPolicyDefault.java:chooseTarget(405)) - Failed to place enough
> replicas, still in need of 1 to reach 2 (unavailableStorages=[DISK],
> storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK],
> creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=false) For
> more information, please enable DEBUG log level on
> org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and
> org.apache.hadoop.net.NetworkTopology
> 2017-08-30 18:02:49,348 [IPC Server handler 0 on 47551] WARN
> protocol.BlockStoragePolicy (BlockStoragePolicy.java:chooseStorageTypes(161))
> - Failed to place enough replicas: expected size is 1 but only 0 storage
> types can be selected (replication=2, selected=[], unavailable=[DISK,
> ARCHIVE], removed=[DISK], policy=BlockStoragePolicy{HOT:7,
> storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]})
> 2017-08-30 18:02:49,348 [IPC Server handler 0 on 47551] WARN
> blockmanagement.BlockPlacementPolicy
> (BlockPlacementPolicyDefault.java:chooseTarget(405)) - Failed to place enough
> replicas, still in need of 1 to reach 2 (unavailableStorages=[DISK, ARCHIVE],
> storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK],
> creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=false) All
> required storage types are unavailable: unavailableStorages=[DISK, ARCHIVE],
> storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK],
> creationFallbacks=[], replicationFallbacks=[ARCHIVE]}
> 2017-08-30 18:02:49,354 [DataStreamer for file /user/jenkins/noheartbeat.dat
> block BP-1775344991-172.17.0.2-1504116157738:blk_1073741825_1001] WARN
> hdfs.DataStreamer (DataStreamer.java:run(844)) - DataStreamer Exception
> java.io.IOException: Failed to replace a bad datanode on the existing
> pipeline due to no more good datanodes being available to try. (Nodes:
> current=[DatanodeInfoWithStorage[127.0.0.1:51925,DS-274e8cc9-280b-4370-b494-6a4f0d67ccf4,DISK]],
>
> original=[DatanodeInfoWithStorage[127.0.0.1:51925,DS-274e8cc9-280b-4370-b494-6a4f0d67ccf4,DISK]]).
> The current failed datanode replacement policy is ALWAYS, and a client may
> configure this via
> 'dfs.client.block.write.replace-datanode-on-failure.policy' in its
> configuration.
> at
> org.apache.hadoop.hdfs.DataStreamer.findNewDatanode(DataStreamer.java:1322)
> at
> org.apache.hadoop.hdfs.DataStreamer.addDatanode2ExistingPipeline(DataStreamer.java:1388)
> at
> org.apache.hadoop.hdfs.DataStreamer.handleDatanodeReplacement(DataStreamer.java:1587)
> at
> org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1488)
> at
> org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1470)
> at
> org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1274)
> at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:684)
> 2017-08-30 18:02:49,355 [main] WARN hdfs.DFSClient
> (DFSOutputStream.java:flushOrSync(729)) - Error while syncing
> java.io.IOException: Failed to replace a bad datanode on the existing
> pipeline due to no more good datanodes being available to try. (Nodes:
> current=[DatanodeInfoWithStorage[127.0.0.1:51925,DS-274e8cc9-280b-4370-b494-6a4f0d67ccf4,DISK]],
>
> original=[DatanodeInfoWithStorage[127.0.0.1:51925,DS-274e8cc9-280b-4370-b494-6a4f0d67ccf4,DISK]]).
> The current failed datanode replacement policy is ALWAYS, and a client may
> configure this via
> 'dfs.client.block.write.replace-datanode-on-failure.policy' in its
> configuration.
> at
> org.apache.hadoop.hdfs.DataStreamer.findNewDatanode(DataStreamer.java:1322)
> at
> org.apache.hadoop.hdfs.DataStreamer.addDatanode2ExistingPipeline(DataStreamer.java:1388)
> at
> org.apache.hadoop.hdfs.DataStreamer.handleDatanodeReplacement(DataStreamer.java:1587)
> at
> org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1488)
> at
> org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1470)
> at
> org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1274)
> at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:684)
> 2017-08-30 18:02:49,357 [main] INFO hdfs.MiniDFSCluster
> (MiniDFSCluster.java:shutdown(1957)) - Shutting down the Mini HDFS Cluster
> 2017-08-30 18:02:49,357 [main] ERROR hdfs.DFSClient
> (DFSClient.java:closeAllFilesBeingWritten(590)) - Failed to close file:
> /user/jenkins/noheartbeat.dat with inode: 16388
> java.io.IOException: Failed to replace a bad datanode on the existing
> pipeline due to no more good datanodes being available to try. (Nodes:
> current=[DatanodeInfoWithStorage[127.0.0.1:51925,DS-274e8cc9-280b-4370-b494-6a4f0d67ccf4,DISK]],
>
> original=[DatanodeInfoWithStorage[127.0.0.1:51925,DS-274e8cc9-280b-4370-b494-6a4f0d67ccf4,DISK]]).
> The current failed datanode replacement policy is ALWAYS, and a client may
> configure this via
> 'dfs.client.block.write.replace-datanode-on-failure.policy' in its
> configuration.
> at
> org.apache.hadoop.hdfs.DataStreamer.findNewDatanode(DataStreamer.java:1322)
> at
> org.apache.hadoop.hdfs.DataStreamer.addDatanode2ExistingPipeline(DataStreamer.java:1388)
> at
> org.apache.hadoop.hdfs.DataStreamer.handleDatanodeReplacement(DataStreamer.java:1587)
> at
> org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1488)
> at
> org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1470)
> at
> org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1274)
> at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:684)
> 2017-08-30 18:02:49,357 [main] INFO hdfs.MiniDFSCluster
> (MiniDFSCluster.java:shutdownDataNode(2005)) - Shutting down DataNode 2
> 2017-08-30 18:02:49,358 [main] WARN datanode.DirectoryScanner
> (DirectoryScanner.java:shutdown(340)) - DirectoryScanner: shutdown has been
> called
> 2017-08-30 18:02:49,358
> [org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@70c0a3d5] INFO
> datanode.DataNode (DataXceiverServer.java:closeAllPeers(281)) - Closing all
> peers.
> 2017-08-30 18:02:49,359
> [VolumeScannerThread(/testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data6)]
> INFO datanode.VolumeScanner (VolumeScanner.java:run(636)) -
> VolumeScanner(/testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data6,
> DS-a58905ea-f4b4-4b7f-8c16-587ae6b9033b) exiting.
> 2017-08-30 18:02:49,359
> [VolumeScannerThread(/testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data5)]
> INFO datanode.VolumeScanner (VolumeScanner.java:run(636)) -
> VolumeScanner(/testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data5,
> DS-3f87b7ae-9ac7-4485-885e-fc9d2475df21) exiting.
> 2017-08-30 18:02:49,366 [main] INFO handler.ContextHandler
> (ContextHandler.java:doStop(910)) - Stopped
> o.e.j.w.WebAppContext@36c0d0bd{/,null,UNAVAILABLE}{/datanode}
> 2017-08-30 18:02:49,368 [main] INFO server.AbstractConnector
> (AbstractConnector.java:doStop(318)) - Stopped
> ServerConnector@5e1fc2aa{HTTP/1.1,[http/1.1]}{localhost:0}
> 2017-08-30 18:02:49,368 [main] INFO handler.ContextHandler
> (ContextHandler.java:doStop(910)) - Stopped
> o.e.j.s.ServletContextHandler@615e3f51{/static,file:///testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test-classes/webapps/static/,UNAVAILABLE}
> 2017-08-30 18:02:49,368 [main] INFO handler.ContextHandler
> (ContextHandler.java:doStop(910)) - Stopped
> o.e.j.s.ServletContextHandler@2e51d054{/logs,file:///testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/log/,UNAVAILABLE}
> 2017-08-30 18:02:49,372 [main] INFO ipc.Server (Server.java:stop(3074)) -
> Stopping server on 56578
> 2017-08-30 18:02:49,372 [IPC Server listener on 56578] INFO ipc.Server
> (Server.java:run(1178)) - Stopping IPC Server listener on 56578
> 2017-08-30 18:02:49,373 [IPC Server Responder] INFO ipc.Server
> (Server.java:run(1312)) - Stopping IPC Server Responder
> 2017-08-30 18:02:49,373 [BP-1775344991-172.17.0.2-1504116157738 heartbeating
> to localhost/127.0.0.1:47551] WARN datanode.IncrementalBlockReportManager
> (IncrementalBlockReportManager.java:waitTillNextIBR(160)) -
> IncrementalBlockReportManager interrupted
> 2017-08-30 18:02:49,374 [BP-1775344991-172.17.0.2-1504116157738 heartbeating
> to localhost/127.0.0.1:47551] WARN datanode.DataNode
> (BPServiceActor.java:run(852)) - Ending block pool service for: Block pool
> BP-1775344991-172.17.0.2-1504116157738 (Datanode Uuid
> 303e97cb-9549-4084-8530-e6f511ebcbe1) service to localhost/127.0.0.1:47551
> 2017-08-30 18:02:49,374 [BP-1775344991-172.17.0.2-1504116157738 heartbeating
> to localhost/127.0.0.1:47551] INFO datanode.DataNode
> (BlockPoolManager.java:remove(102)) - Removed Block pool
> BP-1775344991-172.17.0.2-1504116157738 (Datanode Uuid
> 303e97cb-9549-4084-8530-e6f511ebcbe1)
> 2017-08-30 18:02:49,374 [BP-1775344991-172.17.0.2-1504116157738 heartbeating
> to localhost/127.0.0.1:47551] INFO impl.FsDatasetImpl
> (FsDatasetImpl.java:shutdownBlockPool(2675)) - Removing block pool
> BP-1775344991-172.17.0.2-1504116157738
> 2017-08-30 18:02:49,375
> [refreshUsed-/testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data5/current/BP-1775344991-172.17.0.2-1504116157738]
> WARN fs.CachingGetSpaceUsed (CachingGetSpaceUsed.java:run(183)) - Thread
> Interrupted waiting to refresh disk information: sleep interrupted
> 2017-08-30 18:02:49,375
> [refreshUsed-/testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data6/current/BP-1775344991-172.17.0.2-1504116157738]
> WARN fs.CachingGetSpaceUsed (CachingGetSpaceUsed.java:run(183)) - Thread
> Interrupted waiting to refresh disk information: sleep interrupted
> 2017-08-30 18:02:49,375 [main] INFO impl.FsDatasetAsyncDiskService
> (FsDatasetAsyncDiskService.java:shutdown(191)) - Shutting down all async disk
> service threads
> 2017-08-30 18:02:49,376 [main] INFO impl.FsDatasetAsyncDiskService
> (FsDatasetAsyncDiskService.java:shutdown(199)) - All async disk service
> threads have been shut down
> 2017-08-30 18:02:49,376 [main] INFO impl.RamDiskAsyncLazyPersistService
> (RamDiskAsyncLazyPersistService.java:shutdown(176)) - Shutting down all async
> lazy persist service threads
> 2017-08-30 18:02:49,376 [main] INFO impl.RamDiskAsyncLazyPersistService
> (RamDiskAsyncLazyPersistService.java:shutdown(183)) - All async lazy persist
> service threads have been shut down
> 2017-08-30 18:02:49,377 [main] INFO datanode.DataNode
> (DataNode.java:shutdown(2109)) - Shutdown complete.
> 2017-08-30 18:02:49,377 [main] INFO hdfs.MiniDFSCluster
> (MiniDFSCluster.java:shutdownDataNode(2005)) - Shutting down DataNode 1
> 2017-08-30 18:02:49,377 [main] WARN datanode.DirectoryScanner
> (DirectoryScanner.java:shutdown(340)) - DirectoryScanner: shutdown has been
> called
> 2017-08-30 18:02:49,377
> [org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@44c13103] INFO
> datanode.DataNode (DataXceiverServer.java:closeAllPeers(281)) - Closing all
> peers.
> 2017-08-30 18:02:49,377
> [VolumeScannerThread(/testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data3)]
> INFO datanode.VolumeScanner (VolumeScanner.java:run(636)) -
> VolumeScanner(/testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data3,
> DS-274e8cc9-280b-4370-b494-6a4f0d67ccf4) exiting.
> 2017-08-30 18:02:49,378
> [VolumeScannerThread(/testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data4)]
> INFO datanode.VolumeScanner (VolumeScanner.java:run(636)) -
> VolumeScanner(/testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data4,
> DS-f402c2a1-2faf-489f-9a91-673666976433) exiting.
> 2017-08-30 18:02:49,382 [main] INFO handler.ContextHandler
> (ContextHandler.java:doStop(910)) - Stopped
> o.e.j.w.WebAppContext@42b28ff1{/,null,UNAVAILABLE}{/datanode}
> 2017-08-30 18:02:49,383 [main] INFO server.AbstractConnector
> (AbstractConnector.java:doStop(318)) - Stopped
> ServerConnector@36061cf3{HTTP/1.1,[http/1.1]}{localhost:0}
> 2017-08-30 18:02:49,384 [main] INFO handler.ContextHandler
> (ContextHandler.java:doStop(910)) - Stopped
> o.e.j.s.ServletContextHandler@481e91b6{/static,file:///testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test-classes/webapps/static/,UNAVAILABLE}
> 2017-08-30 18:02:49,384 [main] INFO handler.ContextHandler
> (ContextHandler.java:doStop(910)) - Stopped
> o.e.j.s.ServletContextHandler@6631cb64{/logs,file:///testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/log/,UNAVAILABLE}
> 2017-08-30 18:02:49,385 [main] INFO ipc.Server (Server.java:stop(3074)) -
> Stopping server on 50140
> 2017-08-30 18:02:49,386 [IPC Server listener on 50140] INFO ipc.Server
> (Server.java:run(1178)) - Stopping IPC Server listener on 50140
> 2017-08-30 18:02:49,386 [IPC Server Responder] INFO ipc.Server
> (Server.java:run(1312)) - Stopping IPC Server Responder
> 2017-08-30 18:02:49,386 [BP-1775344991-172.17.0.2-1504116157738 heartbeating
> to localhost/127.0.0.1:47551] WARN datanode.IncrementalBlockReportManager
> (IncrementalBlockReportManager.java:waitTillNextIBR(160)) -
> IncrementalBlockReportManager interrupted
> 2017-08-30 18:02:49,387 [BP-1775344991-172.17.0.2-1504116157738 heartbeating
> to localhost/127.0.0.1:47551] WARN datanode.DataNode
> (BPServiceActor.java:run(852)) - Ending block pool service for: Block pool
> BP-1775344991-172.17.0.2-1504116157738 (Datanode Uuid
> 0bd5d006-5456-41c3-b8c3-c9c2a666a83a) service to localhost/127.0.0.1:47551
> 2017-08-30 18:02:49,387 [BP-1775344991-172.17.0.2-1504116157738 heartbeating
> to localhost/127.0.0.1:47551] INFO datanode.DataNode
> (BlockPoolManager.java:remove(102)) - Removed Block pool
> BP-1775344991-172.17.0.2-1504116157738 (Datanode Uuid
> 0bd5d006-5456-41c3-b8c3-c9c2a666a83a)
> 2017-08-30 18:02:49,387 [BP-1775344991-172.17.0.2-1504116157738 heartbeating
> to localhost/127.0.0.1:47551] INFO impl.FsDatasetImpl
> (FsDatasetImpl.java:shutdownBlockPool(2675)) - Removing block pool
> BP-1775344991-172.17.0.2-1504116157738
> 2017-08-30 18:02:49,388
> [refreshUsed-/testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data3/current/BP-1775344991-172.17.0.2-1504116157738]
> WARN fs.CachingGetSpaceUsed (CachingGetSpaceUsed.java:run(183)) - Thread
> Interrupted waiting to refresh disk information: sleep interrupted
> 2017-08-30 18:02:49,388
> [refreshUsed-/testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data4/current/BP-1775344991-172.17.0.2-1504116157738]
> WARN fs.CachingGetSpaceUsed (CachingGetSpaceUsed.java:run(183)) - Thread
> Interrupted waiting to refresh disk information: sleep interrupted
> 2017-08-30 18:02:49,389 [main] INFO impl.FsDatasetAsyncDiskService
> (FsDatasetAsyncDiskService.java:shutdown(191)) - Shutting down all async disk
> service threads
> 2017-08-30 18:02:49,389 [main] INFO impl.FsDatasetAsyncDiskService
> (FsDatasetAsyncDiskService.java:shutdown(199)) - All async disk service
> threads have been shut down
> 2017-08-30 18:02:49,389 [main] INFO impl.RamDiskAsyncLazyPersistService
> (RamDiskAsyncLazyPersistService.java:shutdown(176)) - Shutting down all async
> lazy persist service threads
> 2017-08-30 18:02:49,389 [main] INFO impl.RamDiskAsyncLazyPersistService
> (RamDiskAsyncLazyPersistService.java:shutdown(183)) - All async lazy persist
> service threads have been shut down
> 2017-08-30 18:02:49,390 [main] INFO datanode.DataNode
> (DataNode.java:shutdown(2109)) - Shutdown complete.
> 2017-08-30 18:02:49,390 [main] INFO hdfs.MiniDFSCluster
> (MiniDFSCluster.java:shutdownDataNode(2005)) - Shutting down DataNode 0
> 2017-08-30 18:02:49,390 [main] WARN datanode.DirectoryScanner
> (DirectoryScanner.java:shutdown(340)) - DirectoryScanner: shutdown has been
> called
> 2017-08-30 18:02:49,390
> [org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@3e17a0a1] INFO
> datanode.DataNode (DataXceiverServer.java:closeAllPeers(281)) - Closing all
> peers.
> 2017-08-30 18:02:49,391
> [VolumeScannerThread(/testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data2)]
> INFO datanode.VolumeScanner (VolumeScanner.java:run(636)) -
> VolumeScanner(/testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data2,
> DS-230ffa8a-d07d-4ac0-86e2-9acc7403dc9d) exiting.
> 2017-08-30 18:02:49,391
> [VolumeScannerThread(/testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data1)]
> INFO datanode.VolumeScanner (VolumeScanner.java:run(636)) -
> VolumeScanner(/testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data1,
> DS-c7f21b9d-2106-4d12-bb20-c414e1397d77) exiting.
> 2017-08-30 18:02:49,398 [main] INFO handler.ContextHandler
> (ContextHandler.java:doStop(910)) - Stopped
> o.e.j.w.WebAppContext@29eda4f8{/,null,UNAVAILABLE}{/datanode}
> 2017-08-30 18:02:49,399 [main] INFO server.AbstractConnector
> (AbstractConnector.java:doStop(318)) - Stopped
> ServerConnector@1b9776f5{HTTP/1.1,[http/1.1]}{localhost:0}
> 2017-08-30 18:02:49,400 [main] INFO handler.ContextHandler
> (ContextHandler.java:doStop(910)) - Stopped
> o.e.j.s.ServletContextHandler@7a7cc52c{/static,file:///testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test-classes/webapps/static/,UNAVAILABLE}
> 2017-08-30 18:02:49,400 [main] INFO handler.ContextHandler
> (ContextHandler.java:doStop(910)) - Stopped
> o.e.j.s.ServletContextHandler@4e1ce44{/logs,file:///testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/log/,UNAVAILABLE}
> 2017-08-30 18:02:49,402 [main] INFO ipc.Server (Server.java:stop(3074)) -
> Stopping server on 58248
> 2017-08-30 18:02:49,403 [IPC Server listener on 58248] INFO ipc.Server
> (Server.java:run(1178)) - Stopping IPC Server listener on 58248
> 2017-08-30 18:02:49,403 [IPC Server Responder] INFO ipc.Server
> (Server.java:run(1312)) - Stopping IPC Server Responder
> 2017-08-30 18:02:49,403 [BP-1775344991-172.17.0.2-1504116157738 heartbeating
> to localhost/127.0.0.1:47551] WARN datanode.IncrementalBlockReportManager
> (IncrementalBlockReportManager.java:waitTillNextIBR(160)) -
> IncrementalBlockReportManager interrupted
> 2017-08-30 18:02:49,404 [BP-1775344991-172.17.0.2-1504116157738 heartbeating
> to localhost/127.0.0.1:47551] WARN datanode.DataNode
> (BPServiceActor.java:run(852)) - Ending block pool service for: Block pool
> BP-1775344991-172.17.0.2-1504116157738 (Datanode Uuid
> d7ee1847-d046-47ea-9a05-7fd25bcb1e86) service to localhost/127.0.0.1:47551
> 2017-08-30 18:02:49,404 [BP-1775344991-172.17.0.2-1504116157738 heartbeating
> to localhost/127.0.0.1:47551] INFO datanode.DataNode
> (BlockPoolManager.java:remove(102)) - Removed Block pool
> BP-1775344991-172.17.0.2-1504116157738 (Datanode Uuid
> d7ee1847-d046-47ea-9a05-7fd25bcb1e86)
> 2017-08-30 18:02:49,404 [BP-1775344991-172.17.0.2-1504116157738 heartbeating
> to localhost/127.0.0.1:47551] INFO impl.FsDatasetImpl
> (FsDatasetImpl.java:shutdownBlockPool(2675)) - Removing block pool
> BP-1775344991-172.17.0.2-1504116157738
> 2017-08-30 18:02:49,405
> [refreshUsed-/testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data1/current/BP-1775344991-172.17.0.2-1504116157738]
> WARN fs.CachingGetSpaceUsed (CachingGetSpaceUsed.java:run(183)) - Thread
> Interrupted waiting to refresh disk information: sleep interrupted
> 2017-08-30 18:02:49,405
> [refreshUsed-/testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/data/data2/current/BP-1775344991-172.17.0.2-1504116157738]
> WARN fs.CachingGetSpaceUsed (CachingGetSpaceUsed.java:run(183)) - Thread
> Interrupted waiting to refresh disk information: sleep interrupted
> 2017-08-30 18:02:49,406 [main] INFO impl.FsDatasetAsyncDiskService
> (FsDatasetAsyncDiskService.java:shutdown(191)) - Shutting down all async disk
> service threads
> 2017-08-30 18:02:49,406 [main] INFO impl.FsDatasetAsyncDiskService
> (FsDatasetAsyncDiskService.java:shutdown(199)) - All async disk service
> threads have been shut down
> 2017-08-30 18:02:49,406 [main] INFO impl.RamDiskAsyncLazyPersistService
> (RamDiskAsyncLazyPersistService.java:shutdown(176)) - Shutting down all async
> lazy persist service threads
> 2017-08-30 18:02:49,407 [main] INFO impl.RamDiskAsyncLazyPersistService
> (RamDiskAsyncLazyPersistService.java:shutdown(183)) - All async lazy persist
> service threads have been shut down
> 2017-08-30 18:02:49,407 [main] INFO datanode.DataNode
> (DataNode.java:shutdown(2109)) - Shutdown complete.
> 2017-08-30 18:02:49,407 [main] INFO hdfs.MiniDFSCluster
> (MiniDFSCluster.java:stopAndJoinNameNode(2038)) - Shutting down the namenode
> 2017-08-30 18:02:49,407 [main] INFO namenode.FSNamesystem
> (FSNamesystem.java:stopActiveServices(1301)) - Stopping services started for
> active state
> 2017-08-30 18:02:49,408 [main] INFO namenode.FSEditLog
> (FSEditLog.java:endCurrentLogSegment(1369)) - Ending log segment 1, 8
> 2017-08-30 18:02:49,408
> [org.apache.hadoop.hdfs.server.namenode.FSNamesystem$LazyPersistFileScrubber@ef1695a]
> INFO namenode.FSNamesystem (FSNamesystem.java:run(3985)) -
> LazyPersistFileScrubber was interrupted, exiting
> 2017-08-30 18:02:49,408
> [org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeEditLogRoller@238bfd6c]
> INFO namenode.FSNamesystem (FSNamesystem.java:run(3894)) -
> NameNodeEditLogRoller was interrupted, exiting
> 2017-08-30 18:02:49,408 [main] INFO namenode.FSEditLog
> (FSEditLog.java:logSyncAll(600)) - logSyncAll toSyncToTxId=9 lastSyncedTxid=8
> mostRecentTxid=9
> 2017-08-30 18:02:49,408 [main] INFO namenode.FSEditLog
> (FSEditLog.java:logSyncAll(605)) - Done logSyncAll lastWrittenTxId=9
> lastSyncedTxid=9 mostRecentTxid=9
> 2017-08-30 18:02:49,408 [main] INFO namenode.FSEditLog
> (FSEditLog.java:printStatistics(770)) - Number of transactions: 9 Total time
> for transactions(ms): 3 Number of transactions batched in Syncs: 4 Number of
> syncs: 6 SyncTimes(ms): 2 1
> 2017-08-30 18:02:49,409 [main] INFO namenode.FileJournalManager
> (FileJournalManager.java:finalizeLogSegment(143)) - Finalizing edits file
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/name-0-1/current/edits_inprogress_0000000000000000001
> ->
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/name-0-1/current/edits_0000000000000000001-0000000000000000009
> 2017-08-30 18:02:49,410 [main] INFO namenode.FileJournalManager
> (FileJournalManager.java:finalizeLogSegment(143)) - Finalizing edits file
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/name-0-2/current/edits_inprogress_0000000000000000001
> ->
> /testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/3/dfs/name-0-2/current/edits_0000000000000000001-0000000000000000009
> 2017-08-30 18:02:49,410 [CacheReplicationMonitor(1864849826)] INFO
> blockmanagement.CacheReplicationMonitor
> (CacheReplicationMonitor.java:run(169)) - Shutting down
> CacheReplicationMonitor
> 2017-08-30 18:02:49,411 [main] INFO ipc.Server (Server.java:stop(3074)) -
> Stopping server on 47551
> 2017-08-30 18:02:49,412 [IPC Server listener on 47551] INFO ipc.Server
> (Server.java:run(1178)) - Stopping IPC Server listener on 47551
> 2017-08-30 18:02:49,414 [IPC Server Responder] INFO ipc.Server
> (Server.java:run(1312)) - Stopping IPC Server Responder
> 2017-08-30 18:02:49,415 [StorageInfoMonitor] INFO
> blockmanagement.BlockManager (BlockManager.java:run(4450)) - Stopping thread.
> 2017-08-30 18:02:49,415 [RedundancyMonitor] INFO
> blockmanagement.BlockManager (BlockManager.java:run(4415)) - Stopping
> RedundancyMonitor.
> 2017-08-30 18:02:49,422 [main] INFO namenode.FSNamesystem
> (FSNamesystem.java:stopActiveServices(1301)) - Stopping services started for
> active state
> 2017-08-30 18:02:49,423 [main] INFO namenode.FSNamesystem
> (FSNamesystem.java:stopStandbyServices(1397)) - Stopping services started for
> standby state
> 2017-08-30 18:02:49,424 [main] INFO handler.ContextHandler
> (ContextHandler.java:doStop(910)) - Stopped
> o.e.j.w.WebAppContext@5633dafd{/,null,UNAVAILABLE}{/hdfs}
> 2017-08-30 18:02:49,425 [main] INFO server.AbstractConnector
> (AbstractConnector.java:doStop(318)) - Stopped
> ServerConnector@5d5160e6{HTTP/1.1,[http/1.1]}{localhost:0}
> 2017-08-30 18:02:49,425 [main] INFO handler.ContextHandler
> (ContextHandler.java:doStop(910)) - Stopped
> o.e.j.s.ServletContextHandler@5e8cda75{/static,file:///testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test-classes/webapps/static/,UNAVAILABLE}
> 2017-08-30 18:02:49,425 [main] INFO handler.ContextHandler
> (ContextHandler.java:doStop(910)) - Stopped
> o.e.j.s.ServletContextHandler@46c3a14d{/logs,file:///testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/log/,UNAVAILABLE}
> 2017-08-30 18:02:49,426 [main] INFO impl.MetricsSystemImpl
> (MetricsSystemImpl.java:stop(210)) - Stopping DataNode metrics system...
> 2017-08-30 18:02:49,428 [main] INFO impl.MetricsSystemImpl
> (MetricsSystemImpl.java:stop(216)) - DataNode metrics system stopped.
> 2017-08-30 18:02:49,428 [main] INFO impl.MetricsSystemImpl
> (MetricsSystemImpl.java:shutdown(607)) - DataNode metrics system shutdown
> complete.
> {noformat}
--
This message was sent by Atlassian JIRA
(v6.4.14#64029)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]