[jira] [Commented] (HDFS-10609) Uncaught InvalidEncryptionKeyException during pipeline recovery may abort downstream applications
[ https://issues.apache.org/jira/browse/HDFS-10609?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15550965#comment-15550965 ] Akira Ajisaka commented on HDFS-10609: -- Hi [~jojochuang], would you update CHANGES.txt when cherry-picking a commit to branch-2.7/2.6? HADOOP-13670 will fix this, so you don't need to update CHANGES.txt for now. > Uncaught InvalidEncryptionKeyException during pipeline recovery may abort > downstream applications > - > > Key: HDFS-10609 > URL: https://issues.apache.org/jira/browse/HDFS-10609 > Project: Hadoop HDFS > Issue Type: Bug > Components: encryption >Affects Versions: 2.6.0 > Environment: CDH5.8.0 >Reporter: Wei-Chiu Chuang >Assignee: Wei-Chiu Chuang > Fix For: 2.8.0, 2.7.4, 3.0.0-alpha2 > > Attachments: HDFS-10609.001.patch, HDFS-10609.002.patch, > HDFS-10609.003.patch, HDFS-10609.004.patch, HDFS-10609.005.patch, > HDFS-10609.branch-2.7.patch > > > In normal operations, if SASL negotiation fails due to > {{InvalidEncryptionKeyException}}, it is typically a benign exception, which > is caught and retried : > {code:title=SaslDataTransferServer#doSaslHandshake} > if (ioe instanceof SaslException && > ioe.getCause() != null && > ioe.getCause() instanceof InvalidEncryptionKeyException) { > // This could just be because the client is long-lived and hasn't gotten > // a new encryption key from the NN in a while. Upon receiving this > // error, the client will get a new encryption key from the NN and retry > // connecting to this DN. > sendInvalidKeySaslErrorMessage(out, ioe.getCause().getMessage()); > } > {code} > {code:title=DFSOutputStream.DataStreamer#createBlockOutputStream} > if (ie instanceof InvalidEncryptionKeyException && refetchEncryptionKey > 0) { > DFSClient.LOG.info("Will fetch a new encryption key and retry, " > + "encryption key was invalid when connecting to " > + nodes[0] + " : " + ie); > {code} > However, if the exception is thrown during pipeline recovery, the > corresponding code does not handle it properly, and the exception is spilled > out to downstream applications, such as SOLR, aborting its operation: > {quote} > 2016-07-06 12:12:51,992 ERROR org.apache.solr.update.HdfsTransactionLog: > Exception closing tlog. > org.apache.hadoop.hdfs.protocol.datatransfer.InvalidEncryptionKeyException: > Can't re-compute encryption key for nonce, since the required block key > (keyID=557709482) doesn't exist. Current key: 1350592619 > at > org.apache.hadoop.hdfs.protocol.datatransfer.sasl.DataTransferSaslUtil.readSaslMessageAndNegotiatedCipherOption(DataTransferSaslUtil.java:417) > at > org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient.doSaslHandshake(SaslDataTransferClient.java:474) > at > org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient.getEncryptedStreams(SaslDataTransferClient.java:299) > at > org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient.send(SaslDataTransferClient.java:242) > at > org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient.checkTrustAndSend(SaslDataTransferClient.java:211) > at > org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient.socketSend(SaslDataTransferClient.java:183) > at > org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.transfer(DFSOutputStream.java:1308) > at > org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.addDatanode2ExistingPipeline(DFSOutputStream.java:1272) > at > org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.setupPipelineForAppendOrRecovery(DFSOutputStream.java:1433) > at > org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.processDatanodeError(DFSOutputStream.java:1147) > at > org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:632) > 2016-07-06 12:12:51,997 ERROR org.apache.solr.update.CommitTracker: auto > commit error...:org.apache.solr.common.SolrException: > org.apache.hadoop.hdfs.protocol.datatransfer.InvalidEncryptionKeyException: > Can't re-compute encryption key for nonce, since the required block key > (keyID=557709482) doesn't exist. Current key: 1350592619 > at > org.apache.solr.update.HdfsTransactionLog.close(HdfsTransactionLog.java:316) > at > org.apache.solr.update.TransactionLog.decref(TransactionLog.java:505) > at org.apache.solr.update.UpdateLog.addOldLog(UpdateLog.java:380) > at org.apache.solr.update.UpdateLog.postCommit(UpdateLog.java:676) > at > org.apache.solr.update.DirectUpdateHandler2.commit(DirectUpdateHandler2.java:623) > at
[jira] [Commented] (HDFS-10965) Add unit test for HDFS command 'dfsadmin -printTopology'
[ https://issues.apache.org/jira/browse/HDFS-10965?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15550943#comment-15550943 ] Hadoop QA commented on HDFS-10965: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 21s{color} | {color:blue} Docker mode activated. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 1 new or modified test files. {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 8m 49s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 0s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 33s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 10s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 14s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 57s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 56s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 47s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 45s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 45s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 26s{color} | {color:green} hadoop-hdfs-project/hadoop-hdfs: The patch generated 0 new + 167 unchanged - 34 fixed = 167 total (was 201) {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 52s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 9s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 50s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 0s{color} | {color:green} the patch passed {color} | | {color:red}-1{color} | {color:red} unit {color} | {color:red} 86m 20s{color} | {color:red} hadoop-hdfs in the patch failed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 17s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black}108m 47s{color} | {color:black} {color} | \\ \\ || Reason || Tests || | Failed junit tests | hadoop.hdfs.server.namenode.TestAddStripedBlockInFBR | | Timed out junit tests | org.apache.hadoop.hdfs.TestLeaseRecovery2 | \\ \\ || Subsystem || Report/Notes || | Docker | Image:yetus/hadoop:9560f25 | | JIRA Issue | HDFS-10965 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12831881/HDFS-10965.002.patch | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite unit findbugs checkstyle | | uname | Linux 11a0b464d44a 3.13.0-96-generic #143-Ubuntu SMP Mon Aug 29 20:15:20 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh | | git revision | trunk / c5ca216 | | Default Java | 1.8.0_101 | | findbugs | v3.0.0 | | unit | https://builds.apache.org/job/PreCommit-HDFS-Build/17036/artifact/patchprocess/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt | | Test Results | https://builds.apache.org/job/PreCommit-HDFS-Build/17036/testReport/ | | modules | C: hadoop-hdfs-project/hadoop-hdfs U: hadoop-hdfs-project/hadoop-hdfs | | Console output | https://builds.apache.org/job/PreCommit-HDFS-Build/17036/console | | Powered by | Apache Yetus 0.4.0-SNAPSHOT http://yetus.apache.org | This message was automatically generated. > Add unit test for HDFS command 'dfsadmin -printTopology' > > > Key: HDFS-10965 > URL: https://issues.apache.org/jira/browse/HDFS-10965 > Project:
[jira] [Updated] (HDFS-10969) Fix typos in hdfs-default.xml
[ https://issues.apache.org/jira/browse/HDFS-10969?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Yiqun Lin updated HDFS-10969: - Attachment: HDFS-10969.001.patch > Fix typos in hdfs-default.xml > - > > Key: HDFS-10969 > URL: https://issues.apache.org/jira/browse/HDFS-10969 > Project: Hadoop HDFS > Issue Type: Bug > Components: documentation >Reporter: Yiqun Lin >Assignee: Yiqun Lin >Priority: Minor > Attachments: HDFS-10969.001.patch > > > There are three typos in file {{hdfs-default.xml}}: > * In {{dfs.datanode.transferTo.allowed}}: tranfers->transfers > * In {{dfs.block.local-path-access.user}}: allowd->allowed > * In {{dfs.datanode.directoryscan.throttle.limit.ms.per.sec}}: > disbled->disabled -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-10969) Fix typos in hdfs-default.xml
[ https://issues.apache.org/jira/browse/HDFS-10969?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Yiqun Lin updated HDFS-10969: - Status: Patch Available (was: Open) Attach a simple patch to make a quick fix. > Fix typos in hdfs-default.xml > - > > Key: HDFS-10969 > URL: https://issues.apache.org/jira/browse/HDFS-10969 > Project: Hadoop HDFS > Issue Type: Bug > Components: documentation >Reporter: Yiqun Lin >Assignee: Yiqun Lin >Priority: Minor > > There are three typos in file {{hdfs-default.xml}}: > * In {{dfs.datanode.transferTo.allowed}}: tranfers->transfers > * In {{dfs.block.local-path-access.user}}: allowd->allowed > * In {{dfs.datanode.directoryscan.throttle.limit.ms.per.sec}}: > disbled->disabled -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-10629) Federation Router
[ https://issues.apache.org/jira/browse/HDFS-10629?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15550914#comment-15550914 ] Hadoop QA commented on HDFS-10629: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 10s{color} | {color:blue} Docker mode activated. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 7 new or modified test files. {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 6m 49s{color} | {color:green} HDFS-10467 passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 43s{color} | {color:green} HDFS-10467 passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 30s{color} | {color:green} HDFS-10467 passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 51s{color} | {color:green} HDFS-10467 passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 11s{color} | {color:green} HDFS-10467 passed {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 43s{color} | {color:green} HDFS-10467 passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 56s{color} | {color:green} HDFS-10467 passed {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 48s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 43s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 43s{color} | {color:green} the patch passed {color} | | {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange} 0m 27s{color} | {color:orange} hadoop-hdfs-project/hadoop-hdfs: The patch generated 4 new + 394 unchanged - 0 fixed = 398 total (was 394) {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 51s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 9s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} xml {color} | {color:green} 0m 2s{color} | {color:green} The patch has no ill-formed XML file. {color} | | {color:red}-1{color} | {color:red} findbugs {color} | {color:red} 1m 53s{color} | {color:red} hadoop-hdfs-project/hadoop-hdfs generated 5 new + 0 unchanged - 0 fixed = 5 total (was 0) {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 52s{color} | {color:green} the patch passed {color} | | {color:red}-1{color} | {color:red} unit {color} | {color:red} 61m 13s{color} | {color:red} hadoop-hdfs in the patch failed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 17s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 80m 26s{color} | {color:black} {color} | \\ \\ || Reason || Tests || | FindBugs | module:hadoop-hdfs-project/hadoop-hdfs | | | Sequence of calls to java.util.concurrent.ConcurrentHashMap may not be atomic in org.apache.hadoop.hdfs.server.federation.router.ConnectionManager.getConnection(UserGroupInformation, String) At ConnectionManager.java:may not be atomic in org.apache.hadoop.hdfs.server.federation.router.ConnectionManager.getConnection(UserGroupInformation, String) At ConnectionManager.java:[line 297] | | | Uninitialized read of minSize in new org.apache.hadoop.hdfs.server.federation.router.ConnectionManager(Configuration, int, int) At ConnectionManager.java:new org.apache.hadoop.hdfs.server.federation.router.ConnectionManager(Configuration, int, int) At ConnectionManager.java:[line 93] | | | org.apache.hadoop.hdfs.server.federation.router.Router.initAndStartRouter(Configuration, boolean) invokes System.exit(...), which shuts down the entire virtual machine At Router.java:shuts down the entire virtual machine At Router.java:[line 123] | | | instanceof will always return true for all non-null values in org.apache.hadoop.hdfs.server.federation.router.RouterRpcClient.invokeMethod(List, Method, Object[]), since all java.io.IOException are instances of java.io.IOException At RouterRpcClient.java:for all non-null
[jira] [Commented] (HDFS-10949) DiskBalancer: deprecate TestDiskBalancer#setVolumeCapacity
[ https://issues.apache.org/jira/browse/HDFS-10949?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15550909#comment-15550909 ] Hadoop QA commented on HDFS-10949: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 14s{color} | {color:blue} Docker mode activated. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 1 new or modified test files. {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 8m 5s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 46s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 27s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 53s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 14s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 51s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 58s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 57s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 48s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 48s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 23s{color} | {color:green} hadoop-hdfs-project/hadoop-hdfs: The patch generated 0 new + 6 unchanged - 1 fixed = 6 total (was 7) {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 54s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 10s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 58s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 2s{color} | {color:green} the patch passed {color} | | {color:red}-1{color} | {color:red} unit {color} | {color:red} 64m 47s{color} | {color:red} hadoop-hdfs in the patch failed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 19s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 86m 8s{color} | {color:black} {color} | \\ \\ || Reason || Tests || | Failed junit tests | hadoop.hdfs.server.blockmanagement.TestRBWBlockInvalidation | \\ \\ || Subsystem || Report/Notes || | Docker | Image:yetus/hadoop:9560f25 | | JIRA Issue | HDFS-10949 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12831882/HDFS-10949.001.patch | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite unit findbugs checkstyle | | uname | Linux 5cb2ac906b4e 3.13.0-92-generic #139-Ubuntu SMP Tue Jun 28 20:42:26 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh | | git revision | trunk / c5ca216 | | Default Java | 1.8.0_101 | | findbugs | v3.0.0 | | unit | https://builds.apache.org/job/PreCommit-HDFS-Build/17037/artifact/patchprocess/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt | | Test Results | https://builds.apache.org/job/PreCommit-HDFS-Build/17037/testReport/ | | modules | C: hadoop-hdfs-project/hadoop-hdfs U: hadoop-hdfs-project/hadoop-hdfs | | Console output | https://builds.apache.org/job/PreCommit-HDFS-Build/17037/console | | Powered by | Apache Yetus 0.4.0-SNAPSHOT http://yetus.apache.org | This message was automatically generated. > DiskBalancer: deprecate TestDiskBalancer#setVolumeCapacity > -- > > Key: HDFS-10949 > URL: https://issues.apache.org/jira/browse/HDFS-10949 > Project: Hadoop HDFS > Issue Type: Sub-task > Components:
[jira] [Updated] (HDFS-10967) Add configuration for BlockPlacementPolicy to avoid near-full DataNodes
[ https://issues.apache.org/jira/browse/HDFS-10967?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Zhe Zhang updated HDFS-10967: - Description: Large production clusters are likely to have heterogeneous nodes in terms of storage capacity, memory, and CPU cores. It is not always possible to proportionally ingest data into DataNodes based on their remaining storage capacity. Therefore it's possible for a subset of DataNodes to be much closer to full capacity than the rest. Notice that this heterogeneity is most likely rack-by-rack -- i.e. _m_ whole racks with low-storage nodes and _n_ whole racks with high-storage nodes. So It'd be very useful if we can lower the chance for those near-full DataNodes to become destinations for the 2nd and 3rd replicas. was: Large production clusters are likely to have heterogeneous nodes in terms of storage capacity, memory, and CPU cores. It is not always possible to proportionally ingest data into DataNodes based on their remaining storage capacity. Therefore it's possible for a subset of DataNodes to be much closer to full capacity than the rest. Notice that this heterogeneity is most likely rack-by-rack -- i.e. _m_ whole racks with low-storage nodes and _n_ whole racks with high-storage nodes. So It'd be very useful if we can deprioritize those near-full DataNodes as destinations for the 2nd and 3rd replicas. > Add configuration for BlockPlacementPolicy to avoid near-full DataNodes > --- > > Key: HDFS-10967 > URL: https://issues.apache.org/jira/browse/HDFS-10967 > Project: Hadoop HDFS > Issue Type: Improvement > Components: namenode >Reporter: Zhe Zhang > Labels: balancer > > Large production clusters are likely to have heterogeneous nodes in terms of > storage capacity, memory, and CPU cores. It is not always possible to > proportionally ingest data into DataNodes based on their remaining storage > capacity. Therefore it's possible for a subset of DataNodes to be much closer > to full capacity than the rest. > Notice that this heterogeneity is most likely rack-by-rack -- i.e. _m_ whole > racks with low-storage nodes and _n_ whole racks with high-storage nodes. So > It'd be very useful if we can lower the chance for those near-full DataNodes > to become destinations for the 2nd and 3rd replicas. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-10967) Add configuration for BlockPlacementPolicy to avoid near-full DataNodes
[ https://issues.apache.org/jira/browse/HDFS-10967?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Zhe Zhang updated HDFS-10967: - Summary: Add configuration for BlockPlacementPolicy to avoid near-full DataNodes (was: Add configuration for BlockPlacementPolicy to deprioritize near-full DataNodes) > Add configuration for BlockPlacementPolicy to avoid near-full DataNodes > --- > > Key: HDFS-10967 > URL: https://issues.apache.org/jira/browse/HDFS-10967 > Project: Hadoop HDFS > Issue Type: Improvement > Components: namenode >Reporter: Zhe Zhang > Labels: balancer > > Large production clusters are likely to have heterogeneous nodes in terms of > storage capacity, memory, and CPU cores. It is not always possible to > proportionally ingest data into DataNodes based on their remaining storage > capacity. Therefore it's possible for a subset of DataNodes to be much closer > to full capacity than the rest. > Notice that this heterogeneity is most likely rack-by-rack -- i.e. _m_ whole > racks with low-storage nodes and _n_ whole racks with high-storage nodes. So > It'd be very useful if we can deprioritize those near-full DataNodes as > destinations for the 2nd and 3rd replicas. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-10629) Federation Router
[ https://issues.apache.org/jira/browse/HDFS-10629?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jason Kace updated HDFS-10629: -- Attachment: HDFS-10629-HDFS-10467-006.patch Updated patch: # Minor fixes and style updates # Moving config keys to to DFSConfigKeys.java > Federation Router > - > > Key: HDFS-10629 > URL: https://issues.apache.org/jira/browse/HDFS-10629 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: hdfs >Reporter: Inigo Goiri >Assignee: Jason Kace > Attachments: HDFS-10629-HDFS-10467-002.patch, > HDFS-10629-HDFS-10467-003.patch, HDFS-10629-HDFS-10467-004.patch, > HDFS-10629-HDFS-10467-005.patch, HDFS-10629-HDFS-10467-006.patch, > HDFS-10629.000.patch, HDFS-10629.001.patch > > > Component that routes calls from the clients to the right Namespace. It > implements {{ClientProtocol}}. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-10949) DiskBalancer: deprecate TestDiskBalancer#setVolumeCapacity
[ https://issues.apache.org/jira/browse/HDFS-10949?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15550769#comment-15550769 ] Xiaobing Zhou commented on HDFS-10949: -- v001 fixed two check style issues. > DiskBalancer: deprecate TestDiskBalancer#setVolumeCapacity > -- > > Key: HDFS-10949 > URL: https://issues.apache.org/jira/browse/HDFS-10949 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: balancer & mover >Reporter: Xiaobing Zhou >Assignee: Xiaobing Zhou >Priority: Minor > Attachments: HDFS-10949.000.patch, HDFS-10949.001.patch > > > HDFS-10824 fixed amnesia of capacity setting of MiniDFSCluster, there's no > reason for existence of TestDiskBalancer#setVolumeCapacity anymore. It should > be deprecated. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-10949) DiskBalancer: deprecate TestDiskBalancer#setVolumeCapacity
[ https://issues.apache.org/jira/browse/HDFS-10949?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Xiaobing Zhou updated HDFS-10949: - Attachment: HDFS-10949.001.patch > DiskBalancer: deprecate TestDiskBalancer#setVolumeCapacity > -- > > Key: HDFS-10949 > URL: https://issues.apache.org/jira/browse/HDFS-10949 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: balancer & mover >Reporter: Xiaobing Zhou >Assignee: Xiaobing Zhou >Priority: Minor > Attachments: HDFS-10949.000.patch, HDFS-10949.001.patch > > > HDFS-10824 fixed amnesia of capacity setting of MiniDFSCluster, there's no > reason for existence of TestDiskBalancer#setVolumeCapacity anymore. It should > be deprecated. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-10965) Add unit test for HDFS command 'dfsadmin -printTopology'
[ https://issues.apache.org/jira/browse/HDFS-10965?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Xiaobing Zhou updated HDFS-10965: - Attachment: HDFS-10965.002.patch > Add unit test for HDFS command 'dfsadmin -printTopology' > > > Key: HDFS-10965 > URL: https://issues.apache.org/jira/browse/HDFS-10965 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: fs, shell, test >Reporter: Xiaobing Zhou >Assignee: Xiaobing Zhou > Attachments: HDFS-10965.000.patch, HDFS-10965.001.patch, > HDFS-10965.002.patch > > > DFSAdmin#printTopology should also be tested. This proposes adding it in > TestDFSAdmin. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-10965) Add unit test for HDFS command 'dfsadmin -printTopology'
[ https://issues.apache.org/jira/browse/HDFS-10965?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15550761#comment-15550761 ] Xiaobing Zhou commented on HDFS-10965: -- v002 fixed another two check style issues. > Add unit test for HDFS command 'dfsadmin -printTopology' > > > Key: HDFS-10965 > URL: https://issues.apache.org/jira/browse/HDFS-10965 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: fs, shell, test >Reporter: Xiaobing Zhou >Assignee: Xiaobing Zhou > Attachments: HDFS-10965.000.patch, HDFS-10965.001.patch, > HDFS-10965.002.patch > > > DFSAdmin#printTopology should also be tested. This proposes adding it in > TestDFSAdmin. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-10968) BlockManager#isNewRack should consider decommissioning nodes
[ https://issues.apache.org/jira/browse/HDFS-10968?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15550672#comment-15550672 ] Hadoop QA commented on HDFS-10968: -- | (/) *{color:green}+1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 17s{color} | {color:blue} Docker mode activated. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 1 new or modified test files. {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 7m 54s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 50s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 30s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 0s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 15s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 55s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 3s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 1m 3s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 54s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 54s{color} | {color:green} the patch passed {color} | | {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange} 0m 29s{color} | {color:orange} hadoop-hdfs-project/hadoop-hdfs: The patch generated 10 new + 130 unchanged - 4 fixed = 140 total (was 134) {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 1s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 11s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 2m 0s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 3s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} unit {color} | {color:green} 81m 26s{color} | {color:green} hadoop-hdfs in the patch passed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 18s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black}103m 47s{color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Docker | Image:yetus/hadoop:9560f25 | | JIRA Issue | HDFS-10968 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12831857/HDFS-10968.000.patch | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite unit findbugs checkstyle | | uname | Linux 97b67b60f168 3.13.0-96-generic #143-Ubuntu SMP Mon Aug 29 20:15:20 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh | | git revision | trunk / c5ca216 | | Default Java | 1.8.0_101 | | findbugs | v3.0.0 | | checkstyle | https://builds.apache.org/job/PreCommit-HDFS-Build/17035/artifact/patchprocess/diff-checkstyle-hadoop-hdfs-project_hadoop-hdfs.txt | | Test Results | https://builds.apache.org/job/PreCommit-HDFS-Build/17035/testReport/ | | modules | C: hadoop-hdfs-project/hadoop-hdfs U: hadoop-hdfs-project/hadoop-hdfs | | Console output | https://builds.apache.org/job/PreCommit-HDFS-Build/17035/console | | Powered by | Apache Yetus 0.4.0-SNAPSHOT http://yetus.apache.org | This message was automatically generated. > BlockManager#isNewRack should consider decommissioning nodes > > > Key: HDFS-10968 > URL: https://issues.apache.org/jira/browse/HDFS-10968 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: erasure-coding, namenode >Affects Versions: 3.0.0-alpha1 >
[jira] [Commented] (HDFS-10683) Make class Token$PrivateToken private
[ https://issues.apache.org/jira/browse/HDFS-10683?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15550648#comment-15550648 ] John Zhuge commented on HDFS-10683: --- Thanks [~jojochuang] for the review and commit! > Make class Token$PrivateToken private > - > > Key: HDFS-10683 > URL: https://issues.apache.org/jira/browse/HDFS-10683 > Project: Hadoop HDFS > Issue Type: Improvement >Affects Versions: 2.9.0 >Reporter: John Zhuge >Assignee: John Zhuge >Priority: Minor > Labels: fs, ha, security, security_token > Fix For: 2.9.0, 3.0.0-alpha2 > > Attachments: HDFS-10683.001.patch, HDFS-10683.002.patch > > > Avoid {{instanceof}} or typecasting of {{Toke.PrivateToken}} by introducing > an interface method in {{Token}}. Make class {{Toke.PrivateToken}} private. > Use a factory method instead. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-10881) Federation State Store Driver API
[ https://issues.apache.org/jira/browse/HDFS-10881?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15550640#comment-15550640 ] Inigo Goiri commented on HDFS-10881: We are going backwards with Jenkins... the failed tests seems spurious though. How to rerun Jenkins? > Federation State Store Driver API > - > > Key: HDFS-10881 > URL: https://issues.apache.org/jira/browse/HDFS-10881 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: fs >Reporter: Jason Kace >Assignee: Jason Kace > Attachments: HDFS-10881-HDFS-10467-001.patch, > HDFS-10881-HDFS-10467-002.patch, HDFS-10881-HDFS-10467-003.patch, > HDFS-10881-HDFS-10467-004.patch, HDFS-10881-HDFS-10467-005.patch, > HDFS-10881-HDFS-10467-006.patch, HDFS-10881-HDFS-10467-007.patch > > > The API interfaces and minimal classes required to support a state store data > backend such as ZooKeeper or a file system. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-10949) DiskBalancer: deprecate TestDiskBalancer#setVolumeCapacity
[ https://issues.apache.org/jira/browse/HDFS-10949?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15550621#comment-15550621 ] Hadoop QA commented on HDFS-10949: -- | (/) *{color:green}+1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 15s{color} | {color:blue} Docker mode activated. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 1 new or modified test files. {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 7m 20s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 46s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 25s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 54s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 11s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 42s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 57s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 50s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 44s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 44s{color} | {color:green} the patch passed {color} | | {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange} 0m 24s{color} | {color:orange} hadoop-hdfs-project/hadoop-hdfs: The patch generated 2 new + 6 unchanged - 1 fixed = 8 total (was 7) {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 48s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 9s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 48s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 54s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} unit {color} | {color:green} 83m 24s{color} | {color:green} hadoop-hdfs in the patch passed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 19s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black}103m 2s{color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Docker | Image:yetus/hadoop:9560f25 | | JIRA Issue | HDFS-10949 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12831850/HDFS-10949.000.patch | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite unit findbugs checkstyle | | uname | Linux 74de7236aa15 3.13.0-96-generic #143-Ubuntu SMP Mon Aug 29 20:15:20 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh | | git revision | trunk / e68c7b9 | | Default Java | 1.8.0_101 | | findbugs | v3.0.0 | | checkstyle | https://builds.apache.org/job/PreCommit-HDFS-Build/17033/artifact/patchprocess/diff-checkstyle-hadoop-hdfs-project_hadoop-hdfs.txt | | Test Results | https://builds.apache.org/job/PreCommit-HDFS-Build/17033/testReport/ | | modules | C: hadoop-hdfs-project/hadoop-hdfs U: hadoop-hdfs-project/hadoop-hdfs | | Console output | https://builds.apache.org/job/PreCommit-HDFS-Build/17033/console | | Powered by | Apache Yetus 0.4.0-SNAPSHOT http://yetus.apache.org | This message was automatically generated. > DiskBalancer: deprecate TestDiskBalancer#setVolumeCapacity > -- > > Key: HDFS-10949 > URL: https://issues.apache.org/jira/browse/HDFS-10949 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: balancer & mover >Reporter: Xiaobing Zhou >Assignee: Xiaobing
[jira] [Commented] (HDFS-10965) Add unit test for HDFS command 'dfsadmin -printTopology'
[ https://issues.apache.org/jira/browse/HDFS-10965?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15550598#comment-15550598 ] Hadoop QA commented on HDFS-10965: -- | (/) *{color:green}+1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 13s{color} | {color:blue} Docker mode activated. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 1 new or modified test files. {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 7m 8s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 47s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 29s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 54s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 12s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 45s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 57s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 49s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 42s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 42s{color} | {color:green} the patch passed {color} | | {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange} 0m 25s{color} | {color:orange} hadoop-hdfs-project/hadoop-hdfs: The patch generated 2 new + 168 unchanged - 34 fixed = 170 total (was 202) {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 49s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 10s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 2m 0s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 0s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} unit {color} | {color:green} 62m 17s{color} | {color:green} hadoop-hdfs in the patch passed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 18s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 82m 20s{color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Docker | Image:yetus/hadoop:9560f25 | | JIRA Issue | HDFS-10965 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12831852/HDFS-10965.001.patch | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite unit findbugs checkstyle | | uname | Linux b5e8eedaae5e 3.13.0-93-generic #140-Ubuntu SMP Mon Jul 18 21:21:05 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh | | git revision | trunk / c5ca216 | | Default Java | 1.8.0_101 | | findbugs | v3.0.0 | | checkstyle | https://builds.apache.org/job/PreCommit-HDFS-Build/17034/artifact/patchprocess/diff-checkstyle-hadoop-hdfs-project_hadoop-hdfs.txt | | Test Results | https://builds.apache.org/job/PreCommit-HDFS-Build/17034/testReport/ | | modules | C: hadoop-hdfs-project/hadoop-hdfs U: hadoop-hdfs-project/hadoop-hdfs | | Console output | https://builds.apache.org/job/PreCommit-HDFS-Build/17034/console | | Powered by | Apache Yetus 0.4.0-SNAPSHOT http://yetus.apache.org | This message was automatically generated. > Add unit test for HDFS command 'dfsadmin -printTopology' > > > Key: HDFS-10965 > URL: https://issues.apache.org/jira/browse/HDFS-10965 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: fs, shell, test >Reporter: Xiaobing Zhou >Assignee: Xiaobing
[jira] [Updated] (HDFS-10881) Federation State Store Driver API
[ https://issues.apache.org/jira/browse/HDFS-10881?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jason Kace updated HDFS-10881: -- Status: Patch Available (was: In Progress) > Federation State Store Driver API > - > > Key: HDFS-10881 > URL: https://issues.apache.org/jira/browse/HDFS-10881 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: fs >Reporter: Jason Kace >Assignee: Jason Kace > Attachments: HDFS-10881-HDFS-10467-001.patch, > HDFS-10881-HDFS-10467-002.patch, HDFS-10881-HDFS-10467-003.patch, > HDFS-10881-HDFS-10467-004.patch, HDFS-10881-HDFS-10467-005.patch, > HDFS-10881-HDFS-10467-006.patch, HDFS-10881-HDFS-10467-007.patch > > > The API interfaces and minimal classes required to support a state store data > backend such as ZooKeeper or a file system. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-10629) Federation Router
[ https://issues.apache.org/jira/browse/HDFS-10629?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jason Kace updated HDFS-10629: -- Status: Patch Available (was: In Progress) > Federation Router > - > > Key: HDFS-10629 > URL: https://issues.apache.org/jira/browse/HDFS-10629 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: hdfs >Reporter: Inigo Goiri >Assignee: Jason Kace > Attachments: HDFS-10629-HDFS-10467-002.patch, > HDFS-10629-HDFS-10467-003.patch, HDFS-10629-HDFS-10467-004.patch, > HDFS-10629-HDFS-10467-005.patch, HDFS-10629.000.patch, HDFS-10629.001.patch > > > Component that routes calls from the clients to the right Namespace. It > implements {{ClientProtocol}}. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-10881) Federation State Store Driver API
[ https://issues.apache.org/jira/browse/HDFS-10881?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Inigo Goiri updated HDFS-10881: --- Status: In Progress (was: Patch Available) > Federation State Store Driver API > - > > Key: HDFS-10881 > URL: https://issues.apache.org/jira/browse/HDFS-10881 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: fs >Reporter: Jason Kace >Assignee: Jason Kace > Attachments: HDFS-10881-HDFS-10467-001.patch, > HDFS-10881-HDFS-10467-002.patch, HDFS-10881-HDFS-10467-003.patch, > HDFS-10881-HDFS-10467-004.patch, HDFS-10881-HDFS-10467-005.patch, > HDFS-10881-HDFS-10467-006.patch, HDFS-10881-HDFS-10467-007.patch > > > The API interfaces and minimal classes required to support a state store data > backend such as ZooKeeper or a file system. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-10881) Federation State Store Driver API
[ https://issues.apache.org/jira/browse/HDFS-10881?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15550543#comment-15550543 ] Hadoop QA commented on HDFS-10881: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 9s{color} | {color:blue} Docker mode activated. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:red}-1{color} | {color:red} test4tests {color} | {color:red} 0m 0s{color} | {color:red} The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch. {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 9m 49s{color} | {color:green} HDFS-10467 passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 44s{color} | {color:green} HDFS-10467 passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 26s{color} | {color:green} HDFS-10467 passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 54s{color} | {color:green} HDFS-10467 passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 13s{color} | {color:green} HDFS-10467 passed {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 44s{color} | {color:green} HDFS-10467 passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 58s{color} | {color:green} HDFS-10467 passed {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 50s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 44s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 44s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 21s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 50s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 10s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 52s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 54s{color} | {color:green} the patch passed {color} | | {color:red}-1{color} | {color:red} unit {color} | {color:red} 65m 26s{color} | {color:red} hadoop-hdfs in the patch failed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 26s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 87m 52s{color} | {color:black} {color} | \\ \\ || Reason || Tests || | Failed junit tests | hadoop.hdfs.security.TestDelegationTokenForProxyUser | | | hadoop.hdfs.TestFileLengthOnClusterRestart | | | hadoop.hdfs.server.namenode.TestAddStripedBlockInFBR | | | hadoop.hdfs.TestRollingUpgrade | \\ \\ || Subsystem || Report/Notes || | Docker | Image:yetus/hadoop:9560f25 | | JIRA Issue | HDFS-10881 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12831842/HDFS-10881-HDFS-10467-007.patch | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite unit findbugs checkstyle | | uname | Linux 6a193e5df66c 3.13.0-95-generic #142-Ubuntu SMP Fri Aug 12 17:00:09 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh | | git revision | HDFS-10467 / 736d33c | | Default Java | 1.8.0_101 | | findbugs | v3.0.0 | | unit | https://builds.apache.org/job/PreCommit-HDFS-Build/17031/artifact/patchprocess/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt | | Test Results | https://builds.apache.org/job/PreCommit-HDFS-Build/17031/testReport/ | | modules | C: hadoop-hdfs-project/hadoop-hdfs U: hadoop-hdfs-project/hadoop-hdfs | | Console output | https://builds.apache.org/job/PreCommit-HDFS-Build/17031/console | | Powered by | Apache Yetus 0.4.0-SNAPSHOT http://yetus.apache.org | This message was automatically generated. > Federation State Store Driver API > - > >
[jira] [Commented] (HDFS-10745) Directly resolve paths into INodesInPath
[ https://issues.apache.org/jira/browse/HDFS-10745?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15550520#comment-15550520 ] Hadoop QA commented on HDFS-10745: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 19s{color} | {color:blue} Docker mode activated. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:red}-1{color} | {color:red} test4tests {color} | {color:red} 0m 0s{color} | {color:red} The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch. {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 7m 45s{color} | {color:green} branch-2.7 passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 1s{color} | {color:green} branch-2.7 passed with JDK v1.8.0_101 {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 2s{color} | {color:green} branch-2.7 passed with JDK v1.7.0_111 {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 30s{color} | {color:green} branch-2.7 passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 1s{color} | {color:green} branch-2.7 passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 16s{color} | {color:green} branch-2.7 passed {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 2m 56s{color} | {color:green} branch-2.7 passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 58s{color} | {color:green} branch-2.7 passed with JDK v1.8.0_101 {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 45s{color} | {color:green} branch-2.7 passed with JDK v1.7.0_111 {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 53s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 57s{color} | {color:green} the patch passed with JDK v1.8.0_101 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 57s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 1s{color} | {color:green} the patch passed with JDK v1.7.0_111 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 1m 1s{color} | {color:green} the patch passed {color} | | {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange} 0m 28s{color} | {color:orange} hadoop-hdfs-project/hadoop-hdfs: The patch generated 7 new + 514 unchanged - 8 fixed = 521 total (was 522) {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 57s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 13s{color} | {color:green} the patch passed {color} | | {color:red}-1{color} | {color:red} whitespace {color} | {color:red} 0m 0s{color} | {color:red} The patch has 9413 line(s) that end in whitespace. Use git apply --whitespace=fix <>. Refer https://git-scm.com/docs/git-apply {color} | | {color:red}-1{color} | {color:red} whitespace {color} | {color:red} 3m 34s{color} | {color:red} The patch 117 line(s) with tabs. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 3m 10s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 0s{color} | {color:green} the patch passed with JDK v1.8.0_101 {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 42s{color} | {color:green} the patch passed with JDK v1.7.0_111 {color} | | {color:red}-1{color} | {color:red} unit {color} | {color:red} 44m 1s{color} | {color:red} hadoop-hdfs in the patch failed with JDK v1.7.0_111. {color} | | {color:red}-1{color} | {color:red} asflicense {color} | {color:red} 0m 18s{color} | {color:red} The patch generated 3 ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black}130m 14s{color} | {color:black} {color} | \\ \\ || Reason || Tests || | JDK v1.8.0_101 Failed junit tests | hadoop.hdfs.server.datanode.TestDataNodeVolumeFailure | | | hadoop.hdfs.server.namenode.TestFileTruncate | | | hadoop.hdfs.web.TestWebHdfsTokens | | | hadoop.hdfs.TestDFSClientExcludedNodes | | | hadoop.hdfs.TestParallelUnixDomainRead | | | hadoop.hdfs.TestHDFSFileSystemContract | |
[jira] [Commented] (HDFS-10826) Correctly report missing EC blocks in FSCK
[ https://issues.apache.org/jira/browse/HDFS-10826?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15550511#comment-15550511 ] Takanobu Asanuma commented on HDFS-10826: - Thank you for committing, [~jingzhao]! Thank you for reviewing, [~ajisakaa] and [~jojochuang]! > Correctly report missing EC blocks in FSCK > -- > > Key: HDFS-10826 > URL: https://issues.apache.org/jira/browse/HDFS-10826 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: erasure-coding >Reporter: Takanobu Asanuma >Assignee: Takanobu Asanuma > Attachments: HDFS-10826.2.patch, HDFS-10826.3.patch, > HDFS-10826.4.patch, HDFS-10826.5.patch, HDFS-10826.WIP.1.patch > > > For RS-6-3, when there is one ec block group and > 1) 0~3 out of 9 internal blocks are missing, the result of fsck is HEALTY. > 2) 4~8 out of 9 internal blocks are missing, the result of fsck is HEALTY. > {noformat} > Erasure Coded Block Groups: > Total size:536870912 B > Total files: 1 > Total block groups (validated):1 (avg. block group size 536870912 B) > > UNRECOVERABLE BLOCK GROUPS: 1 (100.0 %) > > Minimally erasure-coded block groups: 0 (0.0 %) > Over-erasure-coded block groups: 0 (0.0 %) > Under-erasure-coded block groups: 1 (100.0 %) > Unsatisfactory placement block groups: 0 (0.0 %) > Default ecPolicy: RS-DEFAULT-6-3-64k > Average block group size: 5.0 > Missing block groups: 0 > Corrupt block groups: 0 > Missing internal blocks: 4 (44.43 %) > FSCK ended at Wed Aug 31 13:42:05 JST 2016 in 4 milliseconds > The filesystem under path '/' is HEALTHY > {noformat} > 3) 9 out of 9 internal blocks are missing, the result of fsck is CRITICAL. > (Because it is regarded as a missing block group.) > In case 2), the result should be CRITICAL since the ec block group is > unrecoverable. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-10683) Make class Token$PrivateToken private
[ https://issues.apache.org/jira/browse/HDFS-10683?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15550483#comment-15550483 ] Hudson commented on HDFS-10683: --- SUCCESS: Integrated in Jenkins build Hadoop-trunk-Commit #10550 (See [https://builds.apache.org/job/Hadoop-trunk-Commit/10550/]) HDFS-10683. Make class Token$PrivateToken private. Contributed by John (weichiu: rev c5ca2169151a5eec57152775789b6f53664e102c) * (edit) hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/Credentials.java * (edit) hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/security/TestUserGroupInformation.java * (edit) hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/HAUtil.java * (edit) hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/UserGroupInformation.java * (edit) hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/token/Token.java > Make class Token$PrivateToken private > - > > Key: HDFS-10683 > URL: https://issues.apache.org/jira/browse/HDFS-10683 > Project: Hadoop HDFS > Issue Type: Improvement >Affects Versions: 2.9.0 >Reporter: John Zhuge >Assignee: John Zhuge >Priority: Minor > Labels: fs, ha, security, security_token > Fix For: 2.9.0, 3.0.0-alpha2 > > Attachments: HDFS-10683.001.patch, HDFS-10683.002.patch > > > Avoid {{instanceof}} or typecasting of {{Toke.PrivateToken}} by introducing > an interface method in {{Token}}. Make class {{Toke.PrivateToken}} private. > Use a factory method instead. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-10968) BlockManager#isNewRack should consider decommissioning nodes
[ https://issues.apache.org/jira/browse/HDFS-10968?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jing Zhao updated HDFS-10968: - Status: Patch Available (was: Open) > BlockManager#isNewRack should consider decommissioning nodes > > > Key: HDFS-10968 > URL: https://issues.apache.org/jira/browse/HDFS-10968 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: erasure-coding, namenode >Affects Versions: 3.0.0-alpha1 >Reporter: Jing Zhao >Assignee: Jing Zhao > Attachments: HDFS-10968.000.patch > > > For an EC block, it is possible we have enough internal blocks but without > enough racks. The current reconstruction code calls > {{BlockManager#isNewRack}} to check if the target node can increase the total > rack number for the case, which compares the target node's rack with source > node racks: > {code} > for (DatanodeDescriptor src : srcs) { > if (src.getNetworkLocation().equals(target.getNetworkLocation())) { > return false; > } > } > {code} > However here the {{srcs}} may include a decommissioning node, in which case > we should allow the target node to be in the same rack with it. > For e.g., suppose we have 11 nodes: h1 ~ h11, which are located in racks r1, > r1, r2, r2, r3, r3, r4, r4, r5, r5, r6, respectively. In case that an EC > block has 9 live internal blocks on (h1~h8 + h11), and one internal block on > h9 which is to be decommissioned. The current code will not choose h10 for > reconstruction because isNewRack thinks h10 is on the same rack with h9. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-10968) BlockManager#isNewRack should consider decommissioning nodes
[ https://issues.apache.org/jira/browse/HDFS-10968?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jing Zhao updated HDFS-10968: - Attachment: HDFS-10968.000.patch Upload a patch to fix the issue. The patch also includes a unit test that reproduces the example in the description. > BlockManager#isNewRack should consider decommissioning nodes > > > Key: HDFS-10968 > URL: https://issues.apache.org/jira/browse/HDFS-10968 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: erasure-coding, namenode >Affects Versions: 3.0.0-alpha1 >Reporter: Jing Zhao >Assignee: Jing Zhao > Attachments: HDFS-10968.000.patch > > > For an EC block, it is possible we have enough internal blocks but without > enough racks. The current reconstruction code calls > {{BlockManager#isNewRack}} to check if the target node can increase the total > rack number for the case, which compares the target node's rack with source > node racks: > {code} > for (DatanodeDescriptor src : srcs) { > if (src.getNetworkLocation().equals(target.getNetworkLocation())) { > return false; > } > } > {code} > However here the {{srcs}} may include a decommissioning node, in which case > we should allow the target node to be in the same rack with it. > For e.g., suppose we have 11 nodes: h1 ~ h11, which are located in racks r1, > r1, r2, r2, r3, r3, r4, r4, r5, r5, r6, respectively. In case that an EC > block has 9 live internal blocks on (h1~h8 + h11), and one internal block on > h9 which is to be decommissioned. The current code will not choose h10 for > reconstruction because isNewRack thinks h10 is on the same rack with h9. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Created] (HDFS-10968) BlockManager#isNewRack should consider decommissioning nodes
Jing Zhao created HDFS-10968: Summary: BlockManager#isNewRack should consider decommissioning nodes Key: HDFS-10968 URL: https://issues.apache.org/jira/browse/HDFS-10968 Project: Hadoop HDFS Issue Type: Sub-task Components: erasure-coding, namenode Affects Versions: 3.0.0-alpha1 Reporter: Jing Zhao Assignee: Jing Zhao For an EC block, it is possible we have enough internal blocks but without enough racks. The current reconstruction code calls {{BlockManager#isNewRack}} to check if the target node can increase the total rack number for the case, which compares the target node's rack with source node racks: {code} for (DatanodeDescriptor src : srcs) { if (src.getNetworkLocation().equals(target.getNetworkLocation())) { return false; } } {code} However here the {{srcs}} may include a decommissioning node, in which case we should allow the target node to be in the same rack with it. For e.g., suppose we have 11 nodes: h1 ~ h11, which are located in racks r1, r1, r2, r2, r3, r3, r4, r4, r5, r5, r6, respectively. In case that an EC block has 9 live internal blocks on (h1~h8 + h11), and one internal block on h9 which is to be decommissioned. The current code will not choose h10 for reconstruction because isNewRack thinks h10 is on the same rack with h9. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-10683) Make class Token$PrivateToken private
[ https://issues.apache.org/jira/browse/HDFS-10683?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Wei-Chiu Chuang updated HDFS-10683: --- Resolution: Fixed Hadoop Flags: Reviewed Fix Version/s: 3.0.0-alpha2 2.9.0 Status: Resolved (was: Patch Available) Committed to trunk and branch-2. Thanks [~jzhuge]! > Make class Token$PrivateToken private > - > > Key: HDFS-10683 > URL: https://issues.apache.org/jira/browse/HDFS-10683 > Project: Hadoop HDFS > Issue Type: Improvement >Affects Versions: 2.9.0 >Reporter: John Zhuge >Assignee: John Zhuge >Priority: Minor > Labels: fs, ha, security, security_token > Fix For: 2.9.0, 3.0.0-alpha2 > > Attachments: HDFS-10683.001.patch, HDFS-10683.002.patch > > > Avoid {{instanceof}} or typecasting of {{Toke.PrivateToken}} by introducing > an interface method in {{Token}}. Make class {{Toke.PrivateToken}} private. > Use a factory method instead. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-10690) Optimize insertion/removal of replica in ShortCircuitCache
[ https://issues.apache.org/jira/browse/HDFS-10690?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15550451#comment-15550451 ] Fenghua Hu commented on HDFS-10690: --- [~xyao], thank you for the great help! > Optimize insertion/removal of replica in ShortCircuitCache > -- > > Key: HDFS-10690 > URL: https://issues.apache.org/jira/browse/HDFS-10690 > Project: Hadoop HDFS > Issue Type: Improvement > Components: hdfs-client >Affects Versions: 3.0.0-alpha2 >Reporter: Fenghua Hu >Assignee: Fenghua Hu > Fix For: 2.8.0 > > Attachments: HDFS-10690.001.patch, HDFS-10690.002.patch, > HDFS-10690.003.patch, HDFS-10690.004.patch, HDFS-10690.005.patch, > HDFS-10690.006.patch, HDFS-10690.007.patch, HDFS-10690.008.patch, > ShortCircuitCache_LinkedMap.patch > > Original Estimate: 336h > Remaining Estimate: 336h > > Currently in ShortCircuitCache, two TreeMap objects are used to track the > cached replicas. > private final TreeMapevictable = new TreeMap<>(); > private final TreeMap evictableMmapped = new > TreeMap<>(); > TreeMap employs Red-Black tree for sorting. This isn't an issue when using > traditional HDD. But when using high-performance SSD/PCIe Flash, the cost > inserting/removing an entry becomes considerable. > To mitigate it, we designed a new list-based for replica tracking. > The list is a double-linked FIFO. FIFO is time-based, thus insertion is a > very low cost operation. On the other hand, list is not lookup-friendly. To > address this issue, we introduce two references into ShortCircuitReplica > object. > ShortCircuitReplica next = null; > ShortCircuitReplica prev = null; > In this way, lookup is not needed when removing a replica from the list. We > only need to modify its predecessor's and successor's references in the lists. > Our tests showed up to 15-50% performance improvement when using PCIe flash > as storage media. > The original patch is against 2.6.4, now I am porting to Hadoop trunk, and > patch will be posted soon. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-10965) Add unit test for HDFS command 'dfsadmin -printTopology'
[ https://issues.apache.org/jira/browse/HDFS-10965?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15550409#comment-15550409 ] Xiaobing Zhou commented on HDFS-10965: -- v001 patch fixed check style issues. > Add unit test for HDFS command 'dfsadmin -printTopology' > > > Key: HDFS-10965 > URL: https://issues.apache.org/jira/browse/HDFS-10965 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: fs, shell, test >Reporter: Xiaobing Zhou >Assignee: Xiaobing Zhou > Attachments: HDFS-10965.000.patch, HDFS-10965.001.patch > > > DFSAdmin#printTopology should also be tested. This proposes adding it in > TestDFSAdmin. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-10965) Add unit test for HDFS command 'dfsadmin -printTopology'
[ https://issues.apache.org/jira/browse/HDFS-10965?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Xiaobing Zhou updated HDFS-10965: - Attachment: HDFS-10965.001.patch > Add unit test for HDFS command 'dfsadmin -printTopology' > > > Key: HDFS-10965 > URL: https://issues.apache.org/jira/browse/HDFS-10965 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: fs, shell, test >Reporter: Xiaobing Zhou >Assignee: Xiaobing Zhou > Attachments: HDFS-10965.000.patch, HDFS-10965.001.patch > > > DFSAdmin#printTopology should also be tested. This proposes adding it in > TestDFSAdmin. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-10965) Add unit test for HDFS command 'dfsadmin -printTopology'
[ https://issues.apache.org/jira/browse/HDFS-10965?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15550392#comment-15550392 ] Hadoop QA commented on HDFS-10965: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 15s{color} | {color:blue} Docker mode activated. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 1 new or modified test files. {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 8m 34s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 56s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 30s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 58s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 12s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 55s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 58s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 52s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 47s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 47s{color} | {color:green} the patch passed {color} | | {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange} 0m 26s{color} | {color:orange} hadoop-hdfs-project/hadoop-hdfs: The patch generated 5 new + 168 unchanged - 34 fixed = 173 total (was 202) {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 55s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 11s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 2m 0s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 59s{color} | {color:green} the patch passed {color} | | {color:red}-1{color} | {color:red} unit {color} | {color:red} 59m 31s{color} | {color:red} hadoop-hdfs in the patch failed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 20s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 81m 38s{color} | {color:black} {color} | \\ \\ || Reason || Tests || | Failed junit tests | hadoop.hdfs.server.datanode.TestLargeBlockReport | \\ \\ || Subsystem || Report/Notes || | Docker | Image:yetus/hadoop:9560f25 | | JIRA Issue | HDFS-10965 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12831833/HDFS-10965.000.patch | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite unit findbugs checkstyle | | uname | Linux a13f48042c05 3.13.0-93-generic #140-Ubuntu SMP Mon Jul 18 21:21:05 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh | | git revision | trunk / e68c7b9 | | Default Java | 1.8.0_101 | | findbugs | v3.0.0 | | checkstyle | https://builds.apache.org/job/PreCommit-HDFS-Build/17029/artifact/patchprocess/diff-checkstyle-hadoop-hdfs-project_hadoop-hdfs.txt | | unit | https://builds.apache.org/job/PreCommit-HDFS-Build/17029/artifact/patchprocess/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt | | Test Results | https://builds.apache.org/job/PreCommit-HDFS-Build/17029/testReport/ | | modules | C: hadoop-hdfs-project/hadoop-hdfs U: hadoop-hdfs-project/hadoop-hdfs | | Console output | https://builds.apache.org/job/PreCommit-HDFS-Build/17029/console | | Powered by | Apache Yetus 0.4.0-SNAPSHOT http://yetus.apache.org | This message was automatically generated. > Add unit test for HDFS command 'dfsadmin -printTopology' > > > Key: HDFS-10965 >
[jira] [Updated] (HDFS-10949) DiskBalancer: deprecate TestDiskBalancer#setVolumeCapacity
[ https://issues.apache.org/jira/browse/HDFS-10949?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Xiaobing Zhou updated HDFS-10949: - Status: Patch Available (was: Open) > DiskBalancer: deprecate TestDiskBalancer#setVolumeCapacity > -- > > Key: HDFS-10949 > URL: https://issues.apache.org/jira/browse/HDFS-10949 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: balancer & mover >Reporter: Xiaobing Zhou >Assignee: Xiaobing Zhou >Priority: Minor > Attachments: HDFS-10949.000.patch > > > HDFS-10824 fixed amnesia of capacity setting of MiniDFSCluster, there's no > reason for existence of TestDiskBalancer#setVolumeCapacity anymore. It should > be deprecated. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-10949) DiskBalancer: deprecate TestDiskBalancer#setVolumeCapacity
[ https://issues.apache.org/jira/browse/HDFS-10949?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15550391#comment-15550391 ] Xiaobing Zhou commented on HDFS-10949: -- v000 is posted, please kindly review it, thanks. > DiskBalancer: deprecate TestDiskBalancer#setVolumeCapacity > -- > > Key: HDFS-10949 > URL: https://issues.apache.org/jira/browse/HDFS-10949 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: balancer & mover >Reporter: Xiaobing Zhou >Assignee: Xiaobing Zhou >Priority: Minor > Attachments: HDFS-10949.000.patch > > > HDFS-10824 fixed amnesia of capacity setting of MiniDFSCluster, there's no > reason for existence of TestDiskBalancer#setVolumeCapacity anymore. It should > be deprecated. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-10949) DiskBalancer: deprecate TestDiskBalancer#setVolumeCapacity
[ https://issues.apache.org/jira/browse/HDFS-10949?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Xiaobing Zhou updated HDFS-10949: - Attachment: HDFS-10949.000.patch > DiskBalancer: deprecate TestDiskBalancer#setVolumeCapacity > -- > > Key: HDFS-10949 > URL: https://issues.apache.org/jira/browse/HDFS-10949 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: balancer & mover >Reporter: Xiaobing Zhou >Assignee: Xiaobing Zhou >Priority: Minor > Attachments: HDFS-10949.000.patch > > > HDFS-10824 fixed amnesia of capacity setting of MiniDFSCluster, there's no > reason for existence of TestDiskBalancer#setVolumeCapacity anymore. It should > be deprecated. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Assigned] (HDFS-5684) o.a.h.fs.viewfs.ViewFileSystem.MountPoint has no public methods.
[ https://issues.apache.org/jira/browse/HDFS-5684?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Manoj Govindassamy reassigned HDFS-5684: Assignee: Manoj Govindassamy > o.a.h.fs.viewfs.ViewFileSystem.MountPoint has no public methods. > > > Key: HDFS-5684 > URL: https://issues.apache.org/jira/browse/HDFS-5684 > Project: Hadoop HDFS > Issue Type: Bug > Components: hdfs-client >Affects Versions: 2.2.0 >Reporter: Keith Turner >Assignee: Manoj Govindassamy > > o.a.h.fs.viewfs.ViewFileSystem provides the following method > {noformat} > public MountPoint[] getMountPoints() > {noformat} > However MountPoint has no public methods making getMountPoints() useless. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-10683) Make class Token$PrivateToken private
[ https://issues.apache.org/jira/browse/HDFS-10683?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15550370#comment-15550370 ] Wei-Chiu Chuang commented on HDFS-10683: +1 > Make class Token$PrivateToken private > - > > Key: HDFS-10683 > URL: https://issues.apache.org/jira/browse/HDFS-10683 > Project: Hadoop HDFS > Issue Type: Improvement >Affects Versions: 2.9.0 >Reporter: John Zhuge >Assignee: John Zhuge >Priority: Minor > Labels: fs, ha, security, security_token > Attachments: HDFS-10683.001.patch, HDFS-10683.002.patch > > > Avoid {{instanceof}} or typecasting of {{Toke.PrivateToken}} by introducing > an interface method in {{Token}}. Make class {{Toke.PrivateToken}} private. > Use a factory method instead. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Created] (HDFS-10967) Add configuration for BlockPlacementPolicy to deprioritize near-full DataNodes
Zhe Zhang created HDFS-10967: Summary: Add configuration for BlockPlacementPolicy to deprioritize near-full DataNodes Key: HDFS-10967 URL: https://issues.apache.org/jira/browse/HDFS-10967 Project: Hadoop HDFS Issue Type: Improvement Components: namenode Reporter: Zhe Zhang Large production clusters are likely to have heterogeneous nodes in terms of storage capacity, memory, and CPU cores. It is not always possible to proportionally ingest data into DataNodes based on their remaining storage capacity. Therefore it's possible for a subset of DataNodes to be much closer to full capacity than the rest. Notice that this heterogeneity is most likely rack-by-rack -- i.e. _m_ whole racks with low-storage nodes and _n_ whole racks with high-storage nodes. So It'd be very useful if we can deprioritize those near-full DataNodes as destinations for the 2nd and 3rd replicas. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-6532) Intermittent test failure org.apache.hadoop.hdfs.TestCrcCorruption.testCorruptionDuringWrt
[ https://issues.apache.org/jira/browse/HDFS-6532?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Xiao Chen updated HDFS-6532: Attachment: jstack > Intermittent test failure > org.apache.hadoop.hdfs.TestCrcCorruption.testCorruptionDuringWrt > -- > > Key: HDFS-6532 > URL: https://issues.apache.org/jira/browse/HDFS-6532 > Project: Hadoop HDFS > Issue Type: Bug > Components: datanode, hdfs-client >Affects Versions: 2.4.0 >Reporter: Yongjun Zhang >Assignee: Yiqun Lin > Attachments: HDFS-6532.001.patch, HDFS-6532.002.patch, > PreCommit-HDFS-Build #16770 test - testCorruptionDuringWrt [Jenkins].pdf, > TEST-org.apache.hadoop.hdfs.TestCrcCorruption-select_timeout.xml, > TEST-org.apache.hadoop.hdfs.TestCrcCorruption.xml, jstack > > > Per https://builds.apache.org/job/Hadoop-Hdfs-trunk/1774/testReport, we had > the following failure. Local rerun is successful > {code} > Regression > org.apache.hadoop.hdfs.TestCrcCorruption.testCorruptionDuringWrt > Failing for the past 1 build (Since Failed#1774 ) > Took 50 sec. > Error Message > test timed out after 5 milliseconds > Stacktrace > java.lang.Exception: test timed out after 5 milliseconds > at java.lang.Object.wait(Native Method) > at > org.apache.hadoop.hdfs.DFSOutputStream.waitForAckedSeqno(DFSOutputStream.java:2024) > at > org.apache.hadoop.hdfs.DFSOutputStream.flushInternal(DFSOutputStream.java:2008) > at > org.apache.hadoop.hdfs.DFSOutputStream.close(DFSOutputStream.java:2107) > at > org.apache.hadoop.fs.FSDataOutputStream$PositionCache.close(FSDataOutputStream.java:70) > at > org.apache.hadoop.fs.FSDataOutputStream.close(FSDataOutputStream.java:98) > at > org.apache.hadoop.hdfs.TestCrcCorruption.testCorruptionDuringWrt(TestCrcCorruption.java:133) > {code} > See relevant exceptions in log > {code} > 2014-06-14 11:56:15,283 WARN datanode.DataNode > (BlockReceiver.java:verifyChunks(404)) - Checksum error in block > BP-1675558312-67.195.138.30-1402746971712:blk_1073741825_1001 from > /127.0.0.1:41708 > org.apache.hadoop.fs.ChecksumException: Checksum error: > DFSClient_NONMAPREDUCE_-1139495951_8 at 64512 exp: 1379611785 got: -12163112 > at > org.apache.hadoop.util.DataChecksum.verifyChunkedSums(DataChecksum.java:353) > at > org.apache.hadoop.util.DataChecksum.verifyChunkedSums(DataChecksum.java:284) > at > org.apache.hadoop.hdfs.server.datanode.BlockReceiver.verifyChunks(BlockReceiver.java:402) > at > org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:537) > at > org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:734) > at > org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:741) > at > org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:124) > at > org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:71) > at > org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:234) > at java.lang.Thread.run(Thread.java:662) > 2014-06-14 11:56:15,285 WARN datanode.DataNode > (BlockReceiver.java:run(1207)) - IOException in BlockReceiver.run(): > java.io.IOException: Shutting down writer and responder due to a checksum > error in received data. The error response has been sent upstream. > at > org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.sendAckUpstreamUnprotected(BlockReceiver.java:1352) > at > org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.sendAckUpstream(BlockReceiver.java:1278) > at > org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.run(BlockReceiver.java:1199) > at java.lang.Thread.run(Thread.java:662) > ... > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-6532) Intermittent test failure org.apache.hadoop.hdfs.TestCrcCorruption.testCorruptionDuringWrt
[ https://issues.apache.org/jira/browse/HDFS-6532?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15550368#comment-15550368 ] Xiao Chen commented on HDFS-6532: - Thanks for the comment, [~linyiqun]. Took a jstack when right before the test timed out. It has: {noformat} "ResponseProcessor for block BP-341806944-172.17.0.1-1475696115790:blk_1073741825_1001" java.lang.Thread.State: RUNNABLE at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:79) at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:87) at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:98) at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:157) at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:118) at java.io.FilterInputStream.read(FilterInputStream.java:83) at java.io.FilterInputStream.read(FilterInputStream.java:83) at org.apache.hadoop.hdfs.protocolPB.PBHelper.vintPrefixed(PBHelper.java:2247) at org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:235) at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer$ResponseProcessor.run(DFSOutputStream.java:1017) "PacketResponder: BP-341806944-172.17.0.1-1475696115790:blk_1073741825_1001, type=HAS_DOWNSTREAM_IN_PIPELINE" java.lang.Thread.State: RUNNABLE at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:79) at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:87) at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:98) at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:157) at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:118) at java.io.FilterInputStream.read(FilterInputStream.java:83) at java.io.FilterInputStream.read(FilterInputStream.java:83) at org.apache.hadoop.hdfs.protocolPB.PBHelper.vintPrefixed(PBHelper.java:2247) at org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:235) at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.run(BlockReceiver.java:1303) at java.lang.Thread.run(Thread.java:745) "DataXceiver for client DFSClient_NONMAPREDUCE_27347732_8 at /127.0.0.1:36783 [Receiving block BP-341806944-172.17.0.1-1475696115790:blk_1073741825_1001]" java.lang.Thread.State: RUNNABLE at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:79) at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:87) at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:98) at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:157) at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) at java.io.BufferedInputStream.fill(BufferedInputStream.java:235) at java.io.BufferedInputStream.read1(BufferedInputStream.java:275) at java.io.BufferedInputStream.read(BufferedInputStream.java:334) at java.io.DataInputStream.read(DataInputStream.java:149) at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:199) at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:213) at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:134) at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:109) at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:502) at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:900) at
[jira] [Updated] (HDFS-10881) Federation State Store Driver API
[ https://issues.apache.org/jira/browse/HDFS-10881?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jason Kace updated HDFS-10881: -- Attachment: HDFS-10881-HDFS-10467-007.patch Updating patch to include missing file. > Federation State Store Driver API > - > > Key: HDFS-10881 > URL: https://issues.apache.org/jira/browse/HDFS-10881 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: fs >Reporter: Jason Kace >Assignee: Jason Kace > Attachments: HDFS-10881-HDFS-10467-001.patch, > HDFS-10881-HDFS-10467-002.patch, HDFS-10881-HDFS-10467-003.patch, > HDFS-10881-HDFS-10467-004.patch, HDFS-10881-HDFS-10467-005.patch, > HDFS-10881-HDFS-10467-006.patch, HDFS-10881-HDFS-10467-007.patch > > > The API interfaces and minimal classes required to support a state store data > backend such as ZooKeeper or a file system. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-10683) Make class Token$PrivateToken private
[ https://issues.apache.org/jira/browse/HDFS-10683?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15550311#comment-15550311 ] Hadoop QA commented on HDFS-10683: -- | (/) *{color:green}+1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 24s{color} | {color:blue} Docker mode activated. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 1 new or modified test files. {color} | | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 18s{color} | {color:blue} Maven dependency ordering for branch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 8m 36s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 8m 19s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 1m 44s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 2m 8s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 29s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 3m 26s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 2m 7s{color} | {color:green} trunk passed {color} | | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 17s{color} | {color:blue} Maven dependency ordering for patch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 1m 52s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 8m 58s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 8m 58s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 1m 46s{color} | {color:green} root: The patch generated 0 new + 173 unchanged - 2 fixed = 173 total (was 175) {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 2m 16s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 31s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 3m 54s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 38s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} unit {color} | {color:green} 8m 39s{color} | {color:green} hadoop-common in the patch passed. {color} | | {color:green}+1{color} | {color:green} unit {color} | {color:green} 76m 1s{color} | {color:green} hadoop-hdfs in the patch passed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 29s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black}134m 51s{color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Docker | Image:yetus/hadoop:9560f25 | | JIRA Issue | HDFS-10683 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12830110/HDFS-10683.002.patch | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite unit findbugs checkstyle | | uname | Linux 918bd5603957 3.13.0-96-generic #143-Ubuntu SMP Mon Aug 29 20:15:20 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh | | git revision | trunk / 00160f7 | | Default Java | 1.8.0_101 | | findbugs | v3.0.0 | | Test Results | https://builds.apache.org/job/PreCommit-HDFS-Build/17027/testReport/ | | modules | C: hadoop-common-project/hadoop-common hadoop-hdfs-project/hadoop-hdfs U: . | | Console output | https://builds.apache.org/job/PreCommit-HDFS-Build/17027/console | | Powered by | Apache Yetus 0.4.0-SNAPSHOT http://yetus.apache.org | This message was automatically generated. > Make class Token$PrivateToken private > - > > Key: HDFS-10683 > URL:
[jira] [Updated] (HDFS-10881) Federation State Store Driver API
[ https://issues.apache.org/jira/browse/HDFS-10881?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jason Kace updated HDFS-10881: -- Attachment: (was: HDFS-10881-HDFS-10467-007.patch) > Federation State Store Driver API > - > > Key: HDFS-10881 > URL: https://issues.apache.org/jira/browse/HDFS-10881 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: fs >Reporter: Jason Kace >Assignee: Jason Kace > Attachments: HDFS-10881-HDFS-10467-001.patch, > HDFS-10881-HDFS-10467-002.patch, HDFS-10881-HDFS-10467-003.patch, > HDFS-10881-HDFS-10467-004.patch, HDFS-10881-HDFS-10467-005.patch, > HDFS-10881-HDFS-10467-006.patch > > > The API interfaces and minimal classes required to support a state store data > backend such as ZooKeeper or a file system. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-10558) DiskBalancer: Print the full path to plan file
[ https://issues.apache.org/jira/browse/HDFS-10558?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Xiaobing Zhou updated HDFS-10558: - Status: Patch Available (was: Open) > DiskBalancer: Print the full path to plan file > --- > > Key: HDFS-10558 > URL: https://issues.apache.org/jira/browse/HDFS-10558 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: balancer & mover >Affects Versions: HDFS-1312 >Reporter: Anu Engineer >Assignee: Xiaobing Zhou >Priority: Minor > Labels: newbie > Attachments: HDFS-10558-HDFS-1312.000.patch, HDFS-10558.001.patch > > > We should print the full path to plan file when plan command is being run. > That makes it easy to give that path to -execute option. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-10558) DiskBalancer: Print the full path to plan file
[ https://issues.apache.org/jira/browse/HDFS-10558?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Xiaobing Zhou updated HDFS-10558: - Status: Open (was: Patch Available) > DiskBalancer: Print the full path to plan file > --- > > Key: HDFS-10558 > URL: https://issues.apache.org/jira/browse/HDFS-10558 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: balancer & mover >Affects Versions: HDFS-1312 >Reporter: Anu Engineer >Assignee: Xiaobing Zhou >Priority: Minor > Labels: newbie > Attachments: HDFS-10558-HDFS-1312.000.patch, HDFS-10558.001.patch > > > We should print the full path to plan file when plan command is being run. > That makes it easy to give that path to -execute option. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-10881) Federation State Store Driver API
[ https://issues.apache.org/jira/browse/HDFS-10881?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15550250#comment-15550250 ] Hadoop QA commented on HDFS-10881: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 23s{color} | {color:blue} Docker mode activated. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:red}-1{color} | {color:red} test4tests {color} | {color:red} 0m 0s{color} | {color:red} The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch. {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 11m 44s{color} | {color:green} HDFS-10467 passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 58s{color} | {color:green} HDFS-10467 passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 33s{color} | {color:green} HDFS-10467 passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 6s{color} | {color:green} HDFS-10467 passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 17s{color} | {color:green} HDFS-10467 passed {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 58s{color} | {color:green} HDFS-10467 passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 5s{color} | {color:green} HDFS-10467 passed {color} | | {color:red}-1{color} | {color:red} mvninstall {color} | {color:red} 0m 28s{color} | {color:red} hadoop-hdfs in the patch failed. {color} | | {color:red}-1{color} | {color:red} compile {color} | {color:red} 0m 28s{color} | {color:red} hadoop-hdfs in the patch failed. {color} | | {color:red}-1{color} | {color:red} javac {color} | {color:red} 0m 28s{color} | {color:red} hadoop-hdfs in the patch failed. {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 27s{color} | {color:green} the patch passed {color} | | {color:red}-1{color} | {color:red} mvnsite {color} | {color:red} 0m 26s{color} | {color:red} hadoop-hdfs in the patch failed. {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 11s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:red}-1{color} | {color:red} findbugs {color} | {color:red} 0m 27s{color} | {color:red} hadoop-hdfs in the patch failed. {color} | | {color:red}-1{color} | {color:red} javadoc {color} | {color:red} 0m 58s{color} | {color:red} hadoop-hdfs-project_hadoop-hdfs generated 1 new + 7 unchanged - 0 fixed = 8 total (was 7) {color} | | {color:red}-1{color} | {color:red} unit {color} | {color:red} 0m 28s{color} | {color:red} hadoop-hdfs in the patch failed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 21s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 23m 43s{color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Docker | Image:yetus/hadoop:9560f25 | | JIRA Issue | HDFS-10881 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12831831/HDFS-10881-HDFS-10467-007.patch | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite unit findbugs checkstyle | | uname | Linux 3c53e51929d7 3.13.0-96-generic #143-Ubuntu SMP Mon Aug 29 20:15:20 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh | | git revision | HDFS-10467 / 736d33c | | Default Java | 1.8.0_101 | | findbugs | v3.0.0 | | mvninstall | https://builds.apache.org/job/PreCommit-HDFS-Build/17028/artifact/patchprocess/patch-mvninstall-hadoop-hdfs-project_hadoop-hdfs.txt | | compile | https://builds.apache.org/job/PreCommit-HDFS-Build/17028/artifact/patchprocess/patch-compile-hadoop-hdfs-project_hadoop-hdfs.txt | | javac | https://builds.apache.org/job/PreCommit-HDFS-Build/17028/artifact/patchprocess/patch-compile-hadoop-hdfs-project_hadoop-hdfs.txt | | mvnsite | https://builds.apache.org/job/PreCommit-HDFS-Build/17028/artifact/patchprocess/patch-mvnsite-hadoop-hdfs-project_hadoop-hdfs.txt | | findbugs | https://builds.apache.org/job/PreCommit-HDFS-Build/17028/artifact/patchprocess/patch-findbugs-hadoop-hdfs-project_hadoop-hdfs.txt | | javadoc |
[jira] [Updated] (HDFS-10745) Directly resolve paths into INodesInPath
[ https://issues.apache.org/jira/browse/HDFS-10745?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Zhe Zhang updated HDFS-10745: - Attachment: HDFS-10745-branch-2.7.06.patch Thanks for the good catch [~shv]. Updating the patch to address. I also verified that resolvePath and resolvePathForWrite are under lock in other places. > Directly resolve paths into INodesInPath > > > Key: HDFS-10745 > URL: https://issues.apache.org/jira/browse/HDFS-10745 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: hdfs >Reporter: Daryn Sharp >Assignee: Daryn Sharp > Fix For: 2.8.0, 3.0.0-alpha1 > > Attachments: HDFS-10745-branch-2.7.01.patch, > HDFS-10745-branch-2.7.02.patch, HDFS-10745-branch-2.7.03.patch, > HDFS-10745-branch-2.7.04.patch, HDFS-10745-branch-2.7.05.patch, > HDFS-10745-branch-2.7.06.patch, HDFS-10745-branch-2.7.patch, > HDFS-10745.2.patch, HDFS-10745.branch-2.patch, HDFS-10745.patch > > > The intermediate resolution to a string, only to be decomposed by > {{INodesInPath}} back into a byte[][] can be eliminated by resolving directly > to an IIP. The IIP will contain the resolved path if required. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-10965) Add unit test for HDFS command 'dfsadmin -printTopology'
[ https://issues.apache.org/jira/browse/HDFS-10965?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15550204#comment-15550204 ] Xiaobing Zhou commented on HDFS-10965: -- Posted patch v000 for reviews, thanks. > Add unit test for HDFS command 'dfsadmin -printTopology' > > > Key: HDFS-10965 > URL: https://issues.apache.org/jira/browse/HDFS-10965 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: fs, shell, test >Reporter: Xiaobing Zhou >Assignee: Xiaobing Zhou > Attachments: HDFS-10965.000.patch > > > DFSAdmin#printTopology should also be tested. This proposes adding it in > TestDFSAdmin. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-10965) Add unit test for HDFS command 'dfsadmin -printTopology'
[ https://issues.apache.org/jira/browse/HDFS-10965?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Xiaobing Zhou updated HDFS-10965: - Attachment: HDFS-10965.000.patch > Add unit test for HDFS command 'dfsadmin -printTopology' > > > Key: HDFS-10965 > URL: https://issues.apache.org/jira/browse/HDFS-10965 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: fs, shell, test >Reporter: Xiaobing Zhou >Assignee: Xiaobing Zhou > Attachments: HDFS-10965.000.patch > > > DFSAdmin#printTopology should also be tested. This proposes adding it in > TestDFSAdmin. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-10965) Add unit test for HDFS command 'dfsadmin -printTopology'
[ https://issues.apache.org/jira/browse/HDFS-10965?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Xiaobing Zhou updated HDFS-10965: - Status: Patch Available (was: Open) > Add unit test for HDFS command 'dfsadmin -printTopology' > > > Key: HDFS-10965 > URL: https://issues.apache.org/jira/browse/HDFS-10965 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: fs, shell, test >Reporter: Xiaobing Zhou >Assignee: Xiaobing Zhou > Attachments: HDFS-10965.000.patch > > > DFSAdmin#printTopology should also be tested. This proposes adding it in > TestDFSAdmin. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-10881) Federation State Store Driver API
[ https://issues.apache.org/jira/browse/HDFS-10881?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jason Kace updated HDFS-10881: -- Attachment: HDFS-10881-HDFS-10467-007.patch Uploading the final planned version for this patch. This patch is the base for all of the other patches. Please let us know if you have any feedback. This update is a minor refactoring of the previous versions with more comments. It is also more focused on the driver-level interface and omits some of the state store API interfaces that will be in HDFS-10882. > Federation State Store Driver API > - > > Key: HDFS-10881 > URL: https://issues.apache.org/jira/browse/HDFS-10881 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: fs >Reporter: Jason Kace >Assignee: Jason Kace > Attachments: HDFS-10881-HDFS-10467-001.patch, > HDFS-10881-HDFS-10467-002.patch, HDFS-10881-HDFS-10467-003.patch, > HDFS-10881-HDFS-10467-004.patch, HDFS-10881-HDFS-10467-005.patch, > HDFS-10881-HDFS-10467-006.patch, HDFS-10881-HDFS-10467-007.patch > > > The API interfaces and minimal classes required to support a state store data > backend such as ZooKeeper or a file system. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Created] (HDFS-10966) Enhance Dispatcher logic on deciding when to give up a source DataNode
Zhe Zhang created HDFS-10966: Summary: Enhance Dispatcher logic on deciding when to give up a source DataNode Key: HDFS-10966 URL: https://issues.apache.org/jira/browse/HDFS-10966 Project: Hadoop HDFS Issue Type: Bug Components: balancer & mover Reporter: Zhe Zhang Assignee: Mark Wagner When a {{Dispatcher}} thread works on a source DataNode, in each iteration it tries to execute a {{PendingMove}}. If no block is moved after 5 iterations, this source (over-utlized) DataNode is given up for this Balancer iteration (20 mins). This is problematic if the source DataNode was heavily loaded in the beginning of the iteration. It will quickly encounter 5 unsuccessful moves and be abandoned. We should enhance this logic by e.g. using elapsed time instead of number of iterations. {code} // Check if the previous move was successful } else { // source node cannot find a pending block to move, iteration +1 noPendingMoveIteration++; // in case no blocks can be moved for source node's task, // jump out of while-loop after 5 iterations. if (noPendingMoveIteration >= MAX_NO_PENDING_MOVE_ITERATIONS) { LOG.info("Failed to find a pending move " + noPendingMoveIteration + " times. Skipping " + this); resetScheduledSize(); } } {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-10637) Modifications to remove the assumption that FsVolumes are backed by java.io.File.
[ https://issues.apache.org/jira/browse/HDFS-10637?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15550152#comment-15550152 ] Hadoop QA commented on HDFS-10637: -- | (/) *{color:green}+1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 12s{color} | {color:blue} Docker mode activated. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 14 new or modified test files. {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 7m 17s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 46s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 38s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 54s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 12s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 43s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 55s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 48s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 43s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 43s{color} | {color:green} the patch passed {color} | | {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange} 0m 34s{color} | {color:orange} hadoop-hdfs-project/hadoop-hdfs: The patch generated 9 new + 984 unchanged - 11 fixed = 993 total (was 995) {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 51s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 10s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 51s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 52s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} unit {color} | {color:green} 61m 27s{color} | {color:green} hadoop-hdfs in the patch passed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 18s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 81m 32s{color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Docker | Image:yetus/hadoop:9560f25 | | JIRA Issue | HDFS-10637 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12831807/HDFS-10637.011.patch | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite unit findbugs checkstyle | | uname | Linux 04098573a30c 3.13.0-92-generic #139-Ubuntu SMP Tue Jun 28 20:42:26 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh | | git revision | trunk / 00160f7 | | Default Java | 1.8.0_101 | | findbugs | v3.0.0 | | checkstyle | https://builds.apache.org/job/PreCommit-HDFS-Build/17026/artifact/patchprocess/diff-checkstyle-hadoop-hdfs-project_hadoop-hdfs.txt | | Test Results | https://builds.apache.org/job/PreCommit-HDFS-Build/17026/testReport/ | | modules | C: hadoop-hdfs-project/hadoop-hdfs U: hadoop-hdfs-project/hadoop-hdfs | | Console output | https://builds.apache.org/job/PreCommit-HDFS-Build/17026/console | | Powered by | Apache Yetus 0.4.0-SNAPSHOT http://yetus.apache.org | This message was automatically generated. > Modifications to remove the assumption that FsVolumes are backed by > java.io.File. > - > > Key: HDFS-10637 > URL: https://issues.apache.org/jira/browse/HDFS-10637 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: datanode, fs >
[jira] [Commented] (HDFS-10962) TestRequestHedgingProxyProvider is flaky
[ https://issues.apache.org/jira/browse/HDFS-10962?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15550121#comment-15550121 ] Hudson commented on HDFS-10962: --- SUCCESS: Integrated in Jenkins build Hadoop-trunk-Commit #10549 (See [https://builds.apache.org/job/Hadoop-trunk-Commit/10549/]) HDFS-10962. TestRequestHedgingProxyProvider is flaky. (wang: rev e68c7b96c76b243da53ae539e7c760bf2c08f26e) * (edit) hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/ha/RequestHedgingProxyProvider.java * (edit) hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ha/TestRequestHedgingProxyProvider.java > TestRequestHedgingProxyProvider is flaky > > > Key: HDFS-10962 > URL: https://issues.apache.org/jira/browse/HDFS-10962 > Project: Hadoop HDFS > Issue Type: Bug > Components: test >Affects Versions: 2.6.4 >Reporter: Andrew Wang >Assignee: Andrew Wang > Fix For: 2.8.0, 3.0.0-alpha2 > > Attachments: HDFS-10962.001.patch, HDFS-10962.002.patch > > > This test fails occasionally with an error like this: > {noformat} > org.apache.hadoop.hdfs.server.namenode.ha.TestRequestHedgingProxyProvider.testHedgingWhenOneFails > Error Message > Wanted but not invoked: > namenodeProtocols.getStats(); > -> at > org.apache.hadoop.hdfs.server.namenode.ha.TestRequestHedgingProxyProvider.testHedgingWhenOneFails(TestRequestHedgingProxyProvider.java:78) > Actually, there were zero interactions with this mock. > Stacktrace > org.mockito.exceptions.verification.WantedButNotInvoked: > Wanted but not invoked: > namenodeProtocols.getStats(); > -> at > org.apache.hadoop.hdfs.server.namenode.ha.TestRequestHedgingProxyProvider.testHedgingWhenOneFails(TestRequestHedgingProxyProvider.java:78) > Actually, there were zero interactions with this mock. > at > org.apache.hadoop.hdfs.server.namenode.ha.TestRequestHedgingProxyProvider.testHedgingWhenOneFails(TestRequestHedgingProxyProvider.java:78) > Standard Output > 2016-09-26 15:26:02,234 WARN hdfs.DFSUtil > (DFSUtil.java:getAddressesForNameserviceId(689)) - Namenode for > mycluster-26780990 remains unresolved for ID nn1. Check your hdfs-site.xml > file to ensure namenodes are configured properly. > 2016-09-26 15:26:02,340 WARN hdfs.DFSUtil > (DFSUtil.java:getAddressesForNameserviceId(689)) - Namenode for > mycluster-26780990 remains unresolved for ID nn2. Check your hdfs-site.xml > file to ensure namenodes are configured properly. > {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-10962) TestRequestHedgingProxyProvider is flaky
[ https://issues.apache.org/jira/browse/HDFS-10962?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Andrew Wang updated HDFS-10962: --- Resolution: Fixed Fix Version/s: 3.0.0-alpha2 2.8.0 Status: Resolved (was: Patch Available) Committed to trunk, branch-2, branch-2.8. Thanks for reviewing Xiao! > TestRequestHedgingProxyProvider is flaky > > > Key: HDFS-10962 > URL: https://issues.apache.org/jira/browse/HDFS-10962 > Project: Hadoop HDFS > Issue Type: Bug > Components: test >Affects Versions: 2.6.4 >Reporter: Andrew Wang >Assignee: Andrew Wang > Fix For: 2.8.0, 3.0.0-alpha2 > > Attachments: HDFS-10962.001.patch, HDFS-10962.002.patch > > > This test fails occasionally with an error like this: > {noformat} > org.apache.hadoop.hdfs.server.namenode.ha.TestRequestHedgingProxyProvider.testHedgingWhenOneFails > Error Message > Wanted but not invoked: > namenodeProtocols.getStats(); > -> at > org.apache.hadoop.hdfs.server.namenode.ha.TestRequestHedgingProxyProvider.testHedgingWhenOneFails(TestRequestHedgingProxyProvider.java:78) > Actually, there were zero interactions with this mock. > Stacktrace > org.mockito.exceptions.verification.WantedButNotInvoked: > Wanted but not invoked: > namenodeProtocols.getStats(); > -> at > org.apache.hadoop.hdfs.server.namenode.ha.TestRequestHedgingProxyProvider.testHedgingWhenOneFails(TestRequestHedgingProxyProvider.java:78) > Actually, there were zero interactions with this mock. > at > org.apache.hadoop.hdfs.server.namenode.ha.TestRequestHedgingProxyProvider.testHedgingWhenOneFails(TestRequestHedgingProxyProvider.java:78) > Standard Output > 2016-09-26 15:26:02,234 WARN hdfs.DFSUtil > (DFSUtil.java:getAddressesForNameserviceId(689)) - Namenode for > mycluster-26780990 remains unresolved for ID nn1. Check your hdfs-site.xml > file to ensure namenodes are configured properly. > 2016-09-26 15:26:02,340 WARN hdfs.DFSUtil > (DFSUtil.java:getAddressesForNameserviceId(689)) - Namenode for > mycluster-26780990 remains unresolved for ID nn2. Check your hdfs-site.xml > file to ensure namenodes are configured properly. > {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-10957) Retire BKJM from trunk
[ https://issues.apache.org/jira/browse/HDFS-10957?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15550091#comment-15550091 ] Hadoop QA commented on HDFS-10957: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 15s{color} | {color:blue} Docker mode activated. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 11 new or modified test files. {color} | | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 15s{color} | {color:blue} Maven dependency ordering for branch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 6m 54s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 7m 26s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 1m 34s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 3m 8s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 54s{color} | {color:green} trunk passed {color} | | {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue} 0m 0s{color} | {color:blue} Skipped patched modules with no Java source: hadoop-project hadoop-hdfs-project {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 2m 31s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 3m 6s{color} | {color:green} trunk passed {color} | | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 16s{color} | {color:blue} Maven dependency ordering for patch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 2m 51s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 9m 33s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} cc {color} | {color:green} 9m 33s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 9m 33s{color} | {color:green} root generated 0 new + 707 unchanged - 1 fixed = 707 total (was 708) {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 1m 31s{color} | {color:green} root: The patch generated 0 new + 0 unchanged - 138 fixed = 0 total (was 138) {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 2m 42s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 43s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} xml {color} | {color:green} 0m 2s{color} | {color:green} The patch has no ill-formed XML file. {color} | | {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue} 0m 0s{color} | {color:blue} Skipped patched modules with no Java source: hadoop-project hadoop-hdfs-project {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 52s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 12s{color} | {color:green} hadoop-project in the patch passed. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 59s{color} | {color:green} hadoop-hdfs in the patch passed. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 22s{color} | {color:green} hadoop-hdfs-project generated 0 new + 65 unchanged - 61 fixed = 65 total (was 126) {color} | | {color:green}+1{color} | {color:green} unit {color} | {color:green} 0m 11s{color} | {color:green} hadoop-project in the patch passed. {color} | | {color:red}-1{color} | {color:red} unit {color} | {color:red} 76m 5s{color} | {color:red} hadoop-hdfs in the patch failed. {color} | | {color:red}-1{color} | {color:red} unit {color} | {color:red} 78m 11s{color} | {color:red} hadoop-hdfs-project in the patch failed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 24s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} |
[jira] [Updated] (HDFS-10964) Add more unit tests for ACLs
[ https://issues.apache.org/jira/browse/HDFS-10964?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Namit Maheshwari updated HDFS-10964: Description: This Jira proposes to add more unit tests to validate ACLs. Like below: 1. Test Explicit Mask Entries 2. Set Invalid ACLs - Negative scenarios 3. Get default recursive ACL's 4. ACLs for user belonging to common groups was:This Jira proposes to add more unit tests to validate ACLs. > Add more unit tests for ACLs > > > Key: HDFS-10964 > URL: https://issues.apache.org/jira/browse/HDFS-10964 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: test >Reporter: Namit Maheshwari >Assignee: Namit Maheshwari > > This Jira proposes to add more unit tests to validate ACLs. Like below: > 1. Test Explicit Mask Entries > 2. Set Invalid ACLs - Negative scenarios > 3. Get default recursive ACL's > 4. ACLs for user belonging to common groups -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-10965) Add unit test for HDFS command 'dfsadmin printTopology'-
[ https://issues.apache.org/jira/browse/HDFS-10965?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Xiaobing Zhou updated HDFS-10965: - Summary: Add unit test for HDFS command 'dfsadmin printTopology'- (was: Add unit test for HDFS command 'dfsadmin printTopology') > Add unit test for HDFS command 'dfsadmin printTopology'- > > > Key: HDFS-10965 > URL: https://issues.apache.org/jira/browse/HDFS-10965 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: fs, shell, test >Reporter: Xiaobing Zhou >Assignee: Xiaobing Zhou > > DFSAdmin#printTopology should also be tested. This proposes adding it in > TestDFSAdmin. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-10965) Add unit test for HDFS command 'dfsadmin -printTopology'
[ https://issues.apache.org/jira/browse/HDFS-10965?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Xiaobing Zhou updated HDFS-10965: - Summary: Add unit test for HDFS command 'dfsadmin -printTopology' (was: Add unit test for HDFS command 'dfsadmin printTopology'-) > Add unit test for HDFS command 'dfsadmin -printTopology' > > > Key: HDFS-10965 > URL: https://issues.apache.org/jira/browse/HDFS-10965 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: fs, shell, test >Reporter: Xiaobing Zhou >Assignee: Xiaobing Zhou > > DFSAdmin#printTopology should also be tested. This proposes adding it in > TestDFSAdmin. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-10965) Add unit test for HDFS command 'dfsadmin printTopology'
[ https://issues.apache.org/jira/browse/HDFS-10965?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Xiaobing Zhou updated HDFS-10965: - Summary: Add unit test for HDFS command 'dfsadmin printTopology' (was: Add unit test for DFSAdmin#printTopology) > Add unit test for HDFS command 'dfsadmin printTopology' > --- > > Key: HDFS-10965 > URL: https://issues.apache.org/jira/browse/HDFS-10965 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: fs, shell, test >Reporter: Xiaobing Zhou >Assignee: Xiaobing Zhou > > DFSAdmin#printTopology should also be tested. This proposes adding it in > TestDFSAdmin. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-10965) Add unit test for DFSAdmin#printTopology
[ https://issues.apache.org/jira/browse/HDFS-10965?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Xiaobing Zhou updated HDFS-10965: - Component/s: shell fs > Add unit test for DFSAdmin#printTopology > > > Key: HDFS-10965 > URL: https://issues.apache.org/jira/browse/HDFS-10965 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: fs, shell, test >Reporter: Xiaobing Zhou >Assignee: Xiaobing Zhou > > DFSAdmin#printTopology should also be tested. This proposes adding it in > TestDFSAdmin. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Created] (HDFS-10965) Add unit test for DFSAdmin#printTopology
Xiaobing Zhou created HDFS-10965: Summary: Add unit test for DFSAdmin#printTopology Key: HDFS-10965 URL: https://issues.apache.org/jira/browse/HDFS-10965 Project: Hadoop HDFS Issue Type: Sub-task Reporter: Xiaobing Zhou Assignee: Xiaobing Zhou DFSAdmin#printTopology should also be tested. This proposes adding it in TestDFSAdmin. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-10683) Make class Token$PrivateToken private
[ https://issues.apache.org/jira/browse/HDFS-10683?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15549993#comment-15549993 ] Wei-Chiu Chuang commented on HDFS-10683: Manually triggered precommit. You can monitor the job here: https://builds.apache.org/job/PreCommit-HDFS-Build/17027/ > Make class Token$PrivateToken private > - > > Key: HDFS-10683 > URL: https://issues.apache.org/jira/browse/HDFS-10683 > Project: Hadoop HDFS > Issue Type: Improvement >Affects Versions: 2.9.0 >Reporter: John Zhuge >Assignee: John Zhuge >Priority: Minor > Labels: fs, ha, security, security_token > Attachments: HDFS-10683.001.patch, HDFS-10683.002.patch > > > Avoid {{instanceof}} or typecasting of {{Toke.PrivateToken}} by introducing > an interface method in {{Token}}. Make class {{Toke.PrivateToken}} private. > Use a factory method instead. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-10429) DataStreamer interrupted warning always appears when using CLI upload file
[ https://issues.apache.org/jira/browse/HDFS-10429?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15549970#comment-15549970 ] Wei-Chiu Chuang commented on HDFS-10429: We've got a few cases filed for this issue, so I suppose it's relatively easy to reproduce. However I just don't know if it's because I do not have a proper set up for this issue to surface up. [~jingzhao] would you mind to share how you reproduced it? > DataStreamer interrupted warning always appears when using CLI upload file > --- > > Key: HDFS-10429 > URL: https://issues.apache.org/jira/browse/HDFS-10429 > Project: Hadoop HDFS > Issue Type: Bug >Affects Versions: 2.7.3 >Reporter: Zhiyuan Yang >Assignee: Zhiyuan Yang >Priority: Minor > Attachments: HDFS-10429.1.patch, HDFS-10429.2.patch > > > Every time I use 'hdfs dfs -put' upload file, this warning is printed: > {code:java} > 16/05/18 20:57:56 WARN hdfs.DataStreamer: Caught exception > java.lang.InterruptedException > at java.lang.Object.wait(Native Method) > at java.lang.Thread.join(Thread.java:1245) > at java.lang.Thread.join(Thread.java:1319) > at > org.apache.hadoop.hdfs.DataStreamer.closeResponder(DataStreamer.java:871) > at org.apache.hadoop.hdfs.DataStreamer.endBlock(DataStreamer.java:519) > at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:696) > {code} > The reason is this: originally, DataStreamer::closeResponder always prints a > warning about InterruptedException; since HDFS-9812, > DFSOutputStream::closeImpl always forces threads to close, which causes > InterruptedException. > A simple fix is to use debug level log instead of warning level. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-10429) DataStreamer interrupted warning always appears when using CLI upload file
[ https://issues.apache.org/jira/browse/HDFS-10429?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15549949#comment-15549949 ] Zhiyuan Yang commented on HDFS-10429: - Weird, I'm unable to reproduce it neither... Should we mark this as "cannot reproduce" temporarily and reopen it once problem reappears? > DataStreamer interrupted warning always appears when using CLI upload file > --- > > Key: HDFS-10429 > URL: https://issues.apache.org/jira/browse/HDFS-10429 > Project: Hadoop HDFS > Issue Type: Bug >Affects Versions: 2.7.3 >Reporter: Zhiyuan Yang >Assignee: Zhiyuan Yang >Priority: Minor > Attachments: HDFS-10429.1.patch, HDFS-10429.2.patch > > > Every time I use 'hdfs dfs -put' upload file, this warning is printed: > {code:java} > 16/05/18 20:57:56 WARN hdfs.DataStreamer: Caught exception > java.lang.InterruptedException > at java.lang.Object.wait(Native Method) > at java.lang.Thread.join(Thread.java:1245) > at java.lang.Thread.join(Thread.java:1319) > at > org.apache.hadoop.hdfs.DataStreamer.closeResponder(DataStreamer.java:871) > at org.apache.hadoop.hdfs.DataStreamer.endBlock(DataStreamer.java:519) > at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:696) > {code} > The reason is this: originally, DataStreamer::closeResponder always prints a > warning about InterruptedException; since HDFS-9812, > DFSOutputStream::closeImpl always forces threads to close, which causes > InterruptedException. > A simple fix is to use debug level log instead of warning level. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-10936) allow & disallow snapshot command should not print exception trace when folder not found
[ https://issues.apache.org/jira/browse/HDFS-10936?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15549937#comment-15549937 ] Hadoop QA commented on HDFS-10936: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 10s{color} | {color:blue} Docker mode activated. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:red}-1{color} | {color:red} test4tests {color} | {color:red} 0m 0s{color} | {color:red} The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch. {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 8m 17s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 49s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 29s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 57s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 13s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 47s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 0s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 56s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 49s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 49s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 28s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 53s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 11s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 2m 3s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 59s{color} | {color:green} the patch passed {color} | | {color:red}-1{color} | {color:red} unit {color} | {color:red} 62m 16s{color} | {color:red} hadoop-hdfs in the patch failed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 21s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 84m 0s{color} | {color:black} {color} | \\ \\ || Reason || Tests || | Failed junit tests | hadoop.hdfs.server.namenode.TestDiskspaceQuotaUpdate | \\ \\ || Subsystem || Report/Notes || | Docker | Image:yetus/hadoop:9560f25 | | JIRA Issue | HDFS-10936 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12831799/HDFS-10936_02.patch | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite unit findbugs checkstyle | | uname | Linux 1211c2231fb6 3.13.0-95-generic #142-Ubuntu SMP Fri Aug 12 17:00:09 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh | | git revision | trunk / d65b957 | | Default Java | 1.8.0_101 | | findbugs | v3.0.0 | | unit | https://builds.apache.org/job/PreCommit-HDFS-Build/17025/artifact/patchprocess/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt | | Test Results | https://builds.apache.org/job/PreCommit-HDFS-Build/17025/testReport/ | | modules | C: hadoop-hdfs-project/hadoop-hdfs U: hadoop-hdfs-project/hadoop-hdfs | | Console output | https://builds.apache.org/job/PreCommit-HDFS-Build/17025/console | | Powered by | Apache Yetus 0.4.0-SNAPSHOT http://yetus.apache.org | This message was automatically generated. > allow & disallow snapshot command should not print exception trace when > folder not found > > > Key: HDFS-10936 > URL:
[jira] [Created] (HDFS-10964) Add more unit tests for ACLs
Namit Maheshwari created HDFS-10964: --- Summary: Add more unit tests for ACLs Key: HDFS-10964 URL: https://issues.apache.org/jira/browse/HDFS-10964 Project: Hadoop HDFS Issue Type: Sub-task Components: test Reporter: Namit Maheshwari Assignee: Namit Maheshwari This Jira proposes to add more unit tests to validate ACLs. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Resolved] (HDFS-10950) Add unit tests to verify ACLs in safemode
[ https://issues.apache.org/jira/browse/HDFS-10950?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Xiaobing Zhou resolved HDFS-10950. -- Resolution: Invalid Essentially, test_resetacl calls setAcl, getacl_recursive calls get getAclStatus recursively, and setfacl_default calls setAcl again. They can be all covered by setAcl and getAclStatus. There' no point to verify them. Resolved this as invalid. > Add unit tests to verify ACLs in safemode > - > > Key: HDFS-10950 > URL: https://issues.apache.org/jira/browse/HDFS-10950 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: fs, test >Reporter: Xiaobing Zhou >Assignee: Xiaobing Zhou > > This proposes adding unit tests to validate that getting Acls works when > namende is in safemode, while setting Acls fails. Specifically, the following > needs being covered in newly added tests. > test_getfacl_recursive > test_resetacl > test_setfacl_default -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-10637) Modifications to remove the assumption that FsVolumes are backed by java.io.File.
[ https://issues.apache.org/jira/browse/HDFS-10637?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Virajith Jalaparti updated HDFS-10637: -- Status: Patch Available (was: Open) > Modifications to remove the assumption that FsVolumes are backed by > java.io.File. > - > > Key: HDFS-10637 > URL: https://issues.apache.org/jira/browse/HDFS-10637 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: datanode, fs >Reporter: Virajith Jalaparti >Assignee: Virajith Jalaparti > Attachments: HDFS-10637.001.patch, HDFS-10637.002.patch, > HDFS-10637.003.patch, HDFS-10637.004.patch, HDFS-10637.005.patch, > HDFS-10637.006.patch, HDFS-10637.007.patch, HDFS-10637.008.patch, > HDFS-10637.009.patch, HDFS-10637.010.patch, HDFS-10637.011.patch > > > Modifications to {{FsVolumeSpi}} and {{FsVolumeImpl}} to remove references to > {{java.io.File}}. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-10637) Modifications to remove the assumption that FsVolumes are backed by java.io.File.
[ https://issues.apache.org/jira/browse/HDFS-10637?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Virajith Jalaparti updated HDFS-10637: -- Attachment: HDFS-10637.011.patch New patch to fix the failing test case in {{TestDirectoryScanner}}. > Modifications to remove the assumption that FsVolumes are backed by > java.io.File. > - > > Key: HDFS-10637 > URL: https://issues.apache.org/jira/browse/HDFS-10637 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: datanode, fs >Reporter: Virajith Jalaparti >Assignee: Virajith Jalaparti > Attachments: HDFS-10637.001.patch, HDFS-10637.002.patch, > HDFS-10637.003.patch, HDFS-10637.004.patch, HDFS-10637.005.patch, > HDFS-10637.006.patch, HDFS-10637.007.patch, HDFS-10637.008.patch, > HDFS-10637.009.patch, HDFS-10637.010.patch, HDFS-10637.011.patch > > > Modifications to {{FsVolumeSpi}} and {{FsVolumeImpl}} to remove references to > {{java.io.File}}. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-10609) Uncaught InvalidEncryptionKeyException during pipeline recovery may abort downstream applications
[ https://issues.apache.org/jira/browse/HDFS-10609?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Wei-Chiu Chuang updated HDFS-10609: --- Resolution: Fixed Fix Version/s: 2.7.4 Status: Resolved (was: Patch Available) > Uncaught InvalidEncryptionKeyException during pipeline recovery may abort > downstream applications > - > > Key: HDFS-10609 > URL: https://issues.apache.org/jira/browse/HDFS-10609 > Project: Hadoop HDFS > Issue Type: Bug > Components: encryption >Affects Versions: 2.6.0 > Environment: CDH5.8.0 >Reporter: Wei-Chiu Chuang >Assignee: Wei-Chiu Chuang > Fix For: 2.8.0, 2.7.4, 3.0.0-alpha2 > > Attachments: HDFS-10609.001.patch, HDFS-10609.002.patch, > HDFS-10609.003.patch, HDFS-10609.004.patch, HDFS-10609.005.patch, > HDFS-10609.branch-2.7.patch > > > In normal operations, if SASL negotiation fails due to > {{InvalidEncryptionKeyException}}, it is typically a benign exception, which > is caught and retried : > {code:title=SaslDataTransferServer#doSaslHandshake} > if (ioe instanceof SaslException && > ioe.getCause() != null && > ioe.getCause() instanceof InvalidEncryptionKeyException) { > // This could just be because the client is long-lived and hasn't gotten > // a new encryption key from the NN in a while. Upon receiving this > // error, the client will get a new encryption key from the NN and retry > // connecting to this DN. > sendInvalidKeySaslErrorMessage(out, ioe.getCause().getMessage()); > } > {code} > {code:title=DFSOutputStream.DataStreamer#createBlockOutputStream} > if (ie instanceof InvalidEncryptionKeyException && refetchEncryptionKey > 0) { > DFSClient.LOG.info("Will fetch a new encryption key and retry, " > + "encryption key was invalid when connecting to " > + nodes[0] + " : " + ie); > {code} > However, if the exception is thrown during pipeline recovery, the > corresponding code does not handle it properly, and the exception is spilled > out to downstream applications, such as SOLR, aborting its operation: > {quote} > 2016-07-06 12:12:51,992 ERROR org.apache.solr.update.HdfsTransactionLog: > Exception closing tlog. > org.apache.hadoop.hdfs.protocol.datatransfer.InvalidEncryptionKeyException: > Can't re-compute encryption key for nonce, since the required block key > (keyID=557709482) doesn't exist. Current key: 1350592619 > at > org.apache.hadoop.hdfs.protocol.datatransfer.sasl.DataTransferSaslUtil.readSaslMessageAndNegotiatedCipherOption(DataTransferSaslUtil.java:417) > at > org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient.doSaslHandshake(SaslDataTransferClient.java:474) > at > org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient.getEncryptedStreams(SaslDataTransferClient.java:299) > at > org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient.send(SaslDataTransferClient.java:242) > at > org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient.checkTrustAndSend(SaslDataTransferClient.java:211) > at > org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient.socketSend(SaslDataTransferClient.java:183) > at > org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.transfer(DFSOutputStream.java:1308) > at > org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.addDatanode2ExistingPipeline(DFSOutputStream.java:1272) > at > org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.setupPipelineForAppendOrRecovery(DFSOutputStream.java:1433) > at > org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.processDatanodeError(DFSOutputStream.java:1147) > at > org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:632) > 2016-07-06 12:12:51,997 ERROR org.apache.solr.update.CommitTracker: auto > commit error...:org.apache.solr.common.SolrException: > org.apache.hadoop.hdfs.protocol.datatransfer.InvalidEncryptionKeyException: > Can't re-compute encryption key for nonce, since the required block key > (keyID=557709482) doesn't exist. Current key: 1350592619 > at > org.apache.solr.update.HdfsTransactionLog.close(HdfsTransactionLog.java:316) > at > org.apache.solr.update.TransactionLog.decref(TransactionLog.java:505) > at org.apache.solr.update.UpdateLog.addOldLog(UpdateLog.java:380) > at org.apache.solr.update.UpdateLog.postCommit(UpdateLog.java:676) > at > org.apache.solr.update.DirectUpdateHandler2.commit(DirectUpdateHandler2.java:623) > at org.apache.solr.update.CommitTracker.run(CommitTracker.java:216) > at >
[jira] [Updated] (HDFS-10950) Add unit tests to verify ACLs in safemode
[ https://issues.apache.org/jira/browse/HDFS-10950?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Xiaobing Zhou updated HDFS-10950: - Description: This proposes adding unit tests to validate that getting Acls works when namende is in safemode, while setting Acls fails. Specifically, the following needs being covered in newly added tests. test_getfacl_recursive test_resetacl test_setfacl_default was: This proposes adding unit tests to validate that getting Acls works when namende is in safemode, while setting Acls fails. Specifically, the following needs being covered in newly added tests. test_delete_acls test_delete_default_acls test_delete_specific_acls test_getfacl test_getfacl_recursive test_resetacl test_setfacl test_setfacl_default > Add unit tests to verify ACLs in safemode > - > > Key: HDFS-10950 > URL: https://issues.apache.org/jira/browse/HDFS-10950 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: fs, test >Reporter: Xiaobing Zhou >Assignee: Xiaobing Zhou > > This proposes adding unit tests to validate that getting Acls works when > namende is in safemode, while setting Acls fails. Specifically, the following > needs being covered in newly added tests. > test_getfacl_recursive > test_resetacl > test_setfacl_default -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-10609) Uncaught InvalidEncryptionKeyException during pipeline recovery may abort downstream applications
[ https://issues.apache.org/jira/browse/HDFS-10609?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15549850#comment-15549850 ] Wei-Chiu Chuang commented on HDFS-10609: Committing this based on [~andrew.wang]'s +1 and also the patch passed our internal testing. Thanks [~andrew.wang]!. > Uncaught InvalidEncryptionKeyException during pipeline recovery may abort > downstream applications > - > > Key: HDFS-10609 > URL: https://issues.apache.org/jira/browse/HDFS-10609 > Project: Hadoop HDFS > Issue Type: Bug > Components: encryption >Affects Versions: 2.6.0 > Environment: CDH5.8.0 >Reporter: Wei-Chiu Chuang >Assignee: Wei-Chiu Chuang > Fix For: 2.8.0, 3.0.0-alpha2 > > Attachments: HDFS-10609.001.patch, HDFS-10609.002.patch, > HDFS-10609.003.patch, HDFS-10609.004.patch, HDFS-10609.005.patch, > HDFS-10609.branch-2.7.patch > > > In normal operations, if SASL negotiation fails due to > {{InvalidEncryptionKeyException}}, it is typically a benign exception, which > is caught and retried : > {code:title=SaslDataTransferServer#doSaslHandshake} > if (ioe instanceof SaslException && > ioe.getCause() != null && > ioe.getCause() instanceof InvalidEncryptionKeyException) { > // This could just be because the client is long-lived and hasn't gotten > // a new encryption key from the NN in a while. Upon receiving this > // error, the client will get a new encryption key from the NN and retry > // connecting to this DN. > sendInvalidKeySaslErrorMessage(out, ioe.getCause().getMessage()); > } > {code} > {code:title=DFSOutputStream.DataStreamer#createBlockOutputStream} > if (ie instanceof InvalidEncryptionKeyException && refetchEncryptionKey > 0) { > DFSClient.LOG.info("Will fetch a new encryption key and retry, " > + "encryption key was invalid when connecting to " > + nodes[0] + " : " + ie); > {code} > However, if the exception is thrown during pipeline recovery, the > corresponding code does not handle it properly, and the exception is spilled > out to downstream applications, such as SOLR, aborting its operation: > {quote} > 2016-07-06 12:12:51,992 ERROR org.apache.solr.update.HdfsTransactionLog: > Exception closing tlog. > org.apache.hadoop.hdfs.protocol.datatransfer.InvalidEncryptionKeyException: > Can't re-compute encryption key for nonce, since the required block key > (keyID=557709482) doesn't exist. Current key: 1350592619 > at > org.apache.hadoop.hdfs.protocol.datatransfer.sasl.DataTransferSaslUtil.readSaslMessageAndNegotiatedCipherOption(DataTransferSaslUtil.java:417) > at > org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient.doSaslHandshake(SaslDataTransferClient.java:474) > at > org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient.getEncryptedStreams(SaslDataTransferClient.java:299) > at > org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient.send(SaslDataTransferClient.java:242) > at > org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient.checkTrustAndSend(SaslDataTransferClient.java:211) > at > org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient.socketSend(SaslDataTransferClient.java:183) > at > org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.transfer(DFSOutputStream.java:1308) > at > org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.addDatanode2ExistingPipeline(DFSOutputStream.java:1272) > at > org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.setupPipelineForAppendOrRecovery(DFSOutputStream.java:1433) > at > org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.processDatanodeError(DFSOutputStream.java:1147) > at > org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:632) > 2016-07-06 12:12:51,997 ERROR org.apache.solr.update.CommitTracker: auto > commit error...:org.apache.solr.common.SolrException: > org.apache.hadoop.hdfs.protocol.datatransfer.InvalidEncryptionKeyException: > Can't re-compute encryption key for nonce, since the required block key > (keyID=557709482) doesn't exist. Current key: 1350592619 > at > org.apache.solr.update.HdfsTransactionLog.close(HdfsTransactionLog.java:316) > at > org.apache.solr.update.TransactionLog.decref(TransactionLog.java:505) > at org.apache.solr.update.UpdateLog.addOldLog(UpdateLog.java:380) > at org.apache.solr.update.UpdateLog.postCommit(UpdateLog.java:676) > at > org.apache.solr.update.DirectUpdateHandler2.commit(DirectUpdateHandler2.java:623) > at org.apache.solr.update.CommitTracker.run(CommitTracker.java:216) >
[jira] [Commented] (HDFS-10962) TestRequestHedgingProxyProvider is flaky
[ https://issues.apache.org/jira/browse/HDFS-10962?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15549831#comment-15549831 ] Hadoop QA commented on HDFS-10962: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 10s{color} | {color:blue} Docker mode activated. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 1 new or modified test files. {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 6m 53s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 43s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 25s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 50s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 11s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 40s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 55s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 46s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 41s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 41s{color} | {color:green} the patch passed {color} | | {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange} 0m 22s{color} | {color:orange} hadoop-hdfs-project/hadoop-hdfs: The patch generated 2 new + 6 unchanged - 0 fixed = 8 total (was 6) {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 46s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 9s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 45s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 50s{color} | {color:green} the patch passed {color} | | {color:red}-1{color} | {color:red} unit {color} | {color:red} 62m 50s{color} | {color:red} hadoop-hdfs in the patch failed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 20s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 81m 29s{color} | {color:black} {color} | \\ \\ || Reason || Tests || | Failed junit tests | hadoop.hdfs.server.namenode.TestAddStripedBlockInFBR | \\ \\ || Subsystem || Report/Notes || | Docker | Image:yetus/hadoop:9560f25 | | JIRA Issue | HDFS-10962 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12831792/HDFS-10962.002.patch | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite unit findbugs checkstyle | | uname | Linux ad0b8c62a276 3.13.0-95-generic #142-Ubuntu SMP Fri Aug 12 17:00:09 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh | | git revision | trunk / 8867762 | | Default Java | 1.8.0_101 | | findbugs | v3.0.0 | | checkstyle | https://builds.apache.org/job/PreCommit-HDFS-Build/17024/artifact/patchprocess/diff-checkstyle-hadoop-hdfs-project_hadoop-hdfs.txt | | unit | https://builds.apache.org/job/PreCommit-HDFS-Build/17024/artifact/patchprocess/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt | | Test Results | https://builds.apache.org/job/PreCommit-HDFS-Build/17024/testReport/ | | modules | C: hadoop-hdfs-project/hadoop-hdfs U: hadoop-hdfs-project/hadoop-hdfs | | Console output | https://builds.apache.org/job/PreCommit-HDFS-Build/17024/console | | Powered by | Apache Yetus 0.4.0-SNAPSHOT http://yetus.apache.org | This message was automatically generated. > TestRequestHedgingProxyProvider is flaky > > > Key: HDFS-10962 > URL:
[jira] [Commented] (HDFS-10950) Add unit tests to verify ACLs in safemode
[ https://issues.apache.org/jira/browse/HDFS-10950?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15549777#comment-15549777 ] Xiaobing Zhou commented on HDFS-10950: -- [~cnauroth], thanks for bringing them up. I added those cases need to be covered in description. > Add unit tests to verify ACLs in safemode > - > > Key: HDFS-10950 > URL: https://issues.apache.org/jira/browse/HDFS-10950 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: fs, test >Reporter: Xiaobing Zhou >Assignee: Xiaobing Zhou > > This proposes adding unit tests to validate that getting Acls works when > namende is in safemode, while setting Acls fails. Specifically, the following > needs being covered in newly added tests. > test_delete_acls > test_delete_default_acls > test_delete_specific_acls > test_getfacl > test_getfacl_recursive > test_resetacl > test_setfacl > test_setfacl_default -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-10950) Add unit tests to verify ACLs in safemode
[ https://issues.apache.org/jira/browse/HDFS-10950?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Xiaobing Zhou updated HDFS-10950: - Description: This proposes adding unit tests to validate that getting Acls works when namende is in safemode, while setting Acls fails. Specifically, the following needs being covered in newly added tests. test_delete_acls test_delete_default_acls test_delete_specific_acls test_getfacl test_getfacl_recursive test_resetacl test_setfacl test_setfacl_default was:This proposes adding unit tests to validate that getting Acls works when namende is in safemode, while setting Acls fails. > Add unit tests to verify ACLs in safemode > - > > Key: HDFS-10950 > URL: https://issues.apache.org/jira/browse/HDFS-10950 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: fs, test >Reporter: Xiaobing Zhou >Assignee: Xiaobing Zhou > > This proposes adding unit tests to validate that getting Acls works when > namende is in safemode, while setting Acls fails. Specifically, the following > needs being covered in newly added tests. > test_delete_acls > test_delete_default_acls > test_delete_specific_acls > test_getfacl > test_getfacl_recursive > test_resetacl > test_setfacl > test_setfacl_default -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-10898) libhdfs++: Make log levels consistent
[ https://issues.apache.org/jira/browse/HDFS-10898?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] James Clampffer updated HDFS-10898: --- Resolution: Fixed Status: Resolved (was: Patch Available) Thanks for the reviews [~bobhansen] and [~swalk...@yahoo.com]. Committed to HDFS-8707 > libhdfs++: Make log levels consistent > - > > Key: HDFS-10898 > URL: https://issues.apache.org/jira/browse/HDFS-10898 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: hdfs-client >Reporter: James Clampffer >Assignee: James Clampffer >Priority: Trivial > Attachments: HDFS-10898.HDFS-8707.000.patch, > HDFS-10898.HDFS-8707.001.patch > > > Most of the public C++ FileHandle/FileSystem operations have a LOG_TRACE > level message about parameters passed in etc. However many methods use > LOG_DEBUG and a couple use LOG_INFO. > We most likely want FS operations that happen a lot (read/open/seek/stat) to > stick to LOG_DEBUG consistently and only use LOG_INFO for things like > FileSystem::Connect or RpcConnection:: that don't get called often and are > important enough to warrant showing up in the log. LOG_TRACE can be reserved > for things happening deeper inside public methods and methods that aren't > part of the public API. > Related improvements that could be brought into this to avoid opening a ton > of small Jiras: > -Print the "this" pointer address in the log message to make it easier to > correlate objects when there's concurrent work being done. This has been > very helpful in the past but often got stripped out before patches went in. > People just need be aware that operator new may eventually place an object of > the same type at the same address sometime in the future. > -For objects owned by other objects, but created on the fly, include a > pointer back to the parent/creator object if that pointer is already being > tracked (See the nested stucts in BlockReaderImpl). -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-10936) allow & disallow snapshot command should not print exception trace when folder not found
[ https://issues.apache.org/jira/browse/HDFS-10936?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jagadesh Kiran N updated HDFS-10936: Attachment: HDFS-10936_02.patch Updated the patch correcting checkstyle errors. Please review > allow & disallow snapshot command should not print exception trace when > folder not found > > > Key: HDFS-10936 > URL: https://issues.apache.org/jira/browse/HDFS-10936 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Jagadesh Kiran N >Assignee: Jagadesh Kiran N > Attachments: HDFS-10936_00.patch, HDFS-10936_01.patch, > HDFS-10936_02.patch > > > Allow snapshot for a folder which is not present > {code}./hdfs dfsadmin -allowsnapshot /kiran1 {code} > the following exception occurs > {code} > allowsnapshot: Directory does not exist: /kiran1 > at > org.apache.hadoop.hdfs.server.namenode.INodeDirectory.valueOf(INodeDirectory.java:61) > at > org.apache.hadoop.hdfs.server.namenode.snapshot.SnapshotManager.setSnapshottable(SnapshotManager.java:113) > at > org.apache.hadoop.hdfs.server.namenode.FSDirSnapshotOp.allowSnapshot(FSDirSnapshotOp.java:60) > at > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.allowSnapshot(FSNamesystem.java:7406) > at > org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.allowSnapshot(NameNodeRpcServer.java:1766) > at > org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.allowSnapshot(ClientNamenodeProtocolServerSideTranslatorPB.java:1131) > at > org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) > at > org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:616) > at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:973) > at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2143) > at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2139) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:422) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1726) > at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2137) > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-10962) TestRequestHedgingProxyProvider is flaky
[ https://issues.apache.org/jira/browse/HDFS-10962?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15549662#comment-15549662 ] Xiao Chen commented on HDFS-10962: -- Thanks Andrew, looks great. +1 pending jenkins. > TestRequestHedgingProxyProvider is flaky > > > Key: HDFS-10962 > URL: https://issues.apache.org/jira/browse/HDFS-10962 > Project: Hadoop HDFS > Issue Type: Bug > Components: test >Affects Versions: 2.6.4 >Reporter: Andrew Wang >Assignee: Andrew Wang > Attachments: HDFS-10962.001.patch, HDFS-10962.002.patch > > > This test fails occasionally with an error like this: > {noformat} > org.apache.hadoop.hdfs.server.namenode.ha.TestRequestHedgingProxyProvider.testHedgingWhenOneFails > Error Message > Wanted but not invoked: > namenodeProtocols.getStats(); > -> at > org.apache.hadoop.hdfs.server.namenode.ha.TestRequestHedgingProxyProvider.testHedgingWhenOneFails(TestRequestHedgingProxyProvider.java:78) > Actually, there were zero interactions with this mock. > Stacktrace > org.mockito.exceptions.verification.WantedButNotInvoked: > Wanted but not invoked: > namenodeProtocols.getStats(); > -> at > org.apache.hadoop.hdfs.server.namenode.ha.TestRequestHedgingProxyProvider.testHedgingWhenOneFails(TestRequestHedgingProxyProvider.java:78) > Actually, there were zero interactions with this mock. > at > org.apache.hadoop.hdfs.server.namenode.ha.TestRequestHedgingProxyProvider.testHedgingWhenOneFails(TestRequestHedgingProxyProvider.java:78) > Standard Output > 2016-09-26 15:26:02,234 WARN hdfs.DFSUtil > (DFSUtil.java:getAddressesForNameserviceId(689)) - Namenode for > mycluster-26780990 remains unresolved for ID nn1. Check your hdfs-site.xml > file to ensure namenodes are configured properly. > 2016-09-26 15:26:02,340 WARN hdfs.DFSUtil > (DFSUtil.java:getAddressesForNameserviceId(689)) - Namenode for > mycluster-26780990 remains unresolved for ID nn2. Check your hdfs-site.xml > file to ensure namenodes are configured properly. > {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-10797) Disk usage summary of snapshots causes renamed blocks to get counted twice
[ https://issues.apache.org/jira/browse/HDFS-10797?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15549659#comment-15549659 ] Xiao Chen commented on HDFS-10797: -- Thanks for revving Sean. +1 on patch 10. [~jingzhao] do you have any comments? > Disk usage summary of snapshots causes renamed blocks to get counted twice > -- > > Key: HDFS-10797 > URL: https://issues.apache.org/jira/browse/HDFS-10797 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Sean Mackrory >Assignee: Sean Mackrory > Attachments: HDFS-10797.001.patch, HDFS-10797.002.patch, > HDFS-10797.003.patch, HDFS-10797.004.patch, HDFS-10797.005.patch, > HDFS-10797.006.patch, HDFS-10797.007.patch, HDFS-10797.008.patch, > HDFS-10797.009.patch, HDFS-10797.010.patch, HDFS-10797.010.patch > > > DirectoryWithSnapshotFeature.computeContentSummary4Snapshot calculates how > much disk usage is used by a snapshot by tallying up the files in the > snapshot that have since been deleted (that way it won't overlap with regular > files whose disk usage is computed separately). However that is determined > from a diff that shows moved (to Trash or otherwise) or renamed files as a > deletion and a creation operation that may overlap with the list of blocks. > Only the deletion operation is taken into consideration, and this causes > those blocks to get represented twice in the disk usage tallying. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-10962) TestRequestHedgingProxyProvider is flaky
[ https://issues.apache.org/jira/browse/HDFS-10962?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Andrew Wang updated HDFS-10962: --- Attachment: HDFS-10962.002.patch Thanks for reviewing Xiao, I added some TRACE logging that will help confirm or deny my hypothesis if there is a future failure. > TestRequestHedgingProxyProvider is flaky > > > Key: HDFS-10962 > URL: https://issues.apache.org/jira/browse/HDFS-10962 > Project: Hadoop HDFS > Issue Type: Bug > Components: test >Affects Versions: 2.6.4 >Reporter: Andrew Wang >Assignee: Andrew Wang > Attachments: HDFS-10962.001.patch, HDFS-10962.002.patch > > > This test fails occasionally with an error like this: > {noformat} > org.apache.hadoop.hdfs.server.namenode.ha.TestRequestHedgingProxyProvider.testHedgingWhenOneFails > Error Message > Wanted but not invoked: > namenodeProtocols.getStats(); > -> at > org.apache.hadoop.hdfs.server.namenode.ha.TestRequestHedgingProxyProvider.testHedgingWhenOneFails(TestRequestHedgingProxyProvider.java:78) > Actually, there were zero interactions with this mock. > Stacktrace > org.mockito.exceptions.verification.WantedButNotInvoked: > Wanted but not invoked: > namenodeProtocols.getStats(); > -> at > org.apache.hadoop.hdfs.server.namenode.ha.TestRequestHedgingProxyProvider.testHedgingWhenOneFails(TestRequestHedgingProxyProvider.java:78) > Actually, there were zero interactions with this mock. > at > org.apache.hadoop.hdfs.server.namenode.ha.TestRequestHedgingProxyProvider.testHedgingWhenOneFails(TestRequestHedgingProxyProvider.java:78) > Standard Output > 2016-09-26 15:26:02,234 WARN hdfs.DFSUtil > (DFSUtil.java:getAddressesForNameserviceId(689)) - Namenode for > mycluster-26780990 remains unresolved for ID nn1. Check your hdfs-site.xml > file to ensure namenodes are configured properly. > 2016-09-26 15:26:02,340 WARN hdfs.DFSUtil > (DFSUtil.java:getAddressesForNameserviceId(689)) - Namenode for > mycluster-26780990 remains unresolved for ID nn2. Check your hdfs-site.xml > file to ensure namenodes are configured properly. > {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-10797) Disk usage summary of snapshots causes renamed blocks to get counted twice
[ https://issues.apache.org/jira/browse/HDFS-10797?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15549617#comment-15549617 ] Hadoop QA commented on HDFS-10797: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 17s{color} | {color:blue} Docker mode activated. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 1 new or modified test files. {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 6m 56s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 44s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 30s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 52s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 12s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 42s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 55s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 46s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 41s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 41s{color} | {color:green} the patch passed {color} | | {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange} 0m 27s{color} | {color:orange} hadoop-hdfs-project/hadoop-hdfs: The patch generated 1 new + 267 unchanged - 0 fixed = 268 total (was 267) {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 48s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 10s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 46s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 51s{color} | {color:green} the patch passed {color} | | {color:red}-1{color} | {color:red} unit {color} | {color:red} 57m 5s{color} | {color:red} hadoop-hdfs in the patch failed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 19s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 76m 21s{color} | {color:black} {color} | \\ \\ || Reason || Tests || | Failed junit tests | hadoop.hdfs.server.balancer.TestBalancerWithMultipleNameNodes | \\ \\ || Subsystem || Report/Notes || | Docker | Image:yetus/hadoop:9560f25 | | JIRA Issue | HDFS-10797 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12831783/HDFS-10797.010.patch | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite unit findbugs checkstyle | | uname | Linux 13a2e115431f 3.13.0-36-lowlatency #63-Ubuntu SMP PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh | | git revision | trunk / d6be1e7 | | Default Java | 1.8.0_101 | | findbugs | v3.0.0 | | checkstyle | https://builds.apache.org/job/PreCommit-HDFS-Build/17022/artifact/patchprocess/diff-checkstyle-hadoop-hdfs-project_hadoop-hdfs.txt | | unit | https://builds.apache.org/job/PreCommit-HDFS-Build/17022/artifact/patchprocess/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt | | Test Results | https://builds.apache.org/job/PreCommit-HDFS-Build/17022/testReport/ | | modules | C: hadoop-hdfs-project/hadoop-hdfs U: hadoop-hdfs-project/hadoop-hdfs | | Console output | https://builds.apache.org/job/PreCommit-HDFS-Build/17022/console | | Powered by | Apache Yetus 0.4.0-SNAPSHOT http://yetus.apache.org | This message was automatically generated. > Disk usage summary of snapshots causes renamed blocks to get counted twice >
[jira] [Commented] (HDFS-10957) Retire BKJM from trunk
[ https://issues.apache.org/jira/browse/HDFS-10957?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15549609#comment-15549609 ] Vinayakumar B commented on HDFS-10957: -- Thanks [~kihwal], I will update the patch. > Retire BKJM from trunk > -- > > Key: HDFS-10957 > URL: https://issues.apache.org/jira/browse/HDFS-10957 > Project: Hadoop HDFS > Issue Type: Task >Reporter: Vinayakumar B >Assignee: Vinayakumar B > Attachments: HDFS-10957-01.patch, HDFS-10957-02.patch > > > Based on discussions done in mailing list > [here|http://mail-archives.apache.org/mod_mbox/hadoop-hdfs-dev/201607.mbox/%3c1288296033.5942327.1469727453010.javamail.ya...@mail.yahoo.com%3E] > BKJM is no longer required as QJM supports the required functionality with > minimum deployment overhead. > This Jira is for tracking purpose. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-10957) Retire BKJM from trunk
[ https://issues.apache.org/jira/browse/HDFS-10957?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vinayakumar B updated HDFS-10957: - Attachment: HDFS-10957-02.patch Updated the patch > Retire BKJM from trunk > -- > > Key: HDFS-10957 > URL: https://issues.apache.org/jira/browse/HDFS-10957 > Project: Hadoop HDFS > Issue Type: Task >Reporter: Vinayakumar B >Assignee: Vinayakumar B > Attachments: HDFS-10957-01.patch, HDFS-10957-02.patch > > > Based on discussions done in mailing list > [here|http://mail-archives.apache.org/mod_mbox/hadoop-hdfs-dev/201607.mbox/%3c1288296033.5942327.1469727453010.javamail.ya...@mail.yahoo.com%3E] > BKJM is no longer required as QJM supports the required functionality with > minimum deployment overhead. > This Jira is for tracking purpose. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-10826) Correctly report missing EC blocks in FSCK
[ https://issues.apache.org/jira/browse/HDFS-10826?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15549565#comment-15549565 ] Hudson commented on HDFS-10826: --- SUCCESS: Integrated in Jenkins build Hadoop-trunk-Commit #10546 (See [https://builds.apache.org/job/Hadoop-trunk-Commit/10546/]) HDFS-10826. Correctly report missing EC blocks in FSCK. Contributed by (jing9: rev 886776225611ca3cfff32dd94ea24fe618f14464) * (edit) hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFsck.java * (edit) hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/BlockManager.java * (edit) hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/NamenodeFsck.java * (edit) hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestLeaseRecoveryStriped.java > Correctly report missing EC blocks in FSCK > -- > > Key: HDFS-10826 > URL: https://issues.apache.org/jira/browse/HDFS-10826 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: erasure-coding >Reporter: Takanobu Asanuma >Assignee: Takanobu Asanuma > Attachments: HDFS-10826.2.patch, HDFS-10826.3.patch, > HDFS-10826.4.patch, HDFS-10826.5.patch, HDFS-10826.WIP.1.patch > > > For RS-6-3, when there is one ec block group and > 1) 0~3 out of 9 internal blocks are missing, the result of fsck is HEALTY. > 2) 4~8 out of 9 internal blocks are missing, the result of fsck is HEALTY. > {noformat} > Erasure Coded Block Groups: > Total size:536870912 B > Total files: 1 > Total block groups (validated):1 (avg. block group size 536870912 B) > > UNRECOVERABLE BLOCK GROUPS: 1 (100.0 %) > > Minimally erasure-coded block groups: 0 (0.0 %) > Over-erasure-coded block groups: 0 (0.0 %) > Under-erasure-coded block groups: 1 (100.0 %) > Unsatisfactory placement block groups: 0 (0.0 %) > Default ecPolicy: RS-DEFAULT-6-3-64k > Average block group size: 5.0 > Missing block groups: 0 > Corrupt block groups: 0 > Missing internal blocks: 4 (44.43 %) > FSCK ended at Wed Aug 31 13:42:05 JST 2016 in 4 milliseconds > The filesystem under path '/' is HEALTHY > {noformat} > 3) 9 out of 9 internal blocks are missing, the result of fsck is CRITICAL. > (Because it is regarded as a missing block group.) > In case 2), the result should be CRITICAL since the ec block group is > unrecoverable. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-10609) Uncaught InvalidEncryptionKeyException during pipeline recovery may abort downstream applications
[ https://issues.apache.org/jira/browse/HDFS-10609?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15549539#comment-15549539 ] Andrew Wang commented on HDFS-10609: +1 thanks Wei-Chiu > Uncaught InvalidEncryptionKeyException during pipeline recovery may abort > downstream applications > - > > Key: HDFS-10609 > URL: https://issues.apache.org/jira/browse/HDFS-10609 > Project: Hadoop HDFS > Issue Type: Bug > Components: encryption >Affects Versions: 2.6.0 > Environment: CDH5.8.0 >Reporter: Wei-Chiu Chuang >Assignee: Wei-Chiu Chuang > Fix For: 2.8.0, 3.0.0-alpha2 > > Attachments: HDFS-10609.001.patch, HDFS-10609.002.patch, > HDFS-10609.003.patch, HDFS-10609.004.patch, HDFS-10609.005.patch, > HDFS-10609.branch-2.7.patch > > > In normal operations, if SASL negotiation fails due to > {{InvalidEncryptionKeyException}}, it is typically a benign exception, which > is caught and retried : > {code:title=SaslDataTransferServer#doSaslHandshake} > if (ioe instanceof SaslException && > ioe.getCause() != null && > ioe.getCause() instanceof InvalidEncryptionKeyException) { > // This could just be because the client is long-lived and hasn't gotten > // a new encryption key from the NN in a while. Upon receiving this > // error, the client will get a new encryption key from the NN and retry > // connecting to this DN. > sendInvalidKeySaslErrorMessage(out, ioe.getCause().getMessage()); > } > {code} > {code:title=DFSOutputStream.DataStreamer#createBlockOutputStream} > if (ie instanceof InvalidEncryptionKeyException && refetchEncryptionKey > 0) { > DFSClient.LOG.info("Will fetch a new encryption key and retry, " > + "encryption key was invalid when connecting to " > + nodes[0] + " : " + ie); > {code} > However, if the exception is thrown during pipeline recovery, the > corresponding code does not handle it properly, and the exception is spilled > out to downstream applications, such as SOLR, aborting its operation: > {quote} > 2016-07-06 12:12:51,992 ERROR org.apache.solr.update.HdfsTransactionLog: > Exception closing tlog. > org.apache.hadoop.hdfs.protocol.datatransfer.InvalidEncryptionKeyException: > Can't re-compute encryption key for nonce, since the required block key > (keyID=557709482) doesn't exist. Current key: 1350592619 > at > org.apache.hadoop.hdfs.protocol.datatransfer.sasl.DataTransferSaslUtil.readSaslMessageAndNegotiatedCipherOption(DataTransferSaslUtil.java:417) > at > org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient.doSaslHandshake(SaslDataTransferClient.java:474) > at > org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient.getEncryptedStreams(SaslDataTransferClient.java:299) > at > org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient.send(SaslDataTransferClient.java:242) > at > org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient.checkTrustAndSend(SaslDataTransferClient.java:211) > at > org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient.socketSend(SaslDataTransferClient.java:183) > at > org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.transfer(DFSOutputStream.java:1308) > at > org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.addDatanode2ExistingPipeline(DFSOutputStream.java:1272) > at > org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.setupPipelineForAppendOrRecovery(DFSOutputStream.java:1433) > at > org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.processDatanodeError(DFSOutputStream.java:1147) > at > org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:632) > 2016-07-06 12:12:51,997 ERROR org.apache.solr.update.CommitTracker: auto > commit error...:org.apache.solr.common.SolrException: > org.apache.hadoop.hdfs.protocol.datatransfer.InvalidEncryptionKeyException: > Can't re-compute encryption key for nonce, since the required block key > (keyID=557709482) doesn't exist. Current key: 1350592619 > at > org.apache.solr.update.HdfsTransactionLog.close(HdfsTransactionLog.java:316) > at > org.apache.solr.update.TransactionLog.decref(TransactionLog.java:505) > at org.apache.solr.update.UpdateLog.addOldLog(UpdateLog.java:380) > at org.apache.solr.update.UpdateLog.postCommit(UpdateLog.java:676) > at > org.apache.solr.update.DirectUpdateHandler2.commit(DirectUpdateHandler2.java:623) > at org.apache.solr.update.CommitTracker.run(CommitTracker.java:216) > at > java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) > at
[jira] [Commented] (HDFS-10609) Uncaught InvalidEncryptionKeyException during pipeline recovery may abort downstream applications
[ https://issues.apache.org/jira/browse/HDFS-10609?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15549501#comment-15549501 ] Wei-Chiu Chuang commented on HDFS-10609: None of the tests failed in precommit failed in my local tree. All warnings are unrelated to this patch. > Uncaught InvalidEncryptionKeyException during pipeline recovery may abort > downstream applications > - > > Key: HDFS-10609 > URL: https://issues.apache.org/jira/browse/HDFS-10609 > Project: Hadoop HDFS > Issue Type: Bug > Components: encryption >Affects Versions: 2.6.0 > Environment: CDH5.8.0 >Reporter: Wei-Chiu Chuang >Assignee: Wei-Chiu Chuang > Fix For: 2.8.0, 3.0.0-alpha2 > > Attachments: HDFS-10609.001.patch, HDFS-10609.002.patch, > HDFS-10609.003.patch, HDFS-10609.004.patch, HDFS-10609.005.patch, > HDFS-10609.branch-2.7.patch > > > In normal operations, if SASL negotiation fails due to > {{InvalidEncryptionKeyException}}, it is typically a benign exception, which > is caught and retried : > {code:title=SaslDataTransferServer#doSaslHandshake} > if (ioe instanceof SaslException && > ioe.getCause() != null && > ioe.getCause() instanceof InvalidEncryptionKeyException) { > // This could just be because the client is long-lived and hasn't gotten > // a new encryption key from the NN in a while. Upon receiving this > // error, the client will get a new encryption key from the NN and retry > // connecting to this DN. > sendInvalidKeySaslErrorMessage(out, ioe.getCause().getMessage()); > } > {code} > {code:title=DFSOutputStream.DataStreamer#createBlockOutputStream} > if (ie instanceof InvalidEncryptionKeyException && refetchEncryptionKey > 0) { > DFSClient.LOG.info("Will fetch a new encryption key and retry, " > + "encryption key was invalid when connecting to " > + nodes[0] + " : " + ie); > {code} > However, if the exception is thrown during pipeline recovery, the > corresponding code does not handle it properly, and the exception is spilled > out to downstream applications, such as SOLR, aborting its operation: > {quote} > 2016-07-06 12:12:51,992 ERROR org.apache.solr.update.HdfsTransactionLog: > Exception closing tlog. > org.apache.hadoop.hdfs.protocol.datatransfer.InvalidEncryptionKeyException: > Can't re-compute encryption key for nonce, since the required block key > (keyID=557709482) doesn't exist. Current key: 1350592619 > at > org.apache.hadoop.hdfs.protocol.datatransfer.sasl.DataTransferSaslUtil.readSaslMessageAndNegotiatedCipherOption(DataTransferSaslUtil.java:417) > at > org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient.doSaslHandshake(SaslDataTransferClient.java:474) > at > org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient.getEncryptedStreams(SaslDataTransferClient.java:299) > at > org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient.send(SaslDataTransferClient.java:242) > at > org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient.checkTrustAndSend(SaslDataTransferClient.java:211) > at > org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferClient.socketSend(SaslDataTransferClient.java:183) > at > org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.transfer(DFSOutputStream.java:1308) > at > org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.addDatanode2ExistingPipeline(DFSOutputStream.java:1272) > at > org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.setupPipelineForAppendOrRecovery(DFSOutputStream.java:1433) > at > org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.processDatanodeError(DFSOutputStream.java:1147) > at > org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:632) > 2016-07-06 12:12:51,997 ERROR org.apache.solr.update.CommitTracker: auto > commit error...:org.apache.solr.common.SolrException: > org.apache.hadoop.hdfs.protocol.datatransfer.InvalidEncryptionKeyException: > Can't re-compute encryption key for nonce, since the required block key > (keyID=557709482) doesn't exist. Current key: 1350592619 > at > org.apache.solr.update.HdfsTransactionLog.close(HdfsTransactionLog.java:316) > at > org.apache.solr.update.TransactionLog.decref(TransactionLog.java:505) > at org.apache.solr.update.UpdateLog.addOldLog(UpdateLog.java:380) > at org.apache.solr.update.UpdateLog.postCommit(UpdateLog.java:676) > at > org.apache.solr.update.DirectUpdateHandler2.commit(DirectUpdateHandler2.java:623) > at org.apache.solr.update.CommitTracker.run(CommitTracker.java:216) > at >
[jira] [Updated] (HDFS-10826) Correctly report missing EC blocks in FSCK
[ https://issues.apache.org/jira/browse/HDFS-10826?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jing Zhao updated HDFS-10826: - Summary: Correctly report missing EC blocks in FSCK (was: The result of fsck should be CRITICAL when there are unrecoverable ec block groups.) > Correctly report missing EC blocks in FSCK > -- > > Key: HDFS-10826 > URL: https://issues.apache.org/jira/browse/HDFS-10826 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: erasure-coding >Reporter: Takanobu Asanuma >Assignee: Takanobu Asanuma > Attachments: HDFS-10826.2.patch, HDFS-10826.3.patch, > HDFS-10826.4.patch, HDFS-10826.5.patch, HDFS-10826.WIP.1.patch > > > For RS-6-3, when there is one ec block group and > 1) 0~3 out of 9 internal blocks are missing, the result of fsck is HEALTY. > 2) 4~8 out of 9 internal blocks are missing, the result of fsck is HEALTY. > {noformat} > Erasure Coded Block Groups: > Total size:536870912 B > Total files: 1 > Total block groups (validated):1 (avg. block group size 536870912 B) > > UNRECOVERABLE BLOCK GROUPS: 1 (100.0 %) > > Minimally erasure-coded block groups: 0 (0.0 %) > Over-erasure-coded block groups: 0 (0.0 %) > Under-erasure-coded block groups: 1 (100.0 %) > Unsatisfactory placement block groups: 0 (0.0 %) > Default ecPolicy: RS-DEFAULT-6-3-64k > Average block group size: 5.0 > Missing block groups: 0 > Corrupt block groups: 0 > Missing internal blocks: 4 (44.43 %) > FSCK ended at Wed Aug 31 13:42:05 JST 2016 in 4 milliseconds > The filesystem under path '/' is HEALTHY > {noformat} > 3) 9 out of 9 internal blocks are missing, the result of fsck is CRITICAL. > (Because it is regarded as a missing block group.) > In case 2), the result should be CRITICAL since the ec block group is > unrecoverable. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-10826) The result of fsck should be CRITICAL when there are unrecoverable ec block groups.
[ https://issues.apache.org/jira/browse/HDFS-10826?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15549491#comment-15549491 ] Jing Zhao commented on HDFS-10826: -- I've committed the patch into trunk. Thanks for the contribution, [~tasanuma0829]. Thanks for the review, [~ajisakaa] and [~jojochuang]. > The result of fsck should be CRITICAL when there are unrecoverable ec block > groups. > --- > > Key: HDFS-10826 > URL: https://issues.apache.org/jira/browse/HDFS-10826 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: erasure-coding >Reporter: Takanobu Asanuma >Assignee: Takanobu Asanuma > Attachments: HDFS-10826.2.patch, HDFS-10826.3.patch, > HDFS-10826.4.patch, HDFS-10826.5.patch, HDFS-10826.WIP.1.patch > > > For RS-6-3, when there is one ec block group and > 1) 0~3 out of 9 internal blocks are missing, the result of fsck is HEALTY. > 2) 4~8 out of 9 internal blocks are missing, the result of fsck is HEALTY. > {noformat} > Erasure Coded Block Groups: > Total size:536870912 B > Total files: 1 > Total block groups (validated):1 (avg. block group size 536870912 B) > > UNRECOVERABLE BLOCK GROUPS: 1 (100.0 %) > > Minimally erasure-coded block groups: 0 (0.0 %) > Over-erasure-coded block groups: 0 (0.0 %) > Under-erasure-coded block groups: 1 (100.0 %) > Unsatisfactory placement block groups: 0 (0.0 %) > Default ecPolicy: RS-DEFAULT-6-3-64k > Average block group size: 5.0 > Missing block groups: 0 > Corrupt block groups: 0 > Missing internal blocks: 4 (44.43 %) > FSCK ended at Wed Aug 31 13:42:05 JST 2016 in 4 milliseconds > The filesystem under path '/' is HEALTHY > {noformat} > 3) 9 out of 9 internal blocks are missing, the result of fsck is CRITICAL. > (Because it is regarded as a missing block group.) > In case 2), the result should be CRITICAL since the ec block group is > unrecoverable. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-10898) libhdfs++: Make log levels consistent
[ https://issues.apache.org/jira/browse/HDFS-10898?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15549483#comment-15549483 ] Bob Hansen commented on HDFS-10898: --- +1 > libhdfs++: Make log levels consistent > - > > Key: HDFS-10898 > URL: https://issues.apache.org/jira/browse/HDFS-10898 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: hdfs-client >Reporter: James Clampffer >Assignee: James Clampffer >Priority: Trivial > Attachments: HDFS-10898.HDFS-8707.000.patch, > HDFS-10898.HDFS-8707.001.patch > > > Most of the public C++ FileHandle/FileSystem operations have a LOG_TRACE > level message about parameters passed in etc. However many methods use > LOG_DEBUG and a couple use LOG_INFO. > We most likely want FS operations that happen a lot (read/open/seek/stat) to > stick to LOG_DEBUG consistently and only use LOG_INFO for things like > FileSystem::Connect or RpcConnection:: that don't get called often and are > important enough to warrant showing up in the log. LOG_TRACE can be reserved > for things happening deeper inside public methods and methods that aren't > part of the public API. > Related improvements that could be brought into this to avoid opening a ton > of small Jiras: > -Print the "this" pointer address in the log message to make it easier to > correlate objects when there's concurrent work being done. This has been > very helpful in the past but often got stripped out before patches went in. > People just need be aware that operator new may eventually place an object of > the same type at the same address sometime in the future. > -For objects owned by other objects, but created on the fly, include a > pointer back to the parent/creator object if that pointer is already being > tracked (See the nested stucts in BlockReaderImpl). -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-10797) Disk usage summary of snapshots causes renamed blocks to get counted twice
[ https://issues.apache.org/jira/browse/HDFS-10797?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sean Mackrory updated HDFS-10797: - Attachment: HDFS-10797.010.patch > Disk usage summary of snapshots causes renamed blocks to get counted twice > -- > > Key: HDFS-10797 > URL: https://issues.apache.org/jira/browse/HDFS-10797 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Sean Mackrory >Assignee: Sean Mackrory > Attachments: HDFS-10797.001.patch, HDFS-10797.002.patch, > HDFS-10797.003.patch, HDFS-10797.004.patch, HDFS-10797.005.patch, > HDFS-10797.006.patch, HDFS-10797.007.patch, HDFS-10797.008.patch, > HDFS-10797.009.patch, HDFS-10797.010.patch, HDFS-10797.010.patch > > > DirectoryWithSnapshotFeature.computeContentSummary4Snapshot calculates how > much disk usage is used by a snapshot by tallying up the files in the > snapshot that have since been deleted (that way it won't overlap with regular > files whose disk usage is computed separately). However that is determined > from a diff that shows moved (to Trash or otherwise) or renamed files as a > deletion and a creation operation that may overlap with the list of blocks. > Only the deletion operation is taken into consideration, and this causes > those blocks to get represented twice in the disk usage tallying. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-10922) Adding additional unit tests for Trash (II)
[ https://issues.apache.org/jira/browse/HDFS-10922?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Xiaoyu Yao updated HDFS-10922: -- Summary: Adding additional unit tests for Trash (II) (was: Adding additional unit tests for Trash) > Adding additional unit tests for Trash (II) > --- > > Key: HDFS-10922 > URL: https://issues.apache.org/jira/browse/HDFS-10922 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: test >Reporter: Xiaoyu Yao >Assignee: Weiwei Yang > Attachments: HDFS-10922.02.patch, HDFS-10922.03.patch, > HDFS-10922.04.patch, HDFS-10922.05.patch, HDFS-10922.06.patch > > > This ticket is opened to track adding unit tests for Trash. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-10797) Disk usage summary of snapshots causes renamed blocks to get counted twice
[ https://issues.apache.org/jira/browse/HDFS-10797?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sean Mackrory updated HDFS-10797: - Attachment: HDFS-10797.010.patch > Disk usage summary of snapshots causes renamed blocks to get counted twice > -- > > Key: HDFS-10797 > URL: https://issues.apache.org/jira/browse/HDFS-10797 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Sean Mackrory >Assignee: Sean Mackrory > Attachments: HDFS-10797.001.patch, HDFS-10797.002.patch, > HDFS-10797.003.patch, HDFS-10797.004.patch, HDFS-10797.005.patch, > HDFS-10797.006.patch, HDFS-10797.007.patch, HDFS-10797.008.patch, > HDFS-10797.009.patch, HDFS-10797.010.patch > > > DirectoryWithSnapshotFeature.computeContentSummary4Snapshot calculates how > much disk usage is used by a snapshot by tallying up the files in the > snapshot that have since been deleted (that way it won't overlap with regular > files whose disk usage is computed separately). However that is determined > from a diff that shows moved (to Trash or otherwise) or renamed files as a > deletion and a creation operation that may overlap with the list of blocks. > Only the deletion operation is taken into consideration, and this causes > those blocks to get represented twice in the disk usage tallying. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Comment Edited] (HDFS-10922) Adding additional unit tests for Trash
[ https://issues.apache.org/jira/browse/HDFS-10922?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15549350#comment-15549350 ] Xiaoyu Yao edited comment on HDFS-10922 at 10/5/16 5:20 PM: bq. Now working on #4 about namenode restart. One question, the patch modified both common and hdfs projects, is it OK to use just single 1 patch ? Good catch, [~cheersyang]. Let's use this ticket for hdfs only changes in #4. That should be landed in {{o.a.h.hdfs.TestHDFSTrash}} I opened a new ticket HADOOP-13686 for #1-3. I notice the permission test for #1 is added for only hdfs. Can you move that to hadoop-common and attach the updated patch in HADOOP-13686? This way, we will have better coverage than having the test only for hdfs. was (Author: xyao): bq. Now working on #4 about namenode restart. One question, the patch modified both common and hdfs projects, is it OK to use just single 1 patch ? Good catch, [~cheersyang]. Let's have a separate ticket for the hdfs changes for #4. That should be landed in {{o.a.h.hdfs.TestHDFSTrash}} I will move this ticket from hadoop-hdfs to hadoop-common for #1-3 only. I notice the permission test for #1 is added for only hdfs. Can you move the test to hadoop-common as the behavior should not be specific to hdfs? > Adding additional unit tests for Trash > -- > > Key: HDFS-10922 > URL: https://issues.apache.org/jira/browse/HDFS-10922 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: test >Reporter: Xiaoyu Yao >Assignee: Weiwei Yang > Attachments: HDFS-10922.02.patch, HDFS-10922.03.patch, > HDFS-10922.04.patch, HDFS-10922.05.patch, HDFS-10922.06.patch > > > This ticket is opened to track adding unit tests for Trash. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-10963) Reduce log level when network topology cannot find enough datanodes.
[ https://issues.apache.org/jira/browse/HDFS-10963?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15549386#comment-15549386 ] Xiao Chen commented on HDFS-10963: -- Thanks for the review, [~brahmareddy]. > Reduce log level when network topology cannot find enough datanodes. > > > Key: HDFS-10963 > URL: https://issues.apache.org/jira/browse/HDFS-10963 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Xiao Chen >Assignee: Xiao Chen >Priority: Minor > Labels: supportability > Attachments: HDFS-10963.01.patch > > > Credit to [~qwertymaniac] who reported this. > {quote} > The change made by HDFS-10320 now causes all 1-rack cluster to keep printing > below: > WARNorg.apache.hadoop.net.NetworkTopology > Failed to find datanode (scope="" excludedScope="/default"). > {quote} > This was added in HDFS-10320 to replace an exception that caused NN > terminate. A warn log was added. But thinking closely, this should be debug > to reduce confusion. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-10797) Disk usage summary of snapshots causes renamed blocks to get counted twice
[ https://issues.apache.org/jira/browse/HDFS-10797?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15549352#comment-15549352 ] Xiao Chen commented on HDFS-10797: -- Thanks Sean. Agreed correctness is the goal. My comment regarding performance was on the nature of the context based summary, which is added to prevent a {{du}} locking the NN. And as we both agreed, this patch doesn't touch that. Also about {{getIncludedNodes}}, it seems we don't call it anywhere... Maybe we can remove the getter method, and when a use case emerges, let the new change handle it? > Disk usage summary of snapshots causes renamed blocks to get counted twice > -- > > Key: HDFS-10797 > URL: https://issues.apache.org/jira/browse/HDFS-10797 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Sean Mackrory >Assignee: Sean Mackrory > Attachments: HDFS-10797.001.patch, HDFS-10797.002.patch, > HDFS-10797.003.patch, HDFS-10797.004.patch, HDFS-10797.005.patch, > HDFS-10797.006.patch, HDFS-10797.007.patch, HDFS-10797.008.patch, > HDFS-10797.009.patch > > > DirectoryWithSnapshotFeature.computeContentSummary4Snapshot calculates how > much disk usage is used by a snapshot by tallying up the files in the > snapshot that have since been deleted (that way it won't overlap with regular > files whose disk usage is computed separately). However that is determined > from a diff that shows moved (to Trash or otherwise) or renamed files as a > deletion and a creation operation that may overlap with the list of blocks. > Only the deletion operation is taken into consideration, and this causes > those blocks to get represented twice in the disk usage tallying. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-10922) Adding additional unit tests for Trash
[ https://issues.apache.org/jira/browse/HDFS-10922?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15549350#comment-15549350 ] Xiaoyu Yao commented on HDFS-10922: --- bq. Now working on #4 about namenode restart. One question, the patch modified both common and hdfs projects, is it OK to use just single 1 patch ? Good catch, [~cheersyang]. Let's have a separate ticket for the hdfs changes for #4. That should be landed in {{o.a.h.hdfs.TestHDFSTrash}} Please let me know when your changes for #1-3 in this ticket are ready for review. Thanks! > Adding additional unit tests for Trash > -- > > Key: HDFS-10922 > URL: https://issues.apache.org/jira/browse/HDFS-10922 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: test >Reporter: Xiaoyu Yao >Assignee: Weiwei Yang > Attachments: HDFS-10922.02.patch, HDFS-10922.03.patch, > HDFS-10922.04.patch, HDFS-10922.05.patch, HDFS-10922.06.patch > > > This ticket is opened to track adding unit tests for Trash. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Comment Edited] (HDFS-10922) Adding additional unit tests for Trash
[ https://issues.apache.org/jira/browse/HDFS-10922?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15549350#comment-15549350 ] Xiaoyu Yao edited comment on HDFS-10922 at 10/5/16 5:12 PM: bq. Now working on #4 about namenode restart. One question, the patch modified both common and hdfs projects, is it OK to use just single 1 patch ? Good catch, [~cheersyang]. Let's have a separate ticket for the hdfs changes for #4. That should be landed in {{o.a.h.hdfs.TestHDFSTrash}} I will move this ticket from hadoop-hdfs to hadoop-common for #1-3 only. I notice the permission test for #1 is added for only hdfs. Can you move the test to hadoop-common as the behavior should not be specific to hdfs? was (Author: xyao): bq. Now working on #4 about namenode restart. One question, the patch modified both common and hdfs projects, is it OK to use just single 1 patch ? Good catch, [~cheersyang]. Let's have a separate ticket for the hdfs changes for #4. That should be landed in {{o.a.h.hdfs.TestHDFSTrash}} Please let me know when your changes for #1-3 in this ticket are ready for review. Thanks! > Adding additional unit tests for Trash > -- > > Key: HDFS-10922 > URL: https://issues.apache.org/jira/browse/HDFS-10922 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: test >Reporter: Xiaoyu Yao >Assignee: Weiwei Yang > Attachments: HDFS-10922.02.patch, HDFS-10922.03.patch, > HDFS-10922.04.patch, HDFS-10922.05.patch, HDFS-10922.06.patch > > > This ticket is opened to track adding unit tests for Trash. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-10898) libhdfs++: Make log levels consistent
[ https://issues.apache.org/jira/browse/HDFS-10898?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15549337#comment-15549337 ] Hadoop QA commented on HDFS-10898: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 14s{color} | {color:blue} Docker mode activated. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:red}-1{color} | {color:red} test4tests {color} | {color:red} 0m 0s{color} | {color:red} The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch. {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 16m 55s{color} | {color:green} HDFS-8707 passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 6m 8s{color} | {color:green} HDFS-8707 passed with JDK v1.8.0_101 {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 6m 11s{color} | {color:green} HDFS-8707 passed with JDK v1.7.0_111 {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 16s{color} | {color:green} HDFS-8707 passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 14s{color} | {color:green} HDFS-8707 passed {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 9s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 5m 59s{color} | {color:green} the patch passed with JDK v1.8.0_101 {color} | | {color:green}+1{color} | {color:green} cc {color} | {color:green} 5m 59s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 5m 59s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 6m 37s{color} | {color:green} the patch passed with JDK v1.7.0_111 {color} | | {color:green}+1{color} | {color:green} cc {color} | {color:green} 6m 37s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 6m 37s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 12s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 10s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} unit {color} | {color:green} 9m 8s{color} | {color:green} hadoop-hdfs-native-client in the patch passed with JDK v1.7.0_111. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 19s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 63m 35s{color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Docker | Image:yetus/hadoop:78fc6b6 | | JIRA Issue | HDFS-10898 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12831761/HDFS-10898.HDFS-8707.001.patch | | Optional Tests | asflicense compile cc mvnsite javac unit | | uname | Linux ff80de8b5ddb 3.13.0-93-generic #140-Ubuntu SMP Mon Jul 18 21:21:05 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh | | git revision | HDFS-8707 / f7d82e1 | | Default Java | 1.7.0_111 | | Multi-JDK versions | /usr/lib/jvm/java-8-oracle:1.8.0_101 /usr/lib/jvm/java-7-openjdk-amd64:1.7.0_111 | | JDK v1.7.0_111 Test Results | https://builds.apache.org/job/PreCommit-HDFS-Build/17021/testReport/ | | modules | C: hadoop-hdfs-project/hadoop-hdfs-native-client U: hadoop-hdfs-project/hadoop-hdfs-native-client | | Console output | https://builds.apache.org/job/PreCommit-HDFS-Build/17021/console | | Powered by | Apache Yetus 0.4.0-SNAPSHOT http://yetus.apache.org | This message was automatically generated. > libhdfs++: Make log levels consistent > - > > Key: HDFS-10898 > URL: https://issues.apache.org/jira/browse/HDFS-10898 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: hdfs-client >Reporter: James Clampffer >Assignee: James Clampffer >Priority: Trivial > Attachments: