[jira] [Work logged] (HDFS-15714) HDFS Provided Storage Read/Write Mount Support On-the-fly
[ https://issues.apache.org/jira/browse/HDFS-15714?focusedWorklogId=559209=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-559209 ] ASF GitHub Bot logged work on HDFS-15714: - Author: ASF GitHub Bot Created on: 01/Mar/21 06:21 Start Date: 01/Mar/21 06:21 Worklog Time Spent: 10m Work Description: Zhangshunyu commented on pull request #2655: URL: https://github.com/apache/hadoop/pull/2655#issuecomment-787680182 @PHILO-HE I checked the config value and it is 1 This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org Issue Time Tracking --- Worklog Id: (was: 559209) Time Spent: 1h 10m (was: 1h) > HDFS Provided Storage Read/Write Mount Support On-the-fly > - > > Key: HDFS-15714 > URL: https://issues.apache.org/jira/browse/HDFS-15714 > Project: Hadoop HDFS > Issue Type: New Feature > Components: datanode, namenode >Affects Versions: 3.4.0 >Reporter: Feilong He >Assignee: Feilong He >Priority: Major > Labels: pull-request-available > Attachments: HDFS-15714-01.patch, > HDFS_Provided_Storage_Design-V1.pdf, HDFS_Provided_Storage_Performance-V1.pdf > > Time Spent: 1h 10m > Remaining Estimate: 0h > > HDFS Provided Storage (PS) is a feature to tier HDFS over other file systems. > In HDFS-9806, PROVIDED storage type was introduced to HDFS. Through > configuring external storage with PROVIDED tag for DataNode, user can enable > application to access data stored externally from HDFS side. However, there > are two issues need to be addressed. Firstly, mounting external storage > on-the-fly, namely dynamic mount, is lacking. It is necessary to get it > supported to flexibly combine HDFS with an external storage at runtime. > Secondly, PS write is not supported by current HDFS. But in real > applications, it is common to transfer data bi-directionally for read/write > between HDFS and external storage. > Through this JIRA, we are presenting our work for PS write support and > dynamic mount support for both read & write. Please note in the community > several JIRAs have been filed for these topics. Our work is based on these > previous community work, with new design & implementation to support called > writeBack mount and enable admin to add any mount on-the-fly. We appreciate > those folks in the community for their great contribution! See their pending > JIRAs: HDFS-14805 & HDFS-12090. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Work logged] (HDFS-15714) HDFS Provided Storage Read/Write Mount Support On-the-fly
[ https://issues.apache.org/jira/browse/HDFS-15714?focusedWorklogId=559193=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-559193 ] ASF GitHub Bot logged work on HDFS-15714: - Author: ASF GitHub Bot Created on: 01/Mar/21 05:25 Start Date: 01/Mar/21 05:25 Worklog Time Spent: 10m Work Description: PHILO-HE commented on pull request #2655: URL: https://github.com/apache/hadoop/pull/2655#issuecomment-787648343 Hi, @Zhangshunyu, in your HDFS config, did you set `dfs.provided.overreplication.factor` with a number larger than 0? This property is used to control the number of cached block (cached to HDFS from remote storage). If it is set to 0 (default setting), the redundant cached block will be deleted by HDFS, which can trigger replica not found exception. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org Issue Time Tracking --- Worklog Id: (was: 559193) Time Spent: 1h (was: 50m) > HDFS Provided Storage Read/Write Mount Support On-the-fly > - > > Key: HDFS-15714 > URL: https://issues.apache.org/jira/browse/HDFS-15714 > Project: Hadoop HDFS > Issue Type: New Feature > Components: datanode, namenode >Affects Versions: 3.4.0 >Reporter: Feilong He >Assignee: Feilong He >Priority: Major > Labels: pull-request-available > Attachments: HDFS-15714-01.patch, > HDFS_Provided_Storage_Design-V1.pdf, HDFS_Provided_Storage_Performance-V1.pdf > > Time Spent: 1h > Remaining Estimate: 0h > > HDFS Provided Storage (PS) is a feature to tier HDFS over other file systems. > In HDFS-9806, PROVIDED storage type was introduced to HDFS. Through > configuring external storage with PROVIDED tag for DataNode, user can enable > application to access data stored externally from HDFS side. However, there > are two issues need to be addressed. Firstly, mounting external storage > on-the-fly, namely dynamic mount, is lacking. It is necessary to get it > supported to flexibly combine HDFS with an external storage at runtime. > Secondly, PS write is not supported by current HDFS. But in real > applications, it is common to transfer data bi-directionally for read/write > between HDFS and external storage. > Through this JIRA, we are presenting our work for PS write support and > dynamic mount support for both read & write. Please note in the community > several JIRAs have been filed for these topics. Our work is based on these > previous community work, with new design & implementation to support called > writeBack mount and enable admin to add any mount on-the-fly. We appreciate > those folks in the community for their great contribution! See their pending > JIRAs: HDFS-14805 & HDFS-12090. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Work logged] (HDFS-15808) Add metrics for FSNamesystem read/write lock hold long time
[ https://issues.apache.org/jira/browse/HDFS-15808?focusedWorklogId=559168=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-559168 ] ASF GitHub Bot logged work on HDFS-15808: - Author: ASF GitHub Bot Created on: 01/Mar/21 03:26 Start Date: 01/Mar/21 03:26 Worklog Time Spent: 10m Work Description: tomscut commented on pull request #2668: URL: https://github.com/apache/hadoop/pull/2668#issuecomment-787609646 Hi @anuengineer , could you please review again? Let's see if there are any other questions. Thank you. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org Issue Time Tracking --- Worklog Id: (was: 559168) Time Spent: 5h 40m (was: 5.5h) > Add metrics for FSNamesystem read/write lock hold long time > --- > > Key: HDFS-15808 > URL: https://issues.apache.org/jira/browse/HDFS-15808 > Project: Hadoop HDFS > Issue Type: Wish > Components: hdfs >Reporter: tomscut >Assignee: tomscut >Priority: Major > Labels: hdfs, lock, metrics, pull-request-available > Attachments: ExpiredHeartbeat.png, lockLongHoldCount > > Time Spent: 5h 40m > Remaining Estimate: 0h > > To monitor how often read/write locks exceed thresholds, we can add two > metrics(ReadLockLongHoldCount/WriteLockLongHoldCount), which are exposed in > JMX. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Work logged] (HDFS-15856) Make recover the pipeline in same packet exceed times for stream closed configurable.
[ https://issues.apache.org/jira/browse/HDFS-15856?focusedWorklogId=559165=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-559165 ] ASF GitHub Bot logged work on HDFS-15856: - Author: ASF GitHub Bot Created on: 01/Mar/21 03:15 Start Date: 01/Mar/21 03:15 Worklog Time Spent: 10m Work Description: qizhu-lucas commented on a change in pull request #2721: URL: https://github.com/apache/hadoop/pull/2721#discussion_r584422979 ## File path: hadoop-hdfs-project/hadoop-hdfs/src/main/resources/hdfs-default.xml ## @@ -4352,6 +4352,17 @@ + + dfs.client.pipeline.recovery.max-retries + 5 + +If we had to recover the pipeline more than the value Review comment: Good suggestion, i have changed to it. Thanks @jojochuang ## File path: hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/hdfs/DataStreamer.java ## @@ -1263,14 +1265,18 @@ private boolean processDatanodeOrExternalError() throws IOException { packetSendTime.clear(); } -// If we had to recover the pipeline five times in a row for the +// If we had to recover the pipeline more than the value +// defined by maxPipelineRecoveryRetries in a row for the // same packet, this client likely has corrupt data or corrupting // during transmission. -if (!errorState.isRestartingNode() && ++pipelineRecoveryCount > 5) { +if (!errorState.isRestartingNode() && ++pipelineRecoveryCount > +maxPipelineRecoveryRetries) { LOG.warn("Error recovering pipeline for writing " + - block + ". Already retried 5 times for the same packet."); + block + ". Already retried " + maxPipelineRecoveryRetries + + " times for the same packet."); lastException.set(new IOException("Failing write. Tried pipeline " + - "recovery 5 times without success.")); + "recovery "+ maxPipelineRecoveryRetries Review comment: Thanks @jojochuang for review. Fixed it latest pull request. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org Issue Time Tracking --- Worklog Id: (was: 559165) Time Spent: 2h (was: 1h 50m) > Make recover the pipeline in same packet exceed times for stream closed > configurable. > - > > Key: HDFS-15856 > URL: https://issues.apache.org/jira/browse/HDFS-15856 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Qi Zhu >Assignee: Qi Zhu >Priority: Minor > Labels: pull-request-available > Time Spent: 2h > Remaining Estimate: 0h > > Now recover the pipeline five times in a row for the same packet, will close > the stream, but i think it should be configurable for different cluster > needed. > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Work logged] (HDFS-15856) Make recover the pipeline in same packet exceed times for stream closed configurable.
[ https://issues.apache.org/jira/browse/HDFS-15856?focusedWorklogId=559164=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-559164 ] ASF GitHub Bot logged work on HDFS-15856: - Author: ASF GitHub Bot Created on: 01/Mar/21 03:15 Start Date: 01/Mar/21 03:15 Worklog Time Spent: 10m Work Description: qizhu-lucas commented on a change in pull request #2721: URL: https://github.com/apache/hadoop/pull/2721#discussion_r584422934 ## File path: hadoop-hdfs-project/hadoop-hdfs/src/main/resources/hdfs-default.xml ## @@ -4352,6 +4352,17 @@ + + dfs.client.pipeline.recovery.max-retries + 5 + +If we had to recover the pipeline more than the value Review comment: Good suggestion, i have changed to it. Thanks @jojochuang This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org Issue Time Tracking --- Worklog Id: (was: 559164) Time Spent: 1h 50m (was: 1h 40m) > Make recover the pipeline in same packet exceed times for stream closed > configurable. > - > > Key: HDFS-15856 > URL: https://issues.apache.org/jira/browse/HDFS-15856 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Qi Zhu >Assignee: Qi Zhu >Priority: Minor > Labels: pull-request-available > Time Spent: 1h 50m > Remaining Estimate: 0h > > Now recover the pipeline five times in a row for the same packet, will close > the stream, but i think it should be configurable for different cluster > needed. > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Work logged] (HDFS-15856) Make recover the pipeline in same packet exceed times for stream closed configurable.
[ https://issues.apache.org/jira/browse/HDFS-15856?focusedWorklogId=559163=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-559163 ] ASF GitHub Bot logged work on HDFS-15856: - Author: ASF GitHub Bot Created on: 01/Mar/21 03:15 Start Date: 01/Mar/21 03:15 Worklog Time Spent: 10m Work Description: qizhu-lucas commented on a change in pull request #2721: URL: https://github.com/apache/hadoop/pull/2721#discussion_r584422780 ## File path: hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/hdfs/DataStreamer.java ## @@ -1263,14 +1265,18 @@ private boolean processDatanodeOrExternalError() throws IOException { packetSendTime.clear(); } -// If we had to recover the pipeline five times in a row for the +// If we had to recover the pipeline more than the value +// defined by maxPipelineRecoveryRetries in a row for the // same packet, this client likely has corrupt data or corrupting // during transmission. -if (!errorState.isRestartingNode() && ++pipelineRecoveryCount > 5) { +if (!errorState.isRestartingNode() && ++pipelineRecoveryCount > +maxPipelineRecoveryRetries) { LOG.warn("Error recovering pipeline for writing " + - block + ". Already retried 5 times for the same packet."); + block + ". Already retried " + maxPipelineRecoveryRetries + + " times for the same packet."); lastException.set(new IOException("Failing write. Tried pipeline " + - "recovery 5 times without success.")); + "recovery "+ maxPipelineRecoveryRetries Review comment: Thanks @jojochuang for review. Fixed it latest pull request. ## File path: hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/hdfs/DataStreamer.java ## @@ -1263,14 +1265,18 @@ private boolean processDatanodeOrExternalError() throws IOException { packetSendTime.clear(); } -// If we had to recover the pipeline five times in a row for the +// If we had to recover the pipeline more than the value +// defined by maxPipelineRecoveryRetries in a row for the // same packet, this client likely has corrupt data or corrupting // during transmission. -if (!errorState.isRestartingNode() && ++pipelineRecoveryCount > 5) { +if (!errorState.isRestartingNode() && ++pipelineRecoveryCount > +maxPipelineRecoveryRetries) { LOG.warn("Error recovering pipeline for writing " + - block + ". Already retried 5 times for the same packet."); + block + ". Already retried " + maxPipelineRecoveryRetries + + " times for the same packet."); lastException.set(new IOException("Failing write. Tried pipeline " + - "recovery 5 times without success.")); + "recovery "+ maxPipelineRecoveryRetries Review comment: Thanks @jojochuang for review. Fixed it latest pull request. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org Issue Time Tracking --- Worklog Id: (was: 559163) Time Spent: 1h 40m (was: 1.5h) > Make recover the pipeline in same packet exceed times for stream closed > configurable. > - > > Key: HDFS-15856 > URL: https://issues.apache.org/jira/browse/HDFS-15856 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Qi Zhu >Assignee: Qi Zhu >Priority: Minor > Labels: pull-request-available > Time Spent: 1h 40m > Remaining Estimate: 0h > > Now recover the pipeline five times in a row for the same packet, will close > the stream, but i think it should be configurable for different cluster > needed. > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Work logged] (HDFS-15714) HDFS Provided Storage Read/Write Mount Support On-the-fly
[ https://issues.apache.org/jira/browse/HDFS-15714?focusedWorklogId=559160=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-559160 ] ASF GitHub Bot logged work on HDFS-15714: - Author: ASF GitHub Bot Created on: 01/Mar/21 02:54 Start Date: 01/Mar/21 02:54 Worklog Time Spent: 10m Work Description: Zhangshunyu commented on pull request #2655: URL: https://github.com/apache/hadoop/pull/2655#issuecomment-787600101 @PHILO-HE Thanks for the reply. Please find the error msg: ``` ERROR org.apache.hadoop.hdfs.server.datanode.DataNode: Exception in creating BlockSender: {} org.apache.hadoop.hdfs.server.datanode.ReplicaNotFoundException: Replica not found for BP-1079229821-192.168.0.156-1614366591560:blk_1073791203_50379. The block may have been removed recently by the balancer or by intentionally reducing the replication factor. This condition is usually harmless. To be certain, please check the preceding datanode log messages for signs of a more serious issue. ``` we use readOnly mount mode, the data is stored on s3 store, after mount we read the data use readthrough mode, we find that some times the datanode will throw this exception, it seems the read op request send to the wrong datanode? This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org Issue Time Tracking --- Worklog Id: (was: 559160) Time Spent: 50m (was: 40m) > HDFS Provided Storage Read/Write Mount Support On-the-fly > - > > Key: HDFS-15714 > URL: https://issues.apache.org/jira/browse/HDFS-15714 > Project: Hadoop HDFS > Issue Type: New Feature > Components: datanode, namenode >Affects Versions: 3.4.0 >Reporter: Feilong He >Assignee: Feilong He >Priority: Major > Labels: pull-request-available > Attachments: HDFS-15714-01.patch, > HDFS_Provided_Storage_Design-V1.pdf, HDFS_Provided_Storage_Performance-V1.pdf > > Time Spent: 50m > Remaining Estimate: 0h > > HDFS Provided Storage (PS) is a feature to tier HDFS over other file systems. > In HDFS-9806, PROVIDED storage type was introduced to HDFS. Through > configuring external storage with PROVIDED tag for DataNode, user can enable > application to access data stored externally from HDFS side. However, there > are two issues need to be addressed. Firstly, mounting external storage > on-the-fly, namely dynamic mount, is lacking. It is necessary to get it > supported to flexibly combine HDFS with an external storage at runtime. > Secondly, PS write is not supported by current HDFS. But in real > applications, it is common to transfer data bi-directionally for read/write > between HDFS and external storage. > Through this JIRA, we are presenting our work for PS write support and > dynamic mount support for both read & write. Please note in the community > several JIRAs have been filed for these topics. Our work is based on these > previous community work, with new design & implementation to support called > writeBack mount and enable admin to add any mount on-the-fly. We appreciate > those folks in the community for their great contribution! See their pending > JIRAs: HDFS-14805 & HDFS-12090. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Work logged] (HDFS-15714) HDFS Provided Storage Read/Write Mount Support On-the-fly
[ https://issues.apache.org/jira/browse/HDFS-15714?focusedWorklogId=559159=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-559159 ] ASF GitHub Bot logged work on HDFS-15714: - Author: ASF GitHub Bot Created on: 01/Mar/21 02:50 Start Date: 01/Mar/21 02:50 Worklog Time Spent: 10m Work Description: PHILO-HE commented on pull request #2655: URL: https://github.com/apache/hadoop/pull/2655#issuecomment-787599087 Hi, @Zhangshunyu, thanks for your feedback. Could you please give more details? Did you mount an external storage in readOnly mode? Or writeBack mount? This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org Issue Time Tracking --- Worklog Id: (was: 559159) Time Spent: 40m (was: 0.5h) > HDFS Provided Storage Read/Write Mount Support On-the-fly > - > > Key: HDFS-15714 > URL: https://issues.apache.org/jira/browse/HDFS-15714 > Project: Hadoop HDFS > Issue Type: New Feature > Components: datanode, namenode >Affects Versions: 3.4.0 >Reporter: Feilong He >Assignee: Feilong He >Priority: Major > Labels: pull-request-available > Attachments: HDFS-15714-01.patch, > HDFS_Provided_Storage_Design-V1.pdf, HDFS_Provided_Storage_Performance-V1.pdf > > Time Spent: 40m > Remaining Estimate: 0h > > HDFS Provided Storage (PS) is a feature to tier HDFS over other file systems. > In HDFS-9806, PROVIDED storage type was introduced to HDFS. Through > configuring external storage with PROVIDED tag for DataNode, user can enable > application to access data stored externally from HDFS side. However, there > are two issues need to be addressed. Firstly, mounting external storage > on-the-fly, namely dynamic mount, is lacking. It is necessary to get it > supported to flexibly combine HDFS with an external storage at runtime. > Secondly, PS write is not supported by current HDFS. But in real > applications, it is common to transfer data bi-directionally for read/write > between HDFS and external storage. > Through this JIRA, we are presenting our work for PS write support and > dynamic mount support for both read & write. Please note in the community > several JIRAs have been filed for these topics. Our work is based on these > previous community work, with new design & implementation to support called > writeBack mount and enable admin to add any mount on-the-fly. We appreciate > those folks in the community for their great contribution! See their pending > JIRAs: HDFS-14805 & HDFS-12090. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Work logged] (HDFS-15854) Make some parameters configurable for SlowDiskTracker and SlowPeerTracker
[ https://issues.apache.org/jira/browse/HDFS-15854?focusedWorklogId=559150=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-559150 ] ASF GitHub Bot logged work on HDFS-15854: - Author: ASF GitHub Bot Created on: 01/Mar/21 01:21 Start Date: 01/Mar/21 01:21 Worklog Time Spent: 10m Work Description: tomscut commented on pull request #2718: URL: https://github.com/apache/hadoop/pull/2718#issuecomment-787572529 Hi @arp7 @dineshchitlangia , Could you please help to review the code. Thank you. It is related to a previous issue: https://issues.apache.org/jira/browse/HDFS-15814. Those failed unit tests were unrelated to the change. Failed junit tests: hadoop.hdfs.server.datanode.TestNNHandlesBlockReportPerStorage hadoop.hdfs.server.namenode.TestNamenodeCapacityReport This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org Issue Time Tracking --- Worklog Id: (was: 559150) Time Spent: 20m (was: 10m) > Make some parameters configurable for SlowDiskTracker and SlowPeerTracker > - > > Key: HDFS-15854 > URL: https://issues.apache.org/jira/browse/HDFS-15854 > Project: Hadoop HDFS > Issue Type: Wish >Reporter: tomscut >Assignee: tomscut >Priority: Major > Labels: pull-request-available > Time Spent: 20m > Remaining Estimate: 0h > > Make some parameters configurable for SlowDiskTracker and SlowPeerTracker. > Related to https://issues.apache.org/jira/browse/HDFS-15814. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Work logged] (HDFS-15856) Make recover the pipeline in same packet exceed times for stream closed configurable.
[ https://issues.apache.org/jira/browse/HDFS-15856?focusedWorklogId=559142=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-559142 ] ASF GitHub Bot logged work on HDFS-15856: - Author: ASF GitHub Bot Created on: 01/Mar/21 00:55 Start Date: 01/Mar/21 00:55 Worklog Time Spent: 10m Work Description: jojochuang commented on a change in pull request #2721: URL: https://github.com/apache/hadoop/pull/2721#discussion_r584392706 ## File path: hadoop-hdfs-project/hadoop-hdfs/src/main/resources/hdfs-default.xml ## @@ -4352,6 +4352,17 @@ + + dfs.client.pipeline.recovery.max-retries + 5 + +If we had to recover the pipeline more than the value Review comment: Suggest to add a description for this configuration property. Something like "if the DFS client encounters errors in write pipeline, retry up to the number defined by this property before giving up" This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org Issue Time Tracking --- Worklog Id: (was: 559142) Time Spent: 1.5h (was: 1h 20m) > Make recover the pipeline in same packet exceed times for stream closed > configurable. > - > > Key: HDFS-15856 > URL: https://issues.apache.org/jira/browse/HDFS-15856 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Qi Zhu >Assignee: Qi Zhu >Priority: Minor > Labels: pull-request-available > Time Spent: 1.5h > Remaining Estimate: 0h > > Now recover the pipeline five times in a row for the same packet, will close > the stream, but i think it should be configurable for different cluster > needed. > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Work logged] (HDFS-15856) Make recover the pipeline in same packet exceed times for stream closed configurable.
[ https://issues.apache.org/jira/browse/HDFS-15856?focusedWorklogId=559140=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-559140 ] ASF GitHub Bot logged work on HDFS-15856: - Author: ASF GitHub Bot Created on: 01/Mar/21 00:51 Start Date: 01/Mar/21 00:51 Worklog Time Spent: 10m Work Description: jojochuang commented on a change in pull request #2721: URL: https://github.com/apache/hadoop/pull/2721#discussion_r584392237 ## File path: hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/hdfs/DataStreamer.java ## @@ -1263,14 +1265,18 @@ private boolean processDatanodeOrExternalError() throws IOException { packetSendTime.clear(); } -// If we had to recover the pipeline five times in a row for the +// If we had to recover the pipeline more than the value +// defined by maxPipelineRecoveryRetries in a row for the // same packet, this client likely has corrupt data or corrupting // during transmission. -if (!errorState.isRestartingNode() && ++pipelineRecoveryCount > 5) { +if (!errorState.isRestartingNode() && ++pipelineRecoveryCount > +maxPipelineRecoveryRetries) { LOG.warn("Error recovering pipeline for writing " + - block + ". Already retried 5 times for the same packet."); + block + ". Already retried " + maxPipelineRecoveryRetries + + " times for the same packet."); lastException.set(new IOException("Failing write. Tried pipeline " + - "recovery 5 times without success.")); + "recovery "+ maxPipelineRecoveryRetries Review comment: nit: add space between " and + This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org Issue Time Tracking --- Worklog Id: (was: 559140) Time Spent: 1h 20m (was: 1h 10m) > Make recover the pipeline in same packet exceed times for stream closed > configurable. > - > > Key: HDFS-15856 > URL: https://issues.apache.org/jira/browse/HDFS-15856 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Qi Zhu >Assignee: Qi Zhu >Priority: Minor > Labels: pull-request-available > Time Spent: 1h 20m > Remaining Estimate: 0h > > Now recover the pipeline five times in a row for the same packet, will close > the stream, but i think it should be configurable for different cluster > needed. > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-15863) Validation message to be corrected in RBF FairnessPolicyController
[ https://issues.apache.org/jira/browse/HDFS-15863?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17292521#comment-17292521 ] Hadoop QA commented on HDFS-15863: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Logfile || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 23m 31s{color} | {color:blue}{color} | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || || | {color:green}+1{color} | {color:green} dupname {color} | {color:green} 0m 0s{color} | {color:green}{color} | {color:green} No case conflicting files found. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green}{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} {color} | {color:green} 0m 0s{color} | {color:green}test4tests{color} | {color:green} The patch appears to include 1 new or modified test files. {color} | || || || || {color:brown} trunk Compile Tests {color} || || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 34m 27s{color} | {color:green}{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 38s{color} | {color:green}{color} | {color:green} trunk passed with JDK Ubuntu-11.0.10+9-Ubuntu-0ubuntu1.20.04 {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 32s{color} | {color:green}{color} | {color:green} trunk passed with JDK Private Build-1.8.0_282-8u282-b08-0ubuntu1~20.04-b08 {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 22s{color} | {color:green}{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 38s{color} | {color:green}{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 16m 14s{color} | {color:green}{color} | {color:green} branch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 36s{color} | {color:green}{color} | {color:green} trunk passed with JDK Ubuntu-11.0.10+9-Ubuntu-0ubuntu1.20.04 {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 51s{color} | {color:green}{color} | {color:green} trunk passed with JDK Private Build-1.8.0_282-8u282-b08-0ubuntu1~20.04-b08 {color} | | {color:blue}0{color} | {color:blue} spotbugs {color} | {color:blue} 1m 13s{color} | {color:blue}{color} | {color:blue} Used deprecated FindBugs config; considering switching to SpotBugs. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 12s{color} | {color:green}{color} | {color:green} trunk passed {color} | || || || || {color:brown} Patch Compile Tests {color} || || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 31s{color} | {color:green}{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 32s{color} | {color:green}{color} | {color:green} the patch passed with JDK Ubuntu-11.0.10+9-Ubuntu-0ubuntu1.20.04 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 32s{color} | {color:green}{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 27s{color} | {color:green}{color} | {color:green} the patch passed with JDK Private Build-1.8.0_282-8u282-b08-0ubuntu1~20.04-b08 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 27s{color} | {color:green}{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 15s{color} | {color:green}{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 29s{color} | {color:green}{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green}{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 14m 52s{color} | {color:green}{color} | {color:green} patch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 33s{color} | {color:green}{color} | {color:green} the patch passed with JDK Ubuntu-11.0.10+9-Ubuntu-0ubuntu1.20.04 {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 49s{color} | {color:green}{color} | {color:green} the patch
[jira] [Commented] (HDFS-15745) Make DataNodePeerMetrics#LOW_THRESHOLD_MS and MIN_OUTLIER_DETECTION_NODES configurable
[ https://issues.apache.org/jira/browse/HDFS-15745?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17292490#comment-17292490 ] Renukaprasad C commented on HDFS-15745: --- Thanks [~huanghaibin] for reporting & [~ayushtkn] for quick review and commit. Can we merge the same to 3.3 & 3.1 branches as well? > Make DataNodePeerMetrics#LOW_THRESHOLD_MS and MIN_OUTLIER_DETECTION_NODES > configurable > -- > > Key: HDFS-15745 > URL: https://issues.apache.org/jira/browse/HDFS-15745 > Project: Hadoop HDFS > Issue Type: Improvement >Reporter: Haibin Huang >Assignee: Haibin Huang >Priority: Major > Fix For: 3.4.0 > > Attachments: HDFS-15745-001.patch, HDFS-15745-002.patch, > HDFS-15745-003.patch, image-2020-12-22-17-00-50-796.png > > > When i enable DataNodePeerMetrics to find slow slow peer in cluster, i found > there is a lot of slow peer but ReportingNodes's averageDelay is very low, > and these slow peer node are normal. I think the reason of why generating so > many slow peer is that the value of DataNodePeerMetrics#LOW_THRESHOLD_MS is > too small (only 5ms) and it is not configurable. The default value of slow io > warning log threshold is 300ms, i.e. > DFSConfigKeys.DFS_DATANODE_SLOW_IO_WARNING_THRESHOLD_DEFAULT = 300, so > DataNodePeerMetrics#LOW_THRESHOLD_MS should not be less than 300ms, otherwise > namenode will get a lot of invalid slow peer information. > !image-2020-12-22-17-00-50-796.png! -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-15863) Validation message to be corrected in RBF FairnessPolicyController
[ https://issues.apache.org/jira/browse/HDFS-15863?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Renukaprasad C updated HDFS-15863: -- Attachment: HDFS-15863.001.patch Status: Patch Available (was: Open) > Validation message to be corrected in RBF FairnessPolicyController > -- > > Key: HDFS-15863 > URL: https://issues.apache.org/jira/browse/HDFS-15863 > Project: Hadoop HDFS > Issue Type: Bug > Components: rbf >Affects Versions: 3.4.0 >Reporter: Renukaprasad C >Assignee: Renukaprasad C >Priority: Minor > Attachments: HDFS-15863.001.patch > > > org.apache.hadoop.hdfs.server.federation.fairness.StaticRouterRpcFairnessPolicyController#validateCount > When dfs.federation.router.handler.count is lessthan the total dedicated > handlers for all NS, then error message shows 0 & -ve values in error > message, instead can show the actual configured values. > Current message is : "Available handlers -5 lower than min 0 for nsId nn1" > This can be changed to: "Configured handlers > ${DFS_ROUTER_HANDLER_COUNT_KEY}=10 lower than min 15 for nsId nn1", where 10 > is hander count & 15 is sum of dedicated handler count. > Related to: HDFS-14090 -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Work logged] (HDFS-15843) [libhdfs++] Make write cross platform
[ https://issues.apache.org/jira/browse/HDFS-15843?focusedWorklogId=559107=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-559107 ] ASF GitHub Bot logged work on HDFS-15843: - Author: ASF GitHub Bot Created on: 28/Feb/21 18:04 Start Date: 28/Feb/21 18:04 Worklog Time Spent: 10m Work Description: GauthamBanasandra commented on pull request #2710: URL: https://github.com/apache/hadoop/pull/2710#issuecomment-787494012 Thanks for your review and suggestion @smengcl , @jojochuang. I've re-implemented my solution using the `write` system call for Windows and Linux separately, neatly tucked inside the `XPlatform` class . Please do have another look at my PR. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org Issue Time Tracking --- Worklog Id: (was: 559107) Time Spent: 1h 20m (was: 1h 10m) > [libhdfs++] Make write cross platform > - > > Key: HDFS-15843 > URL: https://issues.apache.org/jira/browse/HDFS-15843 > Project: Hadoop HDFS > Issue Type: Improvement > Components: libhdfs++ >Affects Versions: 3.2.2 >Reporter: Gautham Banasandra >Assignee: Gautham Banasandra >Priority: Major > Labels: pull-request-available > Time Spent: 1h 20m > Remaining Estimate: 0h > > We're currently using the *write* function from unistd.h which isn't > cross-platform. We need to replace this with *std::cout.write* instead. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Work logged] (HDFS-15843) [libhdfs++] Make write cross platform
[ https://issues.apache.org/jira/browse/HDFS-15843?focusedWorklogId=559102=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-559102 ] ASF GitHub Bot logged work on HDFS-15843: - Author: ASF GitHub Bot Created on: 28/Feb/21 16:33 Start Date: 28/Feb/21 16:33 Worklog Time Spent: 10m Work Description: GauthamBanasandra commented on a change in pull request #2710: URL: https://github.com/apache/hadoop/pull/2710#discussion_r584322501 ## File path: hadoop-hdfs-project/hadoop-hdfs-native-client/src/main/native/libhdfspp/examples/c/connect_cancel/connect_cancel.c ## @@ -43,10 +42,10 @@ const char *catch_exit = "Exiting the signal handler.\n"; // Print to stdout without calling malloc or otherwise indirectly modify userspace state. // Write calls to stdout may still interleave with stuff coming from elsewhere. static void sighandler_direct_stdout(const char *msg) { - if(!msg) + if(!msg) { return; - ssize_t res = write(1 /*posix stdout fd*/, msg, strlen(msg)); - (void)res; + } + printf("%s", msg); Review comment: I've removed this. Please take a look at the latest commit. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org Issue Time Tracking --- Worklog Id: (was: 559102) Time Spent: 1h 10m (was: 1h) > [libhdfs++] Make write cross platform > - > > Key: HDFS-15843 > URL: https://issues.apache.org/jira/browse/HDFS-15843 > Project: Hadoop HDFS > Issue Type: Improvement > Components: libhdfs++ >Affects Versions: 3.2.2 >Reporter: Gautham Banasandra >Assignee: Gautham Banasandra >Priority: Major > Labels: pull-request-available > Time Spent: 1h 10m > Remaining Estimate: 0h > > We're currently using the *write* function from unistd.h which isn't > cross-platform. We need to replace this with *std::cout.write* instead. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Work logged] (HDFS-13975) TestBalancer#testMaxIterationTime fails sporadically
[ https://issues.apache.org/jira/browse/HDFS-13975?focusedWorklogId=559100=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-559100 ] ASF GitHub Bot logged work on HDFS-13975: - Author: ASF GitHub Bot Created on: 28/Feb/21 16:07 Start Date: 28/Feb/21 16:07 Worklog Time Spent: 10m Work Description: touchida commented on pull request #2726: URL: https://github.com/apache/hadoop/pull/2726#issuecomment-787475633 Both failed and crashed tests are unrelated to this PR. I created [HDFS-15864](https://issues.apache.org/jira/browse/HDFS-15864) for the former, but didn't for the latter, since I'm not sure which caused the crash, the build environment or the test itself. Ready for review. - Failed test ``` [ERROR] Tests run: 29, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 47.231 s <<< FAILURE! - in org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestFsDatasetImpl [ERROR] testDnRestartWithHardLink(org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestFsDatasetImpl) Time elapsed: 1.452 s <<< FAILURE! java.lang.AssertionError at org.junit.Assert.fail(Assert.java:86) at org.junit.Assert.assertTrue(Assert.java:41) at org.junit.Assert.assertTrue(Assert.java:52) at org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestFsDatasetImpl.testDnRestartWithHardLink(TestFsDatasetImpl.java:1377) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:298) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:292) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.lang.Thread.run(Thread.java:748) ``` - Crashed test ``` org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestSpaceReservation org.apache.maven.surefire.booter.SurefireBooterForkException: ExecutionException The forked VM terminated without properly saying goodbye. VM crash or System.exit called? Command was /bin/sh -c cd /home/jenkins/jenkins-agent/workspace/hadoop-multibranch_PR-2726/src/hadoop-hdfs-project/hadoop-hdfs && /usr/lib/jvm/java-8-openjdk-amd64/jre/bin/java -Xmx2048m -XX:+HeapDumpOnOutOfMemoryError -DminiClusterDedicatedDirs=true -jar /home/jenkins/jenkins-agent/workspace/hadoop-multibranch_PR-2726/src/hadoop-hdfs-project/hadoop-hdfs/target/surefire/surefirebooter6106358806783121604.jar /home/jenkins/jenkins-agent/workspace/hadoop-multibranch_PR-2726/src/hadoop-hdfs-project/hadoop-hdfs/target/surefire 2021-02-27T21-36-41_009-jvmRun2 surefire4678365300915584400tmp surefire_5963761349348601015446tmp Error occurred in starting fork, check output in log Process Exit Code: 1 Crashed tests: org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestSpaceReservation at org.apache.maven.plugin.surefire.booterclient.ForkStarter.awaitResultsDone(ForkStarter.java:511) at org.apache.maven.plugin.surefire.booterclient.ForkStarter.runSuitesForkPerTestSet(ForkStarter.java:458) at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:299) at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:247) at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeProvider(AbstractSurefireMojo.java:1149) at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeAfterPreconditionsChecked(AbstractSurefireMojo.java:991) at org.apache.maven.plugin.surefire.AbstractSurefireMojo.execute(AbstractSurefireMojo.java:837) at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo(DefaultBuildPluginManager.java:137) at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:210) at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:156) at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:148) at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:117) at
[jira] [Created] (HDFS-15864) TestFsDatasetImpl#testDnRestartWithHardLink fails intermittently
Toshihiko Uchida created HDFS-15864: --- Summary: TestFsDatasetImpl#testDnRestartWithHardLink fails intermittently Key: HDFS-15864 URL: https://issues.apache.org/jira/browse/HDFS-15864 Project: Hadoop HDFS Issue Type: Sub-task Reporter: Toshihiko Uchida This unit test failed in https://github.com/apache/hadoop/pull/2726 due to an AssertionError. {code} [ERROR] testDnRestartWithHardLink(org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestFsDatasetImpl) Time elapsed: 1.452 s <<< FAILURE! java.lang.AssertionError at org.junit.Assert.fail(Assert.java:86) at org.junit.Assert.assertTrue(Assert.java:41) at org.junit.Assert.assertTrue(Assert.java:52) at org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestFsDatasetImpl.testDnRestartWithHardLink(TestFsDatasetImpl.java:1377) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:298) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:292) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.lang.Thread.run(Thread.java:748) {code} The failure occurred at the following first assertion. {code} cluster.restartDataNode(0); cluster.waitDatanodeFullyStarted(cluster.getDataNodes().get(0), 6); cluster.triggerBlockReports(); assertTrue(Files.exists(Paths.get(newReplicaInfo.getBlockURI(; assertTrue(Files.exists(Paths.get(oldReplicaInfo.getBlockURI(; {code} -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-13975) TestBalancer#testMaxIterationTime fails sporadically
[ https://issues.apache.org/jira/browse/HDFS-13975?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Toshihiko Uchida updated HDFS-13975: Labels: flaky-test pull-request-available (was: pull-request-available) > TestBalancer#testMaxIterationTime fails sporadically > > > Key: HDFS-13975 > URL: https://issues.apache.org/jira/browse/HDFS-13975 > Project: Hadoop HDFS > Issue Type: Sub-task >Affects Versions: 3.2.0 >Reporter: Jason Darrell Lowe >Assignee: Toshihiko Uchida >Priority: Major > Labels: flaky-test, pull-request-available > Time Spent: 10m > Remaining Estimate: 0h > > A number of precommit builds have seen this test fail like this: > {noformat} > java.lang.AssertionError: Unexpected iteration runtime: 4021ms > 3.5s > at org.junit.Assert.fail(Assert.java:88) > at org.junit.Assert.assertTrue(Assert.java:41) > at > org.apache.hadoop.hdfs.server.balancer.TestBalancer.testMaxIterationTime(TestBalancer.java:1649) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:498) > at > org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47) > at > org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) > at > org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44) > at > org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) > at > org.junit.internal.runners.statements.FailOnTimeout$StatementThread.run(FailOnTimeout.java:74) > {noformat} -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-15856) Make recover the pipeline in same packet exceed times for stream closed configurable.
[ https://issues.apache.org/jira/browse/HDFS-15856?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17292354#comment-17292354 ] Hadoop QA commented on HDFS-15856: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Logfile || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 41s{color} | | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || || | {color:green}+1{color} | {color:green} dupname {color} | {color:green} 0m 0s{color} | | {color:green} No case conflicting files found. {color} | | {color:blue}0{color} | {color:blue} codespell {color} | {color:blue} 0m 0s{color} | | {color:blue} codespell was not available. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | | {color:green} The patch does not contain any @author tags. {color} | | {color:red}-1{color} | {color:red} test4tests {color} | {color:red} 0m 0s{color} | | {color:red} The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch. {color} | || || || || {color:brown} trunk Compile Tests {color} || || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 14m 14s{color} | | {color:blue} Maven dependency ordering for branch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 20m 11s{color} | | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 4m 50s{color} | | {color:green} trunk passed with JDK Ubuntu-11.0.10+9-Ubuntu-0ubuntu1.20.04 {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 4m 25s{color} | | {color:green} trunk passed with JDK Private Build-1.8.0_282-8u282-b08-0ubuntu1~20.04-b08 {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 1m 15s{color} | | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 2m 13s{color} | | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 35s{color} | | {color:green} trunk passed with JDK Ubuntu-11.0.10+9-Ubuntu-0ubuntu1.20.04 {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 2m 5s{color} | | {color:green} trunk passed with JDK Private Build-1.8.0_282-8u282-b08-0ubuntu1~20.04-b08 {color} | | {color:blue}0{color} | {color:blue} spotbugs {color} | {color:blue} 3m 2s{color} | | {color:blue} Used deprecated FindBugs config; considering switching to SpotBugs. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 5m 25s{color} | | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 16m 1s{color} | | {color:green} branch has no errors when building and testing our client artifacts. {color} | || || || || {color:brown} Patch Compile Tests {color} || || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 26s{color} | | {color:blue} Maven dependency ordering for patch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 1m 56s{color} | | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 4m 41s{color} | | {color:green} the patch passed with JDK Ubuntu-11.0.10+9-Ubuntu-0ubuntu1.20.04 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 4m 41s{color} | | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 4m 19s{color} | | {color:green} the patch passed with JDK Private Build-1.8.0_282-8u282-b08-0ubuntu1~20.04-b08 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 4m 19s{color} | | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} blanks {color} | {color:green} 0m 0s{color} | | {color:green} The patch has no blanks issues. {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 1m 6s{color} | | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 59s{color} | | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} xml {color} | {color:green} 0m 1s{color} | | {color:green} The patch has no ill-formed XML file. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 19s{color} | | {color:green} the patch passed with JDK Ubuntu-11.0.10+9-Ubuntu-0ubuntu1.20.04 {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 49s{color} | |