[jira] [Work logged] (HDFS-15714) HDFS Provided Storage Read/Write Mount Support On-the-fly

2021-02-28 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15714?focusedWorklogId=559209=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-559209
 ]

ASF GitHub Bot logged work on HDFS-15714:
-

Author: ASF GitHub Bot
Created on: 01/Mar/21 06:21
Start Date: 01/Mar/21 06:21
Worklog Time Spent: 10m 
  Work Description: Zhangshunyu commented on pull request #2655:
URL: https://github.com/apache/hadoop/pull/2655#issuecomment-787680182


   @PHILO-HE I checked the config value and it is 1



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 559209)
Time Spent: 1h 10m  (was: 1h)

> HDFS Provided Storage Read/Write Mount Support On-the-fly
> -
>
> Key: HDFS-15714
> URL: https://issues.apache.org/jira/browse/HDFS-15714
> Project: Hadoop HDFS
>  Issue Type: New Feature
>  Components: datanode, namenode
>Affects Versions: 3.4.0
>Reporter: Feilong He
>Assignee: Feilong He
>Priority: Major
>  Labels: pull-request-available
> Attachments: HDFS-15714-01.patch, 
> HDFS_Provided_Storage_Design-V1.pdf, HDFS_Provided_Storage_Performance-V1.pdf
>
>  Time Spent: 1h 10m
>  Remaining Estimate: 0h
>
> HDFS Provided Storage (PS) is a feature to tier HDFS over other file systems. 
> In HDFS-9806, PROVIDED storage type was introduced to HDFS. Through 
> configuring external storage with PROVIDED tag for DataNode, user can enable 
> application to access data stored externally from HDFS side. However, there 
> are two issues need to be addressed. Firstly, mounting external storage 
> on-the-fly, namely dynamic mount, is lacking. It is necessary to get it 
> supported to flexibly combine HDFS with an external storage at runtime. 
> Secondly, PS write is not supported by current HDFS. But in real 
> applications, it is common to transfer data bi-directionally for read/write 
> between HDFS and external storage.
> Through this JIRA, we are presenting our work for PS write support and 
> dynamic mount support for both read & write. Please note in the community 
> several JIRAs have been filed for these topics. Our work is based on these 
> previous community work, with new design & implementation to support called 
> writeBack mount and enable admin to add any mount on-the-fly. We appreciate 
> those folks in the community for their great contribution! See their pending 
> JIRAs: HDFS-14805 & HDFS-12090.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Work logged] (HDFS-15714) HDFS Provided Storage Read/Write Mount Support On-the-fly

2021-02-28 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15714?focusedWorklogId=559193=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-559193
 ]

ASF GitHub Bot logged work on HDFS-15714:
-

Author: ASF GitHub Bot
Created on: 01/Mar/21 05:25
Start Date: 01/Mar/21 05:25
Worklog Time Spent: 10m 
  Work Description: PHILO-HE commented on pull request #2655:
URL: https://github.com/apache/hadoop/pull/2655#issuecomment-787648343


   Hi, @Zhangshunyu, in your HDFS config, did you set 
`dfs.provided.overreplication.factor` with a number larger than 0? This 
property is used to control the number of cached block (cached to HDFS from 
remote storage). If it is set to 0 (default setting), the redundant cached 
block will be deleted by HDFS, which can trigger replica not found exception.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 559193)
Time Spent: 1h  (was: 50m)

> HDFS Provided Storage Read/Write Mount Support On-the-fly
> -
>
> Key: HDFS-15714
> URL: https://issues.apache.org/jira/browse/HDFS-15714
> Project: Hadoop HDFS
>  Issue Type: New Feature
>  Components: datanode, namenode
>Affects Versions: 3.4.0
>Reporter: Feilong He
>Assignee: Feilong He
>Priority: Major
>  Labels: pull-request-available
> Attachments: HDFS-15714-01.patch, 
> HDFS_Provided_Storage_Design-V1.pdf, HDFS_Provided_Storage_Performance-V1.pdf
>
>  Time Spent: 1h
>  Remaining Estimate: 0h
>
> HDFS Provided Storage (PS) is a feature to tier HDFS over other file systems. 
> In HDFS-9806, PROVIDED storage type was introduced to HDFS. Through 
> configuring external storage with PROVIDED tag for DataNode, user can enable 
> application to access data stored externally from HDFS side. However, there 
> are two issues need to be addressed. Firstly, mounting external storage 
> on-the-fly, namely dynamic mount, is lacking. It is necessary to get it 
> supported to flexibly combine HDFS with an external storage at runtime. 
> Secondly, PS write is not supported by current HDFS. But in real 
> applications, it is common to transfer data bi-directionally for read/write 
> between HDFS and external storage.
> Through this JIRA, we are presenting our work for PS write support and 
> dynamic mount support for both read & write. Please note in the community 
> several JIRAs have been filed for these topics. Our work is based on these 
> previous community work, with new design & implementation to support called 
> writeBack mount and enable admin to add any mount on-the-fly. We appreciate 
> those folks in the community for their great contribution! See their pending 
> JIRAs: HDFS-14805 & HDFS-12090.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Work logged] (HDFS-15808) Add metrics for FSNamesystem read/write lock hold long time

2021-02-28 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15808?focusedWorklogId=559168=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-559168
 ]

ASF GitHub Bot logged work on HDFS-15808:
-

Author: ASF GitHub Bot
Created on: 01/Mar/21 03:26
Start Date: 01/Mar/21 03:26
Worklog Time Spent: 10m 
  Work Description: tomscut commented on pull request #2668:
URL: https://github.com/apache/hadoop/pull/2668#issuecomment-787609646


   Hi @anuengineer , could you please review again? Let's see if there are any 
other questions. Thank you.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 559168)
Time Spent: 5h 40m  (was: 5.5h)

> Add metrics for FSNamesystem read/write lock hold long time
> ---
>
> Key: HDFS-15808
> URL: https://issues.apache.org/jira/browse/HDFS-15808
> Project: Hadoop HDFS
>  Issue Type: Wish
>  Components: hdfs
>Reporter: tomscut
>Assignee: tomscut
>Priority: Major
>  Labels: hdfs, lock, metrics, pull-request-available
> Attachments: ExpiredHeartbeat.png, lockLongHoldCount
>
>  Time Spent: 5h 40m
>  Remaining Estimate: 0h
>
> To monitor how often read/write locks exceed thresholds, we can add two 
> metrics(ReadLockLongHoldCount/WriteLockLongHoldCount), which are exposed in 
> JMX.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Work logged] (HDFS-15856) Make recover the pipeline in same packet exceed times for stream closed configurable.

2021-02-28 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15856?focusedWorklogId=559165=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-559165
 ]

ASF GitHub Bot logged work on HDFS-15856:
-

Author: ASF GitHub Bot
Created on: 01/Mar/21 03:15
Start Date: 01/Mar/21 03:15
Worklog Time Spent: 10m 
  Work Description: qizhu-lucas commented on a change in pull request #2721:
URL: https://github.com/apache/hadoop/pull/2721#discussion_r584422979



##
File path: hadoop-hdfs-project/hadoop-hdfs/src/main/resources/hdfs-default.xml
##
@@ -4352,6 +4352,17 @@
   
 
 
+
+  dfs.client.pipeline.recovery.max-retries
+  5
+  
+If we had to recover the pipeline more than the value

Review comment:
   Good suggestion, i have changed to it.
   Thanks @jojochuang 

##
File path: 
hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/hdfs/DataStreamer.java
##
@@ -1263,14 +1265,18 @@ private boolean processDatanodeOrExternalError() throws 
IOException {
   packetSendTime.clear();
 }
 
-// If we had to recover the pipeline five times in a row for the
+// If we had to recover the pipeline more than the value
+// defined by maxPipelineRecoveryRetries in a row for the
 // same packet, this client likely has corrupt data or corrupting
 // during transmission.
-if (!errorState.isRestartingNode() && ++pipelineRecoveryCount > 5) {
+if (!errorState.isRestartingNode() && ++pipelineRecoveryCount >
+maxPipelineRecoveryRetries) {
   LOG.warn("Error recovering pipeline for writing " +
-  block + ". Already retried 5 times for the same packet.");
+  block + ". Already retried " + maxPipelineRecoveryRetries
+  + " times for the same packet.");
   lastException.set(new IOException("Failing write. Tried pipeline " +
-  "recovery 5 times without success."));
+  "recovery "+ maxPipelineRecoveryRetries

Review comment:
   Thanks @jojochuang for review.
   Fixed it latest pull request.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 559165)
Time Spent: 2h  (was: 1h 50m)

> Make recover the pipeline in same packet exceed times for stream closed 
> configurable.
> -
>
> Key: HDFS-15856
> URL: https://issues.apache.org/jira/browse/HDFS-15856
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: Qi Zhu
>Assignee: Qi Zhu
>Priority: Minor
>  Labels: pull-request-available
>  Time Spent: 2h
>  Remaining Estimate: 0h
>
> Now recover the pipeline five times in a row for the same packet, will close 
> the stream, but i think it should be configurable for different cluster 
> needed.
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Work logged] (HDFS-15856) Make recover the pipeline in same packet exceed times for stream closed configurable.

2021-02-28 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15856?focusedWorklogId=559164=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-559164
 ]

ASF GitHub Bot logged work on HDFS-15856:
-

Author: ASF GitHub Bot
Created on: 01/Mar/21 03:15
Start Date: 01/Mar/21 03:15
Worklog Time Spent: 10m 
  Work Description: qizhu-lucas commented on a change in pull request #2721:
URL: https://github.com/apache/hadoop/pull/2721#discussion_r584422934



##
File path: hadoop-hdfs-project/hadoop-hdfs/src/main/resources/hdfs-default.xml
##
@@ -4352,6 +4352,17 @@
   
 
 
+
+  dfs.client.pipeline.recovery.max-retries
+  5
+  
+If we had to recover the pipeline more than the value

Review comment:
   Good suggestion, i have changed to it.
   Thanks @jojochuang 





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 559164)
Time Spent: 1h 50m  (was: 1h 40m)

> Make recover the pipeline in same packet exceed times for stream closed 
> configurable.
> -
>
> Key: HDFS-15856
> URL: https://issues.apache.org/jira/browse/HDFS-15856
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: Qi Zhu
>Assignee: Qi Zhu
>Priority: Minor
>  Labels: pull-request-available
>  Time Spent: 1h 50m
>  Remaining Estimate: 0h
>
> Now recover the pipeline five times in a row for the same packet, will close 
> the stream, but i think it should be configurable for different cluster 
> needed.
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Work logged] (HDFS-15856) Make recover the pipeline in same packet exceed times for stream closed configurable.

2021-02-28 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15856?focusedWorklogId=559163=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-559163
 ]

ASF GitHub Bot logged work on HDFS-15856:
-

Author: ASF GitHub Bot
Created on: 01/Mar/21 03:15
Start Date: 01/Mar/21 03:15
Worklog Time Spent: 10m 
  Work Description: qizhu-lucas commented on a change in pull request #2721:
URL: https://github.com/apache/hadoop/pull/2721#discussion_r584422780



##
File path: 
hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/hdfs/DataStreamer.java
##
@@ -1263,14 +1265,18 @@ private boolean processDatanodeOrExternalError() throws 
IOException {
   packetSendTime.clear();
 }
 
-// If we had to recover the pipeline five times in a row for the
+// If we had to recover the pipeline more than the value
+// defined by maxPipelineRecoveryRetries in a row for the
 // same packet, this client likely has corrupt data or corrupting
 // during transmission.
-if (!errorState.isRestartingNode() && ++pipelineRecoveryCount > 5) {
+if (!errorState.isRestartingNode() && ++pipelineRecoveryCount >
+maxPipelineRecoveryRetries) {
   LOG.warn("Error recovering pipeline for writing " +
-  block + ". Already retried 5 times for the same packet.");
+  block + ". Already retried " + maxPipelineRecoveryRetries
+  + " times for the same packet.");
   lastException.set(new IOException("Failing write. Tried pipeline " +
-  "recovery 5 times without success."));
+  "recovery "+ maxPipelineRecoveryRetries

Review comment:
   Thanks @jojochuang for review.
   Fixed it latest pull request.

##
File path: 
hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/hdfs/DataStreamer.java
##
@@ -1263,14 +1265,18 @@ private boolean processDatanodeOrExternalError() throws 
IOException {
   packetSendTime.clear();
 }
 
-// If we had to recover the pipeline five times in a row for the
+// If we had to recover the pipeline more than the value
+// defined by maxPipelineRecoveryRetries in a row for the
 // same packet, this client likely has corrupt data or corrupting
 // during transmission.
-if (!errorState.isRestartingNode() && ++pipelineRecoveryCount > 5) {
+if (!errorState.isRestartingNode() && ++pipelineRecoveryCount >
+maxPipelineRecoveryRetries) {
   LOG.warn("Error recovering pipeline for writing " +
-  block + ". Already retried 5 times for the same packet.");
+  block + ". Already retried " + maxPipelineRecoveryRetries
+  + " times for the same packet.");
   lastException.set(new IOException("Failing write. Tried pipeline " +
-  "recovery 5 times without success."));
+  "recovery "+ maxPipelineRecoveryRetries

Review comment:
   Thanks @jojochuang for review.
   Fixed it latest pull request.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 559163)
Time Spent: 1h 40m  (was: 1.5h)

> Make recover the pipeline in same packet exceed times for stream closed 
> configurable.
> -
>
> Key: HDFS-15856
> URL: https://issues.apache.org/jira/browse/HDFS-15856
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: Qi Zhu
>Assignee: Qi Zhu
>Priority: Minor
>  Labels: pull-request-available
>  Time Spent: 1h 40m
>  Remaining Estimate: 0h
>
> Now recover the pipeline five times in a row for the same packet, will close 
> the stream, but i think it should be configurable for different cluster 
> needed.
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Work logged] (HDFS-15714) HDFS Provided Storage Read/Write Mount Support On-the-fly

2021-02-28 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15714?focusedWorklogId=559160=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-559160
 ]

ASF GitHub Bot logged work on HDFS-15714:
-

Author: ASF GitHub Bot
Created on: 01/Mar/21 02:54
Start Date: 01/Mar/21 02:54
Worklog Time Spent: 10m 
  Work Description: Zhangshunyu commented on pull request #2655:
URL: https://github.com/apache/hadoop/pull/2655#issuecomment-787600101


   @PHILO-HE Thanks for the reply. Please find the error msg:
   ```
ERROR org.apache.hadoop.hdfs.server.datanode.DataNode: Exception in 
creating BlockSender: {}
   org.apache.hadoop.hdfs.server.datanode.ReplicaNotFoundException: Replica not 
found for BP-1079229821-192.168.0.156-1614366591560:blk_1073791203_50379. The 
block may have been removed recently by the balancer or by intentionally 
reducing the replication factor. This condition is usually harmless. To be 
certain, please check the preceding datanode log messages for signs of a more 
serious issue.
   ```
   we use readOnly mount mode, the data is stored on s3 store, after mount we 
read the data use readthrough mode, we find that some times the datanode will 
throw this exception, it seems the read op request send to the wrong datanode?
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 559160)
Time Spent: 50m  (was: 40m)

> HDFS Provided Storage Read/Write Mount Support On-the-fly
> -
>
> Key: HDFS-15714
> URL: https://issues.apache.org/jira/browse/HDFS-15714
> Project: Hadoop HDFS
>  Issue Type: New Feature
>  Components: datanode, namenode
>Affects Versions: 3.4.0
>Reporter: Feilong He
>Assignee: Feilong He
>Priority: Major
>  Labels: pull-request-available
> Attachments: HDFS-15714-01.patch, 
> HDFS_Provided_Storage_Design-V1.pdf, HDFS_Provided_Storage_Performance-V1.pdf
>
>  Time Spent: 50m
>  Remaining Estimate: 0h
>
> HDFS Provided Storage (PS) is a feature to tier HDFS over other file systems. 
> In HDFS-9806, PROVIDED storage type was introduced to HDFS. Through 
> configuring external storage with PROVIDED tag for DataNode, user can enable 
> application to access data stored externally from HDFS side. However, there 
> are two issues need to be addressed. Firstly, mounting external storage 
> on-the-fly, namely dynamic mount, is lacking. It is necessary to get it 
> supported to flexibly combine HDFS with an external storage at runtime. 
> Secondly, PS write is not supported by current HDFS. But in real 
> applications, it is common to transfer data bi-directionally for read/write 
> between HDFS and external storage.
> Through this JIRA, we are presenting our work for PS write support and 
> dynamic mount support for both read & write. Please note in the community 
> several JIRAs have been filed for these topics. Our work is based on these 
> previous community work, with new design & implementation to support called 
> writeBack mount and enable admin to add any mount on-the-fly. We appreciate 
> those folks in the community for their great contribution! See their pending 
> JIRAs: HDFS-14805 & HDFS-12090.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Work logged] (HDFS-15714) HDFS Provided Storage Read/Write Mount Support On-the-fly

2021-02-28 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15714?focusedWorklogId=559159=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-559159
 ]

ASF GitHub Bot logged work on HDFS-15714:
-

Author: ASF GitHub Bot
Created on: 01/Mar/21 02:50
Start Date: 01/Mar/21 02:50
Worklog Time Spent: 10m 
  Work Description: PHILO-HE commented on pull request #2655:
URL: https://github.com/apache/hadoop/pull/2655#issuecomment-787599087


   Hi, @Zhangshunyu, thanks for your feedback. Could you please give more 
details? Did you mount an external storage in readOnly mode? Or writeBack mount?



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 559159)
Time Spent: 40m  (was: 0.5h)

> HDFS Provided Storage Read/Write Mount Support On-the-fly
> -
>
> Key: HDFS-15714
> URL: https://issues.apache.org/jira/browse/HDFS-15714
> Project: Hadoop HDFS
>  Issue Type: New Feature
>  Components: datanode, namenode
>Affects Versions: 3.4.0
>Reporter: Feilong He
>Assignee: Feilong He
>Priority: Major
>  Labels: pull-request-available
> Attachments: HDFS-15714-01.patch, 
> HDFS_Provided_Storage_Design-V1.pdf, HDFS_Provided_Storage_Performance-V1.pdf
>
>  Time Spent: 40m
>  Remaining Estimate: 0h
>
> HDFS Provided Storage (PS) is a feature to tier HDFS over other file systems. 
> In HDFS-9806, PROVIDED storage type was introduced to HDFS. Through 
> configuring external storage with PROVIDED tag for DataNode, user can enable 
> application to access data stored externally from HDFS side. However, there 
> are two issues need to be addressed. Firstly, mounting external storage 
> on-the-fly, namely dynamic mount, is lacking. It is necessary to get it 
> supported to flexibly combine HDFS with an external storage at runtime. 
> Secondly, PS write is not supported by current HDFS. But in real 
> applications, it is common to transfer data bi-directionally for read/write 
> between HDFS and external storage.
> Through this JIRA, we are presenting our work for PS write support and 
> dynamic mount support for both read & write. Please note in the community 
> several JIRAs have been filed for these topics. Our work is based on these 
> previous community work, with new design & implementation to support called 
> writeBack mount and enable admin to add any mount on-the-fly. We appreciate 
> those folks in the community for their great contribution! See their pending 
> JIRAs: HDFS-14805 & HDFS-12090.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Work logged] (HDFS-15854) Make some parameters configurable for SlowDiskTracker and SlowPeerTracker

2021-02-28 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15854?focusedWorklogId=559150=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-559150
 ]

ASF GitHub Bot logged work on HDFS-15854:
-

Author: ASF GitHub Bot
Created on: 01/Mar/21 01:21
Start Date: 01/Mar/21 01:21
Worklog Time Spent: 10m 
  Work Description: tomscut commented on pull request #2718:
URL: https://github.com/apache/hadoop/pull/2718#issuecomment-787572529


   Hi @arp7 @dineshchitlangia , Could you please help to review the code. Thank 
you.
   It is related to a previous issue: 
https://issues.apache.org/jira/browse/HDFS-15814.
   
   
   Those failed unit tests were unrelated to the change.
   Failed junit tests:
   hadoop.hdfs.server.datanode.TestNNHandlesBlockReportPerStorage
   hadoop.hdfs.server.namenode.TestNamenodeCapacityReport



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 559150)
Time Spent: 20m  (was: 10m)

> Make some parameters configurable for SlowDiskTracker and SlowPeerTracker
> -
>
> Key: HDFS-15854
> URL: https://issues.apache.org/jira/browse/HDFS-15854
> Project: Hadoop HDFS
>  Issue Type: Wish
>Reporter: tomscut
>Assignee: tomscut
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> Make some parameters configurable for SlowDiskTracker and SlowPeerTracker. 
> Related to https://issues.apache.org/jira/browse/HDFS-15814.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Work logged] (HDFS-15856) Make recover the pipeline in same packet exceed times for stream closed configurable.

2021-02-28 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15856?focusedWorklogId=559142=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-559142
 ]

ASF GitHub Bot logged work on HDFS-15856:
-

Author: ASF GitHub Bot
Created on: 01/Mar/21 00:55
Start Date: 01/Mar/21 00:55
Worklog Time Spent: 10m 
  Work Description: jojochuang commented on a change in pull request #2721:
URL: https://github.com/apache/hadoop/pull/2721#discussion_r584392706



##
File path: hadoop-hdfs-project/hadoop-hdfs/src/main/resources/hdfs-default.xml
##
@@ -4352,6 +4352,17 @@
   
 
 
+
+  dfs.client.pipeline.recovery.max-retries
+  5
+  
+If we had to recover the pipeline more than the value

Review comment:
   Suggest to add a description for this configuration property. Something 
like "if the DFS client encounters errors in write pipeline, retry up to the 
number defined by this property before giving up"





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 559142)
Time Spent: 1.5h  (was: 1h 20m)

> Make recover the pipeline in same packet exceed times for stream closed 
> configurable.
> -
>
> Key: HDFS-15856
> URL: https://issues.apache.org/jira/browse/HDFS-15856
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: Qi Zhu
>Assignee: Qi Zhu
>Priority: Minor
>  Labels: pull-request-available
>  Time Spent: 1.5h
>  Remaining Estimate: 0h
>
> Now recover the pipeline five times in a row for the same packet, will close 
> the stream, but i think it should be configurable for different cluster 
> needed.
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Work logged] (HDFS-15856) Make recover the pipeline in same packet exceed times for stream closed configurable.

2021-02-28 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15856?focusedWorklogId=559140=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-559140
 ]

ASF GitHub Bot logged work on HDFS-15856:
-

Author: ASF GitHub Bot
Created on: 01/Mar/21 00:51
Start Date: 01/Mar/21 00:51
Worklog Time Spent: 10m 
  Work Description: jojochuang commented on a change in pull request #2721:
URL: https://github.com/apache/hadoop/pull/2721#discussion_r584392237



##
File path: 
hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/hdfs/DataStreamer.java
##
@@ -1263,14 +1265,18 @@ private boolean processDatanodeOrExternalError() throws 
IOException {
   packetSendTime.clear();
 }
 
-// If we had to recover the pipeline five times in a row for the
+// If we had to recover the pipeline more than the value
+// defined by maxPipelineRecoveryRetries in a row for the
 // same packet, this client likely has corrupt data or corrupting
 // during transmission.
-if (!errorState.isRestartingNode() && ++pipelineRecoveryCount > 5) {
+if (!errorState.isRestartingNode() && ++pipelineRecoveryCount >
+maxPipelineRecoveryRetries) {
   LOG.warn("Error recovering pipeline for writing " +
-  block + ". Already retried 5 times for the same packet.");
+  block + ". Already retried " + maxPipelineRecoveryRetries
+  + " times for the same packet.");
   lastException.set(new IOException("Failing write. Tried pipeline " +
-  "recovery 5 times without success."));
+  "recovery "+ maxPipelineRecoveryRetries

Review comment:
   nit: add space between " and +





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 559140)
Time Spent: 1h 20m  (was: 1h 10m)

> Make recover the pipeline in same packet exceed times for stream closed 
> configurable.
> -
>
> Key: HDFS-15856
> URL: https://issues.apache.org/jira/browse/HDFS-15856
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: Qi Zhu
>Assignee: Qi Zhu
>Priority: Minor
>  Labels: pull-request-available
>  Time Spent: 1h 20m
>  Remaining Estimate: 0h
>
> Now recover the pipeline five times in a row for the same packet, will close 
> the stream, but i think it should be configurable for different cluster 
> needed.
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-15863) Validation message to be corrected in RBF FairnessPolicyController

2021-02-28 Thread Hadoop QA (Jira)


[ 
https://issues.apache.org/jira/browse/HDFS-15863?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17292521#comment-17292521
 ] 

Hadoop QA commented on HDFS-15863:
--

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime ||  Logfile || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 23m 
31s{color} | {color:blue}{color} | {color:blue} Docker mode activated. {color} |
|| || || || {color:brown} Prechecks {color} || ||
| {color:green}+1{color} | {color:green} dupname {color} | {color:green}  0m  
0s{color} | {color:green}{color} | {color:green} No case conflicting files 
found. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green}{color} | {color:green} The patch does not contain any 
@author tags. {color} |
| {color:green}+1{color} | {color:green} {color} | {color:green}  0m  0s{color} 
| {color:green}test4tests{color} | {color:green} The patch appears to include 1 
new or modified test files. {color} |
|| || || || {color:brown} trunk Compile Tests {color} || ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 34m 
27s{color} | {color:green}{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
38s{color} | {color:green}{color} | {color:green} trunk passed with JDK 
Ubuntu-11.0.10+9-Ubuntu-0ubuntu1.20.04 {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
32s{color} | {color:green}{color} | {color:green} trunk passed with JDK Private 
Build-1.8.0_282-8u282-b08-0ubuntu1~20.04-b08 {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
22s{color} | {color:green}{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  0m 
38s{color} | {color:green}{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 
16m 14s{color} | {color:green}{color} | {color:green} branch has no errors when 
building and testing our client artifacts. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
36s{color} | {color:green}{color} | {color:green} trunk passed with JDK 
Ubuntu-11.0.10+9-Ubuntu-0ubuntu1.20.04 {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
51s{color} | {color:green}{color} | {color:green} trunk passed with JDK Private 
Build-1.8.0_282-8u282-b08-0ubuntu1~20.04-b08 {color} |
| {color:blue}0{color} | {color:blue} spotbugs {color} | {color:blue}  1m 
13s{color} | {color:blue}{color} | {color:blue} Used deprecated FindBugs 
config; considering switching to SpotBugs. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  1m 
12s{color} | {color:green}{color} | {color:green} trunk passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} || ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  0m 
31s{color} | {color:green}{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
32s{color} | {color:green}{color} | {color:green} the patch passed with JDK 
Ubuntu-11.0.10+9-Ubuntu-0ubuntu1.20.04 {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
32s{color} | {color:green}{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
27s{color} | {color:green}{color} | {color:green} the patch passed with JDK 
Private Build-1.8.0_282-8u282-b08-0ubuntu1~20.04-b08 {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
27s{color} | {color:green}{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
15s{color} | {color:green}{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  0m 
29s{color} | {color:green}{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green}{color} | {color:green} The patch has no whitespace 
issues. {color} |
| {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 
14m 52s{color} | {color:green}{color} | {color:green} patch has no errors when 
building and testing our client artifacts. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
33s{color} | {color:green}{color} | {color:green} the patch passed with JDK 
Ubuntu-11.0.10+9-Ubuntu-0ubuntu1.20.04 {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
49s{color} | {color:green}{color} | {color:green} the patch 

[jira] [Commented] (HDFS-15745) Make DataNodePeerMetrics#LOW_THRESHOLD_MS and MIN_OUTLIER_DETECTION_NODES configurable

2021-02-28 Thread Renukaprasad C (Jira)


[ 
https://issues.apache.org/jira/browse/HDFS-15745?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17292490#comment-17292490
 ] 

Renukaprasad C commented on HDFS-15745:
---

Thanks [~huanghaibin] for reporting & [~ayushtkn] for quick review and commit.
Can we merge the same to 3.3 & 3.1 branches as well?

> Make DataNodePeerMetrics#LOW_THRESHOLD_MS and MIN_OUTLIER_DETECTION_NODES 
> configurable
> --
>
> Key: HDFS-15745
> URL: https://issues.apache.org/jira/browse/HDFS-15745
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: Haibin Huang
>Assignee: Haibin Huang
>Priority: Major
> Fix For: 3.4.0
>
> Attachments: HDFS-15745-001.patch, HDFS-15745-002.patch, 
> HDFS-15745-003.patch, image-2020-12-22-17-00-50-796.png
>
>
> When i enable DataNodePeerMetrics to find slow slow peer in cluster, i found 
> there is a lot of slow peer but ReportingNodes's averageDelay is very low, 
> and these slow peer node are normal. I think the reason of why generating so 
> many slow peer is that  the value of DataNodePeerMetrics#LOW_THRESHOLD_MS is 
> too small (only 5ms) and it is not configurable. The default value of slow io 
> warning log threshold is 300ms, i.e. 
> DFSConfigKeys.DFS_DATANODE_SLOW_IO_WARNING_THRESHOLD_DEFAULT = 300, so 
> DataNodePeerMetrics#LOW_THRESHOLD_MS should not be less than 300ms, otherwise 
> namenode will get a lot of invalid slow peer information.
> !image-2020-12-22-17-00-50-796.png!



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-15863) Validation message to be corrected in RBF FairnessPolicyController

2021-02-28 Thread Renukaprasad C (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15863?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Renukaprasad C updated HDFS-15863:
--
Attachment: HDFS-15863.001.patch
Status: Patch Available  (was: Open)

> Validation message to be corrected in RBF FairnessPolicyController
> --
>
> Key: HDFS-15863
> URL: https://issues.apache.org/jira/browse/HDFS-15863
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: rbf
>Affects Versions: 3.4.0
>Reporter: Renukaprasad C
>Assignee: Renukaprasad C
>Priority: Minor
> Attachments: HDFS-15863.001.patch
>
>
> org.apache.hadoop.hdfs.server.federation.fairness.StaticRouterRpcFairnessPolicyController#validateCount
> When dfs.federation.router.handler.count is lessthan the total dedicated 
> handlers for all NS, then error message shows 0 & -ve values in error 
> message, instead can show the actual configured values.
> Current message is : "Available handlers -5 lower than min 0 for nsId nn1"
> This can be changed to: "Configured handlers 
> ${DFS_ROUTER_HANDLER_COUNT_KEY}=10 lower than min 15 for nsId nn1", where 10 
> is hander count & 15 is sum of dedicated handler count.
> Related to: HDFS-14090



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Work logged] (HDFS-15843) [libhdfs++] Make write cross platform

2021-02-28 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15843?focusedWorklogId=559107=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-559107
 ]

ASF GitHub Bot logged work on HDFS-15843:
-

Author: ASF GitHub Bot
Created on: 28/Feb/21 18:04
Start Date: 28/Feb/21 18:04
Worklog Time Spent: 10m 
  Work Description: GauthamBanasandra commented on pull request #2710:
URL: https://github.com/apache/hadoop/pull/2710#issuecomment-787494012


   Thanks for your review and suggestion @smengcl , @jojochuang. I've 
re-implemented my solution using the `write` system call for Windows and Linux 
separately, neatly tucked inside the `XPlatform` class . Please do have 
another look at my PR.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 559107)
Time Spent: 1h 20m  (was: 1h 10m)

> [libhdfs++] Make write cross platform
> -
>
> Key: HDFS-15843
> URL: https://issues.apache.org/jira/browse/HDFS-15843
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: libhdfs++
>Affects Versions: 3.2.2
>Reporter: Gautham Banasandra
>Assignee: Gautham Banasandra
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 1h 20m
>  Remaining Estimate: 0h
>
> We're currently using the *write* function from unistd.h which isn't 
> cross-platform. We need to replace this with *std::cout.write* instead.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Work logged] (HDFS-15843) [libhdfs++] Make write cross platform

2021-02-28 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15843?focusedWorklogId=559102=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-559102
 ]

ASF GitHub Bot logged work on HDFS-15843:
-

Author: ASF GitHub Bot
Created on: 28/Feb/21 16:33
Start Date: 28/Feb/21 16:33
Worklog Time Spent: 10m 
  Work Description: GauthamBanasandra commented on a change in pull request 
#2710:
URL: https://github.com/apache/hadoop/pull/2710#discussion_r584322501



##
File path: 
hadoop-hdfs-project/hadoop-hdfs-native-client/src/main/native/libhdfspp/examples/c/connect_cancel/connect_cancel.c
##
@@ -43,10 +42,10 @@ const char *catch_exit   = "Exiting the signal handler.\n";
 // Print to stdout without calling malloc or otherwise indirectly modify 
userspace state.
 // Write calls to stdout may still interleave with stuff coming from elsewhere.
 static void sighandler_direct_stdout(const char *msg) {
-  if(!msg)
+  if(!msg) {
 return;
-  ssize_t res = write(1 /*posix stdout fd*/, msg, strlen(msg));
-  (void)res;
+  }
+  printf("%s", msg);

Review comment:
   I've removed this. Please take a look at the latest commit.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 559102)
Time Spent: 1h 10m  (was: 1h)

> [libhdfs++] Make write cross platform
> -
>
> Key: HDFS-15843
> URL: https://issues.apache.org/jira/browse/HDFS-15843
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: libhdfs++
>Affects Versions: 3.2.2
>Reporter: Gautham Banasandra
>Assignee: Gautham Banasandra
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 1h 10m
>  Remaining Estimate: 0h
>
> We're currently using the *write* function from unistd.h which isn't 
> cross-platform. We need to replace this with *std::cout.write* instead.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Work logged] (HDFS-13975) TestBalancer#testMaxIterationTime fails sporadically

2021-02-28 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-13975?focusedWorklogId=559100=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-559100
 ]

ASF GitHub Bot logged work on HDFS-13975:
-

Author: ASF GitHub Bot
Created on: 28/Feb/21 16:07
Start Date: 28/Feb/21 16:07
Worklog Time Spent: 10m 
  Work Description: touchida commented on pull request #2726:
URL: https://github.com/apache/hadoop/pull/2726#issuecomment-787475633


   Both failed and crashed tests are unrelated to this PR. I created 
[HDFS-15864](https://issues.apache.org/jira/browse/HDFS-15864) for the former, 
but didn't for the latter, since I'm not sure which caused the crash, the build 
environment or the test itself.
   Ready for review.
   - Failed test
   ```
   [ERROR] Tests run: 29, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 
47.231 s <<< FAILURE! - in 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestFsDatasetImpl
   [ERROR] 
testDnRestartWithHardLink(org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestFsDatasetImpl)
  Time elapsed: 1.452 s  <<< FAILURE!
   java.lang.AssertionError
at org.junit.Assert.fail(Assert.java:86)
at org.junit.Assert.assertTrue(Assert.java:41)
at org.junit.Assert.assertTrue(Assert.java:52)
at 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestFsDatasetImpl.testDnRestartWithHardLink(TestFsDatasetImpl.java:1377)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
at 
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at 
org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:298)
at 
org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:292)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.lang.Thread.run(Thread.java:748)
   ```
   - Crashed test
   ```
   org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestSpaceReservation
   org.apache.maven.surefire.booter.SurefireBooterForkException: 
ExecutionException The forked VM terminated without properly saying goodbye. VM 
crash or System.exit called?
   Command was /bin/sh -c cd 
/home/jenkins/jenkins-agent/workspace/hadoop-multibranch_PR-2726/src/hadoop-hdfs-project/hadoop-hdfs
 && /usr/lib/jvm/java-8-openjdk-amd64/jre/bin/java -Xmx2048m 
-XX:+HeapDumpOnOutOfMemoryError -DminiClusterDedicatedDirs=true -jar 
/home/jenkins/jenkins-agent/workspace/hadoop-multibranch_PR-2726/src/hadoop-hdfs-project/hadoop-hdfs/target/surefire/surefirebooter6106358806783121604.jar
 
/home/jenkins/jenkins-agent/workspace/hadoop-multibranch_PR-2726/src/hadoop-hdfs-project/hadoop-hdfs/target/surefire
 2021-02-27T21-36-41_009-jvmRun2 surefire4678365300915584400tmp 
surefire_5963761349348601015446tmp
   Error occurred in starting fork, check output in log
   Process Exit Code: 1
   Crashed tests:
   org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestSpaceReservation
at 
org.apache.maven.plugin.surefire.booterclient.ForkStarter.awaitResultsDone(ForkStarter.java:511)
at 
org.apache.maven.plugin.surefire.booterclient.ForkStarter.runSuitesForkPerTestSet(ForkStarter.java:458)
at 
org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:299)
at 
org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:247)
at 
org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeProvider(AbstractSurefireMojo.java:1149)
at 
org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeAfterPreconditionsChecked(AbstractSurefireMojo.java:991)
at 
org.apache.maven.plugin.surefire.AbstractSurefireMojo.execute(AbstractSurefireMojo.java:837)
at 
org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo(DefaultBuildPluginManager.java:137)
at 
org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:210)
at 
org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:156)
at 
org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:148)
at 
org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:117)
at 

[jira] [Created] (HDFS-15864) TestFsDatasetImpl#testDnRestartWithHardLink fails intermittently

2021-02-28 Thread Toshihiko Uchida (Jira)
Toshihiko Uchida created HDFS-15864:
---

 Summary: TestFsDatasetImpl#testDnRestartWithHardLink fails 
intermittently
 Key: HDFS-15864
 URL: https://issues.apache.org/jira/browse/HDFS-15864
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Toshihiko Uchida


This unit test failed in https://github.com/apache/hadoop/pull/2726 due to an 
AssertionError.
{code}
[ERROR] 
testDnRestartWithHardLink(org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestFsDatasetImpl)
  Time elapsed: 1.452 s  <<< FAILURE!
java.lang.AssertionError
at org.junit.Assert.fail(Assert.java:86)
at org.junit.Assert.assertTrue(Assert.java:41)
at org.junit.Assert.assertTrue(Assert.java:52)
at 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestFsDatasetImpl.testDnRestartWithHardLink(TestFsDatasetImpl.java:1377)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
at 
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at 
org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:298)
at 
org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:292)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.lang.Thread.run(Thread.java:748)
{code}
The failure occurred at the following first assertion.
{code}
  cluster.restartDataNode(0);
  cluster.waitDatanodeFullyStarted(cluster.getDataNodes().get(0), 6);
  cluster.triggerBlockReports();

  assertTrue(Files.exists(Paths.get(newReplicaInfo.getBlockURI(;
  assertTrue(Files.exists(Paths.get(oldReplicaInfo.getBlockURI(;
{code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-13975) TestBalancer#testMaxIterationTime fails sporadically

2021-02-28 Thread Toshihiko Uchida (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-13975?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Toshihiko Uchida updated HDFS-13975:

Labels: flaky-test pull-request-available  (was: pull-request-available)

> TestBalancer#testMaxIterationTime fails sporadically
> 
>
> Key: HDFS-13975
> URL: https://issues.apache.org/jira/browse/HDFS-13975
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Affects Versions: 3.2.0
>Reporter: Jason Darrell Lowe
>Assignee: Toshihiko Uchida
>Priority: Major
>  Labels: flaky-test, pull-request-available
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> A number of precommit builds have seen this test fail like this:
> {noformat}
> java.lang.AssertionError: Unexpected iteration runtime: 4021ms > 3.5s
>   at org.junit.Assert.fail(Assert.java:88)
>   at org.junit.Assert.assertTrue(Assert.java:41)
>   at 
> org.apache.hadoop.hdfs.server.balancer.TestBalancer.testMaxIterationTime(TestBalancer.java:1649)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:498)
>   at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
>   at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
>   at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
>   at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
>   at 
> org.junit.internal.runners.statements.FailOnTimeout$StatementThread.run(FailOnTimeout.java:74)
> {noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-15856) Make recover the pipeline in same packet exceed times for stream closed configurable.

2021-02-28 Thread Hadoop QA (Jira)


[ 
https://issues.apache.org/jira/browse/HDFS-15856?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17292354#comment-17292354
 ] 

Hadoop QA commented on HDFS-15856:
--

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime ||  Logfile || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
41s{color} |  | {color:blue} Docker mode activated. {color} |
|| || || || {color:brown} Prechecks {color} || ||
| {color:green}+1{color} | {color:green} dupname {color} | {color:green}  0m  
0s{color} |  | {color:green} No case conflicting files found. {color} |
| {color:blue}0{color} | {color:blue} codespell {color} | {color:blue}  0m  
0s{color} |  | {color:blue} codespell was not available. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} |  | {color:green} The patch does not contain any @author tags. 
{color} |
| {color:red}-1{color} | {color:red} test4tests {color} | {color:red}  0m  
0s{color} |  | {color:red} The patch doesn't appear to include any new or 
modified tests. Please justify why no new tests are needed for this patch. Also 
please list what manual steps were performed to verify this patch. {color} |
|| || || || {color:brown} trunk Compile Tests {color} || ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 14m 
14s{color} |  | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 20m 
11s{color} |  | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  4m 
50s{color} |  | {color:green} trunk passed with JDK 
Ubuntu-11.0.10+9-Ubuntu-0ubuntu1.20.04 {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  4m 
25s{color} |  | {color:green} trunk passed with JDK Private 
Build-1.8.0_282-8u282-b08-0ubuntu1~20.04-b08 {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  1m 
15s{color} |  | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  2m 
13s{color} |  | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
35s{color} |  | {color:green} trunk passed with JDK 
Ubuntu-11.0.10+9-Ubuntu-0ubuntu1.20.04 {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  2m  
5s{color} |  | {color:green} trunk passed with JDK Private 
Build-1.8.0_282-8u282-b08-0ubuntu1~20.04-b08 {color} |
| {color:blue}0{color} | {color:blue} spotbugs {color} | {color:blue}  3m  
2s{color} |  | {color:blue} Used deprecated FindBugs config; considering 
switching to SpotBugs. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  5m 
25s{color} |  | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 
16m  1s{color} |  | {color:green} branch has no errors when building and 
testing our client artifacts. {color} |
|| || || || {color:brown} Patch Compile Tests {color} || ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
26s{color} |  | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
56s{color} |  | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  4m 
41s{color} |  | {color:green} the patch passed with JDK 
Ubuntu-11.0.10+9-Ubuntu-0ubuntu1.20.04 {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  4m 
41s{color} |  | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  4m 
19s{color} |  | {color:green} the patch passed with JDK Private 
Build-1.8.0_282-8u282-b08-0ubuntu1~20.04-b08 {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  4m 
19s{color} |  | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} blanks {color} | {color:green}  0m  
0s{color} |  | {color:green} The patch has no blanks issues. {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  1m 
 6s{color} |  | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  1m 
59s{color} |  | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} xml {color} | {color:green}  0m  
1s{color} |  | {color:green} The patch has no ill-formed XML file. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
19s{color} |  | {color:green} the patch passed with JDK 
Ubuntu-11.0.10+9-Ubuntu-0ubuntu1.20.04 {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
49s{color} |  |