[jira] [Commented] (HDFS-15839) RBF: Cannot get method setBalancerBandwidth on Router Client

2021-02-17 Thread Hadoop QA (Jira)


[ 
https://issues.apache.org/jira/browse/HDFS-15839?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17286309#comment-17286309
 ] 

Hadoop QA commented on HDFS-15839:
--

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime ||  Logfile || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
33s{color} | {color:blue}{color} | {color:blue} Docker mode activated. {color} |
| {color:red}-1{color} | {color:red} yetus {color} | {color:red}  0m  7s{color} 
| {color:red}{color} | {color:red} Unprocessed flag(s): 
--findbugs-strict-precheck {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Docker | ClientAPI=1.41 ServerAPI=1.41 base: 
https://ci-hadoop.apache.org/job/PreCommit-HDFS-Build/479/artifact/out/Dockerfile
 |
| JIRA Issue | HDFS-15839 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/13020611/HDFS-15839.patch |
| Console output | 
https://ci-hadoop.apache.org/job/PreCommit-HDFS-Build/479/console |
| versions | git=2.25.1 |
| Powered by | Apache Yetus 0.13.0-SNAPSHOT https://yetus.apache.org |


This message was automatically generated.



> RBF: Cannot get method setBalancerBandwidth on Router Client
> 
>
> Key: HDFS-15839
> URL: https://issues.apache.org/jira/browse/HDFS-15839
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: rbf
>Reporter: Yang Yun
>Assignee: Yang Yun
>Priority: Major
> Attachments: HDFS-15839.patch
>
>
> When call setBalancerBandwidth, throw exeption,
> {code:java}
> 02-18 14:39:59,186 [IPC Server handler 0 on default port 43545] ERROR 
> router.RemoteMethod (RemoteMethod.java:getMethod(146)) - Cannot get method 
> setBalancerBandwidth with types [class java.lang.Long] from 
> ClientProtocoljava.lang.NoSuchMethodException: 
> org.apache.hadoop.hdfs.protocol.ClientProtocol.setBalancerBandwidth(java.lang.Long)
>  at java.lang.Class.getDeclaredMethod(Class.java:2130) at 
> org.apache.hadoop.hdfs.server.federation.router.RemoteMethod.getMethod(RemoteMethod.java:140)
>  at 
> org.apache.hadoop.hdfs.server.federation.router.RouterRpcClient.invokeConcurrent(RouterRpcClient.java:1312)
>  at 
> org.apache.hadoop.hdfs.server.federation.router.RouterRpcClient.invokeConcurrent(RouterRpcClient.java:1250)
>  at 
> org.apache.hadoop.hdfs.server.federation.router.RouterRpcClient.invokeConcurrent(RouterRpcClient.java:1221)
>  at 
> org.apache.hadoop.hdfs.server.federation.router.RouterRpcClient.invokeConcurrent(RouterRpcClient.java:1194)
>  at 
> org.apache.hadoop.hdfs.server.federation.router.RouterClientProtocol.setBalancerBandwidth(RouterClientProtocol.java:1188)
>  at 
> org.apache.hadoop.hdfs.server.federation.router.RouterRpcServer.setBalancerBandwidth(RouterRpcServer.java:1211)
>  at 
> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.setBalancerBandwidth(ClientNamenodeProtocolServerSideTranslatorPB.java:1254)
>  at 
> org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
>  at 
> org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:537)
>  at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1086) at 
> org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1037) at 
> org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:965) at 
> java.security.AccessController.doPrivileged(Native Method) at 
> javax.security.auth.Subject.doAs(Subject.java:422) at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1899)
>  at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2972){code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-15839) RBF: Cannot get method setBalancerBandwidth on Router Client

2021-02-17 Thread Yang Yun (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15839?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yang Yun updated HDFS-15839:

Attachment: HDFS-15839.patch
Status: Patch Available  (was: Open)

> RBF: Cannot get method setBalancerBandwidth on Router Client
> 
>
> Key: HDFS-15839
> URL: https://issues.apache.org/jira/browse/HDFS-15839
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: rbf
>Reporter: Yang Yun
>Assignee: Yang Yun
>Priority: Major
> Attachments: HDFS-15839.patch
>
>
> When call setBalancerBandwidth, throw exeption,
> {code:java}
> 02-18 14:39:59,186 [IPC Server handler 0 on default port 43545] ERROR 
> router.RemoteMethod (RemoteMethod.java:getMethod(146)) - Cannot get method 
> setBalancerBandwidth with types [class java.lang.Long] from 
> ClientProtocoljava.lang.NoSuchMethodException: 
> org.apache.hadoop.hdfs.protocol.ClientProtocol.setBalancerBandwidth(java.lang.Long)
>  at java.lang.Class.getDeclaredMethod(Class.java:2130) at 
> org.apache.hadoop.hdfs.server.federation.router.RemoteMethod.getMethod(RemoteMethod.java:140)
>  at 
> org.apache.hadoop.hdfs.server.federation.router.RouterRpcClient.invokeConcurrent(RouterRpcClient.java:1312)
>  at 
> org.apache.hadoop.hdfs.server.federation.router.RouterRpcClient.invokeConcurrent(RouterRpcClient.java:1250)
>  at 
> org.apache.hadoop.hdfs.server.federation.router.RouterRpcClient.invokeConcurrent(RouterRpcClient.java:1221)
>  at 
> org.apache.hadoop.hdfs.server.federation.router.RouterRpcClient.invokeConcurrent(RouterRpcClient.java:1194)
>  at 
> org.apache.hadoop.hdfs.server.federation.router.RouterClientProtocol.setBalancerBandwidth(RouterClientProtocol.java:1188)
>  at 
> org.apache.hadoop.hdfs.server.federation.router.RouterRpcServer.setBalancerBandwidth(RouterRpcServer.java:1211)
>  at 
> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.setBalancerBandwidth(ClientNamenodeProtocolServerSideTranslatorPB.java:1254)
>  at 
> org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
>  at 
> org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:537)
>  at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1086) at 
> org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1037) at 
> org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:965) at 
> java.security.AccessController.doPrivileged(Native Method) at 
> javax.security.auth.Subject.doAs(Subject.java:422) at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1899)
>  at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2972){code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Created] (HDFS-15839) RBF: Cannot get method setBalancerBandwidth on Router Client

2021-02-17 Thread Yang Yun (Jira)
Yang Yun created HDFS-15839:
---

 Summary: RBF: Cannot get method setBalancerBandwidth on Router 
Client
 Key: HDFS-15839
 URL: https://issues.apache.org/jira/browse/HDFS-15839
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: rbf
Reporter: Yang Yun
Assignee: Yang Yun


When call setBalancerBandwidth, throw exeption,
{code:java}
02-18 14:39:59,186 [IPC Server handler 0 on default port 43545] ERROR 
router.RemoteMethod (RemoteMethod.java:getMethod(146)) - Cannot get method 
setBalancerBandwidth with types [class java.lang.Long] from 
ClientProtocoljava.lang.NoSuchMethodException: 
org.apache.hadoop.hdfs.protocol.ClientProtocol.setBalancerBandwidth(java.lang.Long)
 at java.lang.Class.getDeclaredMethod(Class.java:2130) at 
org.apache.hadoop.hdfs.server.federation.router.RemoteMethod.getMethod(RemoteMethod.java:140)
 at 
org.apache.hadoop.hdfs.server.federation.router.RouterRpcClient.invokeConcurrent(RouterRpcClient.java:1312)
 at 
org.apache.hadoop.hdfs.server.federation.router.RouterRpcClient.invokeConcurrent(RouterRpcClient.java:1250)
 at 
org.apache.hadoop.hdfs.server.federation.router.RouterRpcClient.invokeConcurrent(RouterRpcClient.java:1221)
 at 
org.apache.hadoop.hdfs.server.federation.router.RouterRpcClient.invokeConcurrent(RouterRpcClient.java:1194)
 at 
org.apache.hadoop.hdfs.server.federation.router.RouterClientProtocol.setBalancerBandwidth(RouterClientProtocol.java:1188)
 at 
org.apache.hadoop.hdfs.server.federation.router.RouterRpcServer.setBalancerBandwidth(RouterRpcServer.java:1211)
 at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.setBalancerBandwidth(ClientNamenodeProtocolServerSideTranslatorPB.java:1254)
 at 
org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
 at 
org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:537)
 at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1086) at 
org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1037) at 
org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:965) at 
java.security.AccessController.doPrivileged(Native Method) at 
javax.security.auth.Subject.doAs(Subject.java:422) at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1899)
 at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2972){code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Work logged] (HDFS-15785) Datanode to support using DNS to resolve nameservices to IP addresses to get list of namenodes

2021-02-17 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15785?focusedWorklogId=554083=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-554083
 ]

ASF GitHub Bot logged work on HDFS-15785:
-

Author: ASF GitHub Bot
Created on: 18/Feb/21 06:15
Start Date: 18/Feb/21 06:15
Worklog Time Spent: 10m 
  Work Description: hadoop-yetus commented on pull request #2639:
URL: https://github.com/apache/hadoop/pull/2639#issuecomment-781085217


   :broken_heart: **-1 overall**
   
   
   
   
   
   
   | Vote | Subsystem | Runtime |  Logfile | Comment |
   |::|--:|:|::|:---:|
   | +0 :ok: |  reexec  |  24m  4s |  |  Docker mode activated.  |
    _ Prechecks _ |
   | +1 :green_heart: |  dupname  |   0m  0s |  |  No case conflicting files 
found.  |
   | +1 :green_heart: |  @author  |   0m  0s |  |  The patch does not contain 
any @author tags.  |
   | +1 :green_heart: |   |   0m  0s | [test4tests](test4tests) |  The patch 
appears to include 1 new or modified test files.  |
    _ trunk Compile Tests _ |
   | +0 :ok: |  mvndep  |  13m 55s |  |  Maven dependency ordering for branch  |
   | +1 :green_heart: |  mvninstall  |  23m  1s |  |  trunk passed  |
   | +1 :green_heart: |  compile  |  22m 30s |  |  trunk passed with JDK 
Ubuntu-11.0.10+9-Ubuntu-0ubuntu1.20.04  |
   | +1 :green_heart: |  compile  |  19m  4s |  |  trunk passed with JDK 
Private Build-1.8.0_282-8u282-b08-0ubuntu1~20.04-b08  |
   | +1 :green_heart: |  checkstyle  |   4m 36s |  |  trunk passed  |
   | +1 :green_heart: |  mvnsite  |   4m  2s |  |  trunk passed  |
   | +1 :green_heart: |  shadedclient  |  24m 36s |  |  branch has no errors 
when building and testing our client artifacts.  |
   | +1 :green_heart: |  javadoc  |   2m 45s |  |  trunk passed with JDK 
Ubuntu-11.0.10+9-Ubuntu-0ubuntu1.20.04  |
   | +1 :green_heart: |  javadoc  |   3m 44s |  |  trunk passed with JDK 
Private Build-1.8.0_282-8u282-b08-0ubuntu1~20.04-b08  |
   | +0 :ok: |  spotbugs  |  39m 21s |  |  Both FindBugs and SpotBugs are 
enabled, using SpotBugs.  |
   | +1 :green_heart: |  spotbugs  |   8m 19s |  |  trunk passed  |
    _ Patch Compile Tests _ |
   | +0 :ok: |  mvndep  |   0m 22s |  |  Maven dependency ordering for patch  |
   | +1 :green_heart: |  mvninstall  |   2m 54s |  |  the patch passed  |
   | +1 :green_heart: |  compile  |  21m 53s |  |  the patch passed with JDK 
Ubuntu-11.0.10+9-Ubuntu-0ubuntu1.20.04  |
   | +1 :green_heart: |  javac  |  21m 53s |  |  
root-jdkUbuntu-11.0.10+9-Ubuntu-0ubuntu1.20.04 with JDK 
Ubuntu-11.0.10+9-Ubuntu-0ubuntu1.20.04 generated 0 new + 1957 unchanged - 2 
fixed = 1957 total (was 1959)  |
   | +1 :green_heart: |  compile  |  19m  7s |  |  the patch passed with JDK 
Private Build-1.8.0_282-8u282-b08-0ubuntu1~20.04-b08  |
   | +1 :green_heart: |  javac  |  19m  7s |  |  
root-jdkPrivateBuild-1.8.0_282-8u282-b08-0ubuntu1~20.04-b08 with JDK Private 
Build-1.8.0_282-8u282-b08-0ubuntu1~20.04-b08 generated 0 new + 1853 unchanged - 
2 fixed = 1853 total (was 1855)  |
   | +1 :green_heart: |  checkstyle  |   4m  5s |  |  root: The patch generated 
0 new + 558 unchanged - 2 fixed = 558 total (was 560)  |
   | +1 :green_heart: |  mvnsite  |   3m 58s |  |  the patch passed  |
   | +1 :green_heart: |  whitespace  |   0m  0s |  |  The patch has no 
whitespace issues.  |
   | +1 :green_heart: |  xml  |   0m  1s |  |  The patch has no ill-formed XML 
file.  |
   | +1 :green_heart: |  shadedclient  |  15m 30s |  |  patch has no errors 
when building and testing our client artifacts.  |
   | +1 :green_heart: |  javadoc  |   2m 40s |  |  the patch passed with JDK 
Ubuntu-11.0.10+9-Ubuntu-0ubuntu1.20.04  |
   | +1 :green_heart: |  javadoc  |   3m 50s |  |  the patch passed with JDK 
Private Build-1.8.0_282-8u282-b08-0ubuntu1~20.04-b08  |
   | -1 :x: |  spotbugs  |   2m 18s | 
[/patch-spotbugs-hadoop-common-project_hadoop-common.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-2639/6/artifact/out/patch-spotbugs-hadoop-common-project_hadoop-common.txt)
 |  hadoop-common-project/hadoop-common cannot run computeBugHistory from 
spotbugs  |
   | -1 :x: |  spotbugs  |   2m 41s | 
[/patch-spotbugs-hadoop-hdfs-project_hadoop-hdfs-client.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-2639/6/artifact/out/patch-spotbugs-hadoop-hdfs-project_hadoop-hdfs-client.txt)
 |  hadoop-hdfs-project/hadoop-hdfs-client cannot run computeBugHistory from 
spotbugs  |
   | -1 :x: |  spotbugs  |   3m 21s | 
[/patch-spotbugs-hadoop-hdfs-project_hadoop-hdfs.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-2639/6/artifact/out/patch-spotbugs-hadoop-hdfs-project_hadoop-hdfs.txt)
 |  hadoop-hdfs-project/hadoop-hdfs cannot run computeBugHistory from spotbugs  
|
    _ Other Tests _ |
   | +1 :green_heart: |  unit  |  17m  9s |  |  hadoop-common in the patch 
passed.  |
   | +1 :green_heart: 

[jira] [Commented] (HDFS-15793) Add command to DFSAdmin for Balancer max concurrent threads

2021-02-17 Thread Yang Yun (Jira)


[ 
https://issues.apache.org/jira/browse/HDFS-15793?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17286295#comment-17286295
 ] 

Yang Yun commented on HDFS-15793:
-

Thanks [~ayushtkn] for your review.

Sorry for the last patch lost some functions, updated to HDFS-15793.002.patch 
with following changes,
 * Add logical to process comand 'DNA_BALANCERBANDWIDTHUPDATE' in  
BPOfferService and call updateBalancerMaxConcurrentMovers of DataXceiverServer 
to make change thread bother smoothly.
 * Add more test to check all datanode or call '-getBalancerMaxThreads' to make 
sure that actaully the threads increased/decreased.
 * Fix the bug in router code and add a test for Router.

 

> Add command to DFSAdmin for Balancer max concurrent  threads
> 
>
> Key: HDFS-15793
> URL: https://issues.apache.org/jira/browse/HDFS-15793
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: balancer  mover
>Reporter: Yang Yun
>Assignee: Yang Yun
>Priority: Minor
> Attachments: HDFS-15793.001.patch, HDFS-15793.002.patch
>
>
> We have DFSAdmin command '-setBalancerBandwidth' to dynamically change the 
> max number of bytes per second of network bandwidth to be used by a datanode 
> during balancing.  Also add '-setBalancerMaxThreads' to dynamically change 
> the balancer maxThread number which may be used concurrently for moving 
> blocks.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-15793) Add command to DFSAdmin for Balancer max concurrent threads

2021-02-17 Thread Hadoop QA (Jira)


[ 
https://issues.apache.org/jira/browse/HDFS-15793?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17286294#comment-17286294
 ] 

Hadoop QA commented on HDFS-15793:
--

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime ||  Logfile || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
40s{color} | {color:blue}{color} | {color:blue} Docker mode activated. {color} |
| {color:red}-1{color} | {color:red} yetus {color} | {color:red}  0m  9s{color} 
| {color:red}{color} | {color:red} Unprocessed flag(s): 
--findbugs-strict-precheck {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Docker | ClientAPI=1.41 ServerAPI=1.41 base: 
https://ci-hadoop.apache.org/job/PreCommit-HDFS-Build/478/artifact/out/Dockerfile
 |
| JIRA Issue | HDFS-15793 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/13020605/HDFS-15793.002.patch |
| Console output | 
https://ci-hadoop.apache.org/job/PreCommit-HDFS-Build/478/console |
| versions | git=2.25.1 |
| Powered by | Apache Yetus 0.13.0-SNAPSHOT https://yetus.apache.org |


This message was automatically generated.



> Add command to DFSAdmin for Balancer max concurrent  threads
> 
>
> Key: HDFS-15793
> URL: https://issues.apache.org/jira/browse/HDFS-15793
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: balancer  mover
>Reporter: Yang Yun
>Assignee: Yang Yun
>Priority: Minor
> Attachments: HDFS-15793.001.patch, HDFS-15793.002.patch
>
>
> We have DFSAdmin command '-setBalancerBandwidth' to dynamically change the 
> max number of bytes per second of network bandwidth to be used by a datanode 
> during balancing.  Also add '-setBalancerMaxThreads' to dynamically change 
> the balancer maxThread number which may be used concurrently for moving 
> blocks.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-15793) Add command to DFSAdmin for Balancer max concurrent threads

2021-02-17 Thread Yang Yun (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15793?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yang Yun updated HDFS-15793:

Attachment: HDFS-15793.002.patch
Status: Patch Available  (was: Open)

> Add command to DFSAdmin for Balancer max concurrent  threads
> 
>
> Key: HDFS-15793
> URL: https://issues.apache.org/jira/browse/HDFS-15793
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: balancer  mover
>Reporter: Yang Yun
>Assignee: Yang Yun
>Priority: Minor
> Attachments: HDFS-15793.001.patch, HDFS-15793.002.patch
>
>
> We have DFSAdmin command '-setBalancerBandwidth' to dynamically change the 
> max number of bytes per second of network bandwidth to be used by a datanode 
> during balancing.  Also add '-setBalancerMaxThreads' to dynamically change 
> the balancer maxThread number which may be used concurrently for moving 
> blocks.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-15793) Add command to DFSAdmin for Balancer max concurrent threads

2021-02-17 Thread Yang Yun (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15793?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yang Yun updated HDFS-15793:

Status: Open  (was: Patch Available)

> Add command to DFSAdmin for Balancer max concurrent  threads
> 
>
> Key: HDFS-15793
> URL: https://issues.apache.org/jira/browse/HDFS-15793
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: balancer  mover
>Reporter: Yang Yun
>Assignee: Yang Yun
>Priority: Minor
> Attachments: HDFS-15793.001.patch
>
>
> We have DFSAdmin command '-setBalancerBandwidth' to dynamically change the 
> max number of bytes per second of network bandwidth to be used by a datanode 
> during balancing.  Also add '-setBalancerMaxThreads' to dynamically change 
> the balancer maxThread number which may be used concurrently for moving 
> blocks.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Work logged] (HDFS-15785) Datanode to support using DNS to resolve nameservices to IP addresses to get list of namenodes

2021-02-17 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15785?focusedWorklogId=554070=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-554070
 ]

ASF GitHub Bot logged work on HDFS-15785:
-

Author: ASF GitHub Bot
Created on: 18/Feb/21 05:14
Start Date: 18/Feb/21 05:14
Worklog Time Spent: 10m 
  Work Description: hadoop-yetus commented on pull request #2639:
URL: https://github.com/apache/hadoop/pull/2639#issuecomment-781060345


   :broken_heart: **-1 overall**
   
   
   
   
   
   
   | Vote | Subsystem | Runtime |  Logfile | Comment |
   |::|--:|:|::|:---:|
   | +0 :ok: |  reexec  |   0m 37s |  |  Docker mode activated.  |
    _ Prechecks _ |
   | +1 :green_heart: |  dupname  |   0m  1s |  |  No case conflicting files 
found.  |
   | +1 :green_heart: |  @author  |   0m  0s |  |  The patch does not contain 
any @author tags.  |
   | +1 :green_heart: |   |   0m  0s | [test4tests](test4tests) |  The patch 
appears to include 1 new or modified test files.  |
    _ trunk Compile Tests _ |
   | +0 :ok: |  mvndep  |  13m 56s |  |  Maven dependency ordering for branch  |
   | +1 :green_heart: |  mvninstall  |  22m 14s |  |  trunk passed  |
   | +1 :green_heart: |  compile  |  22m 44s |  |  trunk passed with JDK 
Ubuntu-11.0.10+9-Ubuntu-0ubuntu1.20.04  |
   | +1 :green_heart: |  compile  |  19m 39s |  |  trunk passed with JDK 
Private Build-1.8.0_282-8u282-b08-0ubuntu1~20.04-b08  |
   | +1 :green_heart: |  checkstyle  |   3m 59s |  |  trunk passed  |
   | +1 :green_heart: |  mvnsite  |   4m 13s |  |  trunk passed  |
   | +1 :green_heart: |  shadedclient  |  22m  1s |  |  branch has no errors 
when building and testing our client artifacts.  |
   | +1 :green_heart: |  javadoc  |   3m 12s |  |  trunk passed with JDK 
Ubuntu-11.0.10+9-Ubuntu-0ubuntu1.20.04  |
   | +1 :green_heart: |  javadoc  |   4m 11s |  |  trunk passed with JDK 
Private Build-1.8.0_282-8u282-b08-0ubuntu1~20.04-b08  |
   | +0 :ok: |  spotbugs  |  37m 32s |  |  Both FindBugs and SpotBugs are 
enabled, using SpotBugs.  |
   | +1 :green_heart: |  spotbugs  |   8m  9s |  |  trunk passed  |
    _ Patch Compile Tests _ |
   | +0 :ok: |  mvndep  |   0m 51s |  |  Maven dependency ordering for patch  |
   | +1 :green_heart: |  mvninstall  |   2m 50s |  |  the patch passed  |
   | +1 :green_heart: |  compile  |  19m 58s |  |  the patch passed with JDK 
Ubuntu-11.0.10+9-Ubuntu-0ubuntu1.20.04  |
   | +1 :green_heart: |  javac  |  19m 58s |  |  
root-jdkUbuntu-11.0.10+9-Ubuntu-0ubuntu1.20.04 with JDK 
Ubuntu-11.0.10+9-Ubuntu-0ubuntu1.20.04 generated 0 new + 1957 unchanged - 2 
fixed = 1957 total (was 1959)  |
   | +1 :green_heart: |  compile  |  17m 56s |  |  the patch passed with JDK 
Private Build-1.8.0_282-8u282-b08-0ubuntu1~20.04-b08  |
   | +1 :green_heart: |  javac  |  17m 56s |  |  
root-jdkPrivateBuild-1.8.0_282-8u282-b08-0ubuntu1~20.04-b08 with JDK Private 
Build-1.8.0_282-8u282-b08-0ubuntu1~20.04-b08 generated 0 new + 1852 unchanged - 
2 fixed = 1852 total (was 1854)  |
   | +1 :green_heart: |  checkstyle  |   3m 56s |  |  root: The patch generated 
0 new + 557 unchanged - 2 fixed = 557 total (was 559)  |
   | +1 :green_heart: |  mvnsite  |   4m 11s |  |  the patch passed  |
   | +1 :green_heart: |  whitespace  |   0m  0s |  |  The patch has no 
whitespace issues.  |
   | +1 :green_heart: |  xml  |   0m  1s |  |  The patch has no ill-formed XML 
file.  |
   | +1 :green_heart: |  shadedclient  |  13m 19s |  |  patch has no errors 
when building and testing our client artifacts.  |
   | +1 :green_heart: |  javadoc  |   3m  1s |  |  the patch passed with JDK 
Ubuntu-11.0.10+9-Ubuntu-0ubuntu1.20.04  |
   | +1 :green_heart: |  javadoc  |   4m  5s |  |  the patch passed with JDK 
Private Build-1.8.0_282-8u282-b08-0ubuntu1~20.04-b08  |
   | -1 :x: |  spotbugs  |   2m 21s | 
[/patch-spotbugs-hadoop-common-project_hadoop-common.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-2639/5/artifact/out/patch-spotbugs-hadoop-common-project_hadoop-common.txt)
 |  hadoop-common-project/hadoop-common cannot run computeBugHistory from 
spotbugs  |
   | -1 :x: |  spotbugs  |   2m 37s | 
[/patch-spotbugs-hadoop-hdfs-project_hadoop-hdfs-client.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-2639/5/artifact/out/patch-spotbugs-hadoop-hdfs-project_hadoop-hdfs-client.txt)
 |  hadoop-hdfs-project/hadoop-hdfs-client cannot run computeBugHistory from 
spotbugs  |
   | -1 :x: |  spotbugs  |   3m 17s | 
[/patch-spotbugs-hadoop-hdfs-project_hadoop-hdfs.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-2639/5/artifact/out/patch-spotbugs-hadoop-hdfs-project_hadoop-hdfs.txt)
 |  hadoop-hdfs-project/hadoop-hdfs cannot run computeBugHistory from spotbugs  
|
    _ Other Tests _ |
   | +1 :green_heart: |  unit  |  17m 13s |  |  hadoop-common in the patch 
passed.  |
   | +1 :green_heart: 

[jira] [Work logged] (HDFS-15808) Add metrics for FSNamesystem read/write lock hold long time

2021-02-17 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15808?focusedWorklogId=554034=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-554034
 ]

ASF GitHub Bot logged work on HDFS-15808:
-

Author: ASF GitHub Bot
Created on: 18/Feb/21 02:20
Start Date: 18/Feb/21 02:20
Worklog Time Spent: 10m 
  Work Description: tomscut edited a comment on pull request #2668:
URL: https://github.com/apache/hadoop/pull/2668#issuecomment-780991248


   I found this JIRA in the log. 
https://issues.apache.org/jira/browse/HADOOP-16870
   Hi @aajisaka , Could you please help to take a look at this problem? And how 
I should deal with it? Thank you.
   
   `[2021-02-17T01:59:03.362Z] HEAD is now at 4cf35315838 HADOOP-16870. Use 
spotbugs-maven-plugin instead of findbugs-maven-plugin (#2454)
   
   [2021-02-17T01:59:03.887Z] Already on 'trunk'
   
   [2021-02-17T01:59:03.887Z] Your branch is up to date with 'origin/trunk'.
   
   [2021-02-17T01:59:04.412Z] Already up to date.
   
   [2021-02-17T01:59:04.412Z] Current branch trunk is up to date.
   
   [2021-02-17T01:59:04.412Z] Already on 'trunk'
   
   [2021-02-17T01:59:04.412Z] Your branch is up to date with 'origin/trunk'.
   
   [2021-02-17T01:59:05.444Z] HEAD is now at 4cf35315838 HADOOP-16870. Use 
spotbugs-maven-plugin instead of findbugs-maven-plugin (#2454)
   
   [2021-02-17T01:59:05.444Z] 
   
   [2021-02-17T01:59:05.444Z] Testing 
https://github.com/apache/hadoop/pull/2668 diff on trunk.
   
   [2021-02-17T01:59:05.444Z] 
   
   [2021-02-17T01:59:05.444Z] 
   
   [2021-02-17T01:59:05.444Z] 

   
   [2021-02-17T01:59:05.444Z] 

   
   [2021-02-17T01:59:05.444Z] Re-exec mode detected. 
Continuing.
   
   [2021-02-17T01:59:05.444Z] 

   
   [2021-02-17T01:59:05.444Z] 

   
   [2021-02-17T01:59:05.444Z] 
   
   [2021-02-17T01:59:05.444Z] 
   
   [2021-02-17T01:59:05.444Z] ERROR: Unprocessed flag(s): 
--findbugs-strict-precheck`



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 554034)
Time Spent: 4h  (was: 3h 50m)

> Add metrics for FSNamesystem read/write lock hold long time
> ---
>
> Key: HDFS-15808
> URL: https://issues.apache.org/jira/browse/HDFS-15808
> Project: Hadoop HDFS
>  Issue Type: Wish
>  Components: hdfs
>Reporter: tomscut
>Assignee: tomscut
>Priority: Major
>  Labels: hdfs, lock, metrics, pull-request-available
>  Time Spent: 4h
>  Remaining Estimate: 0h
>
> To monitor how often read/write locks exceed thresholds, we can add two 
> metrics(ReadLockWarning/WriteLockWarning), which are exposed in JMX.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Work logged] (HDFS-15808) Add metrics for FSNamesystem read/write lock hold long time

2021-02-17 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15808?focusedWorklogId=554033=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-554033
 ]

ASF GitHub Bot logged work on HDFS-15808:
-

Author: ASF GitHub Bot
Created on: 18/Feb/21 02:20
Start Date: 18/Feb/21 02:20
Worklog Time Spent: 10m 
  Work Description: tomscut edited a comment on pull request #2668:
URL: https://github.com/apache/hadoop/pull/2668#issuecomment-780991248


   I found this JIRA in the log. 
https://issues.apache.org/jira/browse/HADOOP-16870
   Hi @aajisaka , Could you please help to take a look at this problem? And how 
I should deal with this problem? Thank you.
   
   `[2021-02-17T01:59:03.362Z] HEAD is now at 4cf35315838 HADOOP-16870. Use 
spotbugs-maven-plugin instead of findbugs-maven-plugin (#2454)
   
   [2021-02-17T01:59:03.887Z] Already on 'trunk'
   
   [2021-02-17T01:59:03.887Z] Your branch is up to date with 'origin/trunk'.
   
   [2021-02-17T01:59:04.412Z] Already up to date.
   
   [2021-02-17T01:59:04.412Z] Current branch trunk is up to date.
   
   [2021-02-17T01:59:04.412Z] Already on 'trunk'
   
   [2021-02-17T01:59:04.412Z] Your branch is up to date with 'origin/trunk'.
   
   [2021-02-17T01:59:05.444Z] HEAD is now at 4cf35315838 HADOOP-16870. Use 
spotbugs-maven-plugin instead of findbugs-maven-plugin (#2454)
   
   [2021-02-17T01:59:05.444Z] 
   
   [2021-02-17T01:59:05.444Z] Testing 
https://github.com/apache/hadoop/pull/2668 diff on trunk.
   
   [2021-02-17T01:59:05.444Z] 
   
   [2021-02-17T01:59:05.444Z] 
   
   [2021-02-17T01:59:05.444Z] 

   
   [2021-02-17T01:59:05.444Z] 

   
   [2021-02-17T01:59:05.444Z] Re-exec mode detected. 
Continuing.
   
   [2021-02-17T01:59:05.444Z] 

   
   [2021-02-17T01:59:05.444Z] 

   
   [2021-02-17T01:59:05.444Z] 
   
   [2021-02-17T01:59:05.444Z] 
   
   [2021-02-17T01:59:05.444Z] ERROR: Unprocessed flag(s): 
--findbugs-strict-precheck`



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 554033)
Time Spent: 3h 50m  (was: 3h 40m)

> Add metrics for FSNamesystem read/write lock hold long time
> ---
>
> Key: HDFS-15808
> URL: https://issues.apache.org/jira/browse/HDFS-15808
> Project: Hadoop HDFS
>  Issue Type: Wish
>  Components: hdfs
>Reporter: tomscut
>Assignee: tomscut
>Priority: Major
>  Labels: hdfs, lock, metrics, pull-request-available
>  Time Spent: 3h 50m
>  Remaining Estimate: 0h
>
> To monitor how often read/write locks exceed thresholds, we can add two 
> metrics(ReadLockWarning/WriteLockWarning), which are exposed in JMX.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Work logged] (HDFS-15808) Add metrics for FSNamesystem read/write lock hold long time

2021-02-17 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15808?focusedWorklogId=554032=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-554032
 ]

ASF GitHub Bot logged work on HDFS-15808:
-

Author: ASF GitHub Bot
Created on: 18/Feb/21 02:18
Start Date: 18/Feb/21 02:18
Worklog Time Spent: 10m 
  Work Description: tomscut edited a comment on pull request #2668:
URL: https://github.com/apache/hadoop/pull/2668#issuecomment-780991248


   I found this JIRA in the log. 
https://issues.apache.org/jira/browse/HADOOP-16870
   Hi @aajisaka , Could you please help to take a look at this problem? And how 
I should deal with this problem? Thank you.
   
   `[2021-02-17T01:59:03.362Z] HEAD is now at 4cf35315838 HADOOP-16870. Use 
spotbugs-maven-plugin instead of findbugs-maven-plugin (#2454)
   [2021-02-17T01:59:03.887Z] Already on 'trunk'
   [2021-02-17T01:59:03.887Z] Your branch is up to date with 'origin/trunk'.
   [2021-02-17T01:59:04.412Z] Already up to date.
   [2021-02-17T01:59:04.412Z] Current branch trunk is up to date.
   [2021-02-17T01:59:04.412Z] Already on 'trunk'
   [2021-02-17T01:59:04.412Z] Your branch is up to date with 'origin/trunk'.
   [2021-02-17T01:59:05.444Z] HEAD is now at 4cf35315838 HADOOP-16870. Use 
spotbugs-maven-plugin instead of findbugs-maven-plugin (#2454)
   [2021-02-17T01:59:05.444Z] 
   [2021-02-17T01:59:05.444Z] Testing 
https://github.com/apache/hadoop/pull/2668 diff on trunk.
   [2021-02-17T01:59:05.444Z] 
   [2021-02-17T01:59:05.444Z] 
   [2021-02-17T01:59:05.444Z] 

   [2021-02-17T01:59:05.444Z] 

   [2021-02-17T01:59:05.444Z] Re-exec mode detected. 
Continuing.
   [2021-02-17T01:59:05.444Z] 

   [2021-02-17T01:59:05.444Z] 

   [2021-02-17T01:59:05.444Z] 
   [2021-02-17T01:59:05.444Z] 
   [2021-02-17T01:59:05.444Z] ERROR: Unprocessed flag(s): 
--findbugs-strict-precheck`



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 554032)
Time Spent: 3h 40m  (was: 3.5h)

> Add metrics for FSNamesystem read/write lock hold long time
> ---
>
> Key: HDFS-15808
> URL: https://issues.apache.org/jira/browse/HDFS-15808
> Project: Hadoop HDFS
>  Issue Type: Wish
>  Components: hdfs
>Reporter: tomscut
>Assignee: tomscut
>Priority: Major
>  Labels: hdfs, lock, metrics, pull-request-available
>  Time Spent: 3h 40m
>  Remaining Estimate: 0h
>
> To monitor how often read/write locks exceed thresholds, we can add two 
> metrics(ReadLockWarning/WriteLockWarning), which are exposed in JMX.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Work logged] (HDFS-15808) Add metrics for FSNamesystem read/write lock hold long time

2021-02-17 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15808?focusedWorklogId=554031=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-554031
 ]

ASF GitHub Bot logged work on HDFS-15808:
-

Author: ASF GitHub Bot
Created on: 18/Feb/21 02:13
Start Date: 18/Feb/21 02:13
Worklog Time Spent: 10m 
  Work Description: tomscut edited a comment on pull request #2668:
URL: https://github.com/apache/hadoop/pull/2668#issuecomment-780991248


   I found this JIRA in the log. 
https://issues.apache.org/jira/browse/HADOOP-16870
   Hi @aajisaka , Could you please help to take a look at this problem? And how 
I should deal with this problem? Thank you.
   
   `[2021-02-17T01:59:03.362Z] HEAD is now at 4cf35315838 HADOOP-16870. Use 
spotbugs-maven-plugin instead of findbugs-maven-plugin (#2454)
   
   [2021-02-17T01:59:03.887Z] Already on 'trunk'
   
   [2021-02-17T01:59:03.887Z] Your branch is up to date with 'origin/trunk'.
   
   [2021-02-17T01:59:04.412Z] Already up to date.
   
   [2021-02-17T01:59:04.412Z] Current branch trunk is up to date.
   
   [2021-02-17T01:59:04.412Z] Already on 'trunk'
   
   [2021-02-17T01:59:04.412Z] Your branch is up to date with 'origin/trunk'.
   
   [2021-02-17T01:59:05.444Z] HEAD is now at 4cf35315838 HADOOP-16870. Use 
spotbugs-maven-plugin instead of findbugs-maven-plugin (#2454)
   
   [2021-02-17T01:59:05.444Z] 
   
   [2021-02-17T01:59:05.444Z] Testing 
https://github.com/apache/hadoop/pull/2668 diff on trunk.
   
   [2021-02-17T01:59:05.444Z] 
   
   [2021-02-17T01:59:05.444Z] 
   
   [2021-02-17T01:59:05.444Z] 

   
   [2021-02-17T01:59:05.444Z] 

   
   [2021-02-17T01:59:05.444Z] Re-exec mode detected. 
Continuing.
   
   [2021-02-17T01:59:05.444Z] 

   
   [2021-02-17T01:59:05.444Z] 

   
   [2021-02-17T01:59:05.444Z] 
   
   [2021-02-17T01:59:05.444Z] 
   
   [2021-02-17T01:59:05.444Z] ERROR: Unprocessed flag(s): 
--findbugs-strict-precheck`



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 554031)
Time Spent: 3.5h  (was: 3h 20m)

> Add metrics for FSNamesystem read/write lock hold long time
> ---
>
> Key: HDFS-15808
> URL: https://issues.apache.org/jira/browse/HDFS-15808
> Project: Hadoop HDFS
>  Issue Type: Wish
>  Components: hdfs
>Reporter: tomscut
>Assignee: tomscut
>Priority: Major
>  Labels: hdfs, lock, metrics, pull-request-available
>  Time Spent: 3.5h
>  Remaining Estimate: 0h
>
> To monitor how often read/write locks exceed thresholds, we can add two 
> metrics(ReadLockWarning/WriteLockWarning), which are exposed in JMX.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Work logged] (HDFS-15808) Add metrics for FSNamesystem read/write lock hold long time

2021-02-17 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15808?focusedWorklogId=554030=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-554030
 ]

ASF GitHub Bot logged work on HDFS-15808:
-

Author: ASF GitHub Bot
Created on: 18/Feb/21 02:12
Start Date: 18/Feb/21 02:12
Worklog Time Spent: 10m 
  Work Description: tomscut commented on pull request #2668:
URL: https://github.com/apache/hadoop/pull/2668#issuecomment-780991248


   I found this JIRA in the log. 
https://issues.apache.org/jira/browse/HADOOP-16870
   Hi @aajisaka , Could you please help to take a look at this problem? And how 
I should deal with this problem? Thank you.
   
   `[2021-02-17T01:59:03.362Z] HEAD is now at 4cf35315838 HADOOP-16870. Use 
spotbugs-maven-plugin instead of findbugs-maven-plugin (#2454)
   [2021-02-17T01:59:03.887Z] Already on 'trunk'
   [2021-02-17T01:59:03.887Z] Your branch is up to date with 'origin/trunk'.
   [2021-02-17T01:59:04.412Z] Already up to date.
   [2021-02-17T01:59:04.412Z] Current branch trunk is up to date.
   [2021-02-17T01:59:04.412Z] Already on 'trunk'
   [2021-02-17T01:59:04.412Z] Your branch is up to date with 'origin/trunk'.
   [2021-02-17T01:59:05.444Z] HEAD is now at 4cf35315838 HADOOP-16870. Use 
spotbugs-maven-plugin instead of findbugs-maven-plugin (#2454)
   [2021-02-17T01:59:05.444Z] 
   [2021-02-17T01:59:05.444Z] Testing 
https://github.com/apache/hadoop/pull/2668 diff on trunk.
   [2021-02-17T01:59:05.444Z] 
   [2021-02-17T01:59:05.444Z] 
   [2021-02-17T01:59:05.444Z] 

   [2021-02-17T01:59:05.444Z] 

   [2021-02-17T01:59:05.444Z] Re-exec mode detected. 
Continuing.
   [2021-02-17T01:59:05.444Z] 

   [2021-02-17T01:59:05.444Z] 

   [2021-02-17T01:59:05.444Z] 
   [2021-02-17T01:59:05.444Z] 
   [2021-02-17T01:59:05.444Z] ERROR: Unprocessed flag(s): 
--findbugs-strict-precheck`



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 554030)
Time Spent: 3h 20m  (was: 3h 10m)

> Add metrics for FSNamesystem read/write lock hold long time
> ---
>
> Key: HDFS-15808
> URL: https://issues.apache.org/jira/browse/HDFS-15808
> Project: Hadoop HDFS
>  Issue Type: Wish
>  Components: hdfs
>Reporter: tomscut
>Assignee: tomscut
>Priority: Major
>  Labels: hdfs, lock, metrics, pull-request-available
>  Time Spent: 3h 20m
>  Remaining Estimate: 0h
>
> To monitor how often read/write locks exceed thresholds, we can add two 
> metrics(ReadLockWarning/WriteLockWarning), which are exposed in JMX.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-15816) If NO stale node in last choosing, the chooseTarget don't need to retry with stale nodes.

2021-02-17 Thread Hadoop QA (Jira)


[ 
https://issues.apache.org/jira/browse/HDFS-15816?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17286197#comment-17286197
 ] 

Hadoop QA commented on HDFS-15816:
--

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime ||  Logfile || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
34s{color} | {color:blue}{color} | {color:blue} Docker mode activated. {color} |
| {color:red}-1{color} | {color:red} yetus {color} | {color:red}  0m  8s{color} 
| {color:red}{color} | {color:red} Unprocessed flag(s): 
--findbugs-strict-precheck {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Docker | ClientAPI=1.41 ServerAPI=1.41 base: 
https://ci-hadoop.apache.org/job/PreCommit-HDFS-Build/477/artifact/out/Dockerfile
 |
| JIRA Issue | HDFS-15816 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/13020592/HDFS-15816.002.patch |
| Console output | 
https://ci-hadoop.apache.org/job/PreCommit-HDFS-Build/477/console |
| versions | git=2.25.1 |
| Powered by | Apache Yetus 0.13.0-SNAPSHOT https://yetus.apache.org |


This message was automatically generated.



> If NO stale node in last choosing, the chooseTarget don't need to retry with 
> stale nodes.
> -
>
> Key: HDFS-15816
> URL: https://issues.apache.org/jira/browse/HDFS-15816
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: block placement
>Reporter: Yang Yun
>Assignee: Yang Yun
>Priority: Minor
> Attachments: HDFS-15816.001.patch, HDFS-15816.002.patch
>
>
> If NO stale node in last choosing, the chooseTarget don't need to retry with 
> stale nodes.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-15816) If NO stale node in last choosing, the chooseTarget don't need to retry with stale nodes.

2021-02-17 Thread Yang Yun (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15816?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yang Yun updated HDFS-15816:

Attachment: (was: HDFS-15816.002.patch)

> If NO stale node in last choosing, the chooseTarget don't need to retry with 
> stale nodes.
> -
>
> Key: HDFS-15816
> URL: https://issues.apache.org/jira/browse/HDFS-15816
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: block placement
>Reporter: Yang Yun
>Assignee: Yang Yun
>Priority: Minor
> Attachments: HDFS-15816.001.patch, HDFS-15816.002.patch
>
>
> If NO stale node in last choosing, the chooseTarget don't need to retry with 
> stale nodes.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-15816) If NO stale node in last choosing, the chooseTarget don't need to retry with stale nodes.

2021-02-17 Thread Yang Yun (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15816?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yang Yun updated HDFS-15816:

Attachment: HDFS-15816.002.patch
Status: Patch Available  (was: Open)

> If NO stale node in last choosing, the chooseTarget don't need to retry with 
> stale nodes.
> -
>
> Key: HDFS-15816
> URL: https://issues.apache.org/jira/browse/HDFS-15816
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: block placement
>Reporter: Yang Yun
>Assignee: Yang Yun
>Priority: Minor
> Attachments: HDFS-15816.001.patch, HDFS-15816.002.patch
>
>
> If NO stale node in last choosing, the chooseTarget don't need to retry with 
> stale nodes.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-15816) If NO stale node in last choosing, the chooseTarget don't need to retry with stale nodes.

2021-02-17 Thread Yang Yun (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15816?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yang Yun updated HDFS-15816:

Status: Open  (was: Patch Available)

> If NO stale node in last choosing, the chooseTarget don't need to retry with 
> stale nodes.
> -
>
> Key: HDFS-15816
> URL: https://issues.apache.org/jira/browse/HDFS-15816
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: block placement
>Reporter: Yang Yun
>Assignee: Yang Yun
>Priority: Minor
> Attachments: HDFS-15816.001.patch, HDFS-15816.002.patch
>
>
> If NO stale node in last choosing, the chooseTarget don't need to retry with 
> stale nodes.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-15808) Add metrics for FSNamesystem read/write lock hold long time

2021-02-17 Thread Konstantin Shvachko (Jira)


[ 
https://issues.apache.org/jira/browse/HDFS-15808?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17286133#comment-17286133
 ] 

Konstantin Shvachko commented on HDFS-15808:


Hey [~tomscut]. The patch looks fine, but I doubt the metric will be useful in 
its current form. Monotonically increasing counter doesn't tell you much when 
plotted. Over time it just becomes an incredibly large number, hard to see its 
fluctuations. And you cannot set alerts if the threshold is exceeded often.
See e.g. {{ExpiredHeartbeats}} or {{LastWrittenTransactionId}} - not useful.
I assume you need something like a rate.

> Add metrics for FSNamesystem read/write lock hold long time
> ---
>
> Key: HDFS-15808
> URL: https://issues.apache.org/jira/browse/HDFS-15808
> Project: Hadoop HDFS
>  Issue Type: Wish
>  Components: hdfs
>Reporter: tomscut
>Assignee: tomscut
>Priority: Major
>  Labels: hdfs, lock, metrics, pull-request-available
>  Time Spent: 3h 10m
>  Remaining Estimate: 0h
>
> To monitor how often read/write locks exceed thresholds, we can add two 
> metrics(ReadLockWarning/WriteLockWarning), which are exposed in JMX.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-15815) if required storageType are unavailable, log the failed reason during choosing Datanode

2021-02-17 Thread Ayush Saxena (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15815?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ayush Saxena updated HDFS-15815:

Fix Version/s: 3.4.0
 Hadoop Flags: Reviewed
   Resolution: Fixed
   Status: Resolved  (was: Patch Available)

>  if required storageType are unavailable, log the failed reason during 
> choosing Datanode
> 
>
> Key: HDFS-15815
> URL: https://issues.apache.org/jira/browse/HDFS-15815
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: block placement
>Reporter: Yang Yun
>Assignee: Yang Yun
>Priority: Minor
> Fix For: 3.4.0
>
> Attachments: HDFS-15815.001.patch, HDFS-15815.002.patch, 
> HDFS-15815.003.patch
>
>
> For better debug,  if required storageType are unavailable, log the failed 
> reason "NO_REQUIRED_STORAGE_TYPE" when choosing Datanode.
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-15815) if required storageType are unavailable, log the failed reason during choosing Datanode

2021-02-17 Thread Ayush Saxena (Jira)


[ 
https://issues.apache.org/jira/browse/HDFS-15815?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17286130#comment-17286130
 ] 

Ayush Saxena commented on HDFS-15815:
-

Committed to trunk.

Thanx [~hadoop_yangyun] for the contribution!!!

>  if required storageType are unavailable, log the failed reason during 
> choosing Datanode
> 
>
> Key: HDFS-15815
> URL: https://issues.apache.org/jira/browse/HDFS-15815
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: block placement
>Reporter: Yang Yun
>Assignee: Yang Yun
>Priority: Minor
> Attachments: HDFS-15815.001.patch, HDFS-15815.002.patch, 
> HDFS-15815.003.patch
>
>
> For better debug,  if required storageType are unavailable, log the failed 
> reason "NO_REQUIRED_STORAGE_TYPE" when choosing Datanode.
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-15837) Incorrect bytes causing block corruption after update pipeline and recovery failure

2021-02-17 Thread Kihwal Lee (Jira)


[ 
https://issues.apache.org/jira/browse/HDFS-15837?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17286111#comment-17286111
 ] 

Kihwal Lee commented on HDFS-15837:
---

The successful pipeline update with 3072 bytes was at 03:42:40.  The 
replication of a 3544 byte block and corruption detection was 4 minutes later. 
The client must have wrote more after the recovery and closed the file. 
Replication happens only after closing the file.  It looks like the file was 
closed at around 03:46:40.

I don't think the size difference observed on datanodes are anomaly. Clients 
usually write more data after a pipeline recovery. Is the file still existing? 
What's the size you see when you do "hadoop fs -ls /pat_to_the_file"?  Is the 
block appearing as missing on the UI? 

If there was a corruption issue, I don't think it was because of the size.  The 
fact that the replication happened means the NN was agreeing with the size 
3544. If the datanode had unexpected size of the replica, the transfer wouldn't 
have happened.  The corruption report came from the destination node for the 
replication, 172.21.234.181.  The replication source does not perform any 
checks and there is no feedback mechanism for reporting error back to the 
source node. So even if a corruption was detected during the replication, only 
the target (172.21.234.181) would notice it.  Check the log of 172.21.234.181 
at around 03:46:43. It must have detected a corruption and logged it.

The replica on the source (172.21.226.26) must be the only copy and if it is 
corrupt, it will show up as missing. If the file still exists, you can get to 
the node and run "hdfs debug verifyMeta" command against the block file and the 
meta file to manually check the integrity.

If it is really corrupt, it could be due to 172.21.226.26's local issue.  
Extreme load can cause write failures and pipeline updates, but we haven't seen 
such condition causing data corruption.

> Incorrect bytes causing block corruption after update pipeline and recovery 
> failure
> ---
>
> Key: HDFS-15837
> URL: https://issues.apache.org/jira/browse/HDFS-15837
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: datanode, hdfs
>Affects Versions: 2.8.5
>Reporter: Udit Mehrotra
>Priority: Major
>
> We are seeing cases on HDFS blocks being marked as *bad* after the initial 
> block receive fails during *update pipeline* followed by *HDFS* *recovery* 
> for the block failing as well. Here is the life cycle of a block 
> *{{blk_1342440165_272630578}}* that was ultimately marked as corrupt:
> 1. The block creation starts at name node as part of *update pipeline* 
> process:
> {noformat}
> 2021-01-25 03:41:17,335 INFO 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem (IPC Server handler 61 on 
> 8020): updatePipeline(blk_1342440165_272500939 => blk_1342440165_272630578) 
> success{noformat}
> 2. The block receiver on the data node fails with a socket timeout exception, 
> and so do the retries:
> {noformat}
> 2021-01-25 03:42:22,525 INFO org.apache.hadoop.hdfs.server.datanode.DataNode 
> (PacketResponder: 
> BP-908477295-172.21.224.178-1606768078949:blk_1342440165_272630578, 
> type=HAS_DOWNSTREAM_IN_PIPELINE, downstreams=1:[172.21.246.239:50010]): 
> PacketResponder: 
> BP-908477295-172.21.224.178-1606768078949:blk_1342440165_272630578, 
> type=HAS_DOWNSTREAM_IN_PIPELINE, downstreams=1:[172.21.246.239:50010]
> java.net.SocketTimeoutException: 65000 millis timeout while waiting for 
> channel to be ready for read. ch : java.nio.channels.SocketChannel[connected 
> local=/172.21.226.26:56294 remote=/172.21.246.239:50010]
> at 
> org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:164)
> at 
> org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161)
> at 
> org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131)
> at 
> org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:118)
> at java.io.FilterInputStream.read(FilterInputStream.java:83)
> at java.io.FilterInputStream.read(FilterInputStream.java:83)
> at 
> org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:400)
> at 
> org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:213)
> at 
> org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.run(BlockReceiver.java:1305)
> at java.lang.Thread.run(Thread.java:748)
> 2021-01-25 03:42:22,526 WARN org.apache.hadoop.hdfs.server.datanode.DataNode 
> (PacketResponder: 
> BP-908477295-172.21.224.178-1606768078949:blk_1342440165_272630578, 
> type=HAS_DOWNSTREAM_IN_PIPELINE, downstreams=1:[172.21.246.239:50010]): 
> IOException in BlockReceiver.run(): 
> java.io.IOException: Connection 

[jira] [Work logged] (HDFS-15808) Add metrics for FSNamesystem read/write lock hold long time

2021-02-17 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15808?focusedWorklogId=553632=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-553632
 ]

ASF GitHub Bot logged work on HDFS-15808:
-

Author: ASF GitHub Bot
Created on: 17/Feb/21 14:31
Start Date: 17/Feb/21 14:31
Worklog Time Spent: 10m 
  Work Description: tomscut commented on pull request #2668:
URL: https://github.com/apache/hadoop/pull/2668#issuecomment-780595016


   > This message was automatically generated.
   
   Hi @anuengineer, the failed check looks unrelated to the change. May I ask 
how I should deal with it?
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 553632)
Time Spent: 3h 10m  (was: 3h)

> Add metrics for FSNamesystem read/write lock hold long time
> ---
>
> Key: HDFS-15808
> URL: https://issues.apache.org/jira/browse/HDFS-15808
> Project: Hadoop HDFS
>  Issue Type: Wish
>  Components: hdfs
>Reporter: tomscut
>Assignee: tomscut
>Priority: Major
>  Labels: hdfs, lock, metrics, pull-request-available
>  Time Spent: 3h 10m
>  Remaining Estimate: 0h
>
> To monitor how often read/write locks exceed thresholds, we can add two 
> metrics(ReadLockWarning/WriteLockWarning), which are exposed in JMX.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-15838) Yarn Job execution get failed when LZ4 Compression Codec is used

2021-02-17 Thread Bhavik Patel (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15838?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bhavik Patel updated HDFS-15838:

Attachment: LZ4.png

> Yarn Job execution get failed when LZ4 Compression Codec is used
> 
>
> Key: HDFS-15838
> URL: https://issues.apache.org/jira/browse/HDFS-15838
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs
>Reporter: Bhavik Patel
>Priority: Major
> Attachments: HDFS-15838.001.patch, LZ4.png
>
>
> When we try to compress a file using the LZ4 codec compression type then the 
> yarn job gets failed with the error message :
> {code:java}
> net.jpountz.lz4.LZ4Compressorcompres(Ljava/nio/ByteBuffer;Ljava/nio/ByteBuffer;)V
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-15838) Yarn Job execution get failed when LZ4 Compression Codec is used

2021-02-17 Thread Bhavik Patel (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15838?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bhavik Patel updated HDFS-15838:

Attachment: HDFS-15838.001.patch

> Yarn Job execution get failed when LZ4 Compression Codec is used
> 
>
> Key: HDFS-15838
> URL: https://issues.apache.org/jira/browse/HDFS-15838
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs
>Reporter: Bhavik Patel
>Priority: Major
> Attachments: HDFS-15838.001.patch
>
>
> When we try to compress a file using the LZ4 codec compression type then the 
> yarn job gets failed with the error message :
> {code:java}
> net.jpountz.lz4.LZ4Compressorcompres(Ljava/nio/ByteBuffer;Ljava/nio/ByteBuffer;)V
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Created] (HDFS-15838) Yarn Job execution get failed when LZ4 Compression Codec is used

2021-02-17 Thread Bhavik Patel (Jira)
Bhavik Patel created HDFS-15838:
---

 Summary: Yarn Job execution get failed when LZ4 Compression Codec 
is used
 Key: HDFS-15838
 URL: https://issues.apache.org/jira/browse/HDFS-15838
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: hdfs
Reporter: Bhavik Patel


When we try to compress a file using the LZ4 codec compression type then the 
yarn job gets failed with the error message :
{code:java}
net.jpountz.lz4.LZ4Compressorcompres(Ljava/nio/ByteBuffer;Ljava/nio/ByteBuffer;)V
{code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Work logged] (HDFS-15836) RBF: Fix contract tests after HADOOP-13327

2021-02-17 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15836?focusedWorklogId=553556=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-553556
 ]

ASF GitHub Bot logged work on HDFS-15836:
-

Author: ASF GitHub Bot
Created on: 17/Feb/21 11:19
Start Date: 17/Feb/21 11:19
Worklog Time Spent: 10m 
  Work Description: steveloughran commented on pull request #2702:
URL: https://github.com/apache/hadoop/pull/2702#issuecomment-780487969


   I'm sorry. I thought I'd tested everything, at least through the IDE, but 
was wrong. The new test really tries to determine what supports hflush/hsync 
and what just pretends to by not raising an exception in hflush() but which 
doesn't actually update the output. Apparently the HBase team care about 
durable storage of their WAL. I've given them the probes, the validation *and 
the ability to use a RAID-5 local FS as the WAL destination*



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 553556)
Time Spent: 1h 20m  (was: 1h 10m)

> RBF: Fix contract tests after HADOOP-13327
> --
>
> Key: HDFS-15836
> URL: https://issues.apache.org/jira/browse/HDFS-15836
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: rbf
>Reporter: Akira Ajisaka
>Assignee: Akira Ajisaka
>Priority: Major
>  Labels: pull-request-available
> Fix For: 3.3.1, 3.4.0
>
>  Time Spent: 1h 20m
>  Remaining Estimate: 0h
>
> {noformat}
> [ERROR] Tests run: 16, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 
> 19.094 s <<< FAILURE! - in 
> org.apache.hadoop.fs.contract.router.TestRouterHDFSContractCreate
> [ERROR] 
> testSyncable(org.apache.hadoop.fs.contract.router.TestRouterHDFSContractCreate)
>   Time elapsed: 0.102 s  <<< FAILURE!
> java.lang.AssertionError: Should not have capability: hflush in 
> FSDataOutputStream{wrappedStream=DFSOutputStream:block==null}
>   at org.junit.Assert.fail(Assert.java:88)
>   at org.junit.Assert.assertTrue(Assert.java:41)
>   at org.junit.Assert.assertFalse(Assert.java:64)
>   at 
> org.apache.hadoop.fs.contract.ContractTestUtils.assertCapabilities(ContractTestUtils.java:1553)
>   at 
> org.apache.hadoop.fs.contract.AbstractContractCreateTest.validateSyncableSemantics(AbstractContractCreateTest.java:497)
>   at 
> org.apache.hadoop.fs.contract.AbstractContractCreateTest.testSyncable(AbstractContractCreateTest.java:459)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:498)
>   at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
>   at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
>   at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
>   at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
>   at 
> org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
>   at 
> org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
>   at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:55)
>   at 
> org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:298)
>   at 
> org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:292)
>   at java.util.concurrent.FutureTask.run(FutureTask.java:266)
>   at java.lang.Thread.run(Thread.java:748)
> {noformat}
> https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-2696/1/artifact/out/patch-unit-hadoop-hdfs-project_hadoop-hdfs-rbf.txt



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-15816) If NO stale node in last choosing, the chooseTarget don't need to retry with stale nodes.

2021-02-17 Thread Hadoop QA (Jira)


[ 
https://issues.apache.org/jira/browse/HDFS-15816?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17285780#comment-17285780
 ] 

Hadoop QA commented on HDFS-15816:
--

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime ||  Logfile || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
33s{color} | {color:blue}{color} | {color:blue} Docker mode activated. {color} |
| {color:red}-1{color} | {color:red} yetus {color} | {color:red}  0m  7s{color} 
| {color:red}{color} | {color:red} Unprocessed flag(s): 
--findbugs-strict-precheck {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Docker | ClientAPI=1.41 ServerAPI=1.41 base: 
https://ci-hadoop.apache.org/job/PreCommit-HDFS-Build/476/artifact/out/Dockerfile
 |
| JIRA Issue | HDFS-15816 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/13020557/HDFS-15816.002.patch |
| Console output | 
https://ci-hadoop.apache.org/job/PreCommit-HDFS-Build/476/console |
| versions | git=2.25.1 |
| Powered by | Apache Yetus 0.13.0-SNAPSHOT https://yetus.apache.org |


This message was automatically generated.



> If NO stale node in last choosing, the chooseTarget don't need to retry with 
> stale nodes.
> -
>
> Key: HDFS-15816
> URL: https://issues.apache.org/jira/browse/HDFS-15816
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: block placement
>Reporter: Yang Yun
>Assignee: Yang Yun
>Priority: Minor
> Attachments: HDFS-15816.001.patch, HDFS-15816.002.patch
>
>
> If NO stale node in last choosing, the chooseTarget don't need to retry with 
> stale nodes.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-15816) If NO stale node in last choosing, the chooseTarget don't need to retry with stale nodes.

2021-02-17 Thread Yang Yun (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15816?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yang Yun updated HDFS-15816:

Attachment: HDFS-15816.002.patch
Status: Patch Available  (was: Open)

> If NO stale node in last choosing, the chooseTarget don't need to retry with 
> stale nodes.
> -
>
> Key: HDFS-15816
> URL: https://issues.apache.org/jira/browse/HDFS-15816
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: block placement
>Reporter: Yang Yun
>Assignee: Yang Yun
>Priority: Minor
> Attachments: HDFS-15816.001.patch, HDFS-15816.002.patch
>
>
> If NO stale node in last choosing, the chooseTarget don't need to retry with 
> stale nodes.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-15816) If NO stale node in last choosing, the chooseTarget don't need to retry with stale nodes.

2021-02-17 Thread Yang Yun (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15816?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yang Yun updated HDFS-15816:

Status: Open  (was: Patch Available)

> If NO stale node in last choosing, the chooseTarget don't need to retry with 
> stale nodes.
> -
>
> Key: HDFS-15816
> URL: https://issues.apache.org/jira/browse/HDFS-15816
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: block placement
>Reporter: Yang Yun
>Assignee: Yang Yun
>Priority: Minor
> Attachments: HDFS-15816.001.patch, HDFS-15816.002.patch
>
>
> If NO stale node in last choosing, the chooseTarget don't need to retry with 
> stale nodes.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-15816) If NO stale node in last choosing, the chooseTarget don't need to retry with stale nodes.

2021-02-17 Thread Yang Yun (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15816?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yang Yun updated HDFS-15816:

Attachment: (was: HDFS-15816.002.patch)

> If NO stale node in last choosing, the chooseTarget don't need to retry with 
> stale nodes.
> -
>
> Key: HDFS-15816
> URL: https://issues.apache.org/jira/browse/HDFS-15816
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: block placement
>Reporter: Yang Yun
>Assignee: Yang Yun
>Priority: Minor
> Attachments: HDFS-15816.001.patch, HDFS-15816.002.patch
>
>
> If NO stale node in last choosing, the chooseTarget don't need to retry with 
> stale nodes.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-15816) If NO stale node in last choosing, the chooseTarget don't need to retry with stale nodes.

2021-02-17 Thread Hadoop QA (Jira)


[ 
https://issues.apache.org/jira/browse/HDFS-15816?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17285760#comment-17285760
 ] 

Hadoop QA commented on HDFS-15816:
--

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime ||  Logfile || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
47s{color} | {color:blue}{color} | {color:blue} Docker mode activated. {color} |
| {color:red}-1{color} | {color:red} yetus {color} | {color:red}  0m  7s{color} 
| {color:red}{color} | {color:red} Unprocessed flag(s): 
--findbugs-strict-precheck {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Docker | ClientAPI=1.41 ServerAPI=1.41 base: 
https://ci-hadoop.apache.org/job/PreCommit-HDFS-Build/475/artifact/out/Dockerfile
 |
| JIRA Issue | HDFS-15816 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/13020552/HDFS-15816.002.patch |
| Console output | 
https://ci-hadoop.apache.org/job/PreCommit-HDFS-Build/475/console |
| versions | git=2.25.1 |
| Powered by | Apache Yetus 0.13.0-SNAPSHOT https://yetus.apache.org |


This message was automatically generated.



> If NO stale node in last choosing, the chooseTarget don't need to retry with 
> stale nodes.
> -
>
> Key: HDFS-15816
> URL: https://issues.apache.org/jira/browse/HDFS-15816
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: block placement
>Reporter: Yang Yun
>Assignee: Yang Yun
>Priority: Minor
> Attachments: HDFS-15816.001.patch, HDFS-15816.002.patch
>
>
> If NO stale node in last choosing, the chooseTarget don't need to retry with 
> stale nodes.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-15816) If NO stale node in last choosing, the chooseTarget don't need to retry with stale nodes.

2021-02-17 Thread Yang Yun (Jira)


[ 
https://issues.apache.org/jira/browse/HDFS-15816?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17285754#comment-17285754
 ] 

Yang Yun commented on HDFS-15816:
-

Thanks [~ayushtkn] for your review.

Add a new variable 'ThreadLocal hasStaleNode' to trace if it met node 
in new patch HDFS-15816.002.patch.

> If NO stale node in last choosing, the chooseTarget don't need to retry with 
> stale nodes.
> -
>
> Key: HDFS-15816
> URL: https://issues.apache.org/jira/browse/HDFS-15816
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: block placement
>Reporter: Yang Yun
>Assignee: Yang Yun
>Priority: Minor
> Attachments: HDFS-15816.001.patch, HDFS-15816.002.patch
>
>
> If NO stale node in last choosing, the chooseTarget don't need to retry with 
> stale nodes.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-15816) If NO stale node in last choosing, the chooseTarget don't need to retry with stale nodes.

2021-02-17 Thread Yang Yun (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15816?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yang Yun updated HDFS-15816:

Attachment: HDFS-15816.002.patch
Status: Patch Available  (was: Open)

> If NO stale node in last choosing, the chooseTarget don't need to retry with 
> stale nodes.
> -
>
> Key: HDFS-15816
> URL: https://issues.apache.org/jira/browse/HDFS-15816
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: block placement
>Reporter: Yang Yun
>Assignee: Yang Yun
>Priority: Minor
> Attachments: HDFS-15816.001.patch, HDFS-15816.002.patch
>
>
> If NO stale node in last choosing, the chooseTarget don't need to retry with 
> stale nodes.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-15816) If NO stale node in last choosing, the chooseTarget don't need to retry with stale nodes.

2021-02-17 Thread Yang Yun (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15816?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yang Yun updated HDFS-15816:

Status: Open  (was: Patch Available)

> If NO stale node in last choosing, the chooseTarget don't need to retry with 
> stale nodes.
> -
>
> Key: HDFS-15816
> URL: https://issues.apache.org/jira/browse/HDFS-15816
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: block placement
>Reporter: Yang Yun
>Assignee: Yang Yun
>Priority: Minor
> Attachments: HDFS-15816.001.patch
>
>
> If NO stale node in last choosing, the chooseTarget don't need to retry with 
> stale nodes.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-9536) OOM errors during parallel upgrade to Block-ID based layout

2021-02-17 Thread Luca Toscano (Jira)


[ 
https://issues.apache.org/jira/browse/HDFS-9536?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17285707#comment-17285707
 ] 

Luca Toscano commented on HDFS-9536:


Hi! I have recently upgraded an Hadoop cluster from 2.6.0 (CDH 5.16.2, so not a 
vanilla 2.6) to 2.10.1 (Apache Bigtop 1.5) and I have experienced a lot of 
troubles while upgrading DNs. Before the upgrade we used to run the DNs with 
-Xmx 4G without any issue, but during the upgrade we had to bump the value to 
16G on some nodes to avoid OOMs and allow the cluster to complete the upgrade. 
The nodes causing the most problems were the one with more blocks (order of 
some millions of blocks).

Is this an issue that happens only when upgrading from 2.6.0 (due to 
https://issues.apache.org/jira/browse/HDFS-6482) or can it happen also on later 
versions? If so, how should the heap sizes of DNs be tuned? And would reducing 
`dfs.datanode.block.id.layout.upgrade.threads` help?

Thanks in advance :)

> OOM errors during parallel upgrade to Block-ID based layout
> ---
>
> Key: HDFS-9536
> URL: https://issues.apache.org/jira/browse/HDFS-9536
> Project: Hadoop HDFS
>  Issue Type: Bug
>Reporter: Vinayakumar B
>Assignee: Vinayakumar B
>Priority: Major
>
> This is a follow-up jira for the OOM errors observed during parallel upgrade 
> to Block-ID based datanode layout using HDFS-8578 fix.
> more clue 
> [here|https://issues.apache.org/jira/browse/HDFS-8578?focusedCommentId=15042012=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-15042012]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org