[jira] [Updated] (HDFS-14811) RBF: TestRouterRpc#testErasureCoding is flaky
[ https://issues.apache.org/jira/browse/HDFS-14811?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Akira Ajisaka updated HDFS-14811: - Component/s: rbf > RBF: TestRouterRpc#testErasureCoding is flaky > - > > Key: HDFS-14811 > URL: https://issues.apache.org/jira/browse/HDFS-14811 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: rbf >Reporter: Chen Zhang >Assignee: Chen Zhang >Priority: Major > Fix For: 3.3.1, 3.4.0 > > Attachments: HDFS-14811.001.patch, HDFS-14811.002.patch > > > The Failed reason: > {code:java} > 2019-09-01 18:19:20,940 [IPC Server handler 5 on default port 53140] INFO > blockmanagement.BlockPlacementPolicy > (BlockPlacementPolicyDefault.java:chooseRandom(838)) - [ > Node /default-rack/127.0.0.1:53148 [ > ] > Node /default-rack/127.0.0.1:53161 [ > ] > Node /default-rack/127.0.0.1:53157 [ > Datanode 127.0.0.1:53157 is not chosen since the node is too busy (load: 3 > > 2.6665). > Node /default-rack/127.0.0.1:53143 [ > ] > Node /default-rack/127.0.0.1:53165 [ > ] > 2019-09-01 18:19:20,940 [IPC Server handler 5 on default port 53140] INFO > blockmanagement.BlockPlacementPolicy > (BlockPlacementPolicyDefault.java:chooseRandom(846)) - Not enough replicas > was chosen. Reason: {NODE_TOO_BUSY=1} > 2019-09-01 18:19:20,941 [IPC Server handler 5 on default port 53140] WARN > blockmanagement.BlockPlacementPolicy > (BlockPlacementPolicyDefault.java:chooseTarget(449)) - Failed to place enough > replicas, still in need of 1 to reach 6 (unavailableStorages=[], > storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], > creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) > 2019-09-01 18:19:20,941 [IPC Server handler 5 on default port 53140] WARN > protocol.BlockStoragePolicy (BlockStoragePolicy.java:chooseStorageTypes(161)) > - Failed to place enough replicas: expected size is 1 but only 0 storage > types can be selected (replication=6, selected=[], unavailable=[DISK], > removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], > creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) > 2019-09-01 18:19:20,941 [IPC Server handler 5 on default port 53140] WARN > blockmanagement.BlockPlacementPolicy > (BlockPlacementPolicyDefault.java:chooseTarget(449)) - Failed to place enough > replicas, still in need of 1 to reach 6 (unavailableStorages=[DISK], > storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], > creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All > required storage types are unavailable: unavailableStorages=[DISK], > storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], > creationFallbacks=[], replicationFallbacks=[ARCHIVE]} > 2019-09-01 18:19:20,941 [IPC Server handler 5 on default port 53140] INFO > ipc.Server (Server.java:logException(2982)) - IPC Server handler 5 on default > port 53140, call Call#1270 Retry#0 > org.apache.hadoop.hdfs.protocol.ClientProtocol.addBlock from 127.0.0.1:53202 > java.io.IOException: File /testec/testfile2 could only be written to 5 of the > 6 required nodes for RS-6-3-1024k. There are 6 datanode(s) running and 6 > node(s) are excluded in this operation. > at > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:) > at > org.apache.hadoop.hdfs.server.namenode.FSDirWriteFileOp.chooseTargetForNewBlock(FSDirWriteFileOp.java:294) > at > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:2815) > at > org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:893) > at > org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:574) > at > org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) > at > org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:529) > at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1070) > at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1001) > at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:929) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:422) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1891) > at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2921) > 2019-09-01 18:19:20,942 [IPC Server handler 6 on default port 53197] INFO > ipc.Server (Server.java:logException(2975)) - IPC Server handler 6 on default > port 53197, call Call#1268 Retry#0 >
[jira] [Updated] (HDFS-14811) RBF: TestRouterRpc#testErasureCoding is flaky
[ https://issues.apache.org/jira/browse/HDFS-14811?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Akira Ajisaka updated HDFS-14811: - Parent: HDFS-14603 Issue Type: Sub-task (was: Bug) > RBF: TestRouterRpc#testErasureCoding is flaky > - > > Key: HDFS-14811 > URL: https://issues.apache.org/jira/browse/HDFS-14811 > Project: Hadoop HDFS > Issue Type: Sub-task >Reporter: Chen Zhang >Assignee: Chen Zhang >Priority: Major > Fix For: 3.3.1, 3.4.0 > > Attachments: HDFS-14811.001.patch, HDFS-14811.002.patch > > > The Failed reason: > {code:java} > 2019-09-01 18:19:20,940 [IPC Server handler 5 on default port 53140] INFO > blockmanagement.BlockPlacementPolicy > (BlockPlacementPolicyDefault.java:chooseRandom(838)) - [ > Node /default-rack/127.0.0.1:53148 [ > ] > Node /default-rack/127.0.0.1:53161 [ > ] > Node /default-rack/127.0.0.1:53157 [ > Datanode 127.0.0.1:53157 is not chosen since the node is too busy (load: 3 > > 2.6665). > Node /default-rack/127.0.0.1:53143 [ > ] > Node /default-rack/127.0.0.1:53165 [ > ] > 2019-09-01 18:19:20,940 [IPC Server handler 5 on default port 53140] INFO > blockmanagement.BlockPlacementPolicy > (BlockPlacementPolicyDefault.java:chooseRandom(846)) - Not enough replicas > was chosen. Reason: {NODE_TOO_BUSY=1} > 2019-09-01 18:19:20,941 [IPC Server handler 5 on default port 53140] WARN > blockmanagement.BlockPlacementPolicy > (BlockPlacementPolicyDefault.java:chooseTarget(449)) - Failed to place enough > replicas, still in need of 1 to reach 6 (unavailableStorages=[], > storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], > creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) > 2019-09-01 18:19:20,941 [IPC Server handler 5 on default port 53140] WARN > protocol.BlockStoragePolicy (BlockStoragePolicy.java:chooseStorageTypes(161)) > - Failed to place enough replicas: expected size is 1 but only 0 storage > types can be selected (replication=6, selected=[], unavailable=[DISK], > removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], > creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) > 2019-09-01 18:19:20,941 [IPC Server handler 5 on default port 53140] WARN > blockmanagement.BlockPlacementPolicy > (BlockPlacementPolicyDefault.java:chooseTarget(449)) - Failed to place enough > replicas, still in need of 1 to reach 6 (unavailableStorages=[DISK], > storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], > creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All > required storage types are unavailable: unavailableStorages=[DISK], > storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], > creationFallbacks=[], replicationFallbacks=[ARCHIVE]} > 2019-09-01 18:19:20,941 [IPC Server handler 5 on default port 53140] INFO > ipc.Server (Server.java:logException(2982)) - IPC Server handler 5 on default > port 53140, call Call#1270 Retry#0 > org.apache.hadoop.hdfs.protocol.ClientProtocol.addBlock from 127.0.0.1:53202 > java.io.IOException: File /testec/testfile2 could only be written to 5 of the > 6 required nodes for RS-6-3-1024k. There are 6 datanode(s) running and 6 > node(s) are excluded in this operation. > at > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:) > at > org.apache.hadoop.hdfs.server.namenode.FSDirWriteFileOp.chooseTargetForNewBlock(FSDirWriteFileOp.java:294) > at > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:2815) > at > org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:893) > at > org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:574) > at > org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) > at > org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:529) > at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1070) > at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1001) > at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:929) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:422) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1891) > at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2921) > 2019-09-01 18:19:20,942 [IPC Server handler 6 on default port 53197] INFO > ipc.Server (Server.java:logException(2975)) - IPC Server handler 6 on default > port 53197, call
[jira] [Updated] (HDFS-14811) RBF: TestRouterRpc#testErasureCoding is flaky
[ https://issues.apache.org/jira/browse/HDFS-14811?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Akira Ajisaka updated HDFS-14811: - Fix Version/s: 3.3.1 Cherry-picked to branch-3.3. I ran the test locally and confirmed the test passed before committing. > RBF: TestRouterRpc#testErasureCoding is flaky > - > > Key: HDFS-14811 > URL: https://issues.apache.org/jira/browse/HDFS-14811 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Chen Zhang >Assignee: Chen Zhang >Priority: Major > Fix For: 3.3.1, 3.4.0 > > Attachments: HDFS-14811.001.patch, HDFS-14811.002.patch > > > The Failed reason: > {code:java} > 2019-09-01 18:19:20,940 [IPC Server handler 5 on default port 53140] INFO > blockmanagement.BlockPlacementPolicy > (BlockPlacementPolicyDefault.java:chooseRandom(838)) - [ > Node /default-rack/127.0.0.1:53148 [ > ] > Node /default-rack/127.0.0.1:53161 [ > ] > Node /default-rack/127.0.0.1:53157 [ > Datanode 127.0.0.1:53157 is not chosen since the node is too busy (load: 3 > > 2.6665). > Node /default-rack/127.0.0.1:53143 [ > ] > Node /default-rack/127.0.0.1:53165 [ > ] > 2019-09-01 18:19:20,940 [IPC Server handler 5 on default port 53140] INFO > blockmanagement.BlockPlacementPolicy > (BlockPlacementPolicyDefault.java:chooseRandom(846)) - Not enough replicas > was chosen. Reason: {NODE_TOO_BUSY=1} > 2019-09-01 18:19:20,941 [IPC Server handler 5 on default port 53140] WARN > blockmanagement.BlockPlacementPolicy > (BlockPlacementPolicyDefault.java:chooseTarget(449)) - Failed to place enough > replicas, still in need of 1 to reach 6 (unavailableStorages=[], > storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], > creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) > 2019-09-01 18:19:20,941 [IPC Server handler 5 on default port 53140] WARN > protocol.BlockStoragePolicy (BlockStoragePolicy.java:chooseStorageTypes(161)) > - Failed to place enough replicas: expected size is 1 but only 0 storage > types can be selected (replication=6, selected=[], unavailable=[DISK], > removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], > creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) > 2019-09-01 18:19:20,941 [IPC Server handler 5 on default port 53140] WARN > blockmanagement.BlockPlacementPolicy > (BlockPlacementPolicyDefault.java:chooseTarget(449)) - Failed to place enough > replicas, still in need of 1 to reach 6 (unavailableStorages=[DISK], > storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], > creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All > required storage types are unavailable: unavailableStorages=[DISK], > storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], > creationFallbacks=[], replicationFallbacks=[ARCHIVE]} > 2019-09-01 18:19:20,941 [IPC Server handler 5 on default port 53140] INFO > ipc.Server (Server.java:logException(2982)) - IPC Server handler 5 on default > port 53140, call Call#1270 Retry#0 > org.apache.hadoop.hdfs.protocol.ClientProtocol.addBlock from 127.0.0.1:53202 > java.io.IOException: File /testec/testfile2 could only be written to 5 of the > 6 required nodes for RS-6-3-1024k. There are 6 datanode(s) running and 6 > node(s) are excluded in this operation. > at > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:) > at > org.apache.hadoop.hdfs.server.namenode.FSDirWriteFileOp.chooseTargetForNewBlock(FSDirWriteFileOp.java:294) > at > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:2815) > at > org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:893) > at > org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:574) > at > org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) > at > org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:529) > at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1070) > at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1001) > at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:929) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:422) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1891) > at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2921) > 2019-09-01 18:19:20,942 [IPC Server handler 6 on default port 53197] INFO > ipc.Server (Server.java:logException(2975)) -
[jira] [Updated] (HDFS-14811) RBF: TestRouterRpc#testErasureCoding is flaky
[ https://issues.apache.org/jira/browse/HDFS-14811?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Ayush Saxena updated HDFS-14811: Fix Version/s: 3.4.0 Hadoop Flags: Reviewed Resolution: Fixed Status: Resolved (was: Patch Available) > RBF: TestRouterRpc#testErasureCoding is flaky > - > > Key: HDFS-14811 > URL: https://issues.apache.org/jira/browse/HDFS-14811 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Chen Zhang >Assignee: Chen Zhang >Priority: Major > Fix For: 3.4.0 > > Attachments: HDFS-14811.001.patch, HDFS-14811.002.patch > > > The Failed reason: > {code:java} > 2019-09-01 18:19:20,940 [IPC Server handler 5 on default port 53140] INFO > blockmanagement.BlockPlacementPolicy > (BlockPlacementPolicyDefault.java:chooseRandom(838)) - [ > Node /default-rack/127.0.0.1:53148 [ > ] > Node /default-rack/127.0.0.1:53161 [ > ] > Node /default-rack/127.0.0.1:53157 [ > Datanode 127.0.0.1:53157 is not chosen since the node is too busy (load: 3 > > 2.6665). > Node /default-rack/127.0.0.1:53143 [ > ] > Node /default-rack/127.0.0.1:53165 [ > ] > 2019-09-01 18:19:20,940 [IPC Server handler 5 on default port 53140] INFO > blockmanagement.BlockPlacementPolicy > (BlockPlacementPolicyDefault.java:chooseRandom(846)) - Not enough replicas > was chosen. Reason: {NODE_TOO_BUSY=1} > 2019-09-01 18:19:20,941 [IPC Server handler 5 on default port 53140] WARN > blockmanagement.BlockPlacementPolicy > (BlockPlacementPolicyDefault.java:chooseTarget(449)) - Failed to place enough > replicas, still in need of 1 to reach 6 (unavailableStorages=[], > storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], > creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) > 2019-09-01 18:19:20,941 [IPC Server handler 5 on default port 53140] WARN > protocol.BlockStoragePolicy (BlockStoragePolicy.java:chooseStorageTypes(161)) > - Failed to place enough replicas: expected size is 1 but only 0 storage > types can be selected (replication=6, selected=[], unavailable=[DISK], > removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], > creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) > 2019-09-01 18:19:20,941 [IPC Server handler 5 on default port 53140] WARN > blockmanagement.BlockPlacementPolicy > (BlockPlacementPolicyDefault.java:chooseTarget(449)) - Failed to place enough > replicas, still in need of 1 to reach 6 (unavailableStorages=[DISK], > storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], > creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All > required storage types are unavailable: unavailableStorages=[DISK], > storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], > creationFallbacks=[], replicationFallbacks=[ARCHIVE]} > 2019-09-01 18:19:20,941 [IPC Server handler 5 on default port 53140] INFO > ipc.Server (Server.java:logException(2982)) - IPC Server handler 5 on default > port 53140, call Call#1270 Retry#0 > org.apache.hadoop.hdfs.protocol.ClientProtocol.addBlock from 127.0.0.1:53202 > java.io.IOException: File /testec/testfile2 could only be written to 5 of the > 6 required nodes for RS-6-3-1024k. There are 6 datanode(s) running and 6 > node(s) are excluded in this operation. > at > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:) > at > org.apache.hadoop.hdfs.server.namenode.FSDirWriteFileOp.chooseTargetForNewBlock(FSDirWriteFileOp.java:294) > at > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:2815) > at > org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:893) > at > org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:574) > at > org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) > at > org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:529) > at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1070) > at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1001) > at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:929) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:422) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1891) > at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2921) > 2019-09-01 18:19:20,942 [IPC Server handler 6 on default port 53197] INFO > ipc.Server (Server.java:logException(2975)) - IPC
[jira] [Updated] (HDFS-14811) RBF: TestRouterRpc#testErasureCoding is flaky
[ https://issues.apache.org/jira/browse/HDFS-14811?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Chen Zhang updated HDFS-14811: -- Attachment: HDFS-14811.002.patch > RBF: TestRouterRpc#testErasureCoding is flaky > - > > Key: HDFS-14811 > URL: https://issues.apache.org/jira/browse/HDFS-14811 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Chen Zhang >Assignee: Chen Zhang >Priority: Major > Attachments: HDFS-14811.001.patch, HDFS-14811.002.patch > > > The Failed reason: > {code:java} > 2019-09-01 18:19:20,940 [IPC Server handler 5 on default port 53140] INFO > blockmanagement.BlockPlacementPolicy > (BlockPlacementPolicyDefault.java:chooseRandom(838)) - [ > Node /default-rack/127.0.0.1:53148 [ > ] > Node /default-rack/127.0.0.1:53161 [ > ] > Node /default-rack/127.0.0.1:53157 [ > Datanode 127.0.0.1:53157 is not chosen since the node is too busy (load: 3 > > 2.6665). > Node /default-rack/127.0.0.1:53143 [ > ] > Node /default-rack/127.0.0.1:53165 [ > ] > 2019-09-01 18:19:20,940 [IPC Server handler 5 on default port 53140] INFO > blockmanagement.BlockPlacementPolicy > (BlockPlacementPolicyDefault.java:chooseRandom(846)) - Not enough replicas > was chosen. Reason: {NODE_TOO_BUSY=1} > 2019-09-01 18:19:20,941 [IPC Server handler 5 on default port 53140] WARN > blockmanagement.BlockPlacementPolicy > (BlockPlacementPolicyDefault.java:chooseTarget(449)) - Failed to place enough > replicas, still in need of 1 to reach 6 (unavailableStorages=[], > storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], > creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) > 2019-09-01 18:19:20,941 [IPC Server handler 5 on default port 53140] WARN > protocol.BlockStoragePolicy (BlockStoragePolicy.java:chooseStorageTypes(161)) > - Failed to place enough replicas: expected size is 1 but only 0 storage > types can be selected (replication=6, selected=[], unavailable=[DISK], > removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], > creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) > 2019-09-01 18:19:20,941 [IPC Server handler 5 on default port 53140] WARN > blockmanagement.BlockPlacementPolicy > (BlockPlacementPolicyDefault.java:chooseTarget(449)) - Failed to place enough > replicas, still in need of 1 to reach 6 (unavailableStorages=[DISK], > storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], > creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All > required storage types are unavailable: unavailableStorages=[DISK], > storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], > creationFallbacks=[], replicationFallbacks=[ARCHIVE]} > 2019-09-01 18:19:20,941 [IPC Server handler 5 on default port 53140] INFO > ipc.Server (Server.java:logException(2982)) - IPC Server handler 5 on default > port 53140, call Call#1270 Retry#0 > org.apache.hadoop.hdfs.protocol.ClientProtocol.addBlock from 127.0.0.1:53202 > java.io.IOException: File /testec/testfile2 could only be written to 5 of the > 6 required nodes for RS-6-3-1024k. There are 6 datanode(s) running and 6 > node(s) are excluded in this operation. > at > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:) > at > org.apache.hadoop.hdfs.server.namenode.FSDirWriteFileOp.chooseTargetForNewBlock(FSDirWriteFileOp.java:294) > at > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:2815) > at > org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:893) > at > org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:574) > at > org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) > at > org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:529) > at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1070) > at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1001) > at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:929) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:422) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1891) > at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2921) > 2019-09-01 18:19:20,942 [IPC Server handler 6 on default port 53197] INFO > ipc.Server (Server.java:logException(2975)) - IPC Server handler 6 on default > port 53197, call Call#1268 Retry#0 > org.apache.hadoop.hdfs.protocol.ClientProtocol.addBlock from
[jira] [Updated] (HDFS-14811) RBF: TestRouterRpc#testErasureCoding is flaky
[ https://issues.apache.org/jira/browse/HDFS-14811?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Chen Zhang updated HDFS-14811: -- Status: Patch Available (was: Open) > RBF: TestRouterRpc#testErasureCoding is flaky > - > > Key: HDFS-14811 > URL: https://issues.apache.org/jira/browse/HDFS-14811 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Chen Zhang >Assignee: Chen Zhang >Priority: Major > Attachments: HDFS-14811.001.patch > > > The Failed reason: > {code:java} > 2019-09-01 18:19:20,940 [IPC Server handler 5 on default port 53140] INFO > blockmanagement.BlockPlacementPolicy > (BlockPlacementPolicyDefault.java:chooseRandom(838)) - [ > Node /default-rack/127.0.0.1:53148 [ > ] > Node /default-rack/127.0.0.1:53161 [ > ] > Node /default-rack/127.0.0.1:53157 [ > Datanode 127.0.0.1:53157 is not chosen since the node is too busy (load: 3 > > 2.6665). > Node /default-rack/127.0.0.1:53143 [ > ] > Node /default-rack/127.0.0.1:53165 [ > ] > 2019-09-01 18:19:20,940 [IPC Server handler 5 on default port 53140] INFO > blockmanagement.BlockPlacementPolicy > (BlockPlacementPolicyDefault.java:chooseRandom(846)) - Not enough replicas > was chosen. Reason: {NODE_TOO_BUSY=1} > 2019-09-01 18:19:20,941 [IPC Server handler 5 on default port 53140] WARN > blockmanagement.BlockPlacementPolicy > (BlockPlacementPolicyDefault.java:chooseTarget(449)) - Failed to place enough > replicas, still in need of 1 to reach 6 (unavailableStorages=[], > storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], > creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) > 2019-09-01 18:19:20,941 [IPC Server handler 5 on default port 53140] WARN > protocol.BlockStoragePolicy (BlockStoragePolicy.java:chooseStorageTypes(161)) > - Failed to place enough replicas: expected size is 1 but only 0 storage > types can be selected (replication=6, selected=[], unavailable=[DISK], > removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], > creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) > 2019-09-01 18:19:20,941 [IPC Server handler 5 on default port 53140] WARN > blockmanagement.BlockPlacementPolicy > (BlockPlacementPolicyDefault.java:chooseTarget(449)) - Failed to place enough > replicas, still in need of 1 to reach 6 (unavailableStorages=[DISK], > storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], > creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All > required storage types are unavailable: unavailableStorages=[DISK], > storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], > creationFallbacks=[], replicationFallbacks=[ARCHIVE]} > 2019-09-01 18:19:20,941 [IPC Server handler 5 on default port 53140] INFO > ipc.Server (Server.java:logException(2982)) - IPC Server handler 5 on default > port 53140, call Call#1270 Retry#0 > org.apache.hadoop.hdfs.protocol.ClientProtocol.addBlock from 127.0.0.1:53202 > java.io.IOException: File /testec/testfile2 could only be written to 5 of the > 6 required nodes for RS-6-3-1024k. There are 6 datanode(s) running and 6 > node(s) are excluded in this operation. > at > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:) > at > org.apache.hadoop.hdfs.server.namenode.FSDirWriteFileOp.chooseTargetForNewBlock(FSDirWriteFileOp.java:294) > at > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:2815) > at > org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:893) > at > org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:574) > at > org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) > at > org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:529) > at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1070) > at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1001) > at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:929) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:422) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1891) > at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2921) > 2019-09-01 18:19:20,942 [IPC Server handler 6 on default port 53197] INFO > ipc.Server (Server.java:logException(2975)) - IPC Server handler 6 on default > port 53197, call Call#1268 Retry#0 > org.apache.hadoop.hdfs.protocol.ClientProtocol.addBlock from >
[jira] [Updated] (HDFS-14811) RBF: TestRouterRpc#testErasureCoding is flaky
[ https://issues.apache.org/jira/browse/HDFS-14811?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Chen Zhang updated HDFS-14811: -- Attachment: HDFS-14811.001.patch > RBF: TestRouterRpc#testErasureCoding is flaky > - > > Key: HDFS-14811 > URL: https://issues.apache.org/jira/browse/HDFS-14811 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Chen Zhang >Assignee: Chen Zhang >Priority: Major > Attachments: HDFS-14811.001.patch > > > The Failed reason: > {code:java} > 2019-09-01 18:19:20,940 [IPC Server handler 5 on default port 53140] INFO > blockmanagement.BlockPlacementPolicy > (BlockPlacementPolicyDefault.java:chooseRandom(838)) - [ > Node /default-rack/127.0.0.1:53148 [ > ] > Node /default-rack/127.0.0.1:53161 [ > ] > Node /default-rack/127.0.0.1:53157 [ > Datanode 127.0.0.1:53157 is not chosen since the node is too busy (load: 3 > > 2.6665). > Node /default-rack/127.0.0.1:53143 [ > ] > Node /default-rack/127.0.0.1:53165 [ > ] > 2019-09-01 18:19:20,940 [IPC Server handler 5 on default port 53140] INFO > blockmanagement.BlockPlacementPolicy > (BlockPlacementPolicyDefault.java:chooseRandom(846)) - Not enough replicas > was chosen. Reason: {NODE_TOO_BUSY=1} > 2019-09-01 18:19:20,941 [IPC Server handler 5 on default port 53140] WARN > blockmanagement.BlockPlacementPolicy > (BlockPlacementPolicyDefault.java:chooseTarget(449)) - Failed to place enough > replicas, still in need of 1 to reach 6 (unavailableStorages=[], > storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], > creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) > 2019-09-01 18:19:20,941 [IPC Server handler 5 on default port 53140] WARN > protocol.BlockStoragePolicy (BlockStoragePolicy.java:chooseStorageTypes(161)) > - Failed to place enough replicas: expected size is 1 but only 0 storage > types can be selected (replication=6, selected=[], unavailable=[DISK], > removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], > creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) > 2019-09-01 18:19:20,941 [IPC Server handler 5 on default port 53140] WARN > blockmanagement.BlockPlacementPolicy > (BlockPlacementPolicyDefault.java:chooseTarget(449)) - Failed to place enough > replicas, still in need of 1 to reach 6 (unavailableStorages=[DISK], > storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], > creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All > required storage types are unavailable: unavailableStorages=[DISK], > storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], > creationFallbacks=[], replicationFallbacks=[ARCHIVE]} > 2019-09-01 18:19:20,941 [IPC Server handler 5 on default port 53140] INFO > ipc.Server (Server.java:logException(2982)) - IPC Server handler 5 on default > port 53140, call Call#1270 Retry#0 > org.apache.hadoop.hdfs.protocol.ClientProtocol.addBlock from 127.0.0.1:53202 > java.io.IOException: File /testec/testfile2 could only be written to 5 of the > 6 required nodes for RS-6-3-1024k. There are 6 datanode(s) running and 6 > node(s) are excluded in this operation. > at > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:) > at > org.apache.hadoop.hdfs.server.namenode.FSDirWriteFileOp.chooseTargetForNewBlock(FSDirWriteFileOp.java:294) > at > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:2815) > at > org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:893) > at > org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:574) > at > org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) > at > org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:529) > at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1070) > at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1001) > at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:929) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:422) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1891) > at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2921) > 2019-09-01 18:19:20,942 [IPC Server handler 6 on default port 53197] INFO > ipc.Server (Server.java:logException(2975)) - IPC Server handler 6 on default > port 53197, call Call#1268 Retry#0 > org.apache.hadoop.hdfs.protocol.ClientProtocol.addBlock from >