[ https://issues.apache.org/jira/browse/HDFS-12727?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Rushabh S Shah updated HDFS-12727: ---------------------------------- Description: TestPread timing out on branch-2.8 and not on trunk. {noformat} 2017-10-24 19:47:37,377 WARN impl.BlockReaderFactory (BlockReaderFactory.java:getRemoteBlockReaderFromTcp(758)) - I/O error constructing remote block reader. java.net.ConnectException: Connection refused at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717) at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206) at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:534) at org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:2955) at org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:815) at org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:740) at org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.build(BlockReaderFactory.java:385) at org.apache.hadoop.hdfs.DFSInputStream.getBlockReader(DFSInputStream.java:708) at org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1230) at org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1198) at org.apache.hadoop.hdfs.DFSInputStream.fetchBlockByteRange(DFSInputStream.java:1158) at org.apache.hadoop.hdfs.DFSInputStream.pread(DFSInputStream.java:1535) at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:1501) at org.apache.hadoop.fs.FSInputStream.readFully(FSInputStream.java:121) at org.apache.hadoop.fs.FSDataInputStream.readFully(FSDataInputStream.java:120) at org.apache.hadoop.hdfs.TestPread.datanodeRestartTest(TestPread.java:245) at org.apache.hadoop.hdfs.TestPread.dfsPreadTest(TestPread.java:478) at org.apache.hadoop.hdfs.TestPread.testPreadDFSNoChecksum(TestPread.java:280) {noformat} Few lines in the log before hanging. {noformat} 2017-10-26 20:21:07,938 WARN impl.BlockReaderFactory (BlockReaderFactory.java:getRemoteBlockReaderFromTcp(758)) - I/O error constructing remote block reader. java.net.ConnectException: Connection refused at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717) at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206) at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:534) at org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:2955) at org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:815) at org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:740) at org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.build(BlockReaderFactory.java:385) at org.apache.hadoop.hdfs.DFSInputStream.getBlockReader(DFSInputStream.java:708) at org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1230) at org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1198) at org.apache.hadoop.hdfs.DFSInputStream.access$000(DFSInputStream.java:97) at org.apache.hadoop.hdfs.DFSInputStream$2.call(DFSInputStream.java:1182) at org.apache.hadoop.hdfs.DFSInputStream$2.call(DFSInputStream.java:1174) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745) 2017-10-26 20:21:07,938 WARN hdfs.DFSClient (DFSInputStream.java:actualGetFromOneDataNode(1270)) - Connection failure: Failed to connect to /127.0.0.1:42357 for file /preadtest.dat for block BP-287215640-172.17.0.18-1509049266453:blk_1073741825_1001:java.net.ConnectException: Connection refused java.net.ConnectException: Connection refused at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717) at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206) at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:534) at org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:2955) at org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:815) at org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:740) at org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.build(BlockReaderFactory.java:385) at org.apache.hadoop.hdfs.DFSInputStream.getBlockReader(DFSInputStream.java:708) at org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1230) at org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1198) at org.apache.hadoop.hdfs.DFSInputStream.access$000(DFSInputStream.java:97) at org.apache.hadoop.hdfs.DFSInputStream$2.call(DFSInputStream.java:1182) at org.apache.hadoop.hdfs.DFSInputStream$2.call(DFSInputStream.java:1174) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745) 2017-10-26 20:21:07,939 WARN hdfs.DFSClient (DFSInputStream.java:getBestNodeDNAddrPair(1112)) - No live nodes contain block BP-287215640-172.17.0.18-1509049266453:blk_1073741825_1001 after checking nodes = [DatanodeInfoWithStorage[127.0.0.1:36669,DS-3ce766bc-dad8-4022-b0d8-396a669ee4b8,DISK], DatanodeInfoWithStorage[127.0.0.1:36005,DS-bd7b59f4-a7de-4524-877e-c0f9a10ce5d5,DISK], DatanodeInfoWithStorage[127.0.0.1:42357,DS-a42e4a89-3985-4dc7-ad6a-c0bcf078ccae,DISK]], ignoredNodes = [DatanodeInfoWithStorage[127.0.0.1:36005,DS-bd7b59f4-a7de-4524-877e-c0f9a10ce5d5,DISK], DatanodeInfoWithStorage[127.0.0.1:42357,DS-a42e4a89-3985-4dc7-ad6a-c0bcf078ccae,DISK]] 2017-10-26 20:31:30,435 INFO datanode.DirectoryScanner (DirectoryScanner.java:scan(667)) - BlockPool BP-287215640-172.17.0.18-1509049266453 Total blocks: 12, missing metadata files:0, missing block files:0, missing blocks in memory:0, mismatched blocks:0 {noformat} was: TestPread timing out on branch-2.8 and not on trunk. Few lines before hanging. {noformat} 2017-10-26 20:21:07,938 WARN impl.BlockReaderFactory (BlockReaderFactory.java:getRemoteBlockReaderFromTcp(758)) - I/O error constructing remote block reader. java.net.ConnectException: Connection refused at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717) at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206) at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:534) at org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:2955) at org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:815) at org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:740) at org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.build(BlockReaderFactory.java:385) at org.apache.hadoop.hdfs.DFSInputStream.getBlockReader(DFSInputStream.java:708) at org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1230) at org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1198) at org.apache.hadoop.hdfs.DFSInputStream.access$000(DFSInputStream.java:97) at org.apache.hadoop.hdfs.DFSInputStream$2.call(DFSInputStream.java:1182) at org.apache.hadoop.hdfs.DFSInputStream$2.call(DFSInputStream.java:1174) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745) 2017-10-26 20:21:07,938 WARN hdfs.DFSClient (DFSInputStream.java:actualGetFromOneDataNode(1270)) - Connection failure: Failed to connect to /127.0.0.1:42357 for file /preadtest.dat for block BP-287215640-172.17.0.18-1509049266453:blk_1073741825_1001:java.net.ConnectException: Connection refused java.net.ConnectException: Connection refused at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717) at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206) at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:534) at org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:2955) at org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:815) at org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:740) at org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.build(BlockReaderFactory.java:385) at org.apache.hadoop.hdfs.DFSInputStream.getBlockReader(DFSInputStream.java:708) at org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1230) at org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1198) at org.apache.hadoop.hdfs.DFSInputStream.access$000(DFSInputStream.java:97) at org.apache.hadoop.hdfs.DFSInputStream$2.call(DFSInputStream.java:1182) at org.apache.hadoop.hdfs.DFSInputStream$2.call(DFSInputStream.java:1174) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745) 2017-10-26 20:21:07,939 WARN hdfs.DFSClient (DFSInputStream.java:getBestNodeDNAddrPair(1112)) - No live nodes contain block BP-287215640-172.17.0.18-1509049266453:blk_1073741825_1001 after checking nodes = [DatanodeInfoWithStorage[127.0.0.1:36669,DS-3ce766bc-dad8-4022-b0d8-396a669ee4b8,DISK], DatanodeInfoWithStorage[127.0.0.1:36005,DS-bd7b59f4-a7de-4524-877e-c0f9a10ce5d5,DISK], DatanodeInfoWithStorage[127.0.0.1:42357,DS-a42e4a89-3985-4dc7-ad6a-c0bcf078ccae,DISK]], ignoredNodes = [DatanodeInfoWithStorage[127.0.0.1:36005,DS-bd7b59f4-a7de-4524-877e-c0f9a10ce5d5,DISK], DatanodeInfoWithStorage[127.0.0.1:42357,DS-a42e4a89-3985-4dc7-ad6a-c0bcf078ccae,DISK]] 2017-10-26 20:31:30,435 INFO datanode.DirectoryScanner (DirectoryScanner.java:scan(667)) - BlockPool BP-287215640-172.17.0.18-1509049266453 Total blocks: 12, missing metadata files:0, missing block files:0, missing blocks in memory:0, mismatched blocks:0 {noformat} > TestPread timing out on branch-2.8 > ---------------------------------- > > Key: HDFS-12727 > URL: https://issues.apache.org/jira/browse/HDFS-12727 > Project: Hadoop HDFS > Issue Type: Bug > Reporter: Rushabh S Shah > > TestPread timing out on branch-2.8 and not on trunk. > {noformat} > 2017-10-24 19:47:37,377 WARN impl.BlockReaderFactory > (BlockReaderFactory.java:getRemoteBlockReaderFromTcp(758)) - I/O error > constructing remote block reader. > java.net.ConnectException: Connection refused > at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) > at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717) > at > org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206) > at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:534) > at org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:2955) > at > org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:815) > at > org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:740) > at > org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.build(BlockReaderFactory.java:385) > at > org.apache.hadoop.hdfs.DFSInputStream.getBlockReader(DFSInputStream.java:708) > at > org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1230) > at > org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1198) > at > org.apache.hadoop.hdfs.DFSInputStream.fetchBlockByteRange(DFSInputStream.java:1158) > at org.apache.hadoop.hdfs.DFSInputStream.pread(DFSInputStream.java:1535) > at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:1501) > at org.apache.hadoop.fs.FSInputStream.readFully(FSInputStream.java:121) > at > org.apache.hadoop.fs.FSDataInputStream.readFully(FSDataInputStream.java:120) > at org.apache.hadoop.hdfs.TestPread.datanodeRestartTest(TestPread.java:245) > at org.apache.hadoop.hdfs.TestPread.dfsPreadTest(TestPread.java:478) > at org.apache.hadoop.hdfs.TestPread.testPreadDFSNoChecksum(TestPread.java:280) > {noformat} > Few lines in the log before hanging. > {noformat} > 2017-10-26 20:21:07,938 WARN impl.BlockReaderFactory > (BlockReaderFactory.java:getRemoteBlockReaderFromTcp(758)) - I/O error > constructing remote block reader. > java.net.ConnectException: Connection refused > at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) > at > sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717) > at > org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206) > at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:534) > at > org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:2955) > at > org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:815) > at > org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:740) > at > org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.build(BlockReaderFactory.java:385) > at > org.apache.hadoop.hdfs.DFSInputStream.getBlockReader(DFSInputStream.java:708) > at > org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1230) > at > org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1198) > at > org.apache.hadoop.hdfs.DFSInputStream.access$000(DFSInputStream.java:97) > at > org.apache.hadoop.hdfs.DFSInputStream$2.call(DFSInputStream.java:1182) > at > org.apache.hadoop.hdfs.DFSInputStream$2.call(DFSInputStream.java:1174) > at java.util.concurrent.FutureTask.run(FutureTask.java:266) > at > java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) > at java.util.concurrent.FutureTask.run(FutureTask.java:266) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) > at java.lang.Thread.run(Thread.java:745) > 2017-10-26 20:21:07,938 WARN hdfs.DFSClient > (DFSInputStream.java:actualGetFromOneDataNode(1270)) - Connection failure: > Failed to connect to /127.0.0.1:42357 for file /preadtest.dat for block > BP-287215640-172.17.0.18-1509049266453:blk_1073741825_1001:java.net.ConnectException: > Connection refused > java.net.ConnectException: Connection refused > at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) > at > sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717) > at > org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206) > at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:534) > at > org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:2955) > at > org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:815) > at > org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:740) > at > org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.build(BlockReaderFactory.java:385) > at > org.apache.hadoop.hdfs.DFSInputStream.getBlockReader(DFSInputStream.java:708) > at > org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1230) > at > org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1198) > at > org.apache.hadoop.hdfs.DFSInputStream.access$000(DFSInputStream.java:97) > at > org.apache.hadoop.hdfs.DFSInputStream$2.call(DFSInputStream.java:1182) > at > org.apache.hadoop.hdfs.DFSInputStream$2.call(DFSInputStream.java:1174) > at java.util.concurrent.FutureTask.run(FutureTask.java:266) > at > java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) > at java.util.concurrent.FutureTask.run(FutureTask.java:266) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) > at java.lang.Thread.run(Thread.java:745) > 2017-10-26 20:21:07,939 WARN hdfs.DFSClient > (DFSInputStream.java:getBestNodeDNAddrPair(1112)) - No live nodes contain > block BP-287215640-172.17.0.18-1509049266453:blk_1073741825_1001 after > checking nodes = > [DatanodeInfoWithStorage[127.0.0.1:36669,DS-3ce766bc-dad8-4022-b0d8-396a669ee4b8,DISK], > > DatanodeInfoWithStorage[127.0.0.1:36005,DS-bd7b59f4-a7de-4524-877e-c0f9a10ce5d5,DISK], > > DatanodeInfoWithStorage[127.0.0.1:42357,DS-a42e4a89-3985-4dc7-ad6a-c0bcf078ccae,DISK]], > ignoredNodes = > [DatanodeInfoWithStorage[127.0.0.1:36005,DS-bd7b59f4-a7de-4524-877e-c0f9a10ce5d5,DISK], > > DatanodeInfoWithStorage[127.0.0.1:42357,DS-a42e4a89-3985-4dc7-ad6a-c0bcf078ccae,DISK]] > 2017-10-26 20:31:30,435 INFO datanode.DirectoryScanner > (DirectoryScanner.java:scan(667)) - BlockPool > BP-287215640-172.17.0.18-1509049266453 Total blocks: 12, missing metadata > files:0, missing block files:0, missing blocks in memory:0, mismatched > blocks:0 > {noformat} -- This message was sent by Atlassian JIRA (v6.4.14#64029) --------------------------------------------------------------------- To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org