[
https://issues.apache.org/jira/browse/HDFS-12727?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Rushabh S Shah updated HDFS-12727:
----------------------------------
Description:
TestPread timing out on branch-2.8 and not on trunk.
{noformat}
2017-10-24 19:47:37,377 WARN impl.BlockReaderFactory
(BlockReaderFactory.java:getRemoteBlockReaderFromTcp(758)) - I/O error
constructing remote block reader.
java.net.ConnectException: Connection refused
at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
at
org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:534)
at org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:2955)
at
org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:815)
at
org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:740)
at
org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.build(BlockReaderFactory.java:385)
at org.apache.hadoop.hdfs.DFSInputStream.getBlockReader(DFSInputStream.java:708)
at
org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1230)
at
org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1198)
at
org.apache.hadoop.hdfs.DFSInputStream.fetchBlockByteRange(DFSInputStream.java:1158)
at org.apache.hadoop.hdfs.DFSInputStream.pread(DFSInputStream.java:1535)
at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:1501)
at org.apache.hadoop.fs.FSInputStream.readFully(FSInputStream.java:121)
at org.apache.hadoop.fs.FSDataInputStream.readFully(FSDataInputStream.java:120)
at org.apache.hadoop.hdfs.TestPread.datanodeRestartTest(TestPread.java:245)
at org.apache.hadoop.hdfs.TestPread.dfsPreadTest(TestPread.java:478)
at org.apache.hadoop.hdfs.TestPread.testPreadDFSNoChecksum(TestPread.java:280)
{noformat}
Few lines in the log before hanging.
{noformat}
2017-10-26 20:21:07,938 WARN impl.BlockReaderFactory
(BlockReaderFactory.java:getRemoteBlockReaderFromTcp(758)) - I/O error
constructing remote block reader.
java.net.ConnectException: Connection refused
at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
at
sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
at
org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:534)
at
org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:2955)
at
org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:815)
at
org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:740)
at
org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.build(BlockReaderFactory.java:385)
at
org.apache.hadoop.hdfs.DFSInputStream.getBlockReader(DFSInputStream.java:708)
at
org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1230)
at
org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1198)
at
org.apache.hadoop.hdfs.DFSInputStream.access$000(DFSInputStream.java:97)
at
org.apache.hadoop.hdfs.DFSInputStream$2.call(DFSInputStream.java:1182)
at
org.apache.hadoop.hdfs.DFSInputStream$2.call(DFSInputStream.java:1174)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
2017-10-26 20:21:07,938 WARN hdfs.DFSClient
(DFSInputStream.java:actualGetFromOneDataNode(1270)) - Connection failure:
Failed to connect to /127.0.0.1:42357 for file /preadtest.dat for block
BP-287215640-172.17.0.18-1509049266453:blk_1073741825_1001:java.net.ConnectException:
Connection refused
java.net.ConnectException: Connection refused
at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
at
sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
at
org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:534)
at
org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:2955)
at
org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:815)
at
org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:740)
at
org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.build(BlockReaderFactory.java:385)
at
org.apache.hadoop.hdfs.DFSInputStream.getBlockReader(DFSInputStream.java:708)
at
org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1230)
at
org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1198)
at
org.apache.hadoop.hdfs.DFSInputStream.access$000(DFSInputStream.java:97)
at
org.apache.hadoop.hdfs.DFSInputStream$2.call(DFSInputStream.java:1182)
at
org.apache.hadoop.hdfs.DFSInputStream$2.call(DFSInputStream.java:1174)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
2017-10-26 20:21:07,939 WARN hdfs.DFSClient
(DFSInputStream.java:getBestNodeDNAddrPair(1112)) - No live nodes contain block
BP-287215640-172.17.0.18-1509049266453:blk_1073741825_1001 after checking nodes
=
[DatanodeInfoWithStorage[127.0.0.1:36669,DS-3ce766bc-dad8-4022-b0d8-396a669ee4b8,DISK],
DatanodeInfoWithStorage[127.0.0.1:36005,DS-bd7b59f4-a7de-4524-877e-c0f9a10ce5d5,DISK],
DatanodeInfoWithStorage[127.0.0.1:42357,DS-a42e4a89-3985-4dc7-ad6a-c0bcf078ccae,DISK]],
ignoredNodes =
[DatanodeInfoWithStorage[127.0.0.1:36005,DS-bd7b59f4-a7de-4524-877e-c0f9a10ce5d5,DISK],
DatanodeInfoWithStorage[127.0.0.1:42357,DS-a42e4a89-3985-4dc7-ad6a-c0bcf078ccae,DISK]]
2017-10-26 20:31:30,435 INFO datanode.DirectoryScanner
(DirectoryScanner.java:scan(667)) - BlockPool
BP-287215640-172.17.0.18-1509049266453 Total blocks: 12, missing metadata
files:0, missing block files:0, missing blocks in memory:0, mismatched blocks:0
{noformat}
was:
TestPread timing out on branch-2.8 and not on trunk.
Few lines before hanging.
{noformat}
2017-10-26 20:21:07,938 WARN impl.BlockReaderFactory
(BlockReaderFactory.java:getRemoteBlockReaderFromTcp(758)) - I/O error
constructing remote block reader.
java.net.ConnectException: Connection refused
at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
at
sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
at
org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:534)
at
org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:2955)
at
org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:815)
at
org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:740)
at
org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.build(BlockReaderFactory.java:385)
at
org.apache.hadoop.hdfs.DFSInputStream.getBlockReader(DFSInputStream.java:708)
at
org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1230)
at
org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1198)
at
org.apache.hadoop.hdfs.DFSInputStream.access$000(DFSInputStream.java:97)
at
org.apache.hadoop.hdfs.DFSInputStream$2.call(DFSInputStream.java:1182)
at
org.apache.hadoop.hdfs.DFSInputStream$2.call(DFSInputStream.java:1174)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
2017-10-26 20:21:07,938 WARN hdfs.DFSClient
(DFSInputStream.java:actualGetFromOneDataNode(1270)) - Connection failure:
Failed to connect to /127.0.0.1:42357 for file /preadtest.dat for block
BP-287215640-172.17.0.18-1509049266453:blk_1073741825_1001:java.net.ConnectException:
Connection refused
java.net.ConnectException: Connection refused
at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
at
sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
at
org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:534)
at
org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:2955)
at
org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:815)
at
org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:740)
at
org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.build(BlockReaderFactory.java:385)
at
org.apache.hadoop.hdfs.DFSInputStream.getBlockReader(DFSInputStream.java:708)
at
org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1230)
at
org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1198)
at
org.apache.hadoop.hdfs.DFSInputStream.access$000(DFSInputStream.java:97)
at
org.apache.hadoop.hdfs.DFSInputStream$2.call(DFSInputStream.java:1182)
at
org.apache.hadoop.hdfs.DFSInputStream$2.call(DFSInputStream.java:1174)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
2017-10-26 20:21:07,939 WARN hdfs.DFSClient
(DFSInputStream.java:getBestNodeDNAddrPair(1112)) - No live nodes contain block
BP-287215640-172.17.0.18-1509049266453:blk_1073741825_1001 after checking nodes
=
[DatanodeInfoWithStorage[127.0.0.1:36669,DS-3ce766bc-dad8-4022-b0d8-396a669ee4b8,DISK],
DatanodeInfoWithStorage[127.0.0.1:36005,DS-bd7b59f4-a7de-4524-877e-c0f9a10ce5d5,DISK],
DatanodeInfoWithStorage[127.0.0.1:42357,DS-a42e4a89-3985-4dc7-ad6a-c0bcf078ccae,DISK]],
ignoredNodes =
[DatanodeInfoWithStorage[127.0.0.1:36005,DS-bd7b59f4-a7de-4524-877e-c0f9a10ce5d5,DISK],
DatanodeInfoWithStorage[127.0.0.1:42357,DS-a42e4a89-3985-4dc7-ad6a-c0bcf078ccae,DISK]]
2017-10-26 20:31:30,435 INFO datanode.DirectoryScanner
(DirectoryScanner.java:scan(667)) - BlockPool
BP-287215640-172.17.0.18-1509049266453 Total blocks: 12, missing metadata
files:0, missing block files:0, missing blocks in memory:0, mismatched blocks:0
{noformat}
> TestPread timing out on branch-2.8
> ----------------------------------
>
> Key: HDFS-12727
> URL: https://issues.apache.org/jira/browse/HDFS-12727
> Project: Hadoop HDFS
> Issue Type: Bug
> Reporter: Rushabh S Shah
>
> TestPread timing out on branch-2.8 and not on trunk.
> {noformat}
> 2017-10-24 19:47:37,377 WARN impl.BlockReaderFactory
> (BlockReaderFactory.java:getRemoteBlockReaderFromTcp(758)) - I/O error
> constructing remote block reader.
> java.net.ConnectException: Connection refused
> at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
> at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
> at
> org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
> at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:534)
> at org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:2955)
> at
> org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:815)
> at
> org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:740)
> at
> org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.build(BlockReaderFactory.java:385)
> at
> org.apache.hadoop.hdfs.DFSInputStream.getBlockReader(DFSInputStream.java:708)
> at
> org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1230)
> at
> org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1198)
> at
> org.apache.hadoop.hdfs.DFSInputStream.fetchBlockByteRange(DFSInputStream.java:1158)
> at org.apache.hadoop.hdfs.DFSInputStream.pread(DFSInputStream.java:1535)
> at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:1501)
> at org.apache.hadoop.fs.FSInputStream.readFully(FSInputStream.java:121)
> at
> org.apache.hadoop.fs.FSDataInputStream.readFully(FSDataInputStream.java:120)
> at org.apache.hadoop.hdfs.TestPread.datanodeRestartTest(TestPread.java:245)
> at org.apache.hadoop.hdfs.TestPread.dfsPreadTest(TestPread.java:478)
> at org.apache.hadoop.hdfs.TestPread.testPreadDFSNoChecksum(TestPread.java:280)
> {noformat}
> Few lines in the log before hanging.
> {noformat}
> 2017-10-26 20:21:07,938 WARN impl.BlockReaderFactory
> (BlockReaderFactory.java:getRemoteBlockReaderFromTcp(758)) - I/O error
> constructing remote block reader.
> java.net.ConnectException: Connection refused
> at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
> at
> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
> at
> org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
> at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:534)
> at
> org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:2955)
> at
> org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:815)
> at
> org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:740)
> at
> org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.build(BlockReaderFactory.java:385)
> at
> org.apache.hadoop.hdfs.DFSInputStream.getBlockReader(DFSInputStream.java:708)
> at
> org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1230)
> at
> org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1198)
> at
> org.apache.hadoop.hdfs.DFSInputStream.access$000(DFSInputStream.java:97)
> at
> org.apache.hadoop.hdfs.DFSInputStream$2.call(DFSInputStream.java:1182)
> at
> org.apache.hadoop.hdfs.DFSInputStream$2.call(DFSInputStream.java:1174)
> at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> at
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
> at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> at java.lang.Thread.run(Thread.java:745)
> 2017-10-26 20:21:07,938 WARN hdfs.DFSClient
> (DFSInputStream.java:actualGetFromOneDataNode(1270)) - Connection failure:
> Failed to connect to /127.0.0.1:42357 for file /preadtest.dat for block
> BP-287215640-172.17.0.18-1509049266453:blk_1073741825_1001:java.net.ConnectException:
> Connection refused
> java.net.ConnectException: Connection refused
> at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
> at
> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
> at
> org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
> at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:534)
> at
> org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:2955)
> at
> org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:815)
> at
> org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:740)
> at
> org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.build(BlockReaderFactory.java:385)
> at
> org.apache.hadoop.hdfs.DFSInputStream.getBlockReader(DFSInputStream.java:708)
> at
> org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1230)
> at
> org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1198)
> at
> org.apache.hadoop.hdfs.DFSInputStream.access$000(DFSInputStream.java:97)
> at
> org.apache.hadoop.hdfs.DFSInputStream$2.call(DFSInputStream.java:1182)
> at
> org.apache.hadoop.hdfs.DFSInputStream$2.call(DFSInputStream.java:1174)
> at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> at
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
> at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> at java.lang.Thread.run(Thread.java:745)
> 2017-10-26 20:21:07,939 WARN hdfs.DFSClient
> (DFSInputStream.java:getBestNodeDNAddrPair(1112)) - No live nodes contain
> block BP-287215640-172.17.0.18-1509049266453:blk_1073741825_1001 after
> checking nodes =
> [DatanodeInfoWithStorage[127.0.0.1:36669,DS-3ce766bc-dad8-4022-b0d8-396a669ee4b8,DISK],
>
> DatanodeInfoWithStorage[127.0.0.1:36005,DS-bd7b59f4-a7de-4524-877e-c0f9a10ce5d5,DISK],
>
> DatanodeInfoWithStorage[127.0.0.1:42357,DS-a42e4a89-3985-4dc7-ad6a-c0bcf078ccae,DISK]],
> ignoredNodes =
> [DatanodeInfoWithStorage[127.0.0.1:36005,DS-bd7b59f4-a7de-4524-877e-c0f9a10ce5d5,DISK],
>
> DatanodeInfoWithStorage[127.0.0.1:42357,DS-a42e4a89-3985-4dc7-ad6a-c0bcf078ccae,DISK]]
> 2017-10-26 20:31:30,435 INFO datanode.DirectoryScanner
> (DirectoryScanner.java:scan(667)) - BlockPool
> BP-287215640-172.17.0.18-1509049266453 Total blocks: 12, missing metadata
> files:0, missing block files:0, missing blocks in memory:0, mismatched
> blocks:0
> {noformat}
--
This message was sent by Atlassian JIRA
(v6.4.14#64029)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]