hadoop-hdfs-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Rushabh S Shah (JIRA)" <j...@apache.org>
Subject [jira] [Updated] (HDFS-12727) TestPread timing out on branch-2.8
Date Thu, 26 Oct 2017 21:07:00 GMT

     [ https://issues.apache.org/jira/browse/HDFS-12727?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]

Rushabh S Shah updated HDFS-12727:
----------------------------------
    Description: 
TestPread timing out on branch-2.8 and not on trunk.
{noformat}
2017-10-24 19:47:37,377 WARN impl.BlockReaderFactory (BlockReaderFactory.java:getRemoteBlockReaderFromTcp(758))
- I/O error constructing remote block reader.
java.net.ConnectException: Connection refused
at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:534)
at org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:2955)
at org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:815)
at org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:740)
at org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.build(BlockReaderFactory.java:385)
at org.apache.hadoop.hdfs.DFSInputStream.getBlockReader(DFSInputStream.java:708)
at org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1230)
at org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1198)
at org.apache.hadoop.hdfs.DFSInputStream.fetchBlockByteRange(DFSInputStream.java:1158)
at org.apache.hadoop.hdfs.DFSInputStream.pread(DFSInputStream.java:1535)
at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:1501)
at org.apache.hadoop.fs.FSInputStream.readFully(FSInputStream.java:121)
at org.apache.hadoop.fs.FSDataInputStream.readFully(FSDataInputStream.java:120)
at org.apache.hadoop.hdfs.TestPread.datanodeRestartTest(TestPread.java:245)
at org.apache.hadoop.hdfs.TestPread.dfsPreadTest(TestPread.java:478)
at org.apache.hadoop.hdfs.TestPread.testPreadDFSNoChecksum(TestPread.java:280)
{noformat}


Few lines in the log before hanging.
{noformat}

2017-10-26 20:21:07,938 WARN  impl.BlockReaderFactory (BlockReaderFactory.java:getRemoteBlockReaderFromTcp(758))
- I/O error constructing remote block reader.
java.net.ConnectException: Connection refused
	at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
	at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
	at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
	at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:534)
	at org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:2955)
	at org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:815)
	at org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:740)
	at org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.build(BlockReaderFactory.java:385)
	at org.apache.hadoop.hdfs.DFSInputStream.getBlockReader(DFSInputStream.java:708)
	at org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1230)
	at org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1198)
	at org.apache.hadoop.hdfs.DFSInputStream.access$000(DFSInputStream.java:97)
	at org.apache.hadoop.hdfs.DFSInputStream$2.call(DFSInputStream.java:1182)
	at org.apache.hadoop.hdfs.DFSInputStream$2.call(DFSInputStream.java:1174)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
	at java.lang.Thread.run(Thread.java:745)
2017-10-26 20:21:07,938 WARN  hdfs.DFSClient (DFSInputStream.java:actualGetFromOneDataNode(1270))
- Connection failure: Failed to connect to /127.0.0.1:42357 for file /preadtest.dat for block
BP-287215640-172.17.0.18-1509049266453:blk_1073741825_1001:java.net.ConnectException: Connection
refused
java.net.ConnectException: Connection refused
	at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
	at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
	at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
	at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:534)
	at org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:2955)
	at org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:815)
	at org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:740)
	at org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.build(BlockReaderFactory.java:385)
	at org.apache.hadoop.hdfs.DFSInputStream.getBlockReader(DFSInputStream.java:708)
	at org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1230)
	at org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1198)
	at org.apache.hadoop.hdfs.DFSInputStream.access$000(DFSInputStream.java:97)
	at org.apache.hadoop.hdfs.DFSInputStream$2.call(DFSInputStream.java:1182)
	at org.apache.hadoop.hdfs.DFSInputStream$2.call(DFSInputStream.java:1174)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
	at java.lang.Thread.run(Thread.java:745)
2017-10-26 20:21:07,939 WARN  hdfs.DFSClient (DFSInputStream.java:getBestNodeDNAddrPair(1112))
- No live nodes contain block BP-287215640-172.17.0.18-1509049266453:blk_1073741825_1001 after
checking nodes = [DatanodeInfoWithStorage[127.0.0.1:36669,DS-3ce766bc-dad8-4022-b0d8-396a669ee4b8,DISK],
DatanodeInfoWithStorage[127.0.0.1:36005,DS-bd7b59f4-a7de-4524-877e-c0f9a10ce5d5,DISK], DatanodeInfoWithStorage[127.0.0.1:42357,DS-a42e4a89-3985-4dc7-ad6a-c0bcf078ccae,DISK]],
ignoredNodes = [DatanodeInfoWithStorage[127.0.0.1:36005,DS-bd7b59f4-a7de-4524-877e-c0f9a10ce5d5,DISK],
DatanodeInfoWithStorage[127.0.0.1:42357,DS-a42e4a89-3985-4dc7-ad6a-c0bcf078ccae,DISK]]
2017-10-26 20:31:30,435 INFO  datanode.DirectoryScanner (DirectoryScanner.java:scan(667))
- BlockPool BP-287215640-172.17.0.18-1509049266453 Total blocks: 12, missing metadata files:0,
missing block files:0, missing blocks in memory:0, mismatched blocks:0
{noformat}


  was:
TestPread timing out on branch-2.8 and not on trunk.
Few lines before hanging.
{noformat}

2017-10-26 20:21:07,938 WARN  impl.BlockReaderFactory (BlockReaderFactory.java:getRemoteBlockReaderFromTcp(758))
- I/O error constructing remote block reader.
java.net.ConnectException: Connection refused
	at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
	at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
	at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
	at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:534)
	at org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:2955)
	at org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:815)
	at org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:740)
	at org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.build(BlockReaderFactory.java:385)
	at org.apache.hadoop.hdfs.DFSInputStream.getBlockReader(DFSInputStream.java:708)
	at org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1230)
	at org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1198)
	at org.apache.hadoop.hdfs.DFSInputStream.access$000(DFSInputStream.java:97)
	at org.apache.hadoop.hdfs.DFSInputStream$2.call(DFSInputStream.java:1182)
	at org.apache.hadoop.hdfs.DFSInputStream$2.call(DFSInputStream.java:1174)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
	at java.lang.Thread.run(Thread.java:745)
2017-10-26 20:21:07,938 WARN  hdfs.DFSClient (DFSInputStream.java:actualGetFromOneDataNode(1270))
- Connection failure: Failed to connect to /127.0.0.1:42357 for file /preadtest.dat for block
BP-287215640-172.17.0.18-1509049266453:blk_1073741825_1001:java.net.ConnectException: Connection
refused
java.net.ConnectException: Connection refused
	at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
	at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
	at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
	at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:534)
	at org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:2955)
	at org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:815)
	at org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:740)
	at org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.build(BlockReaderFactory.java:385)
	at org.apache.hadoop.hdfs.DFSInputStream.getBlockReader(DFSInputStream.java:708)
	at org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1230)
	at org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1198)
	at org.apache.hadoop.hdfs.DFSInputStream.access$000(DFSInputStream.java:97)
	at org.apache.hadoop.hdfs.DFSInputStream$2.call(DFSInputStream.java:1182)
	at org.apache.hadoop.hdfs.DFSInputStream$2.call(DFSInputStream.java:1174)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
	at java.lang.Thread.run(Thread.java:745)
2017-10-26 20:21:07,939 WARN  hdfs.DFSClient (DFSInputStream.java:getBestNodeDNAddrPair(1112))
- No live nodes contain block BP-287215640-172.17.0.18-1509049266453:blk_1073741825_1001 after
checking nodes = [DatanodeInfoWithStorage[127.0.0.1:36669,DS-3ce766bc-dad8-4022-b0d8-396a669ee4b8,DISK],
DatanodeInfoWithStorage[127.0.0.1:36005,DS-bd7b59f4-a7de-4524-877e-c0f9a10ce5d5,DISK], DatanodeInfoWithStorage[127.0.0.1:42357,DS-a42e4a89-3985-4dc7-ad6a-c0bcf078ccae,DISK]],
ignoredNodes = [DatanodeInfoWithStorage[127.0.0.1:36005,DS-bd7b59f4-a7de-4524-877e-c0f9a10ce5d5,DISK],
DatanodeInfoWithStorage[127.0.0.1:42357,DS-a42e4a89-3985-4dc7-ad6a-c0bcf078ccae,DISK]]
2017-10-26 20:31:30,435 INFO  datanode.DirectoryScanner (DirectoryScanner.java:scan(667))
- BlockPool BP-287215640-172.17.0.18-1509049266453 Total blocks: 12, missing metadata files:0,
missing block files:0, missing blocks in memory:0, mismatched blocks:0
{noformat}



> TestPread timing out on branch-2.8
> ----------------------------------
>
>                 Key: HDFS-12727
>                 URL: https://issues.apache.org/jira/browse/HDFS-12727
>             Project: Hadoop HDFS
>          Issue Type: Bug
>            Reporter: Rushabh S Shah
>
> TestPread timing out on branch-2.8 and not on trunk.
> {noformat}
> 2017-10-24 19:47:37,377 WARN impl.BlockReaderFactory (BlockReaderFactory.java:getRemoteBlockReaderFromTcp(758))
- I/O error constructing remote block reader.
> java.net.ConnectException: Connection refused
> at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
> at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
> at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
> at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:534)
> at org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:2955)
> at org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:815)
> at org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:740)
> at org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.build(BlockReaderFactory.java:385)
> at org.apache.hadoop.hdfs.DFSInputStream.getBlockReader(DFSInputStream.java:708)
> at org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1230)
> at org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1198)
> at org.apache.hadoop.hdfs.DFSInputStream.fetchBlockByteRange(DFSInputStream.java:1158)
> at org.apache.hadoop.hdfs.DFSInputStream.pread(DFSInputStream.java:1535)
> at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:1501)
> at org.apache.hadoop.fs.FSInputStream.readFully(FSInputStream.java:121)
> at org.apache.hadoop.fs.FSDataInputStream.readFully(FSDataInputStream.java:120)
> at org.apache.hadoop.hdfs.TestPread.datanodeRestartTest(TestPread.java:245)
> at org.apache.hadoop.hdfs.TestPread.dfsPreadTest(TestPread.java:478)
> at org.apache.hadoop.hdfs.TestPread.testPreadDFSNoChecksum(TestPread.java:280)
> {noformat}
> Few lines in the log before hanging.
> {noformat}
> 2017-10-26 20:21:07,938 WARN  impl.BlockReaderFactory (BlockReaderFactory.java:getRemoteBlockReaderFromTcp(758))
- I/O error constructing remote block reader.
> java.net.ConnectException: Connection refused
> 	at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
> 	at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
> 	at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
> 	at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:534)
> 	at org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:2955)
> 	at org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:815)
> 	at org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:740)
> 	at org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.build(BlockReaderFactory.java:385)
> 	at org.apache.hadoop.hdfs.DFSInputStream.getBlockReader(DFSInputStream.java:708)
> 	at org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1230)
> 	at org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1198)
> 	at org.apache.hadoop.hdfs.DFSInputStream.access$000(DFSInputStream.java:97)
> 	at org.apache.hadoop.hdfs.DFSInputStream$2.call(DFSInputStream.java:1182)
> 	at org.apache.hadoop.hdfs.DFSInputStream$2.call(DFSInputStream.java:1174)
> 	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> 	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
> 	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> 	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> 	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> 	at java.lang.Thread.run(Thread.java:745)
> 2017-10-26 20:21:07,938 WARN  hdfs.DFSClient (DFSInputStream.java:actualGetFromOneDataNode(1270))
- Connection failure: Failed to connect to /127.0.0.1:42357 for file /preadtest.dat for block
BP-287215640-172.17.0.18-1509049266453:blk_1073741825_1001:java.net.ConnectException: Connection
refused
> java.net.ConnectException: Connection refused
> 	at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
> 	at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
> 	at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
> 	at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:534)
> 	at org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:2955)
> 	at org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:815)
> 	at org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:740)
> 	at org.apache.hadoop.hdfs.client.impl.BlockReaderFactory.build(BlockReaderFactory.java:385)
> 	at org.apache.hadoop.hdfs.DFSInputStream.getBlockReader(DFSInputStream.java:708)
> 	at org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1230)
> 	at org.apache.hadoop.hdfs.DFSInputStream.actualGetFromOneDataNode(DFSInputStream.java:1198)
> 	at org.apache.hadoop.hdfs.DFSInputStream.access$000(DFSInputStream.java:97)
> 	at org.apache.hadoop.hdfs.DFSInputStream$2.call(DFSInputStream.java:1182)
> 	at org.apache.hadoop.hdfs.DFSInputStream$2.call(DFSInputStream.java:1174)
> 	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> 	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
> 	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> 	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> 	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> 	at java.lang.Thread.run(Thread.java:745)
> 2017-10-26 20:21:07,939 WARN  hdfs.DFSClient (DFSInputStream.java:getBestNodeDNAddrPair(1112))
- No live nodes contain block BP-287215640-172.17.0.18-1509049266453:blk_1073741825_1001 after
checking nodes = [DatanodeInfoWithStorage[127.0.0.1:36669,DS-3ce766bc-dad8-4022-b0d8-396a669ee4b8,DISK],
DatanodeInfoWithStorage[127.0.0.1:36005,DS-bd7b59f4-a7de-4524-877e-c0f9a10ce5d5,DISK], DatanodeInfoWithStorage[127.0.0.1:42357,DS-a42e4a89-3985-4dc7-ad6a-c0bcf078ccae,DISK]],
ignoredNodes = [DatanodeInfoWithStorage[127.0.0.1:36005,DS-bd7b59f4-a7de-4524-877e-c0f9a10ce5d5,DISK],
DatanodeInfoWithStorage[127.0.0.1:42357,DS-a42e4a89-3985-4dc7-ad6a-c0bcf078ccae,DISK]]
> 2017-10-26 20:31:30,435 INFO  datanode.DirectoryScanner (DirectoryScanner.java:scan(667))
- BlockPool BP-287215640-172.17.0.18-1509049266453 Total blocks: 12, missing metadata files:0,
missing block files:0, missing blocks in memory:0, mismatched blocks:0
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

---------------------------------------------------------------------
To unsubscribe, e-mail: hdfs-issues-unsubscribe@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-help@hadoop.apache.org


Mime
View raw message