[
https://issues.apache.org/jira/browse/HDFS-8162?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Steve Loughran resolved HDFS-8162.
----------------------------------
Resolution: Cannot Reproduce
Hadoop logs to wherever log4j tells it to.
the standard hadoop log4j.properties does route to stderr
{code}
log4j.appender.console.target=System.err
{code}
If you are seeing it go to stdout, it's your program that's causing it.
see also http://wiki.apache.org/hadoop/InvalidJiraIssues
> Stack trace routed to standard out
> ----------------------------------
>
> Key: HDFS-8162
> URL: https://issues.apache.org/jira/browse/HDFS-8162
> Project: Hadoop HDFS
> Issue Type: Improvement
> Components: libhdfs
> Affects Versions: 2.5.2
> Reporter: Rod
> Priority: Minor
>
> Calling hdfsOpenFile() can generate a stacktrace printout to standard out,
> which can be problematic for caller program which is making use of standard
> out. libhdfs stacktraces should be routed to standard error.
> Example of stacktrace:
> WARN [main] hdfs.BlockReaderFactory
> (BlockReaderFactory.java:getRemoteBlockReaderFromTcp(693)) - I/O error
> constructing remote block reader.
> org.apache.hadoop.net.ConnectTimeoutException: 60000 millis timeout while
> waiting for channel to be ready for connect. ch :
> java.nio.channels.SocketChannel[connection-pending remote=/x.x.x.10:50010]
> at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:533)
> at
> org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:3101)
> at
> org.apache.hadoop.hdfs.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:755)
> at
> org.apache.hadoop.hdfs.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:670)
> at
> org.apache.hadoop.hdfs.BlockReaderFactory.build(BlockReaderFactory.java:337)
> at
> org.apache.hadoop.hdfs.DFSInputStream.blockSeekTo(DFSInputStream.java:576)
> at
> org.apache.hadoop.hdfs.DFSInputStream.readWithStrategy(DFSInputStream.java:800)
> at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:854)
> at
> org.apache.hadoop.fs.FSDataInputStream.read(FSDataInputStream.java:143)
> 2015-04-16 10:32:13,946 WARN [main] hdfs.DFSClient
> (DFSInputStream.java:blockSeekTo(612)) - Failed to connect to /x.x.x.10:50010
> for block, add to deadNodes and continue.
> org.apache.hadoop.net.ConnectTimeoutException: 60000 millis timeout while
> waiting for channel to be ready for connect. ch :
> java.nio.channels.SocketChannel[connection-pending remote=/x.x.x.10:50010]
> org.apache.hadoop.net.ConnectTimeoutException: 60000 millis timeout while
> waiting for channel to be ready for connect. ch :
> java.nio.channels.SocketChannel[connection-pending remote=/x.x.x.10:50010]
> at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:533)
> at
> org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:3101)
> at
> org.apache.hadoop.hdfs.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:755)
> at
> org.apache.hadoop.hdfs.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:670)
> at
> org.apache.hadoop.hdfs.BlockReaderFactory.build(BlockReaderFactory.java:337)
> at
> org.apache.hadoop.hdfs.DFSInputStream.blockSeekTo(DFSInputStream.java:576)
> at
> org.apache.hadoop.hdfs.DFSInputStream.readWithStrategy(DFSInputStream.java:800)
> at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:854)
> at
> org.apache.hadoop.fs.FSDataInputStream.read(FSDataInputStream.java:143)
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)