[ https://issues.apache.org/jira/browse/HDFS-6953?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Matt Foley reopened HDFS-6953: ------------------------------ > HDFS file append failing in single node configuration > ----------------------------------------------------- > > Key: HDFS-6953 > URL: https://issues.apache.org/jira/browse/HDFS-6953 > Project: Hadoop HDFS > Issue Type: Bug > Environment: Ubuntu 12.01, Apache Hadoop 2.5.0 single node > configuration > Reporter: Vladislav Falfushinsky > Attachments: Main.java, core-site.xml, hdfs-site.xml, test_hdfs.c > > > The following issue happens in both fully distributed and single node setup. > I have looked to the thread(https://issues.apache.org/jira/browse/HDFS-4600) > about simiral issue in multinode cluster and made some changes of my > configuration however it does not changed anything. The configuration files > and application sources are attached. > Steps to reproduce: > $ ./test_hdfs > 2014-08-27 14:23:08,472 WARN [Thread-5] hdfs.DFSClient > (DFSOutputStream.java:run(628)) - DataStreamer Exception > java.io.IOException: Failed to replace a bad datanode on the existing > pipeline due to no more good datanodes being available to try. (Nodes: > current=[127.0.0.1:50010], original=[127.0.0.1:50010]). The current failed > datanode replacement policy is DEFAULT, and a client may configure this via > 'dfs.client.block.write.replace-datanode-on-failure.policy' in its > configuration. > at > org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.findNewDatanode(DFSOutputStream.java:969) > at > org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.addDatanode2ExistingPipeline(DFSOutputStream.java:1035) > at > org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.setupPipelineForAppendOrRecovery(DFSOutputStream.java:1184) > at > org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:532) > FSDataOutputStream#close error: > java.io.IOException: Failed to replace a bad datanode on the existing > pipeline due to no more good datanodes being available to try. (Nodes: > current=[127.0.0.1:50010], original=[127.0.0.1:50010]). The current failed > datanode replacement policy is DEFAULT, and a client may configure this via > 'dfs.client.block.write.replace-datanode-on-failure.policy' in its > configuration. > at > org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.findNewDatanode(DFSOutputStream.java:969) > at > org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.addDatanode2ExistingPipeline(DFSOutputStream.java:1035) > at > org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.setupPipelineForAppendOrRecovery(DFSOutputStream.java:1184) > at > org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:532) > I have tried to run a simple example in java, that uses append function. It > failed too. > I have tried to get hadoop environment settings from java application. It has > shown the default ones. Not the settings that ones that are mentioned in > core-site.xml and hdfs-site.xml files. -- This message was sent by Atlassian JIRA (v6.3.4#6332)