So I'm new to Hadoop and I have been trying unsuccessfully to work through the Quickstart tutorial to get a single node working in pseudo-distributed mode. I can't seem to put data into HDFS using release 0.18.2 under Java 1.6.0_04-b12:
$ bin/hadoop fs -put conf input 08/11/05 18:32:23 INFO dfs.DFSClient: org.apache.hadoop.ipc.RemoteException: java.io.IOException: File /user/slaurent/input/commons-logging.properties could only be replicated to 0 nodes, instead of 1 ... The dfshealth jsp page reports 1 live datanode. The strange thing is that the node listed as "dkz216" with a url of "http://dkz216.neoplus.adsl.tpnet.pl:50075/browseDirectory.jsp?namenodeInfoPort=50070&dir=%2F"... not sure where that came from. No errors in the log files, other than the replication error. However, I do see one other oddity in the datanode logfile: ---hadoop-user-datanode-server.log--- 2008-11-05 18:32:28,317 INFO org.apache.hadoop.dfs.DataNode: dnRegistration = DatanodeRegistration(dkz216.neoplus.adsl.tpnet.pl:50010, storageID=, infoPort=50075, ipcPort=50020) 2008-11-05 18:32:28,317 INFO org.apache.hadoop.ipc.Server: IPC Server handler 2 on 50020: starting 2008-11-05 18:32:28,443 INFO org.apache.hadoop.dfs.DataNode: New storage id DS-2140500399-83.24.29.216-50010-1225931548407 is assigned to data-node 127.0.0.1:50010 2008-11-05 18:32:28,444 INFO org.apache.hadoop.dfs.DataNode: DatanodeRegistration(127.0.0.1:50010, storageID=DS-2140500399-83.24.29.216-50010-1225931548407, infoPort=50075, ipcPort=50020)In DataNode.run, data = FSDataset{dirpath='/tmp/hadoop-slaurent/dfs/data/current'} ---hadoop-user-datanode-server.log--- Here are my config files: ---hadoop-site.xml--- <?xml version="1.0"?> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> <!-- Put site-specific property overrides in this file. --> <configuration> <property> <name>fs.default.name</name> <value>hdfs://127.0.0.1:9000/</value> </property> <property> <name>mapred.job.tracker</name> <value>127.0.0.1:9001</value> </property> <property> <name>dfs.replication</name> <value>1</value> </property> </configuration> ---hadoop-site.xml--- ---masters---- 127.0.0.1 ---masters---- ---slaves--- 127.0.0.1 ---slaves--- I originally started with localhost everywhere but then switched to 127.0.0.1 to see if that helped. No luck. I can't seem to copy any files to HDFS. Any suggestions would be greatly appreciated! -Sean
