So I'm new to Hadoop and I have been trying unsuccessfully to work
through the Quickstart tutorial to get a single node working in
pseudo-distributed mode. I can't seem to put data into HDFS using
release 0.18.2 under Java 1.6.0_04-b12:

$ bin/hadoop fs -put conf input
08/11/05 18:32:23 INFO dfs.DFSClient:
org.apache.hadoop.ipc.RemoteException: java.io.IOException: File
/user/slaurent/input/commons-logging.properties could only be
replicated to 0 nodes, instead of 1
...

The dfshealth jsp page reports 1 live datanode. The strange thing is
that the node listed as "dkz216" with a url of
"http://dkz216.neoplus.adsl.tpnet.pl:50075/browseDirectory.jsp?namenodeInfoPort=50070&dir=%2F";...
not sure where that came from.

No errors in the log files, other than the replication error. However,
I do see one other oddity in the datanode logfile:

---hadoop-user-datanode-server.log---
2008-11-05 18:32:28,317 INFO org.apache.hadoop.dfs.DataNode:
dnRegistration =
DatanodeRegistration(dkz216.neoplus.adsl.tpnet.pl:50010, storageID=,
infoPort=50075, ipcPort=50020)
2008-11-05 18:32:28,317 INFO org.apache.hadoop.ipc.Server: IPC Server
handler 2 on 50020: starting
2008-11-05 18:32:28,443 INFO org.apache.hadoop.dfs.DataNode: New
storage id DS-2140500399-83.24.29.216-50010-1225931548407 is assigned
to data-node 127.0.0.1:50010
2008-11-05 18:32:28,444 INFO org.apache.hadoop.dfs.DataNode:
DatanodeRegistration(127.0.0.1:50010,
storageID=DS-2140500399-83.24.29.216-50010-1225931548407,
infoPort=50075, ipcPort=50020)In DataNode.run, data =
FSDataset{dirpath='/tmp/hadoop-slaurent/dfs/data/current'}
---hadoop-user-datanode-server.log---


Here are my config files:

---hadoop-site.xml---
<?xml version="1.0"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>

<!-- Put site-specific property overrides in this file. -->

<configuration>
  <property>
    <name>fs.default.name</name>
    <value>hdfs://127.0.0.1:9000/</value>
  </property>
  <property>
    <name>mapred.job.tracker</name>
    <value>127.0.0.1:9001</value>
  </property>
  <property>
    <name>dfs.replication</name>
    <value>1</value>
  </property>
</configuration>
---hadoop-site.xml---

---masters----
127.0.0.1
---masters----

---slaves---
127.0.0.1
---slaves---

I originally started with localhost everywhere but then switched to
127.0.0.1 to see if that helped. No luck. I can't seem to copy any
files to HDFS.

Any suggestions would be greatly appreciated!

-Sean

Reply via email to