Just to follow up, the box is up and running now. The command I was trying to start/add just the new box must be wrong. A stop-all/start-all on the cluster got it to start.
On 3/19/10 1:30 PM, "Scott Kester" <skes...@weather.com> wrote: > Just to be sure, I verfied the fs.default.name is set in core-site.xml. The > config files are identical to the other 10 working datanodes as they were > scp'd from a central location. > >> more core-site.xml > <?xml version="1.0"?> > <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> > > <!-- Put site-specific property overrides in this file. --> > > <configuration> > <property> > <name>fs.default.name</name> > <value>hdfs://hadoopxxxxx.xx.weather.com:9000/</value> > <final>true</final> > </property> > > > > On 3/19/10 12:06 PM, "Jean-Daniel Cryans" <jdcry...@apache.org> wrote: > >> I would argue that it could be the same problem. Most of the users get >> it because the fs.default.name conf is setup wrong when they start the >> cluster the first time so they see it on the namenode first but all >> the other components do the same thing. Can you verify that conf on >> your new node? >> >> J-D >> >> On Fri, Mar 19, 2010 at 5:14 AM, SKester <skes...@weather.com> wrote: >>> Google was my first step in trying to track this down. I found the issue >>> you listed, but that is a namenode related problem. Our namenode and the >>> other 10 datanodes are just fine. Only the new box is having this problem. >>> >>> >>> On 3/18/10 4:41 PM, "Jean-Daniel Cryans" <jdcry...@apache.org> wrote: >>> >>>> Google is your friend ;) >>>> >>>> https://issues.apache.org/jira/browse/HADOOP-5687 >>>> >>>> J-D >>>> >>>> On Thu, Mar 18, 2010 at 1:29 PM, Scott <skes...@weather.com> wrote: >>>>> We have a working 10 node cluster and are trying to add an 11th box >>>>> (insert >>>>> Spinal Tap joke here). The box (Centos Linux) was built in an identical >>>>> manner to the other 10 and has the same version of hadoop (0.20.2). The >>>>> configs are the exact same as the other nodes. However when trying to >>>>> start >>>>> the hadoop daemons it throws a NPE. Here is all that is written to the >>>>> logs. Any idea whats causing this? >>>>> >>>>> ************************************************************/ >>>>> 2010-03-18 16:09:42,993 INFO >>>>> org.apache.hadoop.hdfs.server.datanode.DataNode: STARTUP_MSG: >>>>> /************************************************************ >>>>> STARTUP_MSG: Starting DataNode >>>>> STARTUP_MSG: host = hadoop0b10/192.168.60.100 >>>>> STARTUP_MSG: args = [] >>>>> STARTUP_MSG: version = 0.20.2 >>>>> STARTUP_MSG: build = >>>>> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20 -r >>>>> 911707; compiled by 'chrisdo' on Fri Feb >>>>> 19 08:07:34 UTC 2010 >>>>> ************************************************************/ >>>>> 2010-03-18 16:09:43,058 ERROR >>>>> org.apache.hadoop.hdfs.server.datanode.DataNode: >>>>> java.lang.NullPointerException >>>>> at org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:134) >>>>> at >>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:1>>>>> 5 >>>>> 6) >>>>> at >>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:1>>>>> 6 >>>>> 0) >>>>> at >>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.jav>>>>> a >>>>> :2 >>>>> 46) >>>>> at >>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.<init>(DataNode.java:216) >>>>> at >>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java>>>>> : >>>>> 12 >>>>> 83) >>>>> at >>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNo>>>>> d >>>>> e. >>>>> java:1238) >>>>> at >>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.ja>>>>> v >>>>> a: >>>>> 1246) >>>>> at >>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:1368) >>>>> >>>>> 2010-03-18 16:09:43,059 INFO >>>>> org.apache.hadoop.hdfs.server.datanode.DataNode: SHUTDOWN_MSG: >>>>> /************************************************************ >>>>> SHUTDOWN_MSG: Shutting down DataNode at hadoop0b10/192.168.60.100 >>>>> ************************************************************/ >>>>> >>>>> >>> >>> >