Sandy, I have no idea about your issue :( Zander, Your problem is probably about this JIRA issue: http://issues.apache.org/jira/browse/HADOOP-1212
Here is 2 workarounds explained: http://www.michael-noll.com/wiki/Running_Hadoop_On_Ubuntu_Linux_(Multi-Node_Cluster)#java.io.IOException:_Incompatible_namespaceIDs I haven't tried it, hope it helps. Rasit 2009/2/17 zander1013 <zander1...@gmail.com>: > > hi, > > i am not seeing the DataNode run either. but i am seeing an extra process > TaskTracker run. > > here is what hapens when i start the cluster run jps and stop the cluster... > > had...@node0:/usr/local/hadoop$ bin/start-all.sh > starting namenode, logging to > /usr/local/hadoop/bin/../logs/hadoop-hadoop-namenode-node0.out > node0.local: starting datanode, logging to > /usr/local/hadoop/bin/../logs/hadoop-hadoop-datanode-node0.out > node1.local: starting datanode, logging to > /usr/local/hadoop/bin/../logs/hadoop-hadoop-datanode-node1.out > node0.local: starting secondarynamenode, logging to > /usr/local/hadoop/bin/../logs/hadoop-hadoop-secondarynamenode-node0.out > starting jobtracker, logging to > /usr/local/hadoop/bin/../logs/hadoop-hadoop-jobtracker-node0.out > node0.local: starting tasktracker, logging to > /usr/local/hadoop/bin/../logs/hadoop-hadoop-tasktracker-node0.out > node1.local: starting tasktracker, logging to > /usr/local/hadoop/bin/../logs/hadoop-hadoop-tasktracker-node1.out > had...@node0:/usr/local/hadoop$ jps > 13353 TaskTracker > 13126 SecondaryNameNode > 12846 NameNode > 13455 Jps > 13232 JobTracker > had...@node0:/usr/local/hadoop$ bin/stop-all.sh > stopping jobtracker > node0.local: stopping tasktracker > node1.local: stopping tasktracker > stopping namenode > node0.local: no datanode to stop > node1.local: no datanode to stop > node0.local: stopping secondarynamenode > had...@node0:/usr/local/hadoop$ > > here is the tail of the log file for the session above... > ************************************************************/ > 2009-02-16 19:35:13,999 INFO > org.apache.hadoop.hdfs.server.datanode.DataNode: STARTUP_MSG: > /************************************************************ > STARTUP_MSG: Starting DataNode > STARTUP_MSG: host = node1/127.0.1.1 > STARTUP_MSG: args = [] > STARTUP_MSG: version = 0.19.0 > STARTUP_MSG: build = > https://svn.apache.org/repos/asf/hadoop/core/branches/branch-0.19 -r 713890; > compiled by 'ndaley' on Fri Nov 14 03:12:29 UTC 2008 > ************************************************************/ > 2009-02-16 19:35:18,999 ERROR > org.apache.hadoop.hdfs.server.datanode.DataNode: java.io.IOException: > Incompatible namespaceIDs in > /usr/local/hadoop-datastore/hadoop-hadoop/dfs/data: namenode namespaceID = > 1050914495; datanode namespaceID = 722953254 > at > org.apache.hadoop.hdfs.server.datanode.DataStorage.doTransition(DataStorage.java:233) > at > org.apache.hadoop.hdfs.server.datanode.DataStorage.recoverTransitionRead(DataStorage.java:148) > at > org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:287) > at > org.apache.hadoop.hdfs.server.datanode.DataNode.<init>(DataNode.java:205) > at > org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1199) > at > org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1154) > at > org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:1162) > at > org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:1284) > > 2009-02-16 19:35:19,000 INFO > org.apache.hadoop.hdfs.server.datanode.DataNode: SHUTDOWN_MSG: > /************************************************************ > SHUTDOWN_MSG: Shutting down DataNode at node1/127.0.1.1 > ************************************************************/ > > i have not seen DataNode run yet. i have only started and stopped the > cluster a couple of times. > > i tried to reformat datanode and namenode with bin/hadoop datanode -format > and bin/hadoop namenode -format from /usr/local/hadoop dir. > > please advise > > zander > > > > Mithila Nagendra wrote: >> >> Hey Sandy >> I had a similar problem with Hadoop. All I did was I stopped all the >> daemons >> using stop-all.sh. Then formatted the namenode again using hadoop namenode >> -format. After this I went on to restarting everything by using >> start-all.sh >> >> I hope you dont have much data on the datanode, reformatting it would >> erase >> everything out. >> >> Hope this helps! >> Mithila >> >> >> >> On Sat, Feb 14, 2009 at 2:39 AM, james warren <ja...@rockyou.com> wrote: >> >>> Sandy - >>> >>> I suggest you take a look into your NameNode and DataNode logs. From the >>> information posted, these likely would be at >>> >>> >>> /Users/hadoop/hadoop-0.18.2/bin/../logs/hadoop-hadoop-namenode-loteria.cs.tamu.edu.log >>> >>> /Users/hadoop/hadoop-0.18.2/bin/../logs/hadoop-hadoop-jobtracker-loteria.cs.tamu.edu.log >>> >>> If the cause isn't obvious from what you see there, could you please post >>> the last few lines from each log? >>> >>> -jw >>> >>> On Fri, Feb 13, 2009 at 3:28 PM, Sandy <snickerdoodl...@gmail.com> wrote: >>> >>> > Hello, >>> > >>> > I would really appreciate any help I can get on this! I've suddenly ran >>> > into >>> > a very strange error. >>> > >>> > when I do: >>> > bin/start-all >>> > I get: >>> > hadoop$ bin/start-all.sh >>> > starting namenode, logging to >>> > >>> > >>> /Users/hadoop/hadoop-0.18.2/bin/../logs/hadoop-hadoop-namenode-loteria.cs.tamu.edu.out >>> > starting jobtracker, logging to >>> > >>> > >>> /Users/hadoop/hadoop-0.18.2/bin/../logs/hadoop-hadoop-jobtracker-loteria.cs.tamu.edu.out >>> > >>> > No datanode, secondary namenode or jobtracker are being started. >>> > >>> > When I try to upload anything on the dfs, I get a "node in safemode" >>> error >>> > (even after waiting 5 minutes), presumably because it's trying to reach >>> a >>> > datanode that does not exist. The same "safemode" error occurs when I >>> try >>> > to run jobs. >>> > >>> > I have tried bin/stop-all and then bin/start-all again. I get the same >>> > problem! >>> > >>> > This is incredibly strange, since I was previously able to start and >>> run >>> > jobs without any issue using this version on this machine. I am running >>> > jobs >>> > on a single Mac Pro running OS X 10.5 >>> > >>> > I have tried updating to hadoop-0.19.0, and I get the same problem. I >>> have >>> > even tried this using previous versions, and I'm getting the same >>> problem! >>> > >>> > Anyone have any idea why this suddenly could be happening? What am I >>> doing >>> > wrong? >>> > >>> > For convenience, I'm including portions of both conf/hadoop-env.sh and >>> > conf/hadoop-site.xml: >>> > >>> > --- hadoop-env.sh --- >>> > # Set Hadoop-specific environment variables here. >>> > >>> > # The only required environment variable is JAVA_HOME. All others are >>> > # optional. When running a distributed configuration it is best to >>> > # set JAVA_HOME in this file, so that it is correctly defined on >>> > # remote nodes. >>> > >>> > # The java implementation to use. Required. >>> > export >>> > >>> JAVA_HOME=/System/Library/Frameworks/JavaVM.framework/Versions/1.6.0/Home >>> > >>> > # Extra Java CLASSPATH elements. Optional. >>> > # export HADOOP_CLASSPATH= >>> > >>> > # The maximum amount of heap to use, in MB. Default is 1000. >>> > export HADOOP_HEAPSIZE=3000 >>> > ... >>> > --- hadoop-site.xml --- >>> > <configuration> >>> > >>> > <property> >>> > <name>hadoop.tmp.dir</name> >>> > <value>/Users/hadoop/hadoop-0.18.2/hadoop-${user.name}</value> >>> > <description>A base for other temporary directories.</description> >>> > </property> >>> > >>> > <property> >>> > <name>fs.default.name</name> >>> > <value>hdfs://localhost:9000</value> >>> > <description>The name of the default file system. A URI whose >>> > scheme and authority determine the FileSystem implementation. The >>> > uri's scheme determines the config property (fs.SCHEME.impl) naming >>> > the FileSystem implementation class. The uri's authority is used to >>> > determine the host, port, etc. for a filesystem.</description> >>> > </property> >>> > >>> > <property> >>> > <name>mapred.job.tracker</name> >>> > <value>localhost:9001</value> >>> > <description>The host and port that the MapReduce job tracker runs >>> > at. If "local", then jobs are run in-process as a single map >>> > and reduce task. >>> > </description> >>> > </property> >>> > >>> > <property> >>> > <name>mapred.tasktracker.tasks.maximum</name> >>> > <value>1</value> >>> > <description>The maximum number of tasks that will be run >>> simultaneously >>> by >>> > a >>> > a task tracker >>> > </description> >>> > </property> >>> > ... >>> > >>> >> >> > > -- > View this message in context: > http://www.nabble.com/datanode-not-being-started-tp22006929p22049288.html > Sent from the Hadoop core-user mailing list archive at Nabble.com. > > -- M. Raşit ÖZDAŞ