Workaroung 1 resolved the issue. -zanber
Rasit OZDAS wrote: > > Sandy, I have no idea about your issue :( > > Zander, > Your problem is probably about this JIRA issue: > http://issues.apache.org/jira/browse/HADOOP-1212 > > Here is 2 workarounds explained: > http://www.michael-noll.com/wiki/Running_Hadoop_On_Ubuntu_Linux_(Multi-Node_Cluster)#java.io.IOException:_Incompatible_namespaceIDs > > I haven't tried it, hope it helps. > Rasit > > 2009/2/17 zander1013 <zander1...@gmail.com>: >> >> hi, >> >> i am not seeing the DataNode run either. but i am seeing an extra process >> TaskTracker run. >> >> here is what hapens when i start the cluster run jps and stop the >> cluster... >> >> had...@node0:/usr/local/hadoop$ bin/start-all.sh >> starting namenode, logging to >> /usr/local/hadoop/bin/../logs/hadoop-hadoop-namenode-node0.out >> node0.local: starting datanode, logging to >> /usr/local/hadoop/bin/../logs/hadoop-hadoop-datanode-node0.out >> node1.local: starting datanode, logging to >> /usr/local/hadoop/bin/../logs/hadoop-hadoop-datanode-node1.out >> node0.local: starting secondarynamenode, logging to >> /usr/local/hadoop/bin/../logs/hadoop-hadoop-secondarynamenode-node0.out >> starting jobtracker, logging to >> /usr/local/hadoop/bin/../logs/hadoop-hadoop-jobtracker-node0.out >> node0.local: starting tasktracker, logging to >> /usr/local/hadoop/bin/../logs/hadoop-hadoop-tasktracker-node0.out >> node1.local: starting tasktracker, logging to >> /usr/local/hadoop/bin/../logs/hadoop-hadoop-tasktracker-node1.out >> had...@node0:/usr/local/hadoop$ jps >> 13353 TaskTracker >> 13126 SecondaryNameNode >> 12846 NameNode >> 13455 Jps >> 13232 JobTracker >> had...@node0:/usr/local/hadoop$ bin/stop-all.sh >> stopping jobtracker >> node0.local: stopping tasktracker >> node1.local: stopping tasktracker >> stopping namenode >> node0.local: no datanode to stop >> node1.local: no datanode to stop >> node0.local: stopping secondarynamenode >> had...@node0:/usr/local/hadoop$ >> >> here is the tail of the log file for the session above... >> ************************************************************/ >> 2009-02-16 19:35:13,999 INFO >> org.apache.hadoop.hdfs.server.datanode.DataNode: STARTUP_MSG: >> /************************************************************ >> STARTUP_MSG: Starting DataNode >> STARTUP_MSG: host = node1/127.0.1.1 >> STARTUP_MSG: args = [] >> STARTUP_MSG: version = 0.19.0 >> STARTUP_MSG: build = >> https://svn.apache.org/repos/asf/hadoop/core/branches/branch-0.19 -r >> 713890; >> compiled by 'ndaley' on Fri Nov 14 03:12:29 UTC 2008 >> ************************************************************/ >> 2009-02-16 19:35:18,999 ERROR >> org.apache.hadoop.hdfs.server.datanode.DataNode: java.io.IOException: >> Incompatible namespaceIDs in >> /usr/local/hadoop-datastore/hadoop-hadoop/dfs/data: namenode namespaceID >> = >> 1050914495; datanode namespaceID = 722953254 >> at >> org.apache.hadoop.hdfs.server.datanode.DataStorage.doTransition(DataStorage.java:233) >> at >> org.apache.hadoop.hdfs.server.datanode.DataStorage.recoverTransitionRead(DataStorage.java:148) >> at >> org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:287) >> at >> org.apache.hadoop.hdfs.server.datanode.DataNode.<init>(DataNode.java:205) >> at >> org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1199) >> at >> org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1154) >> at >> org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:1162) >> at >> org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:1284) >> >> 2009-02-16 19:35:19,000 INFO >> org.apache.hadoop.hdfs.server.datanode.DataNode: SHUTDOWN_MSG: >> /************************************************************ >> SHUTDOWN_MSG: Shutting down DataNode at node1/127.0.1.1 >> ************************************************************/ >> >> i have not seen DataNode run yet. i have only started and stopped the >> cluster a couple of times. >> >> i tried to reformat datanode and namenode with bin/hadoop datanode >> -format >> and bin/hadoop namenode -format from /usr/local/hadoop dir. >> >> please advise >> >> zander >> >> >> >> Mithila Nagendra wrote: >>> >>> Hey Sandy >>> I had a similar problem with Hadoop. All I did was I stopped all the >>> daemons >>> using stop-all.sh. Then formatted the namenode again using hadoop >>> namenode >>> -format. After this I went on to restarting everything by using >>> start-all.sh >>> >>> I hope you dont have much data on the datanode, reformatting it would >>> erase >>> everything out. >>> >>> Hope this helps! >>> Mithila >>> >>> >>> >>> On Sat, Feb 14, 2009 at 2:39 AM, james warren <ja...@rockyou.com> wrote: >>> >>>> Sandy - >>>> >>>> I suggest you take a look into your NameNode and DataNode logs. From >>>> the >>>> information posted, these likely would be at >>>> >>>> >>>> /Users/hadoop/hadoop-0.18.2/bin/../logs/hadoop-hadoop-namenode-loteria.cs.tamu.edu.log >>>> >>>> /Users/hadoop/hadoop-0.18.2/bin/../logs/hadoop-hadoop-jobtracker-loteria.cs.tamu.edu.log >>>> >>>> If the cause isn't obvious from what you see there, could you please >>>> post >>>> the last few lines from each log? >>>> >>>> -jw >>>> >>>> On Fri, Feb 13, 2009 at 3:28 PM, Sandy <snickerdoodl...@gmail.com> >>>> wrote: >>>> >>>> > Hello, >>>> > >>>> > I would really appreciate any help I can get on this! I've suddenly >>>> ran >>>> > into >>>> > a very strange error. >>>> > >>>> > when I do: >>>> > bin/start-all >>>> > I get: >>>> > hadoop$ bin/start-all.sh >>>> > starting namenode, logging to >>>> > >>>> > >>>> /Users/hadoop/hadoop-0.18.2/bin/../logs/hadoop-hadoop-namenode-loteria.cs.tamu.edu.out >>>> > starting jobtracker, logging to >>>> > >>>> > >>>> /Users/hadoop/hadoop-0.18.2/bin/../logs/hadoop-hadoop-jobtracker-loteria.cs.tamu.edu.out >>>> > >>>> > No datanode, secondary namenode or jobtracker are being started. >>>> > >>>> > When I try to upload anything on the dfs, I get a "node in safemode" >>>> error >>>> > (even after waiting 5 minutes), presumably because it's trying to >>>> reach >>>> a >>>> > datanode that does not exist. The same "safemode" error occurs when >>>> I >>>> try >>>> > to run jobs. >>>> > >>>> > I have tried bin/stop-all and then bin/start-all again. I get the >>>> same >>>> > problem! >>>> > >>>> > This is incredibly strange, since I was previously able to start and >>>> run >>>> > jobs without any issue using this version on this machine. I am >>>> running >>>> > jobs >>>> > on a single Mac Pro running OS X 10.5 >>>> > >>>> > I have tried updating to hadoop-0.19.0, and I get the same problem. I >>>> have >>>> > even tried this using previous versions, and I'm getting the same >>>> problem! >>>> > >>>> > Anyone have any idea why this suddenly could be happening? What am I >>>> doing >>>> > wrong? >>>> > >>>> > For convenience, I'm including portions of both conf/hadoop-env.sh >>>> and >>>> > conf/hadoop-site.xml: >>>> > >>>> > --- hadoop-env.sh --- >>>> > # Set Hadoop-specific environment variables here. >>>> > >>>> > # The only required environment variable is JAVA_HOME. All others >>>> are >>>> > # optional. When running a distributed configuration it is best to >>>> > # set JAVA_HOME in this file, so that it is correctly defined on >>>> > # remote nodes. >>>> > >>>> > # The java implementation to use. Required. >>>> > export >>>> > >>>> JAVA_HOME=/System/Library/Frameworks/JavaVM.framework/Versions/1.6.0/Home >>>> > >>>> > # Extra Java CLASSPATH elements. Optional. >>>> > # export HADOOP_CLASSPATH= >>>> > >>>> > # The maximum amount of heap to use, in MB. Default is 1000. >>>> > export HADOOP_HEAPSIZE=3000 >>>> > ... >>>> > --- hadoop-site.xml --- >>>> > <configuration> >>>> > >>>> > <property> >>>> > <name>hadoop.tmp.dir</name> >>>> > <value>/Users/hadoop/hadoop-0.18.2/hadoop-${user.name}</value> >>>> > <description>A base for other temporary directories.</description> >>>> > </property> >>>> > >>>> > <property> >>>> > <name>fs.default.name</name> >>>> > <value>hdfs://localhost:9000</value> >>>> > <description>The name of the default file system. A URI whose >>>> > scheme and authority determine the FileSystem implementation. The >>>> > uri's scheme determines the config property (fs.SCHEME.impl) naming >>>> > the FileSystem implementation class. The uri's authority is used to >>>> > determine the host, port, etc. for a filesystem.</description> >>>> > </property> >>>> > >>>> > <property> >>>> > <name>mapred.job.tracker</name> >>>> > <value>localhost:9001</value> >>>> > <description>The host and port that the MapReduce job tracker runs >>>> > at. If "local", then jobs are run in-process as a single map >>>> > and reduce task. >>>> > </description> >>>> > </property> >>>> > >>>> > <property> >>>> > <name>mapred.tasktracker.tasks.maximum</name> >>>> > <value>1</value> >>>> > <description>The maximum number of tasks that will be run >>>> simultaneously >>>> by >>>> > a >>>> > a task tracker >>>> > </description> >>>> > </property> >>>> > ... >>>> > >>>> >>> >>> >> >> -- >> View this message in context: >> http://www.nabble.com/datanode-not-being-started-tp22006929p22049288.html >> Sent from the Hadoop core-user mailing list archive at Nabble.com. >> >> > > > > -- > M. Raşit ÖZDAŞ > > -- View this message in context: http://www.nabble.com/datanode-not-being-started-tp22006929p22065291.html Sent from the Hadoop core-user mailing list archive at Nabble.com.