"It" is '/'? On Apr 29, 2013, at 5:09 PM, Mohammad Tariq <[email protected]> wrote:
> make it 755. > > Warm Regards, > Tariq > https://mtariq.jux.com/ > cloudfront.blogspot.com > > > On Tue, Apr 30, 2013 at 3:30 AM, Kevin Burton <[email protected]> > wrote: >> Thank you the HDFS system seems to be up. Now I am having a problem with >> getting the JobTracker and TaskTracker up. According to the logs on the >> JobTracker mapred doesn't have write permission to /. I am not clear on >> what the permissions should be. >> >> Anyway, thank you. >> >> On Apr 29, 2013, at 4:30 PM, Mohammad Tariq <[email protected]> wrote: >> >>> Hello Kevin, >>> >>> Have you reformatted the NN(unsuccessfully)?Was your NN serving >>> some other cluster earlier or your DNs were part of some other >>> cluster?Datanodes bind themselves to namenode through namespaceID and in >>> your case the IDs of DNs and NN seem to be different. As a workaround you >>> could do this : >>> >>> 1- Stop all the daemons. >>> 2- Go to the directory which you have specified as the value of >>> "dfs.name.dir" property in your hdfs-site.xml file. >>> 3- You'll find a directory called "current" inside this directory where a >>> file named "VERSION" will be present. Open this file and copy the value of >>> "namespaceID" form here. >>> 4- Now go to the directory which you have specified as the value of >>> "dfs.data.dir" property in your hdfs-site.xml file. >>> 5- Move inside the "current" directory and open the "VERSION" file here as >>> well. Now replace the value of "namespaceID" present here with the one you >>> had copied earlier. >>> 6- Restart all the daemons. >>> >>> Note : If you have not created dfs.name.dir and dfs.data.dir separately, >>> you could find all this inside your temp directory. >>> >>> HTH >>> >>> Warm Regards, >>> Tariq >>> https://mtariq.jux.com/ >>> cloudfront.blogspot.com >>> >>> >>> On Tue, Apr 30, 2013 at 2:45 AM, <[email protected]> wrote: >>>> I am trying to start up a cluster and in the datanode log on the NameNode >>>> server I get the error: >>>> >>>> 2013-04-29 15:50:20,988 INFO org.apache.hadoop.hdfs.server.common.Storage: >>>> Lock on /data/hadoop/dfs/data/in_use.lock acquired by nodename >>>> 1406@devUbuntu05 >>>> 2013-04-29 15:50:20,990 FATAL >>>> org.apache.hadoop.hdfs.server.datanode.DataNode: Initialization failed for >>>> block pool Block pool BP-1306349046-172.16.26.68-1367256199559 (storage id >>>> DS-403514403-172.16.26.68-50010-1366406077018) service to >>>> devUbuntu05/172.16.26.68:9000 >>>> java.io.IOException: Incompatible clusterIDs in /data/hadoop/dfs/data: >>>> namenode clusterID = CID-23b9f9c7-2c25-411f-8bd2-4d5c9d7c25a1; datanode >>>> clusterID = CID-e3f6b811-c1b4-4778-a31e-14dea8b2cca8 >>>> at >>>> org.apache.hadoop.hdfs.server.datanode.DataStorage.doTransition(DataStorage.java:391) >>>> at >>>> org.apache.hadoop.hdfs.server.datanode.DataStorage.recoverTransitionRead(DataStorage.java:191) >>>> at >>>> org.apache.hadoop.hdfs.server.datanode.DataStorage.recoverTransitionRead(DataStorage.java:219) >>>> >>>> How do I get around this error? What does the error mean? >>>> >>>> Thank you. >>>> >>>> Kevin >
