make it 755. Warm Regards, Tariq https://mtariq.jux.com/ cloudfront.blogspot.com
On Tue, Apr 30, 2013 at 3:30 AM, Kevin Burton <[email protected]>wrote: > Thank you the HDFS system seems to be up. Now I am having a problem with > getting the JobTracker and TaskTracker up. According to the logs on the > JobTracker mapred doesn't have write permission to /. I am not clear on > what the permissions should be. > > Anyway, thank you. > > On Apr 29, 2013, at 4:30 PM, Mohammad Tariq <[email protected]> wrote: > > Hello Kevin, > > Have you reformatted the NN(unsuccessfully)?Was your NN serving > some other cluster earlier or your DNs were part of some other > cluster?Datanodes bind themselves to namenode through namespaceID and in > your case the IDs of DNs and NN seem to be different. As a workaround you > could do this : > > 1- Stop all the daemons. > 2- Go to the directory which you have specified as the value of > "dfs.name.dir" property in your hdfs-site.xml file. > 3- You'll find a directory called "current" inside this directory where a > file named "VERSION" will be present. Open this file and copy the value of > "namespaceID" form here. > 4- Now go to the directory which you have specified as the value of > "dfs.data.dir" property in your hdfs-site.xml file. > 5- Move inside the "current" directory and open the "VERSION" file here as > well. Now replace the value of "namespaceID" present here with the one you > had copied earlier. > 6- Restart all the daemons. > > Note : If you have not created dfs.name.dir and dfs.data.dir separately, > you could find all this inside your temp directory. > > HTH > > Warm Regards, > Tariq > https://mtariq.jux.com/ > cloudfront.blogspot.com > > > On Tue, Apr 30, 2013 at 2:45 AM, <[email protected]> wrote: > >> I am trying to start up a cluster and in the datanode log on the NameNode >> server I get the error: >> >> 2013-04-29 15:50:20,988 INFO >> org.apache.hadoop.hdfs.server.common.Storage: Lock on >> /data/hadoop/dfs/data/in_use.lock acquired by nodename 1406@devUbuntu05 >> 2013-04-29 15:50:20,990 FATAL >> org.apache.hadoop.hdfs.server.datanode.DataNode: Initialization failed for >> block pool Block pool BP-1306349046-172.16.26.68-1367256199559 (storage id >> DS-403514403-172.16.26.68-50010-1366406077018) service to devUbuntu05/ >> 172.16.26.68:9000 >> java.io.IOException: *Incompatible clusterIDs* in /data/hadoop/dfs/data: >> namenode clusterID = CID-23b9f9c7-2c25-411f-8bd2-4d5c9d7c25a1; datanode >> clusterID = CID-e3f6b811-c1b4-4778-a31e-14dea8b2cca8 >> at >> org.apache.hadoop.hdfs.server.datanode.DataStorage.doTransition(DataStorage.java:391) >> at >> org.apache.hadoop.hdfs.server.datanode.DataStorage.recoverTransitionRead(DataStorage.java:191) >> at >> org.apache.hadoop.hdfs.server.datanode.DataStorage.recoverTransitionRead(DataStorage.java:219) >> >> How do I get around this error? What does the error mean? >> >> Thank you. >> >> Kevin >> > >
