Do u Hv right permissions on the new dirs ? Try stopping n starting cluster...
-JJ On May 24, 2011, at 9:13 PM, Mark question <[email protected]> wrote: > Well, you're right ... moving it to hdfs-site.xml had an effect at least. > But now I'm in the NameSpace incompatable error: > > WARN org.apache.hadoop.hdfs.server.common.Util: Path > /tmp/hadoop-mark/dfs/data should be specified as a URI in configuration > files. Please update hdfs configuration. > java.io.IOException: Incompatible namespaceIDs in /tmp/hadoop-maha/dfs/data > > My configuration for this part in hdfs-site.xml: > <configuration> > <property> > <name>dfs.data.dir</name> > <value>/tmp/hadoop-mark/dfs/data</value> > </property> > <property> > <name>dfs.name.dir</name> > <value>/tmp/hadoop-mark/dfs/name</value> > </property> > <property> > <name>hadoop.tmp.dir</name> > <value>/cs/student/mark/tmp/hodhod</value> > </property> > </configuration> > > The reason why I want to change hadoop.tmp.dir is because the student quota > under /tmp is small so I wanted to mount on /cs/student instead for > hadoop.tmp.dir. > > Thanks, > Mark > > On Tue, May 24, 2011 at 7:25 PM, Joey Echeverria <[email protected]> wrote: > >> Try moving the the configuration to hdfs-site.xml. >> >> One word of warning, if you use /tmp to store your HDFS data, you risk >> data loss. On many operating systems, files and directories in /tmp >> are automatically deleted. >> >> -Joey >> >> On Tue, May 24, 2011 at 10:22 PM, Mark question <[email protected]> >> wrote: >>> Hi guys, >>> >>> I'm using an NFS cluster consisting of 30 machines, but only specified 3 >> of >>> the nodes to be my hadoop cluster. So my problem is this. Datanode won't >>> start in one of the nodes because of the following error: >>> >>> org.apache.hadoop.hdfs.server. >>> common.Storage: Cannot lock storage /cs/student/mark/tmp/hodhod/dfs/data. >>> The directory is already locked >>> >>> I think it's because of the NFS property which allows one node to lock it >>> then the second node can't lock it. So I had to change the following >>> configuration: >>> dfs.data.dir to be "/tmp/hadoop-user/dfs/data" >>> >>> But this configuration is overwritten by ${hadoop.tmp.dir}/dfs/data where >> my >>> hadoop.tmp.dir = " /cs/student/mark/tmp" as you might guess from above. >>> >>> Where is this configuration over-written ? I thought my core-site.xml has >>> the final configuration values. >>> Thanks, >>> Mark >>> >> >> >> >> -- >> Joseph Echeverria >> Cloudera, Inc. >> 443.305.9434 >>
