would that be the reason that 54310 port is not open? I just used * iptables -A INPUT -p tcp --dport 54310 -j ACCEPT to open the port. But it seems the same erorr exists. Richard * On Wed, Dec 8, 2010 at 4:56 PM, Richard Zhang <[email protected]>wrote:
> Hi James: > I verified that I have the following permission set for the path: > > ls -l tmp/dir/hadoop-hadoop/dfs/hadoop > total 4 > drwxr-xr-x 2 hadoop hadoop 4096 2010-12-08 15:56 current > Thanks. > Richard > > > > On Wed, Dec 8, 2010 at 4:50 PM, james warren <[email protected]> wrote: > >> Hi Richard - >> >> First thing that comes to mind is a permissions issue. Can you verify >> that >> your directories along the desired namenode path are writable by the >> appropriate user(s)? >> >> HTH, >> -James >> >> On Wed, Dec 8, 2010 at 1:37 PM, Richard Zhang <[email protected] >> >wrote: >> >> > Hi Guys: >> > I am just installation the hadoop 0.21.0 in a single node cluster. >> > I encounter the following error when I run bin/hadoop namenode -format >> > >> > 10/12/08 16:27:22 ERROR namenode.NameNode: >> > java.io.IOException: Cannot create directory >> > /your/path/to/hadoop/tmp/dir/hadoop-hadoop/dfs/name/current >> > at >> > >> > >> org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.clearDirectory(Storage.java:312) >> > at >> > org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1425) >> > at >> > org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1444) >> > at >> > >> org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:1242) >> > at >> > >> > >> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1348) >> > at >> > org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1368) >> > >> > >> > Below is my core-site.xml >> > >> > <configuration> >> > <!-- In: conf/core-site.xml --> >> > <property> >> > <name>hadoop.tmp.dir</name> >> > <value>/your/path/to/hadoop/tmp/dir/hadoop-${user.name}</value> >> > <description>A base for other temporary directories.</description> >> > </property> >> > >> > <property> >> > <name>fs.default.name</name> >> > <value>hdfs://localhost:54310</value> >> > <description>The name of the default file system. A URI whose >> > scheme and authority determine the FileSystem implementation. The >> > uri's scheme determines the config property (fs.SCHEME.impl) naming >> > the FileSystem implementation class. The uri's authority is used to >> > determine the host, port, etc. for a filesystem.</description> >> > </property> >> > </configuration> >> > >> > >> > Below is my hdfs-site.xml >> > *<?xml version="1.0"?> >> > <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> >> > >> > <!-- Put site-specific property overrides in this file. --> >> > >> > <configuration> >> > <!-- In: conf/hdfs-site.xml --> >> > <property> >> > <name>dfs.replication</name> >> > <value>1</value> >> > <description>Default block replication. >> > The actual number of replications can be specified when the file is >> > created. >> > The default is used if replication is not specified in create time. >> > </description> >> > </property> >> > >> > </configuration> >> > >> > >> > below is my mapred-site.xml: >> > <?xml version="1.0"?> >> > <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> >> > >> > <!-- Put site-specific property overrides in this file. --> >> > >> > <configuration> >> > >> > <!-- In: conf/mapred-site.xml --> >> > <property> >> > <name>mapred.job.tracker</name> >> > <value>localhost:54311</value> >> > <description>The host and port that the MapReduce job tracker runs >> > at. If "local", then jobs are run in-process as a single map >> > and reduce task. >> > </description> >> > </property> >> > >> > </configuration> >> > >> > >> > Thanks. >> > Richard >> > * >> > >> > >
