Yeah, I figured that match. What I was referring to is the ending of the paths:
.../hadoop-hadoop/dfs/name/current
.../hadoop-hadoop/dfs/hadoop
They are different
--
  Take care,
Konstantin (Cos) Boudnik



On Wed, Dec 8, 2010 at 15:55, Richard Zhang <[email protected]> wrote:
> Hi:
> "/your/path/to/hadoop"  represents the location where hadoop is installed.
> BTW, I believe this is a file writing permission problem. Because I use the
> same *-site.xml setting to install with root and it works.
> But when I use the dedicated user hadoop, it always introduces this problem.
>
> But I do created manually the directory path and grant with 755.
> Weird....
> Richard.
>
> On Wed, Dec 8, 2010 at 6:51 PM, Konstantin Boudnik <[email protected]> wrote:
>
>> it seems that you are looking at 2 different directories:
>>
>> first post: /your/path/to/hadoop/tmp/dir/hadoop-hadoop/dfs/name/current
>> second: ls -l                              tmp/dir/hadoop-hadoop/dfs/hadoop
>> --
>>   Take care,
>> Konstantin (Cos) Boudnik
>>
>>
>>
>> On Wed, Dec 8, 2010 at 14:19, Richard Zhang <[email protected]>
>> wrote:
>> > would that be the reason that 54310 port is not open?
>> > I just used
>> > * iptables -A INPUT -p tcp --dport 54310 -j ACCEPT
>> > to open the port.
>> > But it seems the same erorr exists.
>> > Richard
>> > *
>> > On Wed, Dec 8, 2010 at 4:56 PM, Richard Zhang <[email protected]
>> >wrote:
>> >
>> >> Hi James:
>> >> I verified that I have the following permission set for the path:
>> >>
>> >> ls -l tmp/dir/hadoop-hadoop/dfs/hadoop
>> >> total 4
>> >> drwxr-xr-x 2 hadoop hadoop 4096 2010-12-08 15:56 current
>> >> Thanks.
>> >> Richard
>> >>
>> >>
>> >>
>> >> On Wed, Dec 8, 2010 at 4:50 PM, james warren <[email protected]> wrote:
>> >>
>> >>> Hi Richard -
>> >>>
>> >>> First thing that comes to mind is a permissions issue.  Can you verify
>> >>> that
>> >>> your directories along the desired namenode path are writable by the
>> >>> appropriate user(s)?
>> >>>
>> >>> HTH,
>> >>> -James
>> >>>
>> >>> On Wed, Dec 8, 2010 at 1:37 PM, Richard Zhang <[email protected]
>> >>> >wrote:
>> >>>
>> >>> > Hi Guys:
>> >>> > I am just installation the hadoop 0.21.0 in a single node cluster.
>> >>> > I encounter the following error when I run bin/hadoop namenode
>> -format
>> >>> >
>> >>> > 10/12/08 16:27:22 ERROR namenode.NameNode:
>> >>> > java.io.IOException: Cannot create directory
>> >>> > /your/path/to/hadoop/tmp/dir/hadoop-hadoop/dfs/name/current
>> >>> >        at
>> >>> >
>> >>> >
>> >>>
>> org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.clearDirectory(Storage.java:312)
>> >>> >        at
>> >>> >
>> org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1425)
>> >>> >        at
>> >>> >
>> org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1444)
>> >>> >        at
>> >>> >
>> >>>
>> org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:1242)
>> >>> >        at
>> >>> >
>> >>> >
>> >>>
>> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1348)
>> >>> >        at
>> >>> >
>> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1368)
>> >>> >
>> >>> >
>> >>> > Below is my core-site.xml
>> >>> >
>> >>> > <configuration>
>> >>> > <!-- In: conf/core-site.xml -->
>> >>> > <property>
>> >>> >  <name>hadoop.tmp.dir</name>
>> >>> >  <value>/your/path/to/hadoop/tmp/dir/hadoop-${user.name}</value>
>> >>> >  <description>A base for other temporary directories.</description>
>> >>> > </property>
>> >>> >
>> >>> > <property>
>> >>> >  <name>fs.default.name</name>
>> >>> >  <value>hdfs://localhost:54310</value>
>> >>> >  <description>The name of the default file system.  A URI whose
>> >>> >  scheme and authority determine the FileSystem implementation.  The
>> >>> >  uri's scheme determines the config property (fs.SCHEME.impl) naming
>> >>> >  the FileSystem implementation class.  The uri's authority is used to
>> >>> >  determine the host, port, etc. for a filesystem.</description>
>> >>> > </property>
>> >>> > </configuration>
>> >>> >
>> >>> >
>> >>> > Below is my hdfs-site.xml
>> >>> > *<?xml version="1.0"?>
>> >>> > <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
>> >>> >
>> >>> > <!-- Put site-specific property overrides in this file. -->
>> >>> >
>> >>> > <configuration>
>> >>> > <!-- In: conf/hdfs-site.xml -->
>> >>> > <property>
>> >>> >  <name>dfs.replication</name>
>> >>> >  <value>1</value>
>> >>> >  <description>Default block replication.
>> >>> >  The actual number of replications can be specified when the file is
>> >>> > created.
>> >>> >  The default is used if replication is not specified in create time.
>> >>> >  </description>
>> >>> > </property>
>> >>> >
>> >>> > </configuration>
>> >>> >
>> >>> >
>> >>> > below is my mapred-site.xml:
>> >>> > <?xml version="1.0"?>
>> >>> > <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
>> >>> >
>> >>> > <!-- Put site-specific property overrides in this file. -->
>> >>> >
>> >>> > <configuration>
>> >>> >
>> >>> > <!-- In: conf/mapred-site.xml -->
>> >>> > <property>
>> >>> >  <name>mapred.job.tracker</name>
>> >>> >  <value>localhost:54311</value>
>> >>> >  <description>The host and port that the MapReduce job tracker runs
>> >>> >  at.  If "local", then jobs are run in-process as a single map
>> >>> >  and reduce task.
>> >>> >  </description>
>> >>> > </property>
>> >>> >
>> >>> > </configuration>
>> >>> >
>> >>> >
>> >>> > Thanks.
>> >>> > Richard
>> >>> > *
>> >>> >
>> >>>
>> >>
>> >>
>> >
>>
>

Reply via email to