I think you need to install and configure ssh
On Thu, Aug 9, 2012 at 4:30 PM, anand sharma <[email protected]> wrote: > Thanks all for reply, yes the user has access to that directory and i have > already formatted the namenode; just for simplicity i am not using ssh as i > am doing things for the first time. > > On Thu, Aug 9, 2012 at 3:58 PM, shashwat shriparv < > [email protected]> wrote: > >> format the filesystem >> >> bin/hadoop namenode -format >> >> then try to start namenode :) >> >> >> On Thu, Aug 9, 2012 at 3:51 PM, Mohammad Tariq <[email protected]>wrote: >> >>> Hello Anand, >>> >>> Is there any specific reason behind not using ssh?? >>> >>> Regards, >>> Mohammad Tariq >>> >>> >>> On Thu, Aug 9, 2012 at 3:46 PM, anand sharma <[email protected]> >>> wrote: >>> > Hi, i am just learning the Hadoop and i am setting the development >>> > environment with CDH3 pseudo distributed mode without any ssh >>> cofiguration >>> > in CentOS 6.2 . i can run the sample programs as usual but when i try >>> and >>> > run namenode this is the error it logs... >>> > >>> > [hive@localhost ~]$ hadoop namenode >>> > 12/08/09 20:56:57 INFO namenode.NameNode: STARTUP_MSG: >>> > /************************************************************ >>> > STARTUP_MSG: Starting NameNode >>> > STARTUP_MSG: host = localhost.localdomain/127.0.0.1 >>> > STARTUP_MSG: args = [] >>> > STARTUP_MSG: version = 0.20.2-cdh3u4 >>> > STARTUP_MSG: build = >>> file:///data/1/tmp/topdir/BUILD/hadoop-0.20.2-cdh3u4 >>> > -r 214dd731e3bdb687cb55988d3f47dd9e248c5690; compiled by 'root' on Mon >>> May >>> > 7 14:01:59 PDT 2012 >>> > ************************************************************/ >>> > 12/08/09 20:56:57 INFO jvm.JvmMetrics: Initializing JVM Metrics with >>> > processName=NameNode, sessionId=null >>> > 12/08/09 20:56:57 INFO metrics.NameNodeMetrics: Initializing >>> > NameNodeMeterics using context >>> > object:org.apache.hadoop.metrics.spi.NoEmitMetricsContext >>> > 12/08/09 20:56:57 INFO util.GSet: VM type = 64-bit >>> > 12/08/09 20:56:57 INFO util.GSet: 2% max memory = 17.77875 MB >>> > 12/08/09 20:56:57 INFO util.GSet: capacity = 2^21 = 2097152 >>> entries >>> > 12/08/09 20:56:57 INFO util.GSet: recommended=2097152, actual=2097152 >>> > 12/08/09 20:56:57 INFO namenode.FSNamesystem: fsOwner=hive >>> (auth:SIMPLE) >>> > 12/08/09 20:56:57 INFO namenode.FSNamesystem: supergroup=supergroup >>> > 12/08/09 20:56:57 INFO namenode.FSNamesystem: isPermissionEnabled=false >>> > 12/08/09 20:56:57 INFO namenode.FSNamesystem: >>> > dfs.block.invalidate.limit=1000 >>> > 12/08/09 20:56:57 INFO namenode.FSNamesystem: >>> isAccessTokenEnabled=false >>> > accessKeyUpdateInterval=0 min(s), accessTokenLifetime=0 min(s) >>> > 12/08/09 20:56:57 INFO metrics.FSNamesystemMetrics: Initializing >>> > FSNamesystemMetrics using context >>> > object:org.apache.hadoop.metrics.spi.NoEmitMetricsContext >>> > 12/08/09 20:56:57 ERROR namenode.FSNamesystem: FSNamesystem >>> initialization >>> > failed. >>> > java.io.FileNotFoundException: >>> > /var/lib/hadoop-0.20/cache/hadoop/dfs/name/in_use.lock (Permission >>> denied) >>> > at java.io.RandomAccessFile.open(Native Method) >>> > at java.io.RandomAccessFile.<init>(RandomAccessFile.java:216) >>> > at >>> > >>> org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.tryLock(Storage.java:614) >>> > at >>> > >>> org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.lock(Storage.java:591) >>> > at >>> > >>> org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.analyzeStorage(Storage.java:449) >>> > at >>> > >>> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:304) >>> > at >>> > >>> org.apache.hadoop.hdfs.server.namenode.FSDirectory.loadFSImage(FSDirectory.java:110) >>> > at >>> > >>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.initialize(FSNamesystem.java:372) >>> > at >>> > >>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.<init>(FSNamesystem.java:335) >>> > at >>> > >>> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:271) >>> > at >>> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:467) >>> > at >>> > >>> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1330) >>> > at >>> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1339) >>> > 12/08/09 20:56:57 ERROR namenode.NameNode: >>> java.io.FileNotFoundException: >>> > /var/lib/hadoop-0.20/cache/hadoop/dfs/name/in_use.lock (Permission >>> denied) >>> > at java.io.RandomAccessFile.open(Native Method) >>> > at java.io.RandomAccessFile.<init>(RandomAccessFile.java:216) >>> > at >>> > >>> org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.tryLock(Storage.java:614) >>> > at >>> > >>> org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.lock(Storage.java:591) >>> > at >>> > >>> org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.analyzeStorage(Storage.java:449) >>> > at >>> > >>> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:304) >>> > at >>> > >>> org.apache.hadoop.hdfs.server.namenode.FSDirectory.loadFSImage(FSDirectory.java:110) >>> > at >>> > >>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.initialize(FSNamesystem.java:372) >>> > at >>> > >>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.<init>(FSNamesystem.java:335) >>> > at >>> > >>> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:271) >>> > at >>> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:467) >>> > at >>> > >>> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1330) >>> > at >>> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1339) >>> > >>> > 12/08/09 20:56:57 INFO namenode.NameNode: SHUTDOWN_MSG: >>> > /************************************************************ >>> > SHUTDOWN_MSG: Shutting down NameNode at localhost.localdomain/ >>> 127.0.0.1 >>> > ************************************************************/ >>> > >>> > >>> >> >> >> >> -- >> >> >> ∞ >> Shashwat Shriparv >> >> >> > -- ∞ Shashwat Shriparv
