I did press Y tried it several times and once more now. Sitaraman
On Tue, Dec 24, 2013 at 8:38 PM, Nitin Pawar <[email protected]>wrote: > see the error .. it says not formatted > did you press Y or y ? > try again :) > > > On Tue, Dec 24, 2013 at 8:35 PM, Sitaraman Vilayannur < > [email protected]> wrote: > >> Hi Nitin, >> Even after formatting using hdfs namenode -format, i keep seeing >> namenode not formatted in the logs when i try to start namenode........ >> 12/24 20:33:26 INFO namenode.FSNamesystem: supergroup=supergroup >> 13/12/24 20:33:26 INFO namenode.FSNamesystem: isPermissionEnabled=true >> 13/12/24 20:33:26 INFO namenode.NameNode: Caching file names occuring >> more than 10 times >> 13/12/24 20:33:26 INFO namenode.NNStorage: Storage directory >> /usr/local/Software/hadoop-0.23.10/data/hdfs/namenode has been successfully >> formatted. >> 13/12/24 20:33:26 INFO namenode.FSImage: Saving image file >> /usr/local/Software/hadoop-0.23.10/data/hdfs/namenode/current/fsimage.ckpt_0000000000000000000 >> using no compression >> 13/12/24 20:33:26 INFO namenode.FSImage: Image file of size 124 saved in >> 0 seconds. >> 13/12/24 20:33:26 INFO namenode.NNStorageRetentionManager: Going to >> retain 1 images with txid >= 0 >> 13/12/24 20:33:26 INFO util.ExitUtil: Exiting with status 0 >> 13/12/24 20:33:26 INFO namenode.NameNode: SHUTDOWN_MSG: >> >> /************************************************************ >> SHUTDOWN_MSG: Shutting down NameNode at localhost.localdomain/127.0.0.1 >> ************************************************************/ >> >> >> 2013-12-24 20:33:46,337 INFO >> org.apache.hadoop.hdfs.server.common.Storage: Lock on >> /usr/local/Software/hadoop-2.2.0/data/hdfs/namenode/in_use.lock acquired by >> nodename [email protected] >> 2013-12-24 20:33:46,339 INFO org.mortbay.log: Stopped >> [email protected]:50070 >> 2013-12-24 20:33:46,340 INFO >> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Stopping NameNode >> metrics system... >> 2013-12-24 20:33:46,340 INFO >> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system >> stopped. >> 2013-12-24 20:33:46,340 INFO >> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system >> shutdown complete. >> 2013-12-24 20:33:46,340 FATAL >> org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode join >> java.io.IOException: NameNode is not formatted. >> at >> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:210) >> >> at >> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:787) >> at >> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:568) >> at >> org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:443) >> at >> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:491) >> at >> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:684) >> at >> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:669) >> at >> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1254) >> at >> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1320) >> 2013-12-24 20:33:46,342 INFO org.apache.hadoop.util.ExitUtil: Exiting >> with status 1 >> 2013-12-24 20:33:46,343 INFO >> org.apache.hadoop.hdfs.server.namenode.NameNode: SHUTDOWN_MSG: >> >> /************************************************************ >> SHUTDOWN_MSG: Shutting down NameNode at localhost.localdomain/127.0.0.1 >> ************************************************************/ >> >> >> >> >> On Tue, Dec 24, 2013 at 3:13 PM, Nitin Pawar <[email protected]>wrote: >> >>> the issue here is you tried one version of hadoop and then changed to a >>> different version. >>> >>> You can not do that directly with hadoop. You need to follow a process >>> while upgrading hadoop versions. >>> >>> For now as you are just starting with hadoop, I would recommend just run >>> a dfs format and start the hdfs again >>> >>> >>> On Tue, Dec 24, 2013 at 2:57 PM, Sitaraman Vilayannur < >>> [email protected]> wrote: >>> >>>> When i run namenode with upgrade option i get the following error and >>>> and namenode dosent start... >>>> 2013-12-24 14:48:38,595 INFO org.apache.hadoop.hdfs.StateChange: >>>> STATE* Network topology has 0 racks and 0 datanodes >>>> 2013-12-24 14:48:38,595 INFO org.apache.hadoop.hdfs.StateChange: >>>> STATE* UnderReplicatedBlocks has 0 blocks >>>> 2013-12-24 14:48:38,631 INFO org.apache.hadoop.ipc.Server: IPC Server >>>> Responder: starting >>>> 2013-12-24 14:48:38,632 INFO org.apache.hadoop.ipc.Server: IPC Server >>>> listener on 9000: starting >>>> 2013-12-24 14:48:38,633 INFO >>>> org.apache.hadoop.hdfs.server.namenode.NameNode: NameNode RPC up at: >>>> 192.168.1.2/192.168.1.2:9000 >>>> 2013-12-24 14:48:38,633 INFO >>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Starting services >>>> required for active state >>>> 2013-12-24 14:50:50,060 ERROR >>>> org.apache.hadoop.hdfs.server.namenode.NameNode: RECEIVED SIGNAL 15: >>>> SIGTERM >>>> 2013-12-24 14:50:50,062 INFO >>>> org.apache.hadoop.hdfs.server.namenode.NameNode: SHUTDOWN_MSG: >>>> /************************************************************ >>>> SHUTDOWN_MSG: Shutting down NameNode at localhost.localdomain/127.0.0.1 >>>> ************************************************************/ >>>> >>>> >>>> On 12/24/13, Sitaraman Vilayannur <[email protected]> >>>> wrote: >>>> > Found it, >>>> > I get the following error on starting namenode in 2.2 >>>> > >>>> 10/contrib/capacity-scheduler/*.jar:/usr/local/Software/hadoop-0.23.10/contrib/capacity-scheduler/*.jar:/usr/local/Software/hadoop-0.23.10/contrib/capacity-scheduler/*.jar >>>> > STARTUP_MSG: build = https://svn.apache.org/repos/asf/hadoop/common >>>> > -r 1529768; compiled by 'hortonmu' on 2013-10-07T06:28Z >>>> > STARTUP_MSG: java = 1.7.0_45 >>>> > ************************************************************/ >>>> > 2013-12-24 13:25:48,876 INFO >>>> > org.apache.hadoop.hdfs.server.namenode.NameNode: registered UNIX >>>> > signal handlers for [TERM, HUP, INT] >>>> > 2013-12-24 13:25:49,042 INFO >>>> > org.apache.hadoop.metrics2.impl.MetricsConfig: loaded properties from >>>> > hadoop-metrics2.properties >>>> > 2013-12-24 13:25:49,102 INFO >>>> > org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot >>>> > period at 10 second(s). >>>> > 2013-12-24 13:25:49,102 INFO >>>> > org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics >>>> > system started >>>> > 2013-12-24 13:25:49,232 WARN org.apache.hadoop.util.NativeCodeLoader: >>>> > Unable to load native-hadoop library for your platform... using >>>> > builtin-java classes where applicable >>>> > 2013-12-24 13:25:49,375 INFO org.mortbay.log: Logging to >>>> > org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via >>>> > org.mortbay.log.Slf4jLog >>>> > 2013-12-24 13:25:49,410 INFO org.apache.hadoop.http.HttpServer: Added >>>> > global filter 'safety' >>>> > (class=org.apache.hadoop.http.HttpServer$QuotingInputFilter) >>>> > 2013-12-24 13:25:49,412 INFO org.apache.hadoop.http.HttpServer: Added >>>> > filter static_user_filter >>>> > >>>> (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) >>>> > to context hdfs >>>> > 2013-12-24 13:25:49,412 INFO org.apache.hadoop.http.HttpServer: Added >>>> > filter static_user_filter >>>> > >>>> (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) >>>> > to context static >>>> > 2013-12-24 13:25:49,412 INFO org.apache.hadoop.http.HttpServer: Added >>>> > filter static_user_filter >>>> > >>>> (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) >>>> > to context logs >>>> > 2013-12-24 13:25:49,422 INFO org.apache.hadoop.http.HttpServer: >>>> > dfs.webhdfs.enabled = false >>>> > 2013-12-24 13:25:49,432 INFO org.apache.hadoop.http.HttpServer: Jetty >>>> > bound to port 50070 >>>> > 2013-12-24 13:25:49,432 INFO org.mortbay.log: jetty-6.1.26 >>>> > 2013-12-24 13:25:49,459 WARN org.mortbay.log: Can't reuse >>>> > /tmp/Jetty_0_0_0_0_50070_hdfs____w2cu08, using >>>> > /tmp/Jetty_0_0_0_0_50070_hdfs____w2cu08_2787234685293301311 >>>> > 2013-12-24 13:25:49,610 INFO org.mortbay.log: Started >>>> > [email protected]:50070 >>>> > 2013-12-24 13:25:49,611 INFO >>>> > org.apache.hadoop.hdfs.server.namenode.NameNode: Web-server up at: >>>> > 0.0.0.0:50070 >>>> > 2013-12-24 13:25:49,628 WARN >>>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Only one image >>>> > storage directory (dfs.namenode.name.dir) configured. Beware of >>>> > dataloss due to lack of redundant storage directories! >>>> > 2013-12-24 13:25:49,628 WARN >>>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Only one >>>> > namespace edits storage directory (dfs.namenode.edits.dir) configured. >>>> > Beware of dataloss due to lack of redundant storage directories! >>>> > 2013-12-24 13:25:49,668 INFO >>>> > org.apache.hadoop.hdfs.server.namenode.HostFileManager: read includes: >>>> > HostSet( >>>> > ) >>>> > 2013-12-24 13:25:49,669 INFO >>>> > org.apache.hadoop.hdfs.server.namenode.HostFileManager: read excludes: >>>> > HostSet( >>>> > ) >>>> > 2013-12-24 13:25:49,670 INFO >>>> > org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: >>>> > dfs.block.invalidate.limit=1000 >>>> > 2013-12-24 13:25:49,672 INFO org.apache.hadoop.util.GSet: Computing >>>> > capacity for map BlocksMap >>>> > 2013-12-24 13:25:49,672 INFO org.apache.hadoop.util.GSet: VM type >>>> = >>>> > 64-bit >>>> > 2013-12-24 13:25:49,673 INFO org.apache.hadoop.util.GSet: 2.0% max >>>> > memory = 889 MB >>>> > 2013-12-24 13:25:49,673 INFO org.apache.hadoop.util.GSet: capacity >>>> > = 2^21 = 2097152 entries >>>> > 2013-12-24 13:25:49,677 INFO >>>> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: >>>> > dfs.block.access.token.enable=false >>>> > 2013-12-24 13:25:49,677 INFO >>>> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: >>>> > defaultReplication = 1 >>>> > 2013-12-24 13:25:49,677 INFO >>>> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: >>>> > maxReplication = 512 >>>> > 2013-12-24 13:25:49,677 INFO >>>> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: >>>> > minReplication = 1 >>>> > 2013-12-24 13:25:49,677 INFO >>>> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: >>>> > maxReplicationStreams = 2 >>>> > 2013-12-24 13:25:49,677 INFO >>>> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: >>>> > shouldCheckForEnoughRacks = false >>>> > 2013-12-24 13:25:49,677 INFO >>>> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: >>>> > replicationRecheckInterval = 3000 >>>> > 2013-12-24 13:25:49,677 INFO >>>> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: >>>> > encryptDataTransfer = false >>>> > 2013-12-24 13:25:49,681 INFO >>>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: fsOwner >>>> > = sitaraman (auth:SIMPLE) >>>> > 2013-12-24 13:25:49,681 INFO >>>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: supergroup >>>> > = supergroup >>>> > 2013-12-24 13:25:49,681 INFO >>>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: >>>> > isPermissionEnabled = true >>>> > 2013-12-24 13:25:49,681 INFO >>>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: HA Enabled: false >>>> > 2013-12-24 13:25:49,682 INFO >>>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Append Enabled: >>>> > true >>>> > 2013-12-24 13:25:49,801 INFO org.apache.hadoop.util.GSet: Computing >>>> > capacity for map INodeMap >>>> > 2013-12-24 13:25:49,801 INFO org.apache.hadoop.util.GSet: VM type >>>> = >>>> > 64-bit >>>> > 2013-12-24 13:25:49,801 INFO org.apache.hadoop.util.GSet: 1.0% max >>>> > memory = 889 MB >>>> > 2013-12-24 13:25:49,801 INFO org.apache.hadoop.util.GSet: capacity >>>> > = 2^20 = 1048576 entries >>>> > 2013-12-24 13:25:49,802 INFO >>>> > org.apache.hadoop.hdfs.server.namenode.NameNode: Caching file names >>>> > occuring more than 10 times >>>> > 2013-12-24 13:25:49,804 INFO >>>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: >>>> > dfs.namenode.safemode.threshold-pct = 0.9990000128746033 >>>> > 2013-12-24 13:25:49,804 INFO >>>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: >>>> > dfs.namenode.safemode.min.datanodes = 0 >>>> > 2013-12-24 13:25:49,804 INFO >>>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: >>>> > dfs.namenode.safemode.extension = 30000 >>>> > 2013-12-24 13:25:49,805 INFO >>>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Retry cache on >>>> > namenode is enabled >>>> > 2013-12-24 13:25:49,805 INFO >>>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Retry cache will >>>> > use 0.03 of total heap and retry cache entry expiry time is 600000 >>>> > millis >>>> > 2013-12-24 13:25:49,807 INFO org.apache.hadoop.util.GSet: Computing >>>> > capacity for map Namenode Retry Cache >>>> > 2013-12-24 13:25:49,807 INFO org.apache.hadoop.util.GSet: VM type >>>> = >>>> > 64-bit >>>> > 2013-12-24 13:25:49,807 INFO org.apache.hadoop.util.GSet: >>>> > 0.029999999329447746% max memory = 889 MB >>>> > 2013-12-24 13:25:49,807 INFO org.apache.hadoop.util.GSet: capacity >>>> > = 2^15 = 32768 entries >>>> > 2013-12-24 13:25:49,816 INFO >>>> > org.apache.hadoop.hdfs.server.common.Storage: Lock on >>>> > /usr/local/Software/hadoop-0.23.10/data/hdfs/namenode/in_use.lock >>>> > acquired by nodename [email protected] >>>> > 2013-12-24 13:25:49,861 INFO org.mortbay.log: Stopped >>>> > [email protected]:50070 >>>> > 2013-12-24 13:25:49,964 INFO >>>> > org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Stopping NameNode >>>> > metrics system... >>>> > 2013-12-24 13:25:49,965 INFO >>>> > org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics >>>> > system stopped. >>>> > 2013-12-24 13:25:49,965 INFO >>>> > org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics >>>> > system shutdown complete. >>>> > 2013-12-24 13:25:49,965 FATAL >>>> > org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode >>>> > join >>>> > java.io.IOException: >>>> > File system image contains an old layout version -39. >>>> > An upgrade to version -47 is required. >>>> > Please restart NameNode with -upgrade option. >>>> > at >>>> > >>>> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:221) >>>> > at >>>> > >>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:787) >>>> > at >>>> > >>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:568) >>>> > at >>>> > >>>> org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:443) >>>> > at >>>> > >>>> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:491) >>>> > at >>>> > >>>> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:684) >>>> > at >>>> > >>>> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:669) >>>> > at >>>> > >>>> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1254) >>>> > at >>>> > >>>> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1320) >>>> > 2013-12-24 13:25:49,967 INFO org.apache.hadoop.util.ExitUtil: Exiting >>>> > with status 1 >>>> > 2013-12-24 13:25:49,968 INFO >>>> > org.apache.hadoop.hdfs.server.namenode.NameNode: SHUTDOWN_MSG: >>>> > /************************************************************ >>>> > SHUTDOWN_MSG: Shutting down NameNode at localhost.localdomain/ >>>> 127.0.0.1 >>>> > ************************************************************/ >>>> > >>>> > On 12/24/13, Sitaraman Vilayannur <[email protected]> >>>> wrote: >>>> >> The line beginning with ulimit that i have appended below, i thought >>>> >> was the log file? >>>> >> >>>> /usr/local/Software/hadoop-2.2.0/logs/hadoop-sitaraman-namenode-localhost.localdomain.out >>>> >> Sitaraman >>>> >> On 12/24/13, Nitin Pawar <[email protected]> wrote: >>>> >>> Without log, very hard to guess what's happening. >>>> >>> >>>> >>> Can you clean up the log directory and then start over and check >>>> for the >>>> >>> logs again. >>>> >>> >>>> >>> >>>> >>> On Tue, Dec 24, 2013 at 11:44 AM, Sitaraman Vilayannur < >>>> >>> [email protected]> wrote: >>>> >>> >>>> >>>> Hi Nitin, >>>> >>>> I moved to the release 2.2.0 on starting node manager it remains >>>> >>>> silent without errors but nodemanager dosent start....while it >>>> does in >>>> >>>> the earlier 0.23 version >>>> >>>> >>>> >>>> >>>> >>>> ./hadoop-daemon.sh start namenode >>>> >>>> starting namenode, logging to >>>> >>>> >>>> >>>> >>>> /usr/local/Software/hadoop-2.2.0/logs/hadoop-sitaraman-namenode-localhost.localdomain.out >>>> >>>> Java HotSpot(TM) 64-Bit Server VM warning: You have loaded library >>>> >>>> /usr/local/Software/hadoop-2.2.0/lib/native/libhadoop.so.1.0.0 >>>> which >>>> >>>> might have disabled stack guard. The VM will try to fix the stack >>>> >>>> guard now. >>>> >>>> It's highly recommended that you fix the library with 'execstack -c >>>> >>>> <libfile>', or link it with '-z noexecstack'. >>>> >>>> [sitaraman@localhost sbin]$ jps >>>> >>>> 13444 Jps >>>> >>>> [sitaraman@localhost sbin]$ vi >>>> >>>> >>>> >>>> >>>> /usr/local/Software/hadoop-2.2.0/logs/hadoop-sitaraman-namenode-localhost.localdomain.out >>>> >>>> >>>> >>>> >>>> >>>> ulimit -a for user sitaraman >>>> >>>> core file size (blocks, -c) 0 >>>> >>>> data seg size (kbytes, -d) unlimited >>>> >>>> scheduling priority (-e) 0 >>>> >>>> file size (blocks, -f) unlimited >>>> >>>> pending signals (-i) 135104 >>>> >>>> max locked memory (kbytes, -l) 32 >>>> >>>> max memory size (kbytes, -m) unlimited >>>> >>>> open files (-n) 1024 >>>> >>>> pipe size (512 bytes, -p) 8 >>>> >>>> POSIX message queues (bytes, -q) 819200 >>>> >>>> real-time priority (-r) 0 >>>> >>>> stack size (kbytes, -s) 10240 >>>> >>>> cpu time (seconds, -t) unlimited >>>> >>>> max user processes (-u) 135104 >>>> >>>> virtual memory (kbytes, -v) unlimited >>>> >>>> file locks (-x) unlimited >>>> >>>> >>>> >>>> >>>> >>>> On 12/24/13, Nitin Pawar <[email protected]> wrote: >>>> >>>> > For now you can ignore this warning, >>>> >>>> > it was your first program so you can try building other things >>>> and >>>> >>>> > slowly >>>> >>>> > run the commands mentioned the log message to fix these small >>>> >>>> > warnings. >>>> >>>> > >>>> >>>> > >>>> >>>> > On Tue, Dec 24, 2013 at 10:07 AM, Sitaraman Vilayannur < >>>> >>>> > [email protected]> wrote: >>>> >>>> > >>>> >>>> >> Thanks Nitin, That worked, >>>> >>>> >> When i run the Pi example, i get the following warning at the >>>> end, >>>> >>>> >> what must i do about this warning....thanks much for your help. >>>> >>>> >> Sitaraman >>>> >>>> >> inished in 20.82 seconds >>>> >>>> >> Java HotSpot(TM) 64-Bit Server VM warning: You have loaded >>>> library >>>> >>>> >> /usr/local/Software/hadoop-0.23.10/lib/native/libhadoop.so.1.0.0 >>>> >>>> >> which >>>> >>>> >> might have disabled stack guard. The VM will try to fix the >>>> stack >>>> >>>> >> guard now. >>>> >>>> >> It's highly recommended that you fix the library with >>>> 'execstack -c >>>> >>>> >> <libfile>', or link it with '-z noexecstack'. >>>> >>>> >> 13/12/24 10:05:19 WARN util.NativeCodeLoader: Unable to load >>>> >>>> >> native-hadoop library for your platform... using builtin-java >>>> >>>> >> classes >>>> >>>> >> where applicable >>>> >>>> >> Estimated value of Pi is 3.14127500000000000000 >>>> >>>> >> [sitaraman@localhost mapreduce]$ >>>> >>>> >> >>>> >>>> >> On 12/23/13, Nitin Pawar <[email protected]> wrote: >>>> >>>> >> > Can you try starting the process as non root user. >>>> >>>> >> > Give proper permissions to the user and start it as a >>>> different >>>> >>>> >> > user. >>>> >>>> >> > >>>> >>>> >> > Thanks, >>>> >>>> >> > Nitin >>>> >>>> >> > >>>> >>>> >> > >>>> >>>> >> > On Mon, Dec 23, 2013 at 2:15 PM, Sitaraman Vilayannur < >>>> >>>> >> > [email protected]> wrote: >>>> >>>> >> > >>>> >>>> >> >> Hi, >>>> >>>> >> >> When i attempt to start nodemanager i get the following >>>> error. >>>> >>>> >> >> Any >>>> >>>> >> >> help >>>> >>>> >> >> appreciated. I was able to start resource manager datanode, >>>> >>>> namenode >>>> >>>> >> >> and >>>> >>>> >> >> secondarynamenode, >>>> >>>> >> >> >>>> >>>> >> >> >>>> >>>> >> >> ./yarn-daemon.sh start nodemanager >>>> >>>> >> >> starting nodemanager, logging to >>>> >>>> >> >> >>>> >>>> >> >>>> >>>> >>>> /usr/local/Software/hadoop-0.23.10/logs/yarn-root-nodemanager-localhost.localdomain.out >>>> >>>> >> >> Unrecognized option: -jvm >>>> >>>> >> >> Error: Could not create the Java Virtual Machine. >>>> >>>> >> >> Error: A fatal exception has occurred. Program will exit. >>>> >>>> >> >> [root@localhost sbin]# emacs >>>> >>>> >> >> >>>> >>>> >> >>>> >>>> >>>> /usr/local/Software/hadoop-0.23.10/logs/yarn-root-nodemanager-localhost.localdomain.out >>>> >>>> >> >> & >>>> >>>> >> >> [4] 29004 >>>> >>>> >> >> [root@localhost sbin]# jps >>>> >>>> >> >> 28402 SecondaryNameNode >>>> >>>> >> >> 30280 Jps >>>> >>>> >> >> 28299 DataNode >>>> >>>> >> >> 6729 Main >>>> >>>> >> >> 26044 ResourceManager >>>> >>>> >> >> 28197 NameNode >>>> >>>> >> >> >>>> >>>> >> > >>>> >>>> >> > >>>> >>>> >> > >>>> >>>> >> > -- >>>> >>>> >> > Nitin Pawar >>>> >>>> >> > >>>> >>>> >> >>>> >>>> > >>>> >>>> > >>>> >>>> > >>>> >>>> > -- >>>> >>>> > Nitin Pawar >>>> >>>> > >>>> >>>> >>>> >>> >>>> >>> >>>> >>> >>>> >>> -- >>>> >>> Nitin Pawar >>>> >>> >>>> >> >>>> > >>>> >>> >>> >>> >>> -- >>> Nitin Pawar >>> >> >> > > > -- > Nitin Pawar >
