Dustine, What version of Ambari are you running? There is a bug 1.2.0 which causes this issue to happen. If thats the case you can upgrade to 1.2.1 (which is currently under vote).
http://incubator.apache.org/ambari/1.2.1/installing-hadoop-using-ambari/content/ambari-chap1.html Has instructions! thanks mahadev On Thu, Mar 14, 2013 at 7:15 PM, Dustine Rene Bernasor <[email protected]> wrote: > Here's the result > > "href" : > "http://192.168.1.51:8080/api/v1/clusters/BigData/hosts/Crawler51.localdomain.com/host_components/NAMENODE?fields=*", > "HostRoles" : { > "configs" : { }, > > "cluster_name" : "BigData", > "desired_configs" : { }, > "desired_state" : "STARTED", > "state" : "START_FAILED", > > "component_name" : "NAMENODE", > "host_name" : "Crawler51.localdomain.com" > }, > "host" : { > "href" : > "http://192.168.1.51:8080/api/v1/clusters/BigData/hosts/Crawler51.localdomain.com" > }, > "component" : [ > { > "href" : > "http://192.168.1.51:8080/api/v1/clusters/BigData/services/HDFS/components/NAMENODE", > "ServiceComponentInfo" : { > "cluster_name" : "BigData", > > "component_name" : "NAMENODE", > "service_name" : "HDFS" > } > } > ] > > > > On 3/15/2013 1:11 AM, Mahadev Konar wrote: >> >> To get more information can you run one more api command? >> >> curl -u admin:admin >> >> http://192.168.1.51:8080/api/v1/clusters/BigData/hosts/Crawler51.localdomain.com/host_components/NAMENODE?fields=* >> >> thanks >> mahadev >> >> >> On Thu, Mar 14, 2013 at 12:55 AM, Dustine Rene Bernasor >> <[email protected]> wrote: >>> >>> Ooops. I didn't notice. >>> >>> Anyway, here's the result >>> >>> { >>> "href" : >>> >>> "http://192.168.1.51:8080/api/v1/clusters/BigData/services/HDFS/components/NAMENODE?fields=*", >>> "metrics" : { >>> "boottime" : 0, >>> "process" : { >>> "proc_total" : 0, >>> "proc_run" : 0 >>> }, >>> "ugi" : { >>> "loginSuccess_num_ops" : 0, >>> "loginFailure_num_ops" : 0, >>> "loginSuccess_avg_time" : 0, >>> "loginFailure_avg_time" : 0 >>> }, >>> "dfs" : { >>> "namenode" : { >>> "fsImageLoadTime" : 0, >>> "FilesRenamed" : 0, >>> "JournalTransactionsBatchedInSync" : 0, >>> "SafemodeTime" : 0, >>> "FilesDeleted" : 0, >>> "DeleteFileOps" : 0, >>> "FilesAppended" : 0 >>> } >>> }, >>> "disk" : { >>> "disk_total" : 0, >>> "disk_free" : 0, >>> "part_max_used" : 0 >>> }, >>> "cpu" : { >>> "cpu_speed" : 0, >>> "cpu_num" : 0, >>> "cpu_wio" : 0, >>> "cpu_idle" : 0, >>> "cpu_nice" : 0, >>> "cpu_aidle" : 0, >>> "cpu_system" : 0, >>> "cpu_user" : 0 >>> }, >>> "rpcdetailed" : { >>> "delete_avg_time" : 0, >>> "rename_avg_time" : 0, >>> "register_num_ops" : 0, >>> "versionRequest_num_ops" : 0, >>> "blocksBeingWrittenReport_avg_time" : 0, >>> "rename_num_ops" : 0, >>> "register_avg_time" : 0, >>> "mkdirs_avg_time" : 0, >>> "setPermission_num_ops" : 0, >>> "delete_num_ops" : 0, >>> "versionRequest_avg_time" : 0, >>> "setOwner_num_ops" : 0, >>> "setSafeMode_avg_time" : 0, >>> "setOwner_avg_time" : 0, >>> "setSafeMode_num_ops" : 0, >>> "blocksBeingWrittenReport_num_ops" : 0, >>> "setReplication_num_ops" : 0, >>> "setPermission_avg_time" : 0, >>> "mkdirs_num_ops" : 0, >>> "setReplication_avg_time" : 0 >>> }, >>> "load" : { >>> "load_fifteen" : 0, >>> "load_one" : 0, >>> "load_five" : 0 >>> }, >>> "network" : { >>> "pkts_out" : 0, >>> "bytes_in" : 0, >>> "bytes_out" : 0, >>> "pkts_in" : 0 >>> }, >>> "memory" : { >>> "mem_total" : 0, >>> "swap_free" : 0, >>> "mem_buffers" : 0, >>> "mem_shared" : 0, >>> "mem_cached" : 0, >>> "mem_free" : 0, >>> "swap_total" : 0 >>> } >>> }, >>> "ServiceComponentInfo" : { >>> "cluster_name" : "BigData", >>> "desired_configs" : { }, >>> "state" : "STARTED", >>> "component_name" : "NAMENODE", >>> "service_name" : "HDFS" >>> }, >>> "host_components" : [ >>> { >>> "href" : >>> >>> "http://192.168.1.51:8080/api/v1/clusters/BigData/hosts/Crawler51.localdomain.com/host_components/NAMENODE", >>> "HostRoles" : { >>> "cluster_name" : "BigData", >>> "component_name" : "NAMENODE", >>> "host_name" : "Crawler51.localdomain.com" >>> } >>> } >>> ] >>> >>> } >>> >>> >>> >>> >>> On 3/14/2013 3:51 PM, Mahadev Konar wrote: >>> >>> Hi Dustine, >>> I had a typo :). Sorry, can you run: >>> >>> curl -u admin:admin >>> >>> http://<ambari-server>:8080/api/v1/clusters/<clustername>/services/HDFS/components/NAMENODE?fields=* >>> >>> thanks >>> mahadev >>> >>> >>> On Mar 14, 2013, at 12:46 AM, Dustine Rene Bernasor >>> <[email protected]> wrote: >>> >>> Start/Stop button's still disabled. >>> >>> Here's the result of the API call >>> >>> <html> >>> <head> >>> <meta http-equiv="Content-Type" content="text/html;charset=ISO-8859-1"/> >>> <title>Error 403 Bad credentials</title> >>> </head> >>> <body> >>> <h2>HTTP ERROR: 403</h2> >>> <p>Problem accessing >>> /api/v1/clusters/BigData/services/HDFS/components/NAMENODE. Reason: >>> <pre> Bad credentials</pre></p> >>> <hr /><i><small>Powered by Jetty://</small></i> >>> >>> >>> </body> >>> </html> >>> >>> >>> >>> On 3/14/2013 3:29 PM, Mahadev Konar wrote: >>> >>> Yes. The start stop button should re activate is some time (usually takes >>> seconds) if it is 1.2.1 release. >>> >>> If not can you make an API call to see what the status of Namenode is: >>> >>> curl -u admin:amdin >>> >>> http://<ambari-server>:8080/api/v1/clusters/<clustername>/services/HDFS/components/NAMENODE?fields=* >>> >>> (see >>> >>> https://github.com/apache/ambari/blob/trunk/ambari-server/docs/api/v1/index.md >>> for more details on API's) >>> >>> mahadev >>> >>> On Mar 14, 2013, at 12:23 AM, Dustine Rene Bernasor >>> <[email protected]> wrote: >>> >>> Hello, >>> >>> Did you mean /var/run/hadoop/hdfs/namenode/formatted? >>> >>> I cannot restart namenode from the UI. HDFS icon keeps on blinking >>> but the Start and Stop buttons are disabled. >>> >>> Thanks. >>> >>> Dustine >>> >>> >>> On 3/14/2013 3:17 PM, Mahadev Konar wrote: >>> >>> Hi Dustine, >>> Are you installing on a cluster that was already installed via >>> Ambari? If yes, then remove this directory in >>> /var/run/hadoop/hdfs/formatted and restart namenode from the the UI >>> and it should work. >>> >>> If not then its a bug and please create jira nad attach logs for >>> Namenode/amabari agent and server. >>> >>> thanks >>> mahadev >>> >>> On Thu, Mar 14, 2013 at 12:13 AM, Dustine Rene Bernasor >>> <[email protected]> wrote: >>> >>> Hello, >>> >>> I was installing Ambari 1.2.1. When I reach step 9, after the services >>> are >>> installed, >>> NameNode cannot be started. >>> >>> The ff. exception appeared in the log >>> >>> 2013-03-14 10:58:00,426 ERROR >>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: FSNamesystem >>> initialization failed. >>> java.io.IOException: NameNode is not formatted. >>> at >>> >>> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:330) >>> at >>> >>> org.apache.hadoop.hdfs.server.namenode.FSDirectory.loadFSImage(FSDirectory.java:100) >>> at >>> >>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.initialize(FSNamesystem.java:411) >>> at >>> >>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.<init>(FSNamesystem.java:379) >>> at >>> >>> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:287) >>> at >>> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:548) >>> at >>> >>> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1431) >>> at >>> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1440) >>> 2013-03-14 10:58:00,427 ERROR >>> org.apache.hadoop.hdfs.server.namenode.NameNode: java.io.IOException: >>> NameNode is not formatted. >>> at >>> >>> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:330) >>> at >>> >>> org.apache.hadoop.hdfs.server.namenode.FSDirectory.loadFSImage(FSDirectory.java:100) >>> at >>> >>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.initialize(FSNamesystem.java:411) >>> at >>> >>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.<init>(FSNamesystem.java:379) >>> at >>> >>> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:287) >>> at >>> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:548) >>> at >>> >>> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1431) >>> at >>> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1440) >>> >>> 2013-03-14 10:58:00,428 INFO >>> org.apache.hadoop.hdfs.server.namenode.NameNode: SHUTDOWN_MSG: >>> /************************************************************ >>> SHUTDOWN_MSG: Shutting down NameNode at >>> Crawler51.localdomain.com/192.168.3.51 >>> ************************************************************/ >>> >>> Thanks. >>> >>> Dustine >>> >>> >>> >>> >>> >
