Ooops. I didn't notice.
Anyway, here's the result
{
"href" :
"http://192.168.1.51:8080/api/v1/clusters/BigData/services/HDFS/components/NAMENODE?fields=*",
"metrics" : {
"boottime" : 0,
"process" : {
"proc_total" : 0,
"proc_run" : 0
},
"ugi" : {
"loginSuccess_num_ops" : 0,
"loginFailure_num_ops" : 0,
"loginSuccess_avg_time" : 0,
"loginFailure_avg_time" : 0
},
"dfs" : {
"namenode" : {
"fsImageLoadTime" : 0,
"FilesRenamed" : 0,
"JournalTransactionsBatchedInSync" : 0,
"SafemodeTime" : 0,
"FilesDeleted" : 0,
"DeleteFileOps" : 0,
"FilesAppended" : 0
}
},
"disk" : {
"disk_total" : 0,
"disk_free" : 0,
"part_max_used" : 0
},
"cpu" : {
"cpu_speed" : 0,
"cpu_num" : 0,
"cpu_wio" : 0,
"cpu_idle" : 0,
"cpu_nice" : 0,
"cpu_aidle" : 0,
"cpu_system" : 0,
"cpu_user" : 0
},
"rpcdetailed" : {
"delete_avg_time" : 0,
"rename_avg_time" : 0,
"register_num_ops" : 0,
"versionRequest_num_ops" : 0,
"blocksBeingWrittenReport_avg_time" : 0,
"rename_num_ops" : 0,
"register_avg_time" : 0,
"mkdirs_avg_time" : 0,
"setPermission_num_ops" : 0,
"delete_num_ops" : 0,
"versionRequest_avg_time" : 0,
"setOwner_num_ops" : 0,
"setSafeMode_avg_time" : 0,
"setOwner_avg_time" : 0,
"setSafeMode_num_ops" : 0,
"blocksBeingWrittenReport_num_ops" : 0,
"setReplication_num_ops" : 0,
"setPermission_avg_time" : 0,
"mkdirs_num_ops" : 0,
"setReplication_avg_time" : 0
},
"load" : {
"load_fifteen" : 0,
"load_one" : 0,
"load_five" : 0
},
"network" : {
"pkts_out" : 0,
"bytes_in" : 0,
"bytes_out" : 0,
"pkts_in" : 0
},
"memory" : {
"mem_total" : 0,
"swap_free" : 0,
"mem_buffers" : 0,
"mem_shared" : 0,
"mem_cached" : 0,
"mem_free" : 0,
"swap_total" : 0
}
},
"ServiceComponentInfo" : {
"cluster_name" : "BigData",
"desired_configs" : { },
"state" : "STARTED",
"component_name" : "NAMENODE",
"service_name" : "HDFS"
},
"host_components" : [
{
"href" :
"http://192.168.1.51:8080/api/v1/clusters/BigData/hosts/Crawler51.localdomain.com/host_components/NAMENODE",
"HostRoles" : {
"cluster_name" : "BigData",
"component_name" : "NAMENODE",
"host_name" : "Crawler51.localdomain.com"
}
}
]
}
On 3/14/2013 3:51 PM, Mahadev Konar wrote:
Hi Dustine,
I had a typo :). Sorry, can you run:
curl -u admin:admin
http://<ambari-server>:8080/api/v1/clusters/<clustername>/services/HDFS/components/NAMENODE?fields=*
thanks
mahadev
On Mar 14, 2013, at 12:46 AM, Dustine Rene Bernasor
<[email protected]> wrote:
Start/Stop button's still disabled.
Here's the result of the API call
<html>
<head>
<meta http-equiv="Content-Type" content="text/html;charset=ISO-8859-1"/>
<title>Error 403 Bad credentials</title>
</head>
<body>
<h2>HTTP ERROR: 403</h2>
<p>Problem accessing
/api/v1/clusters/BigData/services/HDFS/components/NAMENODE. Reason:
<pre> Bad credentials</pre></p>
<hr /><i><small>Powered by Jetty://</small></i>
</body>
</html>
On 3/14/2013 3:29 PM, Mahadev Konar wrote:
Yes. The start stop button should re activate is some time (usually takes
seconds) if it is 1.2.1 release.
If not can you make an API call to see what the status of Namenode is:
curl -u admin:amdin
http://<ambari-server>:8080/api/v1/clusters/<clustername>/services/HDFS/components/NAMENODE?fields=*
(see
https://github.com/apache/ambari/blob/trunk/ambari-server/docs/api/v1/index.md
for more details on API's)
mahadev
On Mar 14, 2013, at 12:23 AM, Dustine Rene Bernasor
<[email protected]> wrote:
Hello,
Did you mean /var/run/hadoop/hdfs/namenode/formatted?
I cannot restart namenode from the UI. HDFS icon keeps on blinking
but the Start and Stop buttons are disabled.
Thanks.
Dustine
On 3/14/2013 3:17 PM, Mahadev Konar wrote:
Hi Dustine,
Are you installing on a cluster that was already installed via
Ambari? If yes, then remove this directory in
/var/run/hadoop/hdfs/formatted and restart namenode from the the UI
and it should work.
If not then its a bug and please create jira nad attach logs for
Namenode/amabari agent and server.
thanks
mahadev
On Thu, Mar 14, 2013 at 12:13 AM, Dustine Rene Bernasor
<[email protected]> wrote:
Hello,
I was installing Ambari 1.2.1. When I reach step 9, after the services are
installed,
NameNode cannot be started.
The ff. exception appeared in the log
2013-03-14 10:58:00,426 ERROR
org.apache.hadoop.hdfs.server.namenode.FSNamesystem: FSNamesystem
initialization failed.
java.io.IOException: NameNode is not formatted.
at
org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:330)
at
org.apache.hadoop.hdfs.server.namenode.FSDirectory.loadFSImage(FSDirectory.java:100)
at
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.initialize(FSNamesystem.java:411)
at
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.<init>(FSNamesystem.java:379)
at
org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:287)
at
org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:548)
at
org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1431)
at
org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1440)
2013-03-14 10:58:00,427 ERROR
org.apache.hadoop.hdfs.server.namenode.NameNode: java.io.IOException:
NameNode is not formatted.
at
org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:330)
at
org.apache.hadoop.hdfs.server.namenode.FSDirectory.loadFSImage(FSDirectory.java:100)
at
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.initialize(FSNamesystem.java:411)
at
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.<init>(FSNamesystem.java:379)
at
org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:287)
at
org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:548)
at
org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1431)
at
org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1440)
2013-03-14 10:58:00,428 INFO
org.apache.hadoop.hdfs.server.namenode.NameNode: SHUTDOWN_MSG:
/************************************************************
SHUTDOWN_MSG: Shutting down NameNode at
Crawler51.localdomain.com/192.168.3.51
************************************************************/
Thanks.
Dustine