Thanks Jonathan for your reply.

Can you please let me know the API version modifying the threshold values?

Regards
Vijay


> On 23 Oct 2015, at 15:24, Jonathan Hurley <[email protected]> wrote:
> 
> The ambari disk usage alerts are meant to check two things: that you have 
> have enough space total and percent free space in /usr/hdp for data created 
> by hadoop and for installing versioned RPMs. Total free space alerts are 
> something that you’ll probably want to fix since it means you have less than 
> a certain amount of total free space left.
> 
> It seems like you’re talking about percent free space. Those can be changed 
> via the thresholds that the script uses. You can’t do this through the Ambari 
> Web Client. You have two options:
> 
> - Use the Ambari APIs to adjust the threshold values - this command is rather 
> long; let me know if you want to try this and I can paste the code to do it.
> 
> - Edit the script directly and set the defaults to higher limits: 
> https://github.com/apache/ambari/blob/branch-2.1/ambari-server/src/main/resources/host_scripts/alert_disk_space.py#L36-L37
>  
> <https://github.com/apache/ambari/blob/branch-2.1/ambari-server/src/main/resources/host_scripts/alert_disk_space.py#L36-L37>
> 
> 
>> On Oct 23, 2015, at 9:26 AM, Vijaya Narayana Reddy Bhoomi Reddy 
>> <[email protected] 
>> <mailto:[email protected]>> wrote:
>> 
>> 
>> Siddharth,
>> 
>> Thanks for your response. As ours was a 4 node cluster, I changed it to 
>> Embedded mode from distributed mode and is working fine. However, I am 
>> facing another issue with regards to Ambari agent disk usage alerts. 
>> Earlier, I had three alerts for three machines where /usr/hdp is utilised 
>> more than 50%.
>> 
>> Initially when I setup the cluster, I had multiple mount points listed under 
>> yarn.nodemanager.local-dirs and yarn.nodemaneger.log-dirs. /usr/hdp was one 
>> amor them Later, I changed these values such that only one value is present 
>> for these (/export/hadoop/yarn/local and /export/hadoop/yarn/log) and 
>> restarted the required components.
>> 
>> However, I am still seeing the Ambari disk usage alert for /usr/hdp. Can you 
>> please let me know how to get rid of these alerts?
>> 
>> Thanks 
>> Vijay
>> 
>> 
>>> On 22 Oct 2015, at 19:02, Siddharth Wagle <[email protected] 
>>> <mailto:[email protected]>> wrote:
>>> 
>>> Hi Vijaya,
>>> 
>>> Please make all of the configs are accurate. 
>>> (https://cwiki.apache.org/confluence/display/AMBARI/AMS+-+distributed+mode 
>>> <https://cwiki.apache.org/confluence/display/AMBARI/AMS+-+distributed+mode>)
>>> 
>>> Can you attach, your ams-site.xml and /etc/ams-hbase/conf/hbase-site.xml ?
>>> 
>>> - Sid
>>> 
>>> ________________________________________
>>> From: Vijaya Narayana Reddy Bhoomi Reddy <[email protected] 
>>> <mailto:[email protected]>>
>>> Sent: Thursday, October 22, 2015 8:36 AM
>>> To: [email protected] <mailto:[email protected]>
>>> Subject: Issue with Ambari Metrics Collector - Distributed mode
>>> 
>>> Hi,
>>> 
>>> I am facing an issue while setting up Ambari Metrics in distributed mode. I 
>>> am setting up HDP 2.3.x using Ambari 2.1.x. Initially when I was setting up 
>>> the cluster, I was shown a warning message that the volume / directory for 
>>> metrics service  is same as the one used by datanode and hence I was 
>>> recommended to change it. So I went ahead and pointed it to hdfs, trying to 
>>> setting up metrics service in distributed mode.
>>> 
>>> However, Ambari Metrics service is not set up properly and it timed out 
>>> while setting up the cluster, showing a warning that Ambari Metrics service 
>>> hasn’t started. I restarted the Metrics collector service multiple times, 
>>> but it would stop again in a few seconds.
>>> 
>>> On further observation, I realised that in the ams-site.xml file, 
>>> timeline.metrics.service.operation.mode was still pointing to “embedded", 
>>> where as hbase-site.xml had all the required properties set correctly. So I 
>>> changed the timeline.metrics.service.operation.mode property to 
>>> “distributed” and restarted the required services as recommended by Ambari. 
>>> However, the restart process is stuck at 68% and eventually timed out. Its 
>>> not able to restart the Metrics Collector service. However, all the metrics 
>>> monitor services are re-started without any issues.
>>> 
>>> Can anyone please throw light on why this happening and what is the 
>>> solution to fix this?
>>> 
>>> Thanks
>>> Vijay
>>> --
>>> The contents of this e-mail are confidential and for the exclusive use of
>>> the intended recipient. If you receive this e-mail in error please delete
>>> it from your system immediately and notify us either by e-mail or
>>> telephone. You should not copy, forward or otherwise disclose the content
>>> of the e-mail. The views expressed in this communication may not
>>> necessarily be the view held by WHISHWORKS.
>>> 
>> 
>> 
>> The contents of this e-mail are confidential and for the exclusive use of 
>> the intended recipient. If you receive this e-mail in error please delete it 
>> from your system immediately and notify us either by e-mail or telephone. 
>> You should not copy, forward or otherwise disclose the content of the 
>> e-mail. The views expressed in this communication may not necessarily be the 
>> view held by WHISHWORKS.
> 


-- 
The contents of this e-mail are confidential and for the exclusive use of 
the intended recipient. If you receive this e-mail in error please delete 
it from your system immediately and notify us either by e-mail or 
telephone. You should not copy, forward or otherwise disclose the content 
of the e-mail. The views expressed in this communication may not 
necessarily be the view held by WHISHWORKS.

Reply via email to