Thanks Jonathan.

Regards
Vijay

> On 23 Oct 2015, at 16:10, Jonathan Hurley <[email protected]> wrote:
> 
> First you need to get the ID of the alert definition in your system
> GET 
> api/v1/clusters/<cluster>/alert_definitions?AlertDefinition/name=ambari_agent_disk_usage
> 
> Once you have the ID, you can do a PUT:
> PUT api/v1/clusters/<cluster>/alert_definitions/<id>
> 
> {
>   "AlertDefinition" : {
>     "source" : {
>       "parameters" : [
>         {
>           "name" : "minimum.free.space",
>           "display_name" : "Minimum Free Space",
>           "units" : "bytes",
>           "value" : 5.0E9,
>           "description" : "The overall amount of free disk space left before 
> an alert is triggered.",
>           "type" : "NUMERIC"
>         },
>         {
>           "name" : "percent.used.space.warning.threshold",
>           "display_name" : "Warning",
>           "units" : "%",
>           "value" : 0.8,
>           "description" : "The percent of disk space consumed before a 
> warning is triggered.",
>           "type" : "PERCENT"
>         },
>         {
>           "name" : "percent.free.space.critical.threshold",
>           "display_name" : "Critical",
>           "units" : "%",
>           "value" : 0.9,
>           "description" : "The percent of disk space consumed before a 
> critical alert is triggered.",
>           "type" : "PERCENT"
>         }
>       ],
>       "path" : "alert_disk_space.py",
>       "type" : "SCRIPT"
>     }
>   }
> }
> 
> This changes the thresholds to 80% for warning and 90% for critical
> 
>> On Oct 23, 2015, at 10:45 AM, Vijaya Narayana Reddy Bhoomi Reddy 
>> <[email protected] 
>> <mailto:[email protected]>> wrote:
>> 
>> Thanks Jonathan for your reply.
>> 
>> Can you please let me know the API version modifying the threshold values?
>> 
>> Regards
>> Vijay
>> 
>> 
>>> On 23 Oct 2015, at 15:24, Jonathan Hurley <[email protected] 
>>> <mailto:[email protected]>> wrote:
>>> 
>>> The ambari disk usage alerts are meant to check two things: that you have 
>>> have enough space total and percent free space in /usr/hdp for data created 
>>> by hadoop and for installing versioned RPMs. Total free space alerts are 
>>> something that you’ll probably want to fix since it means you have less 
>>> than a certain amount of total free space left.
>>> 
>>> It seems like you’re talking about percent free space. Those can be changed 
>>> via the thresholds that the script uses. You can’t do this through the 
>>> Ambari Web Client. You have two options:
>>> 
>>> - Use the Ambari APIs to adjust the threshold values - this command is 
>>> rather long; let me know if you want to try this and I can paste the code 
>>> to do it.
>>> 
>>> - Edit the script directly and set the defaults to higher limits: 
>>> https://github.com/apache/ambari/blob/branch-2.1/ambari-server/src/main/resources/host_scripts/alert_disk_space.py#L36-L37
>>>  
>>> <https://github.com/apache/ambari/blob/branch-2.1/ambari-server/src/main/resources/host_scripts/alert_disk_space.py#L36-L37>
>>> 
>>> 
>>>> On Oct 23, 2015, at 9:26 AM, Vijaya Narayana Reddy Bhoomi Reddy 
>>>> <[email protected] 
>>>> <mailto:[email protected]>> wrote:
>>>> 
>>>> 
>>>> Siddharth,
>>>> 
>>>> Thanks for your response. As ours was a 4 node cluster, I changed it to 
>>>> Embedded mode from distributed mode and is working fine. However, I am 
>>>> facing another issue with regards to Ambari agent disk usage alerts. 
>>>> Earlier, I had three alerts for three machines where /usr/hdp is utilised 
>>>> more than 50%.
>>>> 
>>>> Initially when I setup the cluster, I had multiple mount points listed 
>>>> under yarn.nodemanager.local-dirs and yarn.nodemaneger.log-dirs. /usr/hdp 
>>>> was one amor them Later, I changed these values such that only one value 
>>>> is present for these (/export/hadoop/yarn/local and 
>>>> /export/hadoop/yarn/log) and restarted the required components.
>>>> 
>>>> However, I am still seeing the Ambari disk usage alert for /usr/hdp. Can 
>>>> you please let me know how to get rid of these alerts?
>>>> 
>>>> Thanks 
>>>> Vijay
>>>> 
>>>> 
>>>>> On 22 Oct 2015, at 19:02, Siddharth Wagle <[email protected] 
>>>>> <mailto:[email protected]>> wrote:
>>>>> 
>>>>> Hi Vijaya,
>>>>> 
>>>>> Please make all of the configs are accurate. 
>>>>> (https://cwiki.apache.org/confluence/display/AMBARI/AMS+-+distributed+mode
>>>>>  
>>>>> <https://cwiki.apache.org/confluence/display/AMBARI/AMS+-+distributed+mode>)
>>>>> 
>>>>> Can you attach, your ams-site.xml and /etc/ams-hbase/conf/hbase-site.xml ?
>>>>> 
>>>>> - Sid
>>>>> 
>>>>> ________________________________________
>>>>> From: Vijaya Narayana Reddy Bhoomi Reddy 
>>>>> <[email protected] 
>>>>> <mailto:[email protected]>>
>>>>> Sent: Thursday, October 22, 2015 8:36 AM
>>>>> To: [email protected] <mailto:[email protected]>
>>>>> Subject: Issue with Ambari Metrics Collector - Distributed mode
>>>>> 
>>>>> Hi,
>>>>> 
>>>>> I am facing an issue while setting up Ambari Metrics in distributed mode. 
>>>>> I am setting up HDP 2.3.x using Ambari 2.1.x. Initially when I was 
>>>>> setting up the cluster, I was shown a warning message that the volume / 
>>>>> directory for metrics service  is same as the one used by datanode and 
>>>>> hence I was recommended to change it. So I went ahead and pointed it to 
>>>>> hdfs, trying to setting up metrics service in distributed mode.
>>>>> 
>>>>> However, Ambari Metrics service is not set up properly and it timed out 
>>>>> while setting up the cluster, showing a warning that Ambari Metrics 
>>>>> service hasn’t started. I restarted the Metrics collector service 
>>>>> multiple times, but it would stop again in a few seconds.
>>>>> 
>>>>> On further observation, I realised that in the ams-site.xml file, 
>>>>> timeline.metrics.service.operation.mode was still pointing to “embedded", 
>>>>> where as hbase-site.xml had all the required properties set correctly. So 
>>>>> I changed the timeline.metrics.service.operation.mode property to 
>>>>> “distributed” and restarted the required services as recommended by 
>>>>> Ambari. However, the restart process is stuck at 68% and eventually timed 
>>>>> out. Its not able to restart the Metrics Collector service. However, all 
>>>>> the metrics monitor services are re-started without any issues.
>>>>> 
>>>>> Can anyone please throw light on why this happening and what is the 
>>>>> solution to fix this?
>>>>> 
>>>>> Thanks
>>>>> Vijay
>>>>> --
>>>>> The contents of this e-mail are confidential and for the exclusive use of
>>>>> the intended recipient. If you receive this e-mail in error please delete
>>>>> it from your system immediately and notify us either by e-mail or
>>>>> telephone. You should not copy, forward or otherwise disclose the content
>>>>> of the e-mail. The views expressed in this communication may not
>>>>> necessarily be the view held by WHISHWORKS.
>>>>> 
>>>> 
>>>> 
>>>> The contents of this e-mail are confidential and for the exclusive use of 
>>>> the intended recipient. If you receive this e-mail in error please delete 
>>>> it from your system immediately and notify us either by e-mail or 
>>>> telephone. You should not copy, forward or otherwise disclose the content 
>>>> of the e-mail. The views expressed in this communication may not 
>>>> necessarily be the view held by WHISHWORKS.
>>> 
>> 
>> 
>> The contents of this e-mail are confidential and for the exclusive use of 
>> the intended recipient. If you receive this e-mail in error please delete it 
>> from your system immediately and notify us either by e-mail or telephone. 
>> You should not copy, forward or otherwise disclose the content of the 
>> e-mail. The views expressed in this communication may not necessarily be the 
>> view held by WHISHWORKS.
> 


-- 
The contents of this e-mail are confidential and for the exclusive use of 
the intended recipient. If you receive this e-mail in error please delete 
it from your system immediately and notify us either by e-mail or 
telephone. You should not copy, forward or otherwise disclose the content 
of the e-mail. The views expressed in this communication may not 
necessarily be the view held by WHISHWORKS.

Reply via email to