[
https://issues.apache.org/jira/browse/AMBARI-14140?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15041272#comment-15041272
]
Hadoop QA commented on AMBARI-14140:
------------------------------------
{color:red}-1 overall{color}. Here are the results of testing the latest
attachment
http://issues.apache.org/jira/secure/attachment/12775655/AMBARI-14140.patch
against trunk revision .
{color:red}-1 patch{color}. The patch command could not apply the patch.
Console output:
https://builds.apache.org/job/Ambari-trunk-test-patch/4482//console
This message is automatically generated.
> Test and Adopt FIFO compaction policy for AMS high load tables
> --------------------------------------------------------------
>
> Key: AMBARI-14140
> URL: https://issues.apache.org/jira/browse/AMBARI-14140
> Project: Ambari
> Issue Type: Bug
> Components: ambari-metrics
> Reporter: Aravindan Vijayan
> Assignee: Aravindan Vijayan
> Priority: Critical
> Fix For: 2.2.0
>
> Attachments: AMBARI-14140.patch
>
>
> FIFO compaction policy selects only files which have all cells expired. The
> column family MUST have non-default TTL.
> Essentially, FIFO compactor does only one job: collects expired store files.
> Because we do not do any real compaction, we do not use CPU and IO (disk and
> network), we do not evict hot data from a block cache. The result: improved
> throughput and latency both write and read.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)