[
https://issues.apache.org/jira/browse/YARN-3816?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14654298#comment-14654298
]
Junping Du commented on YARN-3816:
----------------------------------
Sorry for coming late on this. Thanks [~sjlee0], [~vrushalic] and [~gtCarrera9]
for review and comments!
bq. I propose to introduce the second dimension to the metrics explicitly. This
second dimension nearly maps to "toAggregate" (and/or the REP/SUM distinction)
in your patch. But I think it's probably better to introduce the metric types
explicitly as another enum or by subclassing TimelineMetric. Let me know what
you think.
Do you suggest to use gauge and counter type to replace "toAggregate"? But no
matter counter or gauge type of metrics, we may need to do aggregation. e.g.
CPU usage as guage, or map task number (launched, failed, etc.) as counter
(assume value is tack-on instead of accumulated). The idea to involve
"toAggregate" in metric is for client to indicate if this metric value should
be added/aggregated with other values or is a final value. If a client put a
metrics value that is already aggregated (like HDFS bytes written/read),
collector won't apply any aggregation logic on it.
bq. I'm still very confused by the usage of the word "aggregate". In this
patch, "aggregate" really means accumulating values of a metric along the time
dimension, which is completely different than the notion of aggregation we have
used all along. The aggregation has always been about rolling up values from
children to parents. Can we choose a different word to describe this aspect of
accumulating values along the time dimension, and avoid using "aggregation" for
this? "Accumulate"? "Cumulative"? Any suggestion?
Actually, v2 patch has both. In TimelineCollector, AggregatedMetrics mean
rolling up values from children to parents while AggregatedArea means
accumulating aggregated values of a metric along the time dimension. It may not
be necessary to separate calculating AggregatedArea out as a separated method.
Isn't it? It is a bit rush of naming for poc but we can have some better one
later.
bq. For example, consider HDFS bytes written. The time accumulation is already
built into it (see (1)). If you further accumulate this along the time
dimension, it becomes quadratic (doubly integrated) in time. I don't see how
that can be useful.
You are right. For some cases as you mentioned here, time accumulation is not
very useful. So beside "toAggregate", we may also need another flag (like:
"toAccumulate") to indicate if metric value need to be accumulated along the
time dimension? As we don't have assumption that all counters are already
accumulated over time or not, the client has flexibility to put accumulated or
tack-on values for counter. Thoughts?
bq. I think it would be OK to do this and not the average/max of the previous
discussion. I'd like to hear what others think about this.
Either way should work as we have area value at all timestamps, we can
recalculate average/max later if necessary. Would like to hear others' comments
too.
bq. Can we introduce a configuration that disables this time accumulation
feature? As we discussed, some may not want to have this feature enabled and
are perfectly happy with simple aggregation (from children to parents). It
would be good to isolate this part and be able to enable/disable it.
We surely can disable accumulation in system level. We also can disable
accumulation in metrics level (like proposed above) even accumulation is
enabled in system level.
bq. For timeseries, we need to decide what aggregation means. One option is
that we could normalize the values to a minute level granularity. For example,
add up values per min across each time. So anything that occurred within a
minute will be assigned to the top of that minute: eg if something happening at
2 min 10 seconds is considered to have occurred at 2 min. That way we can sum
up across flows/users/runs etc.
The other option is we only record /store accumulated values at different
timestamp and we do delta calculation later if necessary. This can address more
time granularity as query could apply on different granularity.
> [Aggregation] App-level Aggregation for YARN system metrics
> -----------------------------------------------------------
>
> Key: YARN-3816
> URL: https://issues.apache.org/jira/browse/YARN-3816
> Project: Hadoop YARN
> Issue Type: Sub-task
> Components: timelineserver
> Reporter: Junping Du
> Assignee: Junping Du
> Attachments: Application Level Aggregation of Timeline Data.pdf,
> YARN-3816-poc-v1.patch, YARN-3816-poc-v2.patch
>
>
> We need application level aggregation of Timeline data:
> - To present end user aggregated states for each application, include:
> resource (CPU, Memory) consumption across all containers, number of
> containers launched/completed/failed, etc. We need this for apps while they
> are running as well as when they are done.
> - Also, framework specific metrics, e.g. HDFS_BYTES_READ, should be
> aggregated to show details of states in framework level.
> - Other level (Flow/User/Queue) aggregation can be more efficient to be based
> on Application-level aggregations rather than raw entity-level data as much
> less raws need to scan (with filter out non-aggregated entities, like:
> events, configurations, etc.).
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)