[
https://issues.apache.org/jira/browse/CHUKWA-369?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12750580#action_12750580
]
Ari Rabkin commented on CHUKWA-369:
-----------------------------------
Yes; I'm measuring that. Chunks only get duplicated when a collector crashes;
the number of duplicate chunks is basically just the amount of data lost in
.chukwa files. So for a single collector failure, it's (write rate) * (period
between rotations). This means that the fraction of duplicate data is just
(period between rotations) / (mean time between failures)
So if you assume that collectors crash once a week on average, and that the
rotation rate is every five minutes, then the fraction of duplicate data is
0.05%.
And my measurements bear this out.
> proposed reliability mechanism
> ------------------------------
>
> Key: CHUKWA-369
> URL: https://issues.apache.org/jira/browse/CHUKWA-369
> Project: Hadoop Chukwa
> Issue Type: New Feature
> Components: data collection
> Affects Versions: 0.3.0
> Reporter: Ari Rabkin
> Assignee: Ari Rabkin
> Fix For: 0.3.0
>
> Attachments: delayedAcks.patch
>
>
> We like to say that Chukwa is a system for reliable log collection. It isn't,
> quite, since we don't handle collector crashes. Here's a proposed
> reliability mechanism.
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.