[
https://issues.apache.org/jira/browse/CHUKWA-369?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12744743#action_12744743
]
Ari Rabkin commented on CHUKWA-369:
-----------------------------------
Alright. It wouldn't be too hard to split "collectors" from "nodes that check
for file completion" -- could have only a handful of the latter, drastically
cutting down the load on the filesystem. Would that address your concern?
But I'm curious. Did you really need 100 collectors? My understanding is that
that was a very substantial overprovisioning. Each one can do 20 MB/sec. Do
you really have 2 GB/sec of monitoring data?
> proposed reliability mechanism
> ------------------------------
>
> Key: CHUKWA-369
> URL: https://issues.apache.org/jira/browse/CHUKWA-369
> Project: Hadoop Chukwa
> Issue Type: New Feature
> Components: data collection
> Affects Versions: 0.3.0
> Reporter: Ari Rabkin
> Assignee: Ari Rabkin
> Fix For: 0.3.0
>
> Attachments: delayedAcks.patch
>
>
> We like to say that Chukwa is a system for reliable log collection. It isn't,
> quite, since we don't handle collector crashes. Here's a proposed
> reliability mechanism.
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.