[
https://issues.apache.org/jira/browse/CASSANDRA-919?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12850045#action_12850045
]
Johan Oskarsson commented on CASSANDRA-919:
-------------------------------------------
There's quite a big difference between retrying one rpc request and retrying a
whole map task. A map task has a big overhead, in the common case a whole new
JVM has to be started per task. It would also have to refetch a significant
chunk of data while one rpc retry is a only a few thousand rows.
So if there's a short period where a few rpcs fail (due to gc:ing or similar)
retrying them instead of the task will speed up the overall job by a noticeable
amount of time. It's also worth noting that the hdfs client used by the
standard MapReduce jobs has a retry mechanism.
> Add thrift request retries to Hadoop input format
> -------------------------------------------------
>
> Key: CASSANDRA-919
> URL: https://issues.apache.org/jira/browse/CASSANDRA-919
> Project: Cassandra
> Issue Type: Improvement
> Components: Contrib
> Reporter: Johan Oskarsson
> Assignee: Johan Oskarsson
> Priority: Trivial
> Fix For: 0.7
>
> Attachments: CASSANDRA-919.patch
>
>
> In order to decrease overhead of restarting a map task and increase
> reliability of the record reader we should retry the get_range_slices
> requests if they fail.
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.