Harsh J created HDFS-4257:
-----------------------------
Summary: The ReplaceDatanodeOnFailure policies could have a
forgiving option
Key: HDFS-4257
URL: https://issues.apache.org/jira/browse/HDFS-4257
Project: Hadoop HDFS
Issue Type: Bug
Components: hdfs-client
Affects Versions: 2.0.2-alpha
Reporter: Harsh J
Priority: Minor
Similar question has previously come over HDFS-3091 and friends, but the
essential problem is: "Why can't I write to my cluster of 3 nodes, when I just
have 1 node available at a point in time.".
The policies cover the 4 options, with {{Default}} being default:
{{Disable}} -> Disables the whole replacement concept by throwing out an error.
{{Never}} -> Never replaces a DN upon pipeline failures (not too desirable in
many cases).
{{Default}} -> Replace based on a few conditions, but whose minimum never
touches 1. We always fail if only one DN remains and none others can be added.
{{Always}} -> Replace no matter what. Fail if can't replace.
Would it not make sense to have an option similar to Always/Default, where
despite _trying_, if it isn't possible to have > 1 DN in the pipeline, do not
fail. I think that is what the former write behavior was, and what fit with the
minimum replication factor allowed value.
Why is it grossly wrong to pass a write from a client for a block with just 1
remaining replica in the pipeline (the minimum of 1 grows with the replication
factor demanded from the write), when replication is taken care of immediately
afterwards? How often have we seen missing blocks arise out of allowing this +
facing a big rack(s) failure or so?
--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira