[ https://issues.apache.org/jira/browse/HDFS-1751?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13006976#comment-13006976 ]
Boris Shkolnik commented on HDFS-1751: -------------------------------------- Non code questions: 1. why do we need limit on minimum component length. Some code notes: 1. Please use @Override when overriding methods. 2. verifyFSLimits should throw FSLimitException (most specific exception). Also update the java doc for this method. 3. Do we really need so specific exceptions, can't we just use FSLimitException with different messages. 4. Please use junit 4 for test. > Intrinsic limits for HDFS files, directories > -------------------------------------------- > > Key: HDFS-1751 > URL: https://issues.apache.org/jira/browse/HDFS-1751 > Project: Hadoop HDFS > Issue Type: New Feature > Components: data-node > Affects Versions: 0.22.0 > Reporter: Daryn Sharp > Assignee: Daryn Sharp > Fix For: 0.23.0 > > Attachments: HDFS-1751.patch > > > Enforce a configurable limit on: > the length of a path component > the number of names in a directory > The intention is to prevent a too-long name or a too-full directory. This is > not about RPC buffers, the length of command lines, etc. There may be good > reasons for those kinds of limits, but that is not the intended scope of this > feature. Consequently, a reasonable implementation might be to extend the > existing quota checker so that it faults the creation of a name that violates > the limits. This strategy of faulting new creation evades the problem of > existing names or directories that violate the limits. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira