[ https://issues.apache.org/jira/browse/HDFS-5517?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15583770#comment-15583770 ]
Hadoop QA commented on HDFS-5517: --------------------------------- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 0s{color} | {color:blue} Docker mode activated. {color} | | {color:red}-1{color} | {color:red} patch {color} | {color:red} 0m 7s{color} | {color:red} HDFS-5517 does not apply to trunk. Rebase required? Wrong Branch? See https://wiki.apache.org/hadoop/HowToContribute for help. {color} | \\ \\ || Subsystem || Report/Notes || | JIRA Issue | HDFS-5517 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12614116/HDFS-5517.patch | | Console output | https://builds.apache.org/job/PreCommit-HDFS-Build/17189/console | | Powered by | Apache Yetus 0.4.0-SNAPSHOT http://yetus.apache.org | This message was automatically generated. > Lower the default maximum number of blocks per file > --------------------------------------------------- > > Key: HDFS-5517 > URL: https://issues.apache.org/jira/browse/HDFS-5517 > Project: Hadoop HDFS > Issue Type: Bug > Components: namenode > Affects Versions: 2.2.0 > Reporter: Aaron T. Myers > Assignee: Aaron T. Myers > Labels: BB2015-05-TBR > Attachments: HDFS-5517.patch > > > We introduced the maximum number of blocks per file in HDFS-4305, but we set > the default to 1MM. In practice this limit is so high as to never be hit, > whereas we know that an individual file with 10s of thousands of blocks can > cause problems. We should lower the default value, in my opinion to 10k. -- This message was sent by Atlassian JIRA (v6.3.4#6332) --------------------------------------------------------------------- To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org