[ https://issues.apache.org/jira/browse/MAPREDUCE-7457?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17777394#comment-17777394 ]
ASF GitHub Bot commented on MAPREDUCE-7457: ------------------------------------------- slfan1989 commented on code in PR #6155: URL: https://github.com/apache/hadoop/pull/6155#discussion_r1365883869 ########## hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/MapTask.java: ########## @@ -984,10 +987,16 @@ public void init(MapOutputCollector.Context context MRJobConfig.DEFAULT_IO_SORT_MB); indexCacheMemoryLimit = job.getInt(JobContext.INDEX_CACHE_MEMORY_LIMIT, INDEX_CACHE_MEMORY_LIMIT_DEFAULT); + spillFilesCountLimit = job.getInt(JobContext.SPILL_FILES_COUNT_LIMIT, + SPILL_FILES_COUNT_LIMIT_DEFAULT); Review Comment: We need to pay attention to indentation > Limit number of spill files getting created > ------------------------------------------- > > Key: MAPREDUCE-7457 > URL: https://issues.apache.org/jira/browse/MAPREDUCE-7457 > Project: Hadoop Map/Reduce > Issue Type: Improvement > Reporter: Mudit Sharma > Priority: Critical > Labels: pull-request-available > > Hi, > > We have been facing some issues where many of our cluster node disks go full > because of some rogue applications creating a lot of spill data > We wanted to fail the app if more than a threshold amount of spill files are > written > Please let us know if any such capability is supported > > If the capability is not there, we are proposing it to support it via a > config, we have added a PR for the same: > [https://github.com/apache/hadoop/pull/6155] please let us know your > thoughts on it -- This message was sent by Atlassian Jira (v8.20.10#820010) --------------------------------------------------------------------- To unsubscribe, e-mail: mapreduce-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: mapreduce-issues-h...@hadoop.apache.org