[ https://issues.apache.org/jira/browse/MAPREDUCE-6621?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Junping Du updated MAPREDUCE-6621: ---------------------------------- Fix Version/s: 2.8.0 > Memory Leak in JobClient#submitJobInternal() > -------------------------------------------- > > Key: MAPREDUCE-6621 > URL: https://issues.apache.org/jira/browse/MAPREDUCE-6621 > Project: Hadoop Map/Reduce > Issue Type: Bug > Reporter: Xuan Gong > Assignee: Xuan Gong > Fix For: 2.8.0, 2.7.3, 2.6.4, 3.0.0-alpha1 > > Attachments: MAPREDUCE-6621.1.patch, MAPREDUCE-6621.2.patch > > > In JobClient: > {code} > public RunningJob submitJobInternal(final JobConf conf) > throws FileNotFoundException, IOException { > try { > conf.setBooleanIfUnset("mapred.mapper.new-api", false); > conf.setBooleanIfUnset("mapred.reducer.new-api", false); > Job job = clientUgi.doAs(new PrivilegedExceptionAction<Job> () { > @Override > public Job run() throws IOException, ClassNotFoundException, > InterruptedException { > Job job = Job.getInstance(conf); > job.submit(); > return job; > } > }); > // update our Cluster instance with the one created by Job for > submission > // (we can't pass our Cluster instance to Job, since Job wraps the > config > // instance, and the two configs would then diverge) > cluster = job.getCluster(); > return new NetworkedJob(job); > } catch (InterruptedException ie) { > throw new IOException("interrupted", ie); > } > } > {code} > We will replace the cluster object with the cluster object from Job, but the > previous old cluster object would never be closed. -- This message was sent by Atlassian JIRA (v6.3.4#6332) --------------------------------------------------------------------- To unsubscribe, e-mail: mapreduce-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: mapreduce-issues-h...@hadoop.apache.org