Author: nigel
Date: Thu Nov 13 17:58:43 2008
New Revision: 713903
URL: http://svn.apache.org/viewvc?rev=713903&view=rev
Log:
Preparing for release 0.19.0 RC1
Modified:
hadoop/core/branches/branch-0.19/CHANGES.txt
hadoop/core/branches/branch-0.19/build.xml
hadoop/core/branches/branch-0.19/docs/changes.html
hadoop/core/branches/branch-0.19/docs/releasenotes.html
Modified: hadoop/core/branches/branch-0.19/CHANGES.txt
URL:
http://svn.apache.org/viewvc/hadoop/core/branches/branch-0.19/CHANGES.txt?rev=713903&r1=713902&r2=713903&view=diff
==============================================================================
--- hadoop/core/branches/branch-0.19/CHANGES.txt (original)
+++ hadoop/core/branches/branch-0.19/CHANGES.txt Thu Nov 13 17:58:43 2008
@@ -1,6 +1,6 @@
Hadoop Change Log
-Release 0.19.0 - Unreleased
+Release 0.19.0 - 2008-11-18
INCOMPATIBLE CHANGES
Modified: hadoop/core/branches/branch-0.19/build.xml
URL:
http://svn.apache.org/viewvc/hadoop/core/branches/branch-0.19/build.xml?rev=713903&r1=713902&r2=713903&view=diff
==============================================================================
--- hadoop/core/branches/branch-0.19/build.xml (original)
+++ hadoop/core/branches/branch-0.19/build.xml Thu Nov 13 17:58:43 2008
@@ -26,7 +26,7 @@
<property name="Name" value="Hadoop"/>
<property name="name" value="hadoop"/>
- <property name="version" value="0.19.0-dev"/>
+ <property name="version" value="0.19.1-dev"/>
<property name="final.name" value="${name}-${version}"/>
<property name="year" value="2008"/>
<property name="libhdfs.version" value="1"/>
Modified: hadoop/core/branches/branch-0.19/docs/changes.html
URL:
http://svn.apache.org/viewvc/hadoop/core/branches/branch-0.19/docs/changes.html?rev=713903&r1=713902&r2=713903&view=diff
==============================================================================
--- hadoop/core/branches/branch-0.19/docs/changes.html (original)
+++ hadoop/core/branches/branch-0.19/docs/changes.html Thu Nov 13 17:58:43 2008
@@ -36,7 +36,7 @@
function collapse() {
for (var i = 0; i < document.getElementsByTagName("ul").length; i++) {
var list = document.getElementsByTagName("ul")[i];
- if (list.id != 'release_0.19.0_-_unreleased_' && list.id !=
'release_0.18.2_-_unreleased_') {
+ if (list.id != 'release_0.19.0_-_2008-11-18_' && list.id !=
'release_0.18.3_-_unreleased_') {
list.style.display = "none";
}
}
@@ -52,12 +52,12 @@
<a href="http://hadoop.apache.org/core/"><img class="logoImage" alt="Hadoop"
src="images/hadoop-logo.jpg" title="Scalable Computing Platform"></a>
<h1>Hadoop Change Log</h1>
-<h2><a href="javascript:toggleList('release_0.19.0_-_unreleased_')">Release
0.19.0 - Unreleased
+<h2><a href="javascript:toggleList('release_0.19.0_-_2008-11-18_')">Release
0.19.0 - 2008-11-18
</a></h2>
-<ul id="release_0.19.0_-_unreleased_">
- <li><a
href="javascript:toggleList('release_0.19.0_-_unreleased_._incompatible_changes_')">
INCOMPATIBLE CHANGES
-</a> (21)
- <ol id="release_0.19.0_-_unreleased_._incompatible_changes_">
+<ul id="release_0.19.0_-_2008-11-18_">
+ <li><a
href="javascript:toggleList('release_0.19.0_-_2008-11-18_._incompatible_changes_')">
INCOMPATIBLE CHANGES
+</a> (22)
+ <ol id="release_0.19.0_-_2008-11-18_._incompatible_changes_">
<li><a
href="http://issues.apache.org/jira/browse/HADOOP-3595">HADOOP-3595</a>. Remove
deprecated methods for mapred.combine.once
functionality, which was necessary to providing backwards
compatible combiner semantics for 0.18.<br />(cdouglas via omalley)</li>
@@ -124,11 +124,12 @@
DFS command line report reflects the same change. Config parameter
dfs.datanode.du.pct is no longer used and is removed from the
hadoop-default.xml.<br />(Suresh Srinivas via hairong)</li>
+ <li><a
href="http://issues.apache.org/jira/browse/HADOOP-4116">HADOOP-4116</a>.
Balancer should provide better resource management.<br />(hairong)</li>
</ol>
</li>
- <li><a
href="javascript:toggleList('release_0.19.0_-_unreleased_._new_features_')">
NEW FEATURES
-</a> (40)
- <ol id="release_0.19.0_-_unreleased_._new_features_">
+ <li><a
href="javascript:toggleList('release_0.19.0_-_2008-11-18_._new_features_')">
NEW FEATURES
+</a> (39)
+ <ol id="release_0.19.0_-_2008-11-18_._new_features_">
<li><a
href="http://issues.apache.org/jira/browse/HADOOP-3341">HADOOP-3341</a>. Allow
streaming jobs to specify the field separator for map
and reduce input and output. The new configuration values are:
stream.map.input.field.separator
@@ -209,13 +210,11 @@
<li><a
href="http://issues.apache.org/jira/browse/HADOOP-3019">HADOOP-3019</a>. A new
library to support total order partitions.<br />(cdouglas via omalley)</li>
<li><a
href="http://issues.apache.org/jira/browse/HADOOP-3924">HADOOP-3924</a>. Added
a 'KILLED' job status.<br />(Subramaniam Krishnan via
acmurthy)</li>
- <li><a
href="http://issues.apache.org/jira/browse/HADOOP-2421">HADOOP-2421</a>. Add
jdiff output to documentation, listing all API
-changes from the prior release.<br />(cutting)</li>
</ol>
</li>
- <li><a
href="javascript:toggleList('release_0.19.0_-_unreleased_._improvements_')">
IMPROVEMENTS
+ <li><a
href="javascript:toggleList('release_0.19.0_-_2008-11-18_._improvements_')">
IMPROVEMENTS
</a> (78)
- <ol id="release_0.19.0_-_unreleased_._improvements_">
+ <ol id="release_0.19.0_-_2008-11-18_._improvements_">
<li><a
href="http://issues.apache.org/jira/browse/HADOOP-4205">HADOOP-4205</a>. hive:
metastore and ql to use the refactored SerDe library.<br />(zshao)</li>
<li><a
href="http://issues.apache.org/jira/browse/HADOOP-4106">HADOOP-4106</a>.
libhdfs: add time, permission and user attribute support
(part 2).<br />(Pete Wyckoff through zshao)</li>
@@ -345,8 +344,6 @@
<li><a
href="http://issues.apache.org/jira/browse/HADOOP-4301">HADOOP-4301</a>. Adds
forrest doc for the skip bad records feature.<br />(Sharad Agarwal via
ddas)</li>
<li><a
href="http://issues.apache.org/jira/browse/HADOOP-4354">HADOOP-4354</a>.
Separate TestDatanodeDeath.testDatanodeDeath() into 4 tests.<br
/>(szetszwo)</li>
<li><a
href="http://issues.apache.org/jira/browse/HADOOP-3790">HADOOP-3790</a>. Add
more unit tests for testing HDFS file append.<br />(szetszwo)</li>
- <li><a
href="http://issues.apache.org/jira/browse/HADOOP-4150">HADOOP-4150</a>.
Include librecordio in hadoop releases.<br />(Giridharan Kesavan
-via acmurthy)</li>
<li><a
href="http://issues.apache.org/jira/browse/HADOOP-4321">HADOOP-4321</a>.
Include documentation for the capacity scheduler.<br />(Hemanth
Yamijala via omalley)</li>
<li><a
href="http://issues.apache.org/jira/browse/HADOOP-4424">HADOOP-4424</a>. Change
menu layout for Hadoop documentation (Boris Shkolnik
@@ -355,11 +352,12 @@
<li><a
href="http://issues.apache.org/jira/browse/HADOOP-4438">HADOOP-4438</a>. Update
forrest documentation to include missing FsShell
commands.<br />(Suresh Srinivas via cdouglas)</li>
<li><a
href="http://issues.apache.org/jira/browse/HADOOP-4105">HADOOP-4105</a>. Add
forrest documentation for libhdfs.<br />(Pete Wyckoff via cutting)</li>
+ <li><a
href="http://issues.apache.org/jira/browse/HADOOP-4510">HADOOP-4510</a>. Make
getTaskOutputPath public.<br />(Chris Wensel via omalley)</li>
</ol>
</li>
- <li><a
href="javascript:toggleList('release_0.19.0_-_unreleased_._optimizations_')">
OPTIMIZATIONS
+ <li><a
href="javascript:toggleList('release_0.19.0_-_2008-11-18_._optimizations_')">
OPTIMIZATIONS
</a> (11)
- <ol id="release_0.19.0_-_unreleased_._optimizations_">
+ <ol id="release_0.19.0_-_2008-11-18_._optimizations_">
<li><a
href="http://issues.apache.org/jira/browse/HADOOP-3556">HADOOP-3556</a>.
Removed lock contention in MD5Hash by changing the
singleton MessageDigester by an instance per Thread using
ThreadLocal.<br />(Iv?n de Prado via omalley)</li>
@@ -386,9 +384,9 @@
org.apache.hadoop.mapred package private instead of public.<br
/>(omalley)</li>
</ol>
</li>
- <li><a
href="javascript:toggleList('release_0.19.0_-_unreleased_._bug_fixes_')"> BUG
FIXES
-</a> (146)
- <ol id="release_0.19.0_-_unreleased_._bug_fixes_">
+ <li><a
href="javascript:toggleList('release_0.19.0_-_2008-11-18_._bug_fixes_')"> BUG
FIXES
+</a> (153)
+ <ol id="release_0.19.0_-_2008-11-18_._bug_fixes_">
<li><a
href="http://issues.apache.org/jira/browse/HADOOP-3563">HADOOP-3563</a>.
Refactor the distributed upgrade code so that it is
easier to identify datanode and namenode related code.<br />(dhruba)</li>
<li><a
href="http://issues.apache.org/jira/browse/HADOOP-3640">HADOOP-3640</a>. Fix
the read method in the NativeS3InputStream.<br />(tomwhite via
@@ -608,8 +606,6 @@
<li><a
href="http://issues.apache.org/jira/browse/HADOOP-4018">HADOOP-4018</a>. The
number of tasks for a single job cannot exceed a
pre-configured maximum value.<br />(dhruba)</li>
<li><a
href="http://issues.apache.org/jira/browse/HADOOP-4288">HADOOP-4288</a>. Fixes
a NPE problem in CapacityScheduler.<br />(Amar Kamat via ddas)</li>
- <li><a
href="http://issues.apache.org/jira/browse/HADOOP-3883">HADOOP-3883</a>. Limit
namenode to assign at most one generation stamp for
-a particular block within a short period.<br />(szetszwo)</li>
<li><a
href="http://issues.apache.org/jira/browse/HADOOP-4014">HADOOP-4014</a>. Create
hard links with 'fsutil hardlink' on Windows.<br />(shv)</li>
<li><a
href="http://issues.apache.org/jira/browse/HADOOP-4393">HADOOP-4393</a>. Merged
org.apache.hadoop.fs.permission.AccessControlException
and org.apache.hadoop.security.AccessControlIOException into a single
@@ -619,7 +615,6 @@
<li><a
href="http://issues.apache.org/jira/browse/HADOOP-4361">HADOOP-4361</a>. Makes
sure that jobs killed from command line are killed
fast (i.e., there is a slot to run the cleanup task soon).<br />(Amareshwari
Sriramadasu via ddas)</li>
<li><a
href="http://issues.apache.org/jira/browse/HADOOP-4400">HADOOP-4400</a>. Add
"hdfs://" to fs.default.name on quickstart.html.<br />(Jeff Hammerbacher via
omalley)</li>
- <li><a
href="http://issues.apache.org/jira/browse/HADOOP-4403">HADOOP-4403</a>. Make
TestLeaseRecovery and TestFileCreation more robust.<br />(szetszwo)</li>
<li><a
href="http://issues.apache.org/jira/browse/HADOOP-4378">HADOOP-4378</a>. Fix
TestJobQueueInformation to use SleepJob rather than
WordCount via TestMiniMRWithDFS.<br />(Sreekanth Ramakrishnan via
acmurthy)</li>
<li><a
href="http://issues.apache.org/jira/browse/HADOOP-4376">HADOOP-4376</a>. Fix
formatting in hadoop-default.xml for
@@ -665,36 +660,95 @@
<li><a
href="http://issues.apache.org/jira/browse/HADOOP-4358">HADOOP-4358</a>. No
need to truncate access time in INode. Also fixes NPE
in CreateEditsLog.<br />(Raghu Angadi)</li>
<li><a
href="http://issues.apache.org/jira/browse/HADOOP-4387">HADOOP-4387</a>.
TestHDFSFileSystemContract fails on windows nightly builds.<br />(Raghu
Angadi)</li>
+ <li><a
href="http://issues.apache.org/jira/browse/HADOOP-4466">HADOOP-4466</a>. Ensure
that SequenceFileOutputFormat isn't tied to Writables
+and can be used with other Serialization frameworks.<br />(Chris Wensel via
+acmurthy)</li>
+ <li><a
href="http://issues.apache.org/jira/browse/HADOOP-4525">HADOOP-4525</a>. Fix
ipc.server.ipcnodelay originally missed in in <a
href="http://issues.apache.org/jira/browse/HADOOP-2232">HADOOP-2232</a>.<br
/>(cdouglas via Clint Morgan)</li>
+ <li><a
href="http://issues.apache.org/jira/browse/HADOOP-4498">HADOOP-4498</a>. Ensure
that JobHistory correctly escapes the job name so that
+regex patterns work.<br />(Chris Wensel via acmurthy)</li>
+ <li><a
href="http://issues.apache.org/jira/browse/HADOOP-4446">HADOOP-4446</a>. Modify
guaranteed capacity labels in capacity scheduler's UI
+to reflect the information being displayed.<br />(Sreekanth Ramakrishnan via
+yhemanth)</li>
+ <li><a
href="http://issues.apache.org/jira/browse/HADOOP-4282">HADOOP-4282</a>. Some
user facing URLs are not filtered by user filters.<br />(szetszwo)</li>
+ <li><a
href="http://issues.apache.org/jira/browse/HADOOP-4595">HADOOP-4595</a>. Fixes
two race conditions - one to do with updating free slot count,
+and another to do with starting the MapEventsFetcher thread.<br />(ddas)</li>
+ <li><a
href="http://issues.apache.org/jira/browse/HADOOP-4552">HADOOP-4552</a>. Fix a
deadlock in RPC server.<br />(Raghu Angadi)</li>
+ <li><a
href="http://issues.apache.org/jira/browse/HADOOP-4471">HADOOP-4471</a>. Sort
running jobs by priority in the capacity scheduler.<br />(Amar Kamat via
yhemanth)</li>
+ <li><a
href="http://issues.apache.org/jira/browse/HADOOP-4500">HADOOP-4500</a>. Fix
MultiFileSplit to get the FileSystem from the relevant
+path rather than the JobClient.<br />(Joydeep Sen Sarma via dhruba)</li>
</ol>
</li>
</ul>
-<h2><a href="javascript:toggleList('release_0.18.2_-_unreleased_')">Release
0.18.2 - Unreleased
+<h2><a href="javascript:toggleList('release_0.18.3_-_unreleased_')">Release
0.18.3 - Unreleased
</a></h2>
-<ul id="release_0.18.2_-_unreleased_">
- <li><a
href="javascript:toggleList('release_0.18.2_-_unreleased_._bug_fixes_')"> BUG
FIXES
-</a> (11)
- <ol id="release_0.18.2_-_unreleased_._bug_fixes_">
- <li><a
href="http://issues.apache.org/jira/browse/HADOOP-4116">HADOOP-4116</a>.
Balancer should provide better resource management.<br />(hairong)</li>
+<ul id="release_0.18.3_-_unreleased_">
+ <li><a
href="javascript:toggleList('release_0.18.3_-_unreleased_._improvements_')">
IMPROVEMENTS
+</a> (1)
+ <ol id="release_0.18.3_-_unreleased_._improvements_">
+ <li><a
href="http://issues.apache.org/jira/browse/HADOOP-4150">HADOOP-4150</a>.
Include librecordio in hadoop releases.<br />(Giridharan Kesavan
+via acmurthy)</li>
+ </ol>
+ </li>
+ <li><a
href="javascript:toggleList('release_0.18.3_-_unreleased_._bug_fixes_')"> BUG
FIXES
+</a> (7)
+ <ol id="release_0.18.3_-_unreleased_._bug_fixes_">
+ <li><a
href="http://issues.apache.org/jira/browse/HADOOP-4499">HADOOP-4499</a>.
DFSClient should invoke checksumOk only once.<br />(Raghu Angadi)</li>
+ <li><a
href="http://issues.apache.org/jira/browse/HADOOP-4597">HADOOP-4597</a>.
Calculate mis-replicated blocks when safe-mode is turned
+off manually.<br />(shv)</li>
+ <li><a
href="http://issues.apache.org/jira/browse/HADOOP-3121">HADOOP-3121</a>. lsr
should keep listing the remaining items but not
+terminate if there is any IOException.<br />(szetszwo)</li>
+ <li><a
href="http://issues.apache.org/jira/browse/HADOOP-4610">HADOOP-4610</a>. Always
calculate mis-replicated blocks when safe-mode is
+turned off.<br />(shv)</li>
+ <li><a
href="http://issues.apache.org/jira/browse/HADOOP-3883">HADOOP-3883</a>. Limit
namenode to assign at most one generation stamp for
+a particular block within a short period.<br />(szetszwo)</li>
+ <li><a
href="http://issues.apache.org/jira/browse/HADOOP-4556">HADOOP-4556</a>. Block
went missing.<br />(hairong)</li>
+ <li><a
href="http://issues.apache.org/jira/browse/HADOOP-4643">HADOOP-4643</a>.
NameNode should exclude excessive replicas when counting
+live replicas for a block.<br />(hairong)</li>
+ </ol>
+ </li>
+</ul>
+<h2><a href="javascript:toggleList('older')">Older Releases</a></h2>
+<ul id="older">
+<h3><a href="javascript:toggleList('release_0.18.2_-_2008-11-03_')">Release
0.18.2 - 2008-11-03
+</a></h3>
+<ul id="release_0.18.2_-_2008-11-03_">
+ <li><a
href="javascript:toggleList('release_0.18.2_-_2008-11-03_._bug_fixes_')"> BUG
FIXES
+</a> (16)
+ <ol id="release_0.18.2_-_2008-11-03_._bug_fixes_">
<li><a
href="http://issues.apache.org/jira/browse/HADOOP-3614">HADOOP-3614</a>. Fix a
bug that Datanode may use an old GenerationStamp to get
meta file.<br />(szetszwo)</li>
<li><a
href="http://issues.apache.org/jira/browse/HADOOP-4314">HADOOP-4314</a>.
Simulated datanodes should not include blocks that are still
being written in their block report.<br />(Raghu Angadi)</li>
- <li><a
href="http://issues.apache.org/jira/browse/HADOOP-4228">HADOOP-4228</a>. dfs
datanoe metrics, bytes_read and bytes_written, overflow
+ <li><a
href="http://issues.apache.org/jira/browse/HADOOP-4228">HADOOP-4228</a>. dfs
datanode metrics, bytes_read and bytes_written, overflow
due to incorrect type used.<br />(hairong)</li>
<li><a
href="http://issues.apache.org/jira/browse/HADOOP-4395">HADOOP-4395</a>. The
FSEditLog loading is incorrect for the case OP_SET_OWNER.<br />(szetszwo)</li>
<li><a
href="http://issues.apache.org/jira/browse/HADOOP-4351">HADOOP-4351</a>.
FSNamesystem.getBlockLocationsInternal throws
ArrayIndexOutOfBoundsException.<br />(hairong)</li>
+ <li><a
href="http://issues.apache.org/jira/browse/HADOOP-4403">HADOOP-4403</a>. Make
TestLeaseRecovery and TestFileCreation more robust.<br />(szetszwo)</li>
<li><a
href="http://issues.apache.org/jira/browse/HADOOP-4292">HADOOP-4292</a>. Do not
support append() for LocalFileSystem.<br />(hairong)</li>
<li><a
href="http://issues.apache.org/jira/browse/HADOOP-4399">HADOOP-4399</a>. Make
fuse-dfs multi-thread access safe.<br />(Pete Wyckoff via dhruba)</li>
<li><a
href="http://issues.apache.org/jira/browse/HADOOP-4369">HADOOP-4369</a>. Use
setMetric(...) instead of incrMetric(...) for metrics
averages.<br />(Brian Bockelman via szetszwo)</li>
<li><a
href="http://issues.apache.org/jira/browse/HADOOP-4469">HADOOP-4469</a>. Rename
and add the ant task jar file to the tar file.<br />(nigel)</li>
- <li><a
href="http://issues.apache.org/jira/browse/HADOOP-3914">HADOOP-3914</a>.
DFSClient sends Checksum Ok only once for a block.<br />(Christain Kunz via
hairong)</li>
+ <li><a
href="http://issues.apache.org/jira/browse/HADOOP-3914">HADOOP-3914</a>.
DFSClient sends Checksum Ok only once for a block.<br />(Christian Kunz via
hairong)</li>
+ <li><a
href="http://issues.apache.org/jira/browse/HADOOP-4467">HADOOP-4467</a>.
SerializationFactory now uses the current context ClassLoader
+allowing for user supplied Serialization instances.<br />(Chris Wensel via
+acmurthy)</li>
+ <li><a
href="http://issues.apache.org/jira/browse/HADOOP-4517">HADOOP-4517</a>.
Release FSDataset lock before joining ongoing create threads.<br
/>(szetszwo)</li>
+ <li><a
href="http://issues.apache.org/jira/browse/HADOOP-4526">HADOOP-4526</a>. fsck
failing with NullPointerException.<br />(hairong)</li>
+ <li><a
href="http://issues.apache.org/jira/browse/HADOOP-4483">HADOOP-4483</a> Honor
the max parameter in DatanodeDescriptor.getBlockArray(..)<br />(Ahad Rana and
Hairong Kuang via szetszwo)</li>
+ <li><a
href="http://issues.apache.org/jira/browse/HADOOP-4340">HADOOP-4340</a>.
Correctly set the exit code from JobShell.main so that the
+'hadoop jar' command returns the right code to the user.<br />(acmurthy)</li>
+ </ol>
+ </li>
+ <li><a
href="javascript:toggleList('release_0.18.2_-_2008-11-03_._new_features_')">
NEW FEATURES
+</a> (1)
+ <ol id="release_0.18.2_-_2008-11-03_._new_features_">
+ <li><a
href="http://issues.apache.org/jira/browse/HADOOP-2421">HADOOP-2421</a>. Add
jdiff output to documentation, listing all API
+changes from the prior release.<br />(cutting)</li>
</ol>
</li>
</ul>
-<h2><a href="javascript:toggleList('older')">Older Releases</a></h2>
-<ul id="older">
<h3><a href="javascript:toggleList('release_0.18.1_-_2008-09-17_')">Release
0.18.1 - 2008-09-17
</a></h3>
<ul id="release_0.18.1_-_2008-09-17_">
Modified: hadoop/core/branches/branch-0.19/docs/releasenotes.html
URL:
http://svn.apache.org/viewvc/hadoop/core/branches/branch-0.19/docs/releasenotes.html?rev=713903&r1=713902&r2=713903&view=diff
==============================================================================
--- hadoop/core/branches/branch-0.19/docs/releasenotes.html (original)
+++ hadoop/core/branches/branch-0.19/docs/releasenotes.html Thu Nov 13 17:58:43
2008
@@ -9,7 +9,7 @@
The table below is sorted by Component.
<ul><a name="changes">
<h2>Changes Since Hadoop 0.18.2</h2>
- <table border="1""100%" cellpadding="4">
+<table border="1">
<tr bgcolor="#DDDDDD">
<th align="left">Issue</th><th align="left">Component</th><th
align="left">Notes</th>
</tr>
@@ -17,17 +17,7 @@
<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-2325">HADOOP-2325</a></td><td>build</td><td>Hadoop
now requires Java 6.</td>
</tr>
<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-4150">HADOOP-4150</a></td><td>build</td><td>librecordio
has a build dependency with xerces-c parser. so to get this build; need to
set the path of xerces-c binary to xercescroot and librecordio to true
-<br/>
-exact command would be like this :
-<br/>
-
-<br/>
-ant -Dlibrecordio=true -Dxercescroot=<path to the xerces-c root>
-<br/>
-
-<br/>
-This patch includes targets for compiling, packaging and testing the
librecordio library.</td>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3730">HADOOP-3730</a></td><td>conf</td><td> Added
a JobConf constructor that disables loading default configurations so as to
take all default values from the JobTracker's configuration.</td>
</tr>
<tr>
<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3722">HADOOP-3722</a></td><td>conf</td><td>Changed
streaming StreamJob and Submitter to implement Tool and Configurable, and to
use GenericOptionsParser arguments -fs, -jt, -conf, -D, -libjars, -files, and
-archives. Deprecated -jobconf, -cacheArchive, -dfs, -cacheArchive,
-additionalconfspec, from streaming and pipes in favor of the generic options.
Removed from streaming -config, -mapred.job.tracker, and -cluster.</td>
@@ -36,6 +26,11 @@
<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3646">HADOOP-3646</a></td><td>conf</td><td>Introduced
support for bzip2 compressed files.</td>
</tr>
<tr>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3445">HADOOP-3445</a></td><td>contrib/capacity-sched</td><td>Introduced
Capacity Task Scheduler.
+<br/>
+</td>
+</tr>
+<tr>
<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3719">HADOOP-3719</a></td><td>contrib/chukwa</td><td>Introduced
Chukwa data collection and analysis framework.</td>
</tr>
<tr>
@@ -45,25 +40,22 @@
<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-4117">HADOOP-4117</a></td><td>contrib/ec2</td><td>Changed
scripts to pass initialization script for EC2 instances at boot time (as EC2
user data) rather than embedding initialization information in the EC2 image.
This change makes it easy to customize the hadoop-site.xml file for your
cluster before launch, by editing the hadoop-ec2-init-remote.sh script, or by
setting the environment variable USER_DATA_FILE in hadoop-ec2-env.sh to run a
script of your choice.</td>
</tr>
<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-4106">HADOOP-4106</a></td><td>contrib/fuse-dfs</td><td>Added
time, permission and user attribute support to libhdfs.</td>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3485">HADOOP-3485</a></td><td>contrib/fuse-dfs</td><td>Introduce
write support for Fuse; requires Linux kernel 2.6.15 or better.</td>
</tr>
<tr>
<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3796">HADOOP-3796</a></td><td>contrib/fuse-dfs</td><td>Changed
Fuse configuration to use mount options.</td>
</tr>
<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3908">HADOOP-3908</a></td><td>contrib/fuse-dfs</td><td>Improved
Fuse-dfs better error message if llibhdfs.so doesn't exist.</td>
-</tr>
-<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3485">HADOOP-3485</a></td><td>contrib/fuse-dfs</td><td>Introduce
write support for Fuse; requires Linux kernel 2.6.15 or better.</td>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-4106">HADOOP-4106</a></td><td>contrib/fuse-dfs</td><td>Added
time, permission and user attribute support to libhdfs.</td>
</tr>
<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-4086">HADOOP-4086</a></td><td>contrib/hive</td><td>Add
LIMIT to Hive QL.</td>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3908">HADOOP-3908</a></td><td>contrib/fuse-dfs</td><td>Improved
Fuse-dfs better error message if llibhdfs.so doesn't exist.</td>
</tr>
<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-4205">HADOOP-4205</a></td><td>contrib/hive</td><td>Improved
Hive metastore and ql to use the refactored SerDe library.</td>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-4086">HADOOP-4086</a></td><td>contrib/hive</td><td>Added
LIMIT to Hive query language.</td>
</tr>
<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-4084">HADOOP-4084</a></td><td>contrib/hive</td><td>Introduced
"EXPLAIN" plan for Hive.</td>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-4138">HADOOP-4138</a></td><td>contrib/hive</td><td>Introduced
new SerDe library for src/contrib/hive.</td>
</tr>
<tr>
<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3601">HADOOP-3601</a></td><td>contrib/hive</td><td>Introduced
Hive Data Warehouse built on top of Hadoop that enables structuring Hadoop
files as tables and partitions and allows users to query this data through a
SQL like language using a command line interface.
@@ -71,7 +63,10 @@
</td>
</tr>
<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-4138">HADOOP-4138</a></td><td>contrib/hive</td><td>Introduced
new SerDe library for src/contrib/hive.</td>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-4205">HADOOP-4205</a></td><td>contrib/hive</td><td>Improved
Hive metastore and ql to use the refactored SerDe library.</td>
+</tr>
+<tr>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-4084">HADOOP-4084</a></td><td>contrib/hive</td><td>Introduced
"EXPLAIN" plan for Hive.</td>
</tr>
<tr>
<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3695">HADOOP-3695</a></td><td>contrib/hod</td><td>Added
an ability in HOD to start multiple workers (TaskTrackers and/or DataNodes)
per node to assist testing and simulation of scale. A configuration variable
ringmaster.workers_per_ring was added to specify the number of workers to
start.</td>
@@ -83,82 +78,54 @@
<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-2302">HADOOP-2302</a></td><td>contrib/streaming</td><td>Introduced
numerical key comparison for streaming.</td>
</tr>
<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-4281">HADOOP-4281</a></td><td>dfs</td><td>Changed
command "hadoop dfsadmin -report" to be consistent with Web UI for
both Namenode and Datanode reports. "Total raw bytes" is changed to
"Configured Capacity". "Present Capacity" is newly added to
indicate the present capacity of the DFS. "Remaining raw bytes" is
changed to "DFS Remaining". "Used raw bytes" is changed to
"DFS Used". "% used" is changed to "DFS Used%".
Applications that parse command output should be reviewed.</td>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-4430">HADOOP-4430</a></td><td>dfs</td><td>Changed
reporting in the NameNode Web UI to more closely reflect the behavior of the
re-balancer. Removed no longer used config parameter dfs.datanode.du.pct from
hadoop-default.xml.
+<br/>
+</td>
</tr>
<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-1869">HADOOP-1869</a></td><td>dfs</td><td>Added
HDFS file access times. By default, access times will be precise to the most
recent hour boundary. A configuration parameter dfs.access.time.precision
(milliseconds) is used to control this precision. Setting a value of 0 will
disable persisting access times for HDFS files.</td>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-4281">HADOOP-4281</a></td><td>dfs</td><td>Changed
command "hadoop dfsadmin -report" to be consistent with Web UI for
both Namenode and Datanode reports. "Total raw bytes" is changed to
"Configured Capacity". "Present Capacity" is newly added to
indicate the present capacity of the DFS. "Remaining raw bytes" is
changed to "DFS Remaining". "Used raw bytes" is changed to
"DFS Used". "% used" is changed to "DFS Used%".
Applications that parse command output should be reviewed.</td>
</tr>
<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-4423">HADOOP-4423</a></td><td>dfs</td><td>Added
a new parameter keepLength to ClientDatanodeProtocol.recoverBlock(...)</td>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-2885">HADOOP-2885</a></td><td>dfs</td><td>Restructured
the package hadoop.dfs.</td>
</tr>
<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3992">HADOOP-3992</a></td><td>dfs</td><td>Added
a synthetic load generation facility to the test directory.</td>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-4007">HADOOP-4007</a></td><td>dfs</td><td>Changed
ClientProtocol getStatus and getListing to use the type FileStatus. Removed
type DFSFileInfo.</td>
</tr>
<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-4007">HADOOP-4007</a></td><td>dfs</td><td>Changed
ClientProtocol getStatus and getListing to use the type FileStatus. Removed
type DFSFileInfo.</td>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-2816">HADOOP-2816</a></td><td>dfs</td><td>Improved
space reporting for NameNode Web UI. Applications that parse the Web UI output
should be reviewed.</td>
</tr>
<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3732">HADOOP-3732</a></td><td>dfs</td><td>Delay
intialization of datanode block verification till the verification thread is
started.
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-1700">HADOOP-1700</a></td><td>dfs</td><td>Introduced
append operation for HDFS files.
<br/>
</td>
</tr>
<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3981">HADOOP-3981</a></td><td>dfs</td><td>Implemented
MD5-of-xxxMD5-of-yyyCRC32 which is a distributed file checksum algorithm for
HDFS, where xxx is the number of CRCs per block and yyy is the number of bytes
per CRC.
-<br/>
-
-<br/>
-Changed DistCp to use file checksum for comparing files if both source and
destination FileSystem(s) support getFileChecksum(...).</td>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3938">HADOOP-3938</a></td><td>dfs</td><td>Introducted
byte space quotas for directories. The count shell command modified to report
both name and byte quotas.</td>
</tr>
<tr>
<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-4176">HADOOP-4176</a></td><td>dfs</td><td>Implemented
getFileChecksum(Path) in HftpFileSystemfor distcp support.</td>
</tr>
<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-2816">HADOOP-2816</a></td><td>dfs</td><td>Improved
space reporting for NameNode Web UI. Applications that parse the Web UI output
should be reviewed.</td>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-1869">HADOOP-1869</a></td><td>dfs</td><td>Added
HDFS file access times. By default, access times will be precise to the most
recent hour boundary. A configuration parameter dfs.access.time.precision
(milliseconds) is used to control this precision. Setting a value of 0 will
disable persisting access times for HDFS files.</td>
</tr>
<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-4430">HADOOP-4430</a></td><td>dfs</td><td>Incompatible
changes:
-<br/>
-This change modifies/retains the changes made in 2816 as follows:
-<br/>
-1) Present Capacity added in 2816 is removed from the Web UI
-<br/>
-2) Change of Total Capacity to Configured Capacity and its definition from
2816 is retained in the Web UI
-<br/>
-3) Data node protocol change to report Configured Capacity instead of Total
Capacity is retained.
-<br/>
-4) DFS Used% was calculated as a percentage of Present Capacity. It is changed
to percentage of Configured Capacity.
-<br/>
-
-<br/>
-Other incompatible changes:
-<br/>
-1) Config parameter dfs.datanode.du.pct is no longer used and is removed from
the hadoop-default.xml.
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3981">HADOOP-3981</a></td><td>dfs</td><td>Implemented
MD5-of-xxxMD5-of-yyyCRC32 which is a distributed file checksum algorithm for
HDFS, where xxx is the number of CRCs per block and yyy is the number of bytes
per CRC.
<br/>
<br/>
-2) Namenode Web UI has the following addional changes:
-<br/>
- The following parameters are added to both Cluster Summary
and Datanode information:
-<br/>
- * Non DFS Used - This indicates the disk space taken by non
DFS file
-<br/>
- * DFS remaining % - This is remaining % of Configured
Capacity available for DFS use
-<br/>
-</td>
+Changed DistCp to use file checksum for comparing files if both source and
destination FileSystem(s) support getFileChecksum(...).</td>
</tr>
<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-1700">HADOOP-1700</a></td><td>dfs</td><td>Introduced
append operation for HDFS files.
-<br/>
-</td>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3992">HADOOP-3992</a></td><td>dfs</td><td>Added
a synthetic load generation facility to the test directory.</td>
</tr>
<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3938">HADOOP-3938</a></td><td>dfs</td><td>Introducted
byte space quotas for directories. The count shell command modified to report
both name and byte quotas.</td>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3792">HADOOP-3792</a></td><td>fs</td><td>Changed
exit code from hadoop.fs.FsShell -test to match the usual Unix convention.</td>
</tr>
<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-4358">HADOOP-4358</a></td><td>dfs</td><td>No
need to truncate access time in INode. Also fixes NPE in CreateEditsLog</td>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-4227">HADOOP-4227</a></td><td>fs</td><td>Removed
the deprecated class org.apache.hadoop.fs.ShellCommand.</td>
</tr>
<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-2885">HADOOP-2885</a></td><td>dfs</td><td>Restructured
the package hadoop.dfs.</td>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3498">HADOOP-3498</a></td><td>fs</td><td>Extended
file globbing alternation to cross path components. For example, {/a/b,/c/d}
expands to a path that matches the files /a/b and /c/d.</td>
</tr>
<tr>
<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3911">HADOOP-3911</a></td><td>fs</td><td>Added
a check to fsck options to make sure -files is not the first option so as to
resolve conflicts with GenericOptionsParser.</td>
@@ -167,13 +134,10 @@
<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3941">HADOOP-3941</a></td><td>fs</td><td>Added
new FileSystem APIs: FileChecksum and FileSystem.getFileChecksum(Path).</td>
</tr>
<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3792">HADOOP-3792</a></td><td>fs</td><td>Changed
exit code from hadoop.fs.FsShell -test to match the usual Unix convention.</td>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-4466">HADOOP-4466</a></td><td>io</td><td>Ensure
that SequenceFileOutputFormat isn't tied to Writables and can be used with
other Serialization frameworks.</td>
</tr>
<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3498">HADOOP-3498</a></td><td>fs</td><td>Extended
file globbing alternation to cross path components. For example, {/a/b,/c/d}
expands to a path that matches the files /a/b and /c/d.</td>
-</tr>
-<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-4227">HADOOP-4227</a></td><td>fs</td><td>Removed
the deprecated class org.apache.hadoop.fs.ShellCommand.</td>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-2664">HADOOP-2664</a></td><td>io</td><td>Introduced
LZOP codec.</td>
</tr>
<tr>
<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3549">HADOOP-3549</a></td><td>libhdfs</td><td>Improved
error reporting for libhdfs so permission problems now return EACCES.</td>
@@ -184,47 +148,35 @@
</td>
</tr>
<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3445">HADOOP-3445</a></td><td>mapred</td><td>Introduced
Capacity Task Scheduler.</td>
-</tr>
-<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3595">HADOOP-3595</a></td><td>mapred</td><td> Removed
deprecated methods for mapred.combine.once functionality.</td>
-</tr>
-<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-4018">HADOOP-4018</a></td><td>mapred</td><td>A
new configuration parameter named mapred.max.tasks.per.job that specifies the
maximum number of tasks per job.
-<br/>
-
-<br/>
-</td>
-</tr>
-<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-4261">HADOOP-4261</a></td><td>mapred</td><td>A
setup task (for doing OutputCommitter.setupJob()) will be launched, when the
job is in PREP state and after initializing tasks. Once the setup task
completes, the job will be moved to RUNNING state.</td>
-</tr>
-<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3019">HADOOP-3019</a></td><td>mapred</td><td>Added
a partitioner that effects a total order of output data, and an input sampler
for generating the partition keyset for TotalOrderPartitioner for when the
map's input keytype and distribution approximates its output.</td>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-4190">HADOOP-4190</a></td><td>mapred</td><td>Changed
job history format to add a dot at end of each line.</td>
</tr>
<tr>
<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3970">HADOOP-3970</a></td><td>mapred</td><td>Added
getEscapedCompactString() and fromEscapedCompactString() to Counters.java to
represent counters as Strings and to reconstruct the counters from the
Strings.</td>
</tr>
<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3412">HADOOP-3412</a></td><td>mapred</td><td>Added
the ability to chose between many schedulers, and to limit the number of
running tasks per job.</td>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3245">HADOOP-3245</a></td><td>mapred</td><td>Introduced
recovery of jobs when JobTracker restarts. This facility is off by default.
Introduced config parameters mapred.jobtracker.restart.recover,
mapred.jobtracker.job.history.block.size, and
mapred.jobtracker.job.history.buffer.size.</td>
</tr>
<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3581">HADOOP-3581</a></td><td>mapred</td><td>Added
the ability to kill process trees transgressing memory limits. TaskTracker
uses the configuration parameters introduced in <a
href="https://issues.apache.org:443/jira/browse/HADOOP-3759" title="Provide
ability to run memory intensive jobs without affecting other running tasks on
the nodes"><strike>HADOOP-3759</strike></a>. In addition,
mapred.tasktracker.taskmemorymanager.monitoring-interval specifies the interval
for which TT waits between cycles of monitoring tasks' memory usage, and
mapred.tasktracker.procfsbasedprocesstree.sleeptime-before-sigkill specifies
the time TT waits for sending a SIGKILL to a process-tree that has overrun
memory limits, after it has been sent a SIGTERM.</td>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3829">HADOOP-3829</a></td><td>mapred</td><td>Introduced
new config parameter
org.apache.hadoop.mapred.SkipBadRecords.setMapperMaxSkipRecords to set range of
records to be skipped in the neighborhood of a failed record.</td>
</tr>
<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3684">HADOOP-3684</a></td><td>mapred</td><td>Allowed
user to overwrite clone function in a subclass of TaggedMapOutput class.</td>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-4018">HADOOP-4018</a></td><td>mapred</td><td>Introduced
new configuration parameter mapred.max.tasks.per.job to specifie the maximum
number of tasks per job.
+<br/>
+
+<br/>
+</td>
</tr>
<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3930">HADOOP-3930</a></td><td>mapred</td><td>Changed
TaskScheduler to expose API for Web UI and Command Line Tool.</td>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-249">HADOOP-249</a></td><td>mapred</td><td>Enabled
task JVMs to be reused via the job config mapred.job.reuse.jvm.num.tasks.</td>
</tr>
<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-4190">HADOOP-4190</a></td><td>mapred</td><td>Changed
job history format to add a dot at end of each line.</td>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3746">HADOOP-3746</a></td><td>mapred</td><td>Introduced
Fair Scheduler.</td>
</tr>
<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3478">HADOOP-3478</a></td><td>mapred</td><td>Changed
reducers to fetch maps in the same order for a given host to speed up
identification of the faulty maps; reducers still randomize the host selection
to distribute load.</td>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-153">HADOOP-153</a></td><td>mapred</td><td>Introduced
record skipping where tasks fail on certain records.
(org.apache.hadoop.mapred.SkipBadRecords)</td>
</tr>
<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3829">HADOOP-3829</a></td><td>mapred</td><td>If
the number of skipped records in the neighborhood of a bad record are not
acceptable to the user, then narrow down the skipped range to the user
acceptable value. Use
org.apache.hadoop.mapred.SkipBadRecords.setMapperMaxSkipRecords for setting the
acceptable value.</td>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3652">HADOOP-3652</a></td><td>mapred</td><td>Removed
deprecated org.apache.hadoop.mapred.OutputFormatBase.</td>
</tr>
<tr>
<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3702">HADOOP-3702</a></td><td>mapred</td><td>Introduced
ChainMapper and the ChainReducer classes to allow composing chains of Maps and
Reduces in a single Map/Reduce job, something like MAP+ REDUCE MAP*.
@@ -232,22 +184,19 @@
</td>
</tr>
<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3746">HADOOP-3746</a></td><td>mapred</td><td>Introduced
Fair Scheduler.</td>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3595">HADOOP-3595</a></td><td>mapred</td><td> Removed
deprecated methods for mapred.combine.once functionality.</td>
</tr>
<tr>
<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3149">HADOOP-3149</a></td><td>mapred</td><td>Introduced
MultipleOutputs class so Map/Reduce jobs can write data to different output
files. Each output can use a different OutputFormat. Outpufiles are created
within the job output directory. FileOutputFormat.getPathForCustomFile()
creates a filename under the outputdir that is named with the task ID and task
type (i.e. myfile-r-00001).</td>
</tr>
<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-153">HADOOP-153</a></td><td>mapred</td><td>Introduced
record skipping where tasks fail on certain records.
(org.apache.hadoop.mapred.SkipBadRecords)</td>
-</tr>
-<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3245">HADOOP-3245</a></td><td>mapred</td><td>Introduced
recovery of jobs when JobTracker restarts. This facility is off by default.
Introduced config parameters mapred.jobtracker.restart.recover,
mapred.jobtracker.job.history.block.size, and
mapred.jobtracker.job.history.buffer.size.</td>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3684">HADOOP-3684</a></td><td>mapred</td><td>Allowed
user to overwrite clone function in a subclass of TaggedMapOutput class.</td>
</tr>
<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-249">HADOOP-249</a></td><td>mapred</td><td>Jobs
can enable task JVMs to be reused via the job config
mapred.job.reuse.jvm.num.tasks. If this is 1 (the default), then JVMs are not
reused (1 task per JVM). If it is -1, there is no limit to the number of tasks
a JVM can run (of the same job). One can also specify some value greater than
1. Also a JobConf API has been added - setNumTasksToExecutePerJvm.</td>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3930">HADOOP-3930</a></td><td>mapred</td><td>Changed
TaskScheduler to expose API for Web UI and Command Line Tool.</td>
</tr>
<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-4293">HADOOP-4293</a></td><td>mapred</td><td>Made
Configuration Writable and rename the old write method to writeXml.</td>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3478">HADOOP-3478</a></td><td>mapred</td><td>Changed
reducers to fetch maps in the same order for a given host to speed up
identification of the faulty maps; reducers still randomize the host selection
to distribute load.</td>
</tr>
<tr>
<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3150">HADOOP-3150</a></td><td>mapred</td><td>Moved
task file promotion to the Task. When the task has finished, it will do a
commit and is declared SUCCEDED. Job cleanup is done by a separate task. Job is
declared SUCCEDED/FAILED after the cleanup task has finished. Added public
classes org.apache.hadoop.mapred.JobContext, TaskAttemptContext,
OutputCommitter and FileOutputCommiitter. Added public APIs: public
OutputCommitter getOutputCommitter() and
@@ -255,10 +204,19 @@
public void setOutputCommitter(Class<? extends OutputCommitter>
theClass) in org.apache.hadoop.mapred.JobConf</td>
</tr>
<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-4267">HADOOP-4267</a></td><td>mapred</td><td>Occasional
exceptions during shutting down HSQLDB is logged but not rethrown.</td>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3019">HADOOP-3019</a></td><td>mapred</td><td>Added
a partitioner that effects a total order of output data, and an input sampler
for generating the partition keyset for TotalOrderPartitioner for when the
map's input keytype and distribution approximates its output.</td>
</tr>
<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3652">HADOOP-3652</a></td><td>mapred</td><td>Removed
deprecated org.apache.hadoop.mapred.OutputFormatBase.</td>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3581">HADOOP-3581</a></td><td>mapred</td><td>Added
the ability to kill process trees transgressing memory limits. TaskTracker
uses the configuration parameters introduced in <a
href="https://issues.apache.org:443/jira/browse/HADOOP-3759" title="Provide
ability to run memory intensive jobs without affecting other running tasks on
the nodes"><strike>HADOOP-3759</strike></a>. In addition,
mapred.tasktracker.taskmemorymanager.monitoring-interval specifies the interval
for which TT waits between cycles of monitoring tasks' memory usage, and
mapred.tasktracker.procfsbasedprocesstree.sleeptime-before-sigkill specifies
the time TT waits for sending a SIGKILL to a process-tree that has overrun
memory limits, after it has been sent a SIGTERM.</td>
+</tr>
+<tr>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3828">HADOOP-3828</a></td><td>mapred</td><td>Skipped
records can optionally be written to the HDFS. Refer
org.apache.hadoop.mapred.SkipBadRecords.setSkipOutputPath for setting the
output path.</td>
+</tr>
+<tr>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-4293">HADOOP-4293</a></td><td>mapred</td><td>Made
Configuration Writable and rename the old write method to writeXml.</td>
+</tr>
+<tr>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3412">HADOOP-3412</a></td><td>mapred</td><td>Added
the ability to chose between many schedulers, and to limit the number of
running tasks per job.</td>
</tr>
<tr>
<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3667">HADOOP-3667</a></td><td>mapred</td><td>Removed
the following deprecated methods from JobConf:
@@ -280,31 +238,22 @@
setOutputPath(Path)</td>
</tr>
<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3828">HADOOP-3828</a></td><td>mapred</td><td>Skipped
records can optionally be written to the HDFS. Refer
org.apache.hadoop.mapred.SkipBadRecords.setSkipOutputPath for setting the
output path.</td>
-</tr>
-<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3961">HADOOP-3961</a></td><td>mapred</td><td>resource
estimation now only kicks in after 10% of maps have completed.</td>
-</tr>
-<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3730">HADOOP-3730</a></td><td>mapred</td><td> Added
a JobConf constructor that disables loading default configurations so as to
take all default values from the JobTracker's configuration.</td>
-</tr>
-<tr>
<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3585">HADOOP-3585</a></td><td>metrics</td><td>Added
FailMon as a contrib project for hardware failure monitoring and analysis,
under /src/contrib/failmon. Created User Manual and Quick Start Guide.</td>
</tr>
<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-4228">HADOOP-4228</a></td><td>metrics</td><td>Changed
bytes_read, bytes_written to type long to prevent metrics overflow.</td>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3062">HADOOP-3062</a></td><td>metrics</td><td>Introduced
additional log records for data transfers.</td>
</tr>
<tr>
<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3714">HADOOP-3714</a></td><td>scripts</td><td>Adds
a new contrib, bash-tab-completion, which enables bash tab completion for the
bin/hadoop script. See the README file in the contrib directory for the
installation.</td>
</tr>
<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3939">HADOOP-3939</a></td><td>tools/distcp</td><td>Added
a new option -delete to DistCp so that if the files/directories exist in dst
but not in src will be deleted. It uses FsShell to do delete, so that it will
use trash if the trash is enable.</td>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3889">HADOOP-3889</a></td><td>tools/distcp</td><td>Changed
DistCp error messages when there is a RemoteException. Changed the
corresponding return value from -999 to -3.</td>
</tr>
<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3873">HADOOP-3873</a></td><td>tools/distcp</td><td>Added
two new options -filelimit <n> and -sizelimit <n> to DistCp for
limiting the total number of files and the total size in bytes,
respectively.</td>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3939">HADOOP-3939</a></td><td>tools/distcp</td><td>Added
a new option -delete to DistCp so that if the files/directories exist in dst
but not in src will be deleted. It uses FsShell to do delete, so that it will
use trash if the trash is enable.</td>
</tr>
<tr>
-<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3889">HADOOP-3889</a></td><td>tools/distcp</td><td>Changed
DistCp error messages when there is a RemoteException. Changed the
corresponding return value from -999 to -3.</td>
+<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3873">HADOOP-3873</a></td><td>tools/distcp</td><td>Added
two new options -filelimit <n> and -sizelimit <n> to DistCp for
limiting the total number of files and the total size in bytes,
respectively.</td>
</tr>
<tr>
<td><a
href="https://issues.apache.org:443/jira/browse/HADOOP-3854">HADOOP-3854</a></td><td>util</td><td>Added
a configuration property hadoop.http.filter.initializers and a class
org.apache.hadoop.http.FilterInitializer for supporting servlet filter.
Cluster administrator could possibly configure customized filters for their web
site.</td>