Author: nigel
Date: Thu Nov 13 17:58:43 2008
New Revision: 713903

URL: http://svn.apache.org/viewvc?rev=713903&view=rev
Log:
Preparing for release 0.19.0 RC1

Modified:
    hadoop/core/branches/branch-0.19/CHANGES.txt
    hadoop/core/branches/branch-0.19/build.xml
    hadoop/core/branches/branch-0.19/docs/changes.html
    hadoop/core/branches/branch-0.19/docs/releasenotes.html

Modified: hadoop/core/branches/branch-0.19/CHANGES.txt
URL: 
http://svn.apache.org/viewvc/hadoop/core/branches/branch-0.19/CHANGES.txt?rev=713903&r1=713902&r2=713903&view=diff
==============================================================================
--- hadoop/core/branches/branch-0.19/CHANGES.txt (original)
+++ hadoop/core/branches/branch-0.19/CHANGES.txt Thu Nov 13 17:58:43 2008
@@ -1,6 +1,6 @@
 Hadoop Change Log
 
-Release 0.19.0 - Unreleased
+Release 0.19.0 - 2008-11-18
 
   INCOMPATIBLE CHANGES
 

Modified: hadoop/core/branches/branch-0.19/build.xml
URL: 
http://svn.apache.org/viewvc/hadoop/core/branches/branch-0.19/build.xml?rev=713903&r1=713902&r2=713903&view=diff
==============================================================================
--- hadoop/core/branches/branch-0.19/build.xml (original)
+++ hadoop/core/branches/branch-0.19/build.xml Thu Nov 13 17:58:43 2008
@@ -26,7 +26,7 @@
  
   <property name="Name" value="Hadoop"/>
   <property name="name" value="hadoop"/>
-  <property name="version" value="0.19.0-dev"/>
+  <property name="version" value="0.19.1-dev"/>
   <property name="final.name" value="${name}-${version}"/>
   <property name="year" value="2008"/>
   <property name="libhdfs.version" value="1"/>

Modified: hadoop/core/branches/branch-0.19/docs/changes.html
URL: 
http://svn.apache.org/viewvc/hadoop/core/branches/branch-0.19/docs/changes.html?rev=713903&r1=713902&r2=713903&view=diff
==============================================================================
--- hadoop/core/branches/branch-0.19/docs/changes.html (original)
+++ hadoop/core/branches/branch-0.19/docs/changes.html Thu Nov 13 17:58:43 2008
@@ -36,7 +36,7 @@
     function collapse() {
       for (var i = 0; i < document.getElementsByTagName("ul").length; i++) {
         var list = document.getElementsByTagName("ul")[i];
-        if (list.id != 'release_0.19.0_-_unreleased_' && list.id != 
'release_0.18.2_-_unreleased_') {
+        if (list.id != 'release_0.19.0_-_2008-11-18_' && list.id != 
'release_0.18.3_-_unreleased_') {
           list.style.display = "none";
         }
       }
@@ -52,12 +52,12 @@
 <a href="http://hadoop.apache.org/core/";><img class="logoImage" alt="Hadoop" 
src="images/hadoop-logo.jpg" title="Scalable Computing Platform"></a>
 <h1>Hadoop Change Log</h1>
 
-<h2><a href="javascript:toggleList('release_0.19.0_-_unreleased_')">Release 
0.19.0 - Unreleased
+<h2><a href="javascript:toggleList('release_0.19.0_-_2008-11-18_')">Release 
0.19.0 - 2008-11-18
 </a></h2>
-<ul id="release_0.19.0_-_unreleased_">
-  <li><a 
href="javascript:toggleList('release_0.19.0_-_unreleased_._incompatible_changes_')">
  INCOMPATIBLE CHANGES
-</a>&nbsp;&nbsp;&nbsp;(21)
-    <ol id="release_0.19.0_-_unreleased_._incompatible_changes_">
+<ul id="release_0.19.0_-_2008-11-18_">
+  <li><a 
href="javascript:toggleList('release_0.19.0_-_2008-11-18_._incompatible_changes_')">
  INCOMPATIBLE CHANGES
+</a>&nbsp;&nbsp;&nbsp;(22)
+    <ol id="release_0.19.0_-_2008-11-18_._incompatible_changes_">
       <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-3595";>HADOOP-3595</a>. Remove 
deprecated methods for mapred.combine.once
 functionality, which was necessary to providing backwards
 compatible combiner semantics for 0.18.<br />(cdouglas via omalley)</li>
@@ -124,11 +124,12 @@
 DFS command line report reflects the same change. Config parameter
 dfs.datanode.du.pct is no longer used and is removed from the
 hadoop-default.xml.<br />(Suresh Srinivas via hairong)</li>
+      <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4116";>HADOOP-4116</a>. 
Balancer should provide better resource management.<br />(hairong)</li>
     </ol>
   </li>
-  <li><a 
href="javascript:toggleList('release_0.19.0_-_unreleased_._new_features_')">  
NEW FEATURES
-</a>&nbsp;&nbsp;&nbsp;(40)
-    <ol id="release_0.19.0_-_unreleased_._new_features_">
+  <li><a 
href="javascript:toggleList('release_0.19.0_-_2008-11-18_._new_features_')">  
NEW FEATURES
+</a>&nbsp;&nbsp;&nbsp;(39)
+    <ol id="release_0.19.0_-_2008-11-18_._new_features_">
       <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-3341";>HADOOP-3341</a>. Allow 
streaming jobs to specify the field separator for map
 and reduce input and output. The new configuration values are:
   stream.map.input.field.separator
@@ -209,13 +210,11 @@
       <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-3019";>HADOOP-3019</a>. A new 
library to support total order partitions.<br />(cdouglas via omalley)</li>
       <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-3924";>HADOOP-3924</a>. Added 
a 'KILLED' job status.<br />(Subramaniam Krishnan via
 acmurthy)</li>
-      <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-2421";>HADOOP-2421</a>.  Add 
jdiff output to documentation, listing all API
-changes from the prior release.<br />(cutting)</li>
     </ol>
   </li>
-  <li><a 
href="javascript:toggleList('release_0.19.0_-_unreleased_._improvements_')">  
IMPROVEMENTS
+  <li><a 
href="javascript:toggleList('release_0.19.0_-_2008-11-18_._improvements_')">  
IMPROVEMENTS
 </a>&nbsp;&nbsp;&nbsp;(78)
-    <ol id="release_0.19.0_-_unreleased_._improvements_">
+    <ol id="release_0.19.0_-_2008-11-18_._improvements_">
       <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4205";>HADOOP-4205</a>. hive: 
metastore and ql to use the refactored SerDe library.<br />(zshao)</li>
       <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4106";>HADOOP-4106</a>. 
libhdfs: add time, permission and user attribute support
 (part 2).<br />(Pete Wyckoff through zshao)</li>
@@ -345,8 +344,6 @@
       <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4301";>HADOOP-4301</a>. Adds 
forrest doc for the skip bad records feature.<br />(Sharad Agarwal via 
ddas)</li>
       <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4354";>HADOOP-4354</a>. 
Separate TestDatanodeDeath.testDatanodeDeath() into 4 tests.<br 
/>(szetszwo)</li>
       <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-3790";>HADOOP-3790</a>. Add 
more unit tests for testing HDFS file append.<br />(szetszwo)</li>
-      <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4150";>HADOOP-4150</a>. 
Include librecordio in hadoop releases.<br />(Giridharan Kesavan
-via acmurthy)</li>
       <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4321";>HADOOP-4321</a>. 
Include documentation for the capacity scheduler.<br />(Hemanth
 Yamijala via omalley)</li>
       <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4424";>HADOOP-4424</a>. Change 
menu layout for Hadoop documentation (Boris Shkolnik
@@ -355,11 +352,12 @@
       <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4438";>HADOOP-4438</a>. Update 
forrest documentation to include missing FsShell
 commands.<br />(Suresh Srinivas via cdouglas)</li>
       <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4105";>HADOOP-4105</a>.  Add 
forrest documentation for libhdfs.<br />(Pete Wyckoff via cutting)</li>
+      <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4510";>HADOOP-4510</a>. Make 
getTaskOutputPath public.<br />(Chris Wensel via omalley)</li>
     </ol>
   </li>
-  <li><a 
href="javascript:toggleList('release_0.19.0_-_unreleased_._optimizations_')">  
OPTIMIZATIONS
+  <li><a 
href="javascript:toggleList('release_0.19.0_-_2008-11-18_._optimizations_')">  
OPTIMIZATIONS
 </a>&nbsp;&nbsp;&nbsp;(11)
-    <ol id="release_0.19.0_-_unreleased_._optimizations_">
+    <ol id="release_0.19.0_-_2008-11-18_._optimizations_">
       <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-3556";>HADOOP-3556</a>. 
Removed lock contention in MD5Hash by changing the
 singleton MessageDigester by an instance per Thread using
 ThreadLocal.<br />(Iv?n de Prado via omalley)</li>
@@ -386,9 +384,9 @@
 org.apache.hadoop.mapred  package private instead of public.<br 
/>(omalley)</li>
     </ol>
   </li>
-  <li><a 
href="javascript:toggleList('release_0.19.0_-_unreleased_._bug_fixes_')">  BUG 
FIXES
-</a>&nbsp;&nbsp;&nbsp;(146)
-    <ol id="release_0.19.0_-_unreleased_._bug_fixes_">
+  <li><a 
href="javascript:toggleList('release_0.19.0_-_2008-11-18_._bug_fixes_')">  BUG 
FIXES
+</a>&nbsp;&nbsp;&nbsp;(153)
+    <ol id="release_0.19.0_-_2008-11-18_._bug_fixes_">
       <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-3563";>HADOOP-3563</a>.  
Refactor the distributed upgrade code so that it is
 easier to identify datanode and namenode related code.<br />(dhruba)</li>
       <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-3640";>HADOOP-3640</a>. Fix 
the read method in the NativeS3InputStream.<br />(tomwhite via
@@ -608,8 +606,6 @@
       <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4018";>HADOOP-4018</a>. The 
number of tasks for a single job cannot exceed a
 pre-configured maximum value.<br />(dhruba)</li>
       <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4288";>HADOOP-4288</a>. Fixes 
a NPE problem in CapacityScheduler.<br />(Amar Kamat via ddas)</li>
-      <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-3883";>HADOOP-3883</a>. Limit 
namenode to assign at most one generation stamp for
-a particular block within a short period.<br />(szetszwo)</li>
       <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4014";>HADOOP-4014</a>. Create 
hard links with 'fsutil hardlink' on Windows.<br />(shv)</li>
       <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4393";>HADOOP-4393</a>. Merged 
org.apache.hadoop.fs.permission.AccessControlException
 and org.apache.hadoop.security.AccessControlIOException into a single
@@ -619,7 +615,6 @@
       <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4361";>HADOOP-4361</a>. Makes 
sure that jobs killed from command line are killed
 fast (i.e., there is a slot to run the cleanup task soon).<br />(Amareshwari 
Sriramadasu via ddas)</li>
       <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4400";>HADOOP-4400</a>. Add 
"hdfs://" to fs.default.name on quickstart.html.<br />(Jeff Hammerbacher via 
omalley)</li>
-      <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4403";>HADOOP-4403</a>. Make 
TestLeaseRecovery and TestFileCreation more robust.<br />(szetszwo)</li>
       <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4378";>HADOOP-4378</a>. Fix 
TestJobQueueInformation to use SleepJob rather than
 WordCount via TestMiniMRWithDFS.<br />(Sreekanth Ramakrishnan via 
acmurthy)</li>
       <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4376";>HADOOP-4376</a>. Fix 
formatting in hadoop-default.xml for
@@ -665,36 +660,95 @@
       <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4358";>HADOOP-4358</a>. No 
need to truncate access time in INode. Also fixes NPE
 in CreateEditsLog.<br />(Raghu Angadi)</li>
       <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4387";>HADOOP-4387</a>. 
TestHDFSFileSystemContract fails on windows nightly builds.<br />(Raghu 
Angadi)</li>
+      <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4466";>HADOOP-4466</a>. Ensure 
that SequenceFileOutputFormat isn't tied to Writables
+and can be used with other Serialization frameworks.<br />(Chris Wensel via
+acmurthy)</li>
+      <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4525";>HADOOP-4525</a>. Fix 
ipc.server.ipcnodelay originally missed in in <a 
href="http://issues.apache.org/jira/browse/HADOOP-2232";>HADOOP-2232</a>.<br 
/>(cdouglas via Clint Morgan)</li>
+      <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4498";>HADOOP-4498</a>. Ensure 
that JobHistory correctly escapes the job name so that
+regex patterns work.<br />(Chris Wensel via acmurthy)</li>
+      <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4446";>HADOOP-4446</a>. Modify 
guaranteed capacity labels in capacity scheduler's UI
+to reflect the information being displayed.<br />(Sreekanth Ramakrishnan via
+yhemanth)</li>
+      <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4282";>HADOOP-4282</a>. Some 
user facing URLs are not filtered by user filters.<br />(szetszwo)</li>
+      <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4595";>HADOOP-4595</a>. Fixes 
two race conditions - one to do with updating free slot count,
+and another to do with starting the MapEventsFetcher thread.<br />(ddas)</li>
+      <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4552";>HADOOP-4552</a>. Fix a 
deadlock in RPC server.<br />(Raghu Angadi)</li>
+      <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4471";>HADOOP-4471</a>. Sort 
running jobs by priority in the capacity scheduler.<br />(Amar Kamat via 
yhemanth)</li>
+      <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4500";>HADOOP-4500</a>. Fix 
MultiFileSplit to get the FileSystem from the relevant
+path rather than the JobClient.<br />(Joydeep Sen Sarma via dhruba)</li>
     </ol>
   </li>
 </ul>
-<h2><a href="javascript:toggleList('release_0.18.2_-_unreleased_')">Release 
0.18.2 - Unreleased
+<h2><a href="javascript:toggleList('release_0.18.3_-_unreleased_')">Release 
0.18.3 - Unreleased
 </a></h2>
-<ul id="release_0.18.2_-_unreleased_">
-  <li><a 
href="javascript:toggleList('release_0.18.2_-_unreleased_._bug_fixes_')">  BUG 
FIXES
-</a>&nbsp;&nbsp;&nbsp;(11)
-    <ol id="release_0.18.2_-_unreleased_._bug_fixes_">
-      <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4116";>HADOOP-4116</a>. 
Balancer should provide better resource management.<br />(hairong)</li>
+<ul id="release_0.18.3_-_unreleased_">
+  <li><a 
href="javascript:toggleList('release_0.18.3_-_unreleased_._improvements_')">  
IMPROVEMENTS
+</a>&nbsp;&nbsp;&nbsp;(1)
+    <ol id="release_0.18.3_-_unreleased_._improvements_">
+      <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4150";>HADOOP-4150</a>. 
Include librecordio in hadoop releases.<br />(Giridharan Kesavan
+via acmurthy)</li>
+    </ol>
+  </li>
+  <li><a 
href="javascript:toggleList('release_0.18.3_-_unreleased_._bug_fixes_')">  BUG 
FIXES
+</a>&nbsp;&nbsp;&nbsp;(7)
+    <ol id="release_0.18.3_-_unreleased_._bug_fixes_">
+      <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4499";>HADOOP-4499</a>. 
DFSClient should invoke checksumOk only once.<br />(Raghu Angadi)</li>
+      <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4597";>HADOOP-4597</a>. 
Calculate mis-replicated blocks when safe-mode is turned
+off manually.<br />(shv)</li>
+      <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-3121";>HADOOP-3121</a>. lsr 
should keep listing the remaining items but not
+terminate if there is any IOException.<br />(szetszwo)</li>
+      <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4610";>HADOOP-4610</a>. Always 
calculate mis-replicated blocks when safe-mode is
+turned off.<br />(shv)</li>
+      <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-3883";>HADOOP-3883</a>. Limit 
namenode to assign at most one generation stamp for
+a particular block within a short period.<br />(szetszwo)</li>
+      <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4556";>HADOOP-4556</a>. Block 
went missing.<br />(hairong)</li>
+      <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4643";>HADOOP-4643</a>. 
NameNode should exclude excessive replicas when counting
+live replicas for a block.<br />(hairong)</li>
+    </ol>
+  </li>
+</ul>
+<h2><a href="javascript:toggleList('older')">Older Releases</a></h2>
+<ul id="older">
+<h3><a href="javascript:toggleList('release_0.18.2_-_2008-11-03_')">Release 
0.18.2 - 2008-11-03
+</a></h3>
+<ul id="release_0.18.2_-_2008-11-03_">
+  <li><a 
href="javascript:toggleList('release_0.18.2_-_2008-11-03_._bug_fixes_')">  BUG 
FIXES
+</a>&nbsp;&nbsp;&nbsp;(16)
+    <ol id="release_0.18.2_-_2008-11-03_._bug_fixes_">
       <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-3614";>HADOOP-3614</a>. Fix a 
bug that Datanode may use an old GenerationStamp to get
 meta file.<br />(szetszwo)</li>
       <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4314";>HADOOP-4314</a>. 
Simulated datanodes should not include blocks that are still
 being written in their block report.<br />(Raghu Angadi)</li>
-      <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4228";>HADOOP-4228</a>. dfs 
datanoe metrics, bytes_read and bytes_written, overflow
+      <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4228";>HADOOP-4228</a>. dfs 
datanode metrics, bytes_read and bytes_written, overflow
 due to incorrect type used.<br />(hairong)</li>
       <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4395";>HADOOP-4395</a>. The 
FSEditLog loading is incorrect for the case OP_SET_OWNER.<br />(szetszwo)</li>
       <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4351";>HADOOP-4351</a>. 
FSNamesystem.getBlockLocationsInternal throws
 ArrayIndexOutOfBoundsException.<br />(hairong)</li>
+      <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4403";>HADOOP-4403</a>. Make 
TestLeaseRecovery and TestFileCreation more robust.<br />(szetszwo)</li>
       <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4292";>HADOOP-4292</a>. Do not 
support append() for LocalFileSystem.<br />(hairong)</li>
       <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4399";>HADOOP-4399</a>. Make 
fuse-dfs multi-thread access safe.<br />(Pete Wyckoff via dhruba)</li>
       <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4369";>HADOOP-4369</a>. Use 
setMetric(...) instead of incrMetric(...) for metrics
 averages.<br />(Brian Bockelman via szetszwo)</li>
       <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4469";>HADOOP-4469</a>. Rename 
and add the ant task jar file to the tar file.<br />(nigel)</li>
-      <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-3914";>HADOOP-3914</a>. 
DFSClient sends Checksum Ok only once for a block.<br />(Christain Kunz via 
hairong)</li>
+      <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-3914";>HADOOP-3914</a>. 
DFSClient sends Checksum Ok only once for a block.<br />(Christian Kunz via 
hairong)</li>
+      <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4467";>HADOOP-4467</a>. 
SerializationFactory now uses the current context ClassLoader
+allowing for user supplied Serialization instances.<br />(Chris Wensel via
+acmurthy)</li>
+      <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4517";>HADOOP-4517</a>. 
Release FSDataset lock before joining ongoing create threads.<br 
/>(szetszwo)</li>
+      <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4526";>HADOOP-4526</a>. fsck 
failing with NullPointerException.<br />(hairong)</li>
+      <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4483";>HADOOP-4483</a> Honor 
the max parameter in DatanodeDescriptor.getBlockArray(..)<br />(Ahad Rana and 
Hairong Kuang via szetszwo)</li>
+      <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-4340";>HADOOP-4340</a>. 
Correctly set the exit code from JobShell.main so that the
+'hadoop jar' command returns the right code to the user.<br />(acmurthy)</li>
+    </ol>
+  </li>
+  <li><a 
href="javascript:toggleList('release_0.18.2_-_2008-11-03_._new_features_')">  
NEW FEATURES
+</a>&nbsp;&nbsp;&nbsp;(1)
+    <ol id="release_0.18.2_-_2008-11-03_._new_features_">
+      <li><a 
href="http://issues.apache.org/jira/browse/HADOOP-2421";>HADOOP-2421</a>.  Add 
jdiff output to documentation, listing all API
+changes from the prior release.<br />(cutting)</li>
     </ol>
   </li>
 </ul>
-<h2><a href="javascript:toggleList('older')">Older Releases</a></h2>
-<ul id="older">
 <h3><a href="javascript:toggleList('release_0.18.1_-_2008-09-17_')">Release 
0.18.1 - 2008-09-17
 </a></h3>
 <ul id="release_0.18.1_-_2008-09-17_">

Modified: hadoop/core/branches/branch-0.19/docs/releasenotes.html
URL: 
http://svn.apache.org/viewvc/hadoop/core/branches/branch-0.19/docs/releasenotes.html?rev=713903&r1=713902&r2=713903&view=diff
==============================================================================
--- hadoop/core/branches/branch-0.19/docs/releasenotes.html (original)
+++ hadoop/core/branches/branch-0.19/docs/releasenotes.html Thu Nov 13 17:58:43 
2008
@@ -9,7 +9,7 @@
 The table below is sorted by Component.
 <ul><a name="changes">
 <h2>Changes Since Hadoop 0.18.2</h2>
-  <table border="1""100%" cellpadding="4">
+<table border="1">
 <tr bgcolor="#DDDDDD">
 <th align="left">Issue</th><th align="left">Component</th><th 
align="left">Notes</th>
 </tr>
@@ -17,17 +17,7 @@
 <td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-2325";>HADOOP-2325</a></td><td>build</td><td>Hadoop
 now requires Java 6.</td>
 </tr>
 <tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-4150";>HADOOP-4150</a></td><td>build</td><td>librecordio
 has a build dependency with xerces-c parser.  so to get this build; need to 
set the path of xerces-c binary to xercescroot and librecordio to true 
-<br/>
-exact command would be like this :
-<br/>
-
-<br/>
-ant  -Dlibrecordio=true -Dxercescroot=&lt;path to the xerces-c root&gt; 
-<br/>
-
-<br/>
-This patch includes targets for compiling, packaging and testing the 
librecordio library.</td>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3730";>HADOOP-3730</a></td><td>conf</td><td>&nbsp;Added
 a JobConf constructor that disables loading  default configurations so as to 
take all default values from the JobTracker's configuration.</td>
 </tr>
 <tr>
 <td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3722";>HADOOP-3722</a></td><td>conf</td><td>Changed
 streaming StreamJob and Submitter to implement Tool and Configurable, and to 
use GenericOptionsParser arguments  -fs, -jt, -conf, -D, -libjars, -files, and 
-archives. Deprecated -jobconf, -cacheArchive, -dfs, -cacheArchive, 
-additionalconfspec,  from streaming and pipes in favor of the generic options. 
Removed from streaming  -config, -mapred.job.tracker, and -cluster.</td>
@@ -36,6 +26,11 @@
 <td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3646";>HADOOP-3646</a></td><td>conf</td><td>Introduced
 support for bzip2 compressed files.</td>
 </tr>
 <tr>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3445";>HADOOP-3445</a></td><td>contrib/capacity-sched</td><td>Introduced
 Capacity Task Scheduler.
+<br/>
+</td>
+</tr>
+<tr>
 <td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3719";>HADOOP-3719</a></td><td>contrib/chukwa</td><td>Introduced
 Chukwa data collection and analysis framework.</td>
 </tr>
 <tr>
@@ -45,25 +40,22 @@
 <td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-4117";>HADOOP-4117</a></td><td>contrib/ec2</td><td>Changed
 scripts to pass initialization script for EC2 instances at boot time (as EC2 
user data) rather than embedding initialization information in the EC2 image. 
This change makes it easy to customize the hadoop-site.xml file for your 
cluster before launch, by editing the hadoop-ec2-init-remote.sh script, or by 
setting the environment variable USER_DATA_FILE in hadoop-ec2-env.sh to run a 
script of your choice.</td>
 </tr>
 <tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-4106";>HADOOP-4106</a></td><td>contrib/fuse-dfs</td><td>Added
 time, permission and user attribute support to libhdfs.</td>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3485";>HADOOP-3485</a></td><td>contrib/fuse-dfs</td><td>Introduce
 write support for Fuse; requires Linux kernel 2.6.15 or better.</td>
 </tr>
 <tr>
 <td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3796";>HADOOP-3796</a></td><td>contrib/fuse-dfs</td><td>Changed
 Fuse configuration to use mount options.</td>
 </tr>
 <tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3908";>HADOOP-3908</a></td><td>contrib/fuse-dfs</td><td>Improved
 Fuse-dfs better error message if llibhdfs.so doesn't exist.</td>
-</tr>
-<tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3485";>HADOOP-3485</a></td><td>contrib/fuse-dfs</td><td>Introduce
 write support for Fuse; requires Linux kernel 2.6.15 or better.</td>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-4106";>HADOOP-4106</a></td><td>contrib/fuse-dfs</td><td>Added
 time, permission and user attribute support to libhdfs.</td>
 </tr>
 <tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-4086";>HADOOP-4086</a></td><td>contrib/hive</td><td>Add
 LIMIT to Hive QL.</td>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3908";>HADOOP-3908</a></td><td>contrib/fuse-dfs</td><td>Improved
 Fuse-dfs better error message if llibhdfs.so doesn't exist.</td>
 </tr>
 <tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-4205";>HADOOP-4205</a></td><td>contrib/hive</td><td>Improved
 Hive metastore and ql to use the refactored SerDe library.</td>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-4086";>HADOOP-4086</a></td><td>contrib/hive</td><td>Added
 LIMIT to Hive query language.</td>
 </tr>
 <tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-4084";>HADOOP-4084</a></td><td>contrib/hive</td><td>Introduced
 &quot;EXPLAIN&quot; plan for Hive.</td>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-4138";>HADOOP-4138</a></td><td>contrib/hive</td><td>Introduced
 new SerDe library for src/contrib/hive.</td>
 </tr>
 <tr>
 <td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3601";>HADOOP-3601</a></td><td>contrib/hive</td><td>Introduced
 Hive Data Warehouse built on top of Hadoop that enables structuring Hadoop 
files as tables and partitions and allows users to query this data through a 
SQL like language using a command line interface.
@@ -71,7 +63,10 @@
 </td>
 </tr>
 <tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-4138";>HADOOP-4138</a></td><td>contrib/hive</td><td>Introduced
 new SerDe library for src/contrib/hive.</td>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-4205";>HADOOP-4205</a></td><td>contrib/hive</td><td>Improved
 Hive metastore and ql to use the refactored SerDe library.</td>
+</tr>
+<tr>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-4084";>HADOOP-4084</a></td><td>contrib/hive</td><td>Introduced
 &quot;EXPLAIN&quot; plan for Hive.</td>
 </tr>
 <tr>
 <td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3695";>HADOOP-3695</a></td><td>contrib/hod</td><td>Added
 an ability in HOD to start multiple workers (TaskTrackers and/or DataNodes) 
per node to assist testing and simulation of scale. A configuration variable 
ringmaster.workers_per_ring was added to specify the number of workers to 
start.</td>
@@ -83,82 +78,54 @@
 <td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-2302";>HADOOP-2302</a></td><td>contrib/streaming</td><td>Introduced
 numerical key comparison for streaming.</td>
 </tr>
 <tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-4281";>HADOOP-4281</a></td><td>dfs</td><td>Changed
 command &quot;hadoop dfsadmin -report&quot; to be consistent with Web UI for 
both Namenode and Datanode reports. &quot;Total raw bytes&quot; is changed to 
&quot;Configured Capacity&quot;. &quot;Present Capacity&quot; is newly added to 
indicate the present capacity of the DFS. &quot;Remaining raw bytes&quot; is 
changed to &quot;DFS Remaining&quot;. &quot;Used raw bytes&quot; is changed to 
&quot;DFS Used&quot;. &quot;% used&quot; is changed to &quot;DFS Used%&quot;. 
Applications that parse command output should be reviewed.</td>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-4430";>HADOOP-4430</a></td><td>dfs</td><td>Changed
 reporting in the NameNode Web UI to more closely reflect the behavior of the 
re-balancer. Removed no longer used config parameter dfs.datanode.du.pct from 
hadoop-default.xml.
+<br/>
+</td>
 </tr>
 <tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-1869";>HADOOP-1869</a></td><td>dfs</td><td>Added
 HDFS file access times. By default, access times will be precise to the most 
recent hour boundary. A configuration parameter dfs.access.time.precision 
(milliseconds) is used to control this precision. Setting a value of 0 will 
disable persisting access times for HDFS files.</td>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-4281";>HADOOP-4281</a></td><td>dfs</td><td>Changed
 command &quot;hadoop dfsadmin -report&quot; to be consistent with Web UI for 
both Namenode and Datanode reports. &quot;Total raw bytes&quot; is changed to 
&quot;Configured Capacity&quot;. &quot;Present Capacity&quot; is newly added to 
indicate the present capacity of the DFS. &quot;Remaining raw bytes&quot; is 
changed to &quot;DFS Remaining&quot;. &quot;Used raw bytes&quot; is changed to 
&quot;DFS Used&quot;. &quot;% used&quot; is changed to &quot;DFS Used%&quot;. 
Applications that parse command output should be reviewed.</td>
 </tr>
 <tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-4423";>HADOOP-4423</a></td><td>dfs</td><td>Added
 a new parameter keepLength to ClientDatanodeProtocol.recoverBlock(...)</td>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-2885";>HADOOP-2885</a></td><td>dfs</td><td>Restructured
 the package hadoop.dfs.</td>
 </tr>
 <tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3992";>HADOOP-3992</a></td><td>dfs</td><td>Added
 a synthetic load generation facility to the test directory.</td>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-4007";>HADOOP-4007</a></td><td>dfs</td><td>Changed
 ClientProtocol getStatus and getListing to use the type FileStatus. Removed 
type DFSFileInfo.</td>
 </tr>
 <tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-4007";>HADOOP-4007</a></td><td>dfs</td><td>Changed
 ClientProtocol getStatus and getListing to use the type FileStatus. Removed 
type DFSFileInfo.</td>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-2816";>HADOOP-2816</a></td><td>dfs</td><td>Improved
 space reporting for NameNode Web UI. Applications that parse the Web UI output 
should be reviewed.</td>
 </tr>
 <tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3732";>HADOOP-3732</a></td><td>dfs</td><td>Delay
 intialization of datanode block verification till the verification thread is 
started.
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-1700";>HADOOP-1700</a></td><td>dfs</td><td>Introduced
 append operation for HDFS files.
 <br/>
 </td>
 </tr>
 <tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3981";>HADOOP-3981</a></td><td>dfs</td><td>Implemented
 MD5-of-xxxMD5-of-yyyCRC32 which is a distributed file checksum algorithm for 
HDFS, where xxx is the number of CRCs per block and yyy is the number of bytes 
per CRC.
-<br/>
-
-<br/>
-Changed DistCp to use file checksum for comparing files if both source and 
destination FileSystem(s) support getFileChecksum(...).</td>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3938";>HADOOP-3938</a></td><td>dfs</td><td>Introducted
 byte space quotas for directories. The count shell command modified to report 
both name and byte quotas.</td>
 </tr>
 <tr>
 <td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-4176";>HADOOP-4176</a></td><td>dfs</td><td>Implemented
 getFileChecksum(Path) in HftpFileSystemfor distcp support.</td>
 </tr>
 <tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-2816";>HADOOP-2816</a></td><td>dfs</td><td>Improved
 space reporting for NameNode Web UI. Applications that parse the Web UI output 
should be reviewed.</td>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-1869";>HADOOP-1869</a></td><td>dfs</td><td>Added
 HDFS file access times. By default, access times will be precise to the most 
recent hour boundary. A configuration parameter dfs.access.time.precision 
(milliseconds) is used to control this precision. Setting a value of 0 will 
disable persisting access times for HDFS files.</td>
 </tr>
 <tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-4430";>HADOOP-4430</a></td><td>dfs</td><td>Incompatible
 changes:
-<br/>
-This change modifies/retains the changes made in 2816 as follows:
-<br/>
-1) Present Capacity added in 2816 is removed from the Web UI
-<br/>
-2) Change of Total Capacity to Configured Capacity and its definition from 
2816 is retained in the Web UI
-<br/>
-3) Data node protocol change to report Configured Capacity instead of Total 
Capacity is retained.
-<br/>
-4) DFS Used% was calculated as a percentage of Present Capacity. It is changed 
to percentage of Configured Capacity.
-<br/>
-
-<br/>
-Other incompatible changes:
-<br/>
-1) Config parameter dfs.datanode.du.pct is no longer used and is removed from 
the hadoop-default.xml.
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3981";>HADOOP-3981</a></td><td>dfs</td><td>Implemented
 MD5-of-xxxMD5-of-yyyCRC32 which is a distributed file checksum algorithm for 
HDFS, where xxx is the number of CRCs per block and yyy is the number of bytes 
per CRC.
 <br/>
 
 <br/>
-2) Namenode Web UI has the following addional changes:
-<br/>
-&nbsp;&nbsp;&nbsp;The following parameters are added to both Cluster Summary 
and Datanode information:
-<br/>
-&nbsp;&nbsp;&nbsp;* Non DFS Used - This indicates the disk space taken by non 
DFS file
-<br/>
-&nbsp;&nbsp;&nbsp;* DFS remaining % - This is remaining % of Configured 
Capacity available for DFS use
-<br/>
-</td>
+Changed DistCp to use file checksum for comparing files if both source and 
destination FileSystem(s) support getFileChecksum(...).</td>
 </tr>
 <tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-1700";>HADOOP-1700</a></td><td>dfs</td><td>Introduced
 append operation for HDFS files.
-<br/>
-</td>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3992";>HADOOP-3992</a></td><td>dfs</td><td>Added
 a synthetic load generation facility to the test directory.</td>
 </tr>
 <tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3938";>HADOOP-3938</a></td><td>dfs</td><td>Introducted
 byte space quotas for directories. The count shell command modified to report 
both name and byte quotas.</td>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3792";>HADOOP-3792</a></td><td>fs</td><td>Changed
 exit code from hadoop.fs.FsShell -test to match the usual Unix convention.</td>
 </tr>
 <tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-4358";>HADOOP-4358</a></td><td>dfs</td><td>No
 need to truncate access time in INode. Also fixes NPE in CreateEditsLog</td>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-4227";>HADOOP-4227</a></td><td>fs</td><td>Removed
 the deprecated class org.apache.hadoop.fs.ShellCommand.</td>
 </tr>
 <tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-2885";>HADOOP-2885</a></td><td>dfs</td><td>Restructured
 the package hadoop.dfs.</td>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3498";>HADOOP-3498</a></td><td>fs</td><td>Extended
 file globbing alternation to cross path components. For example, {/a/b,/c/d} 
expands to a path that matches the files /a/b and /c/d.</td>
 </tr>
 <tr>
 <td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3911";>HADOOP-3911</a></td><td>fs</td><td>Added
 a check to fsck options to make sure -files is not the first option so as to 
resolve conflicts with GenericOptionsParser.</td>
@@ -167,13 +134,10 @@
 <td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3941";>HADOOP-3941</a></td><td>fs</td><td>Added
 new FileSystem APIs: FileChecksum and FileSystem.getFileChecksum(Path).</td>
 </tr>
 <tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3792";>HADOOP-3792</a></td><td>fs</td><td>Changed
 exit code from hadoop.fs.FsShell -test to match the usual Unix convention.</td>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-4466";>HADOOP-4466</a></td><td>io</td><td>Ensure
 that SequenceFileOutputFormat isn't tied to Writables and can be used with 
other Serialization frameworks.</td>
 </tr>
 <tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3498";>HADOOP-3498</a></td><td>fs</td><td>Extended
 file globbing alternation to cross path components. For example, {/a/b,/c/d} 
expands to a path that matches the files /a/b and /c/d.</td>
-</tr>
-<tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-4227";>HADOOP-4227</a></td><td>fs</td><td>Removed
 the deprecated class org.apache.hadoop.fs.ShellCommand.</td>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-2664";>HADOOP-2664</a></td><td>io</td><td>Introduced
 LZOP codec.</td>
 </tr>
 <tr>
 <td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3549";>HADOOP-3549</a></td><td>libhdfs</td><td>Improved
 error reporting for libhdfs so permission problems now return EACCES.</td>
@@ -184,47 +148,35 @@
 </td>
 </tr>
 <tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3445";>HADOOP-3445</a></td><td>mapred</td><td>Introduced
 Capacity Task Scheduler.</td>
-</tr>
-<tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3595";>HADOOP-3595</a></td><td>mapred</td><td>&nbsp;Removed
 deprecated methods for mapred.combine.once functionality.</td>
-</tr>
-<tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-4018";>HADOOP-4018</a></td><td>mapred</td><td>A
 new configuration parameter named mapred.max.tasks.per.job that specifies the 
maximum number of tasks per job.
-<br/>
-
-<br/>
-</td>
-</tr>
-<tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-4261";>HADOOP-4261</a></td><td>mapred</td><td>A
 setup task (for doing OutputCommitter.setupJob()) will be launched, when the 
job is in PREP state and after initializing tasks. Once the setup task 
completes, the job will be moved to RUNNING state.</td>
-</tr>
-<tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3019";>HADOOP-3019</a></td><td>mapred</td><td>Added
 a partitioner that effects a total order of output data, and an input sampler 
for generating the partition keyset for TotalOrderPartitioner for when the 
map's input keytype and distribution approximates its output.</td>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-4190";>HADOOP-4190</a></td><td>mapred</td><td>Changed
 job history format to add a dot at end of each line.</td>
 </tr>
 <tr>
 <td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3970";>HADOOP-3970</a></td><td>mapred</td><td>Added
 getEscapedCompactString() and fromEscapedCompactString() to Counters.java to 
represent counters as Strings and to reconstruct the counters from the 
Strings.</td>
 </tr>
 <tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3412";>HADOOP-3412</a></td><td>mapred</td><td>Added
 the ability to chose between many schedulers, and to limit the number of 
running tasks per job.</td>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3245";>HADOOP-3245</a></td><td>mapred</td><td>Introduced
 recovery of jobs when JobTracker restarts. This facility is off by default. 
Introduced config parameters mapred.jobtracker.restart.recover, 
mapred.jobtracker.job.history.block.size, and 
mapred.jobtracker.job.history.buffer.size.</td>
 </tr>
 <tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3581";>HADOOP-3581</a></td><td>mapred</td><td>Added
 the ability to kill process trees transgressing memory limits. TaskTracker 
uses the configuration parameters introduced in <a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3759"; title="Provide 
ability to run memory intensive jobs without affecting other running tasks on 
the nodes"><strike>HADOOP-3759</strike></a>. In addition, 
mapred.tasktracker.taskmemorymanager.monitoring-interval specifies the interval 
for which TT waits between cycles of monitoring tasks' memory usage, and 
mapred.tasktracker.procfsbasedprocesstree.sleeptime-before-sigkill specifies 
the time TT waits for sending a SIGKILL to a process-tree that has overrun 
memory limits, after it has been sent a SIGTERM.</td>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3829";>HADOOP-3829</a></td><td>mapred</td><td>Introduced
 new config parameter 
org.apache.hadoop.mapred.SkipBadRecords.setMapperMaxSkipRecords to set range of 
records to be skipped in the neighborhood of a failed record.</td>
 </tr>
 <tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3684";>HADOOP-3684</a></td><td>mapred</td><td>Allowed
 user to overwrite clone function in a subclass of TaggedMapOutput class.</td>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-4018";>HADOOP-4018</a></td><td>mapred</td><td>Introduced
 new configuration parameter mapred.max.tasks.per.job to specifie the maximum 
number of tasks per job.
+<br/>
+
+<br/>
+</td>
 </tr>
 <tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3930";>HADOOP-3930</a></td><td>mapred</td><td>Changed
 TaskScheduler to expose API for Web UI and Command Line Tool.</td>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-249";>HADOOP-249</a></td><td>mapred</td><td>Enabled
 task JVMs to be reused via the job config mapred.job.reuse.jvm.num.tasks.</td>
 </tr>
 <tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-4190";>HADOOP-4190</a></td><td>mapred</td><td>Changed
 job history format to add a dot at end of each line.</td>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3746";>HADOOP-3746</a></td><td>mapred</td><td>Introduced
 Fair Scheduler.</td>
 </tr>
 <tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3478";>HADOOP-3478</a></td><td>mapred</td><td>Changed
 reducers to fetch maps in the same order for a given host to speed up 
identification of the faulty maps; reducers still randomize the host selection 
to distribute load.</td>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-153";>HADOOP-153</a></td><td>mapred</td><td>Introduced
 record skipping where tasks fail on certain records. 
(org.apache.hadoop.mapred.SkipBadRecords)</td>
 </tr>
 <tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3829";>HADOOP-3829</a></td><td>mapred</td><td>If
 the number of skipped records in the neighborhood of a bad record are not 
acceptable to the user, then narrow down the skipped range to the user 
acceptable value. Use 
org.apache.hadoop.mapred.SkipBadRecords.setMapperMaxSkipRecords for setting the 
acceptable value.</td>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3652";>HADOOP-3652</a></td><td>mapred</td><td>Removed
 deprecated org.apache.hadoop.mapred.OutputFormatBase.</td>
 </tr>
 <tr>
 <td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3702";>HADOOP-3702</a></td><td>mapred</td><td>Introduced
 ChainMapper and the ChainReducer classes to allow composing chains of Maps and 
Reduces in a single Map/Reduce job, something like MAP+ REDUCE MAP*.
@@ -232,22 +184,19 @@
 </td>
 </tr>
 <tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3746";>HADOOP-3746</a></td><td>mapred</td><td>Introduced
 Fair Scheduler.</td>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3595";>HADOOP-3595</a></td><td>mapred</td><td>&nbsp;Removed
 deprecated methods for mapred.combine.once functionality.</td>
 </tr>
 <tr>
 <td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3149";>HADOOP-3149</a></td><td>mapred</td><td>Introduced
 MultipleOutputs class so Map/Reduce jobs can write data to different output 
files. Each output can use a different OutputFormat. Outpufiles are created 
within the job output directory. FileOutputFormat.getPathForCustomFile() 
creates a filename under the outputdir that is named with the task ID and task 
type (i.e. myfile-r-00001).</td>
 </tr>
 <tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-153";>HADOOP-153</a></td><td>mapred</td><td>Introduced
 record skipping where tasks fail on certain records. 
(org.apache.hadoop.mapred.SkipBadRecords)</td>
-</tr>
-<tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3245";>HADOOP-3245</a></td><td>mapred</td><td>Introduced
 recovery of jobs when JobTracker restarts. This facility is off by default. 
Introduced config parameters mapred.jobtracker.restart.recover, 
mapred.jobtracker.job.history.block.size, and 
mapred.jobtracker.job.history.buffer.size.</td>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3684";>HADOOP-3684</a></td><td>mapred</td><td>Allowed
 user to overwrite clone function in a subclass of TaggedMapOutput class.</td>
 </tr>
 <tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-249";>HADOOP-249</a></td><td>mapred</td><td>Jobs
 can enable task JVMs to be reused via the job config 
mapred.job.reuse.jvm.num.tasks. If this is 1 (the default), then JVMs are not 
reused (1 task per JVM). If it is -1, there is no limit to the number of tasks 
a JVM can run (of the same job). One can also specify some value greater than 
1. Also a JobConf API has been added - setNumTasksToExecutePerJvm.</td>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3930";>HADOOP-3930</a></td><td>mapred</td><td>Changed
 TaskScheduler to expose API for Web UI and Command Line Tool.</td>
 </tr>
 <tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-4293";>HADOOP-4293</a></td><td>mapred</td><td>Made
 Configuration Writable and rename the old write method to writeXml.</td>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3478";>HADOOP-3478</a></td><td>mapred</td><td>Changed
 reducers to fetch maps in the same order for a given host to speed up 
identification of the faulty maps; reducers still randomize the host selection 
to distribute load.</td>
 </tr>
 <tr>
 <td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3150";>HADOOP-3150</a></td><td>mapred</td><td>Moved
 task file promotion to the Task. When the task has finished, it will do a 
commit and is declared SUCCEDED. Job cleanup is done by a separate task. Job is 
declared SUCCEDED/FAILED after the cleanup task has finished. Added public 
classes org.apache.hadoop.mapred.JobContext, TaskAttemptContext, 
OutputCommitter and FileOutputCommiitter. Added public APIs:   public 
OutputCommitter getOutputCommitter() and 
@@ -255,10 +204,19 @@
 public void setOutputCommitter(Class&lt;? extends OutputCommitter&gt; 
theClass) in org.apache.hadoop.mapred.JobConf</td>
 </tr>
 <tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-4267";>HADOOP-4267</a></td><td>mapred</td><td>Occasional
 exceptions during shutting down HSQLDB is logged but not rethrown.</td>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3019";>HADOOP-3019</a></td><td>mapred</td><td>Added
 a partitioner that effects a total order of output data, and an input sampler 
for generating the partition keyset for TotalOrderPartitioner for when the 
map's input keytype and distribution approximates its output.</td>
 </tr>
 <tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3652";>HADOOP-3652</a></td><td>mapred</td><td>Removed
 deprecated org.apache.hadoop.mapred.OutputFormatBase.</td>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3581";>HADOOP-3581</a></td><td>mapred</td><td>Added
 the ability to kill process trees transgressing memory limits. TaskTracker 
uses the configuration parameters introduced in <a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3759"; title="Provide 
ability to run memory intensive jobs without affecting other running tasks on 
the nodes"><strike>HADOOP-3759</strike></a>. In addition, 
mapred.tasktracker.taskmemorymanager.monitoring-interval specifies the interval 
for which TT waits between cycles of monitoring tasks' memory usage, and 
mapred.tasktracker.procfsbasedprocesstree.sleeptime-before-sigkill specifies 
the time TT waits for sending a SIGKILL to a process-tree that has overrun 
memory limits, after it has been sent a SIGTERM.</td>
+</tr>
+<tr>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3828";>HADOOP-3828</a></td><td>mapred</td><td>Skipped
 records can optionally be written to the HDFS. Refer 
org.apache.hadoop.mapred.SkipBadRecords.setSkipOutputPath for setting the 
output path.</td>
+</tr>
+<tr>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-4293";>HADOOP-4293</a></td><td>mapred</td><td>Made
 Configuration Writable and rename the old write method to writeXml.</td>
+</tr>
+<tr>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3412";>HADOOP-3412</a></td><td>mapred</td><td>Added
 the ability to chose between many schedulers, and to limit the number of 
running tasks per job.</td>
 </tr>
 <tr>
 <td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3667";>HADOOP-3667</a></td><td>mapred</td><td>Removed
 the following deprecated methods from JobConf:
@@ -280,31 +238,22 @@
 &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;setOutputPath(Path)</td>
 </tr>
 <tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3828";>HADOOP-3828</a></td><td>mapred</td><td>Skipped
 records can optionally be written to the HDFS. Refer 
org.apache.hadoop.mapred.SkipBadRecords.setSkipOutputPath for setting the 
output path.</td>
-</tr>
-<tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3961";>HADOOP-3961</a></td><td>mapred</td><td>resource
 estimation now only kicks in after 10% of maps have completed.</td>
-</tr>
-<tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3730";>HADOOP-3730</a></td><td>mapred</td><td>&nbsp;Added
 a JobConf constructor that disables loading  default configurations so as to 
take all default values from the JobTracker's configuration.</td>
-</tr>
-<tr>
 <td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3585";>HADOOP-3585</a></td><td>metrics</td><td>Added
 FailMon as a contrib project for hardware failure monitoring and analysis, 
under /src/contrib/failmon. Created User Manual and Quick Start Guide.</td>
 </tr>
 <tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-4228";>HADOOP-4228</a></td><td>metrics</td><td>Changed
 bytes_read, bytes_written to type long to prevent metrics overflow.</td>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3062";>HADOOP-3062</a></td><td>metrics</td><td>Introduced
 additional log records for data transfers.</td>
 </tr>
 <tr>
 <td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3714";>HADOOP-3714</a></td><td>scripts</td><td>Adds
 a new contrib, bash-tab-completion, which enables bash tab completion for the 
bin/hadoop script. See the README file in the contrib directory for the 
installation.</td>
 </tr>
 <tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3939";>HADOOP-3939</a></td><td>tools/distcp</td><td>Added
 a new option -delete to DistCp so that if the files/directories exist in dst 
but not in src will be deleted.  It uses FsShell to do delete, so that it will 
use trash if  the trash is enable.</td>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3889";>HADOOP-3889</a></td><td>tools/distcp</td><td>Changed
 DistCp error messages when there is a RemoteException.  Changed the 
corresponding return value from -999 to -3.</td>
 </tr>
 <tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3873";>HADOOP-3873</a></td><td>tools/distcp</td><td>Added
 two new options -filelimit &lt;n&gt; and -sizelimit &lt;n&gt; to DistCp for 
limiting the total number of files and the total size in bytes, 
respectively.</td>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3939";>HADOOP-3939</a></td><td>tools/distcp</td><td>Added
 a new option -delete to DistCp so that if the files/directories exist in dst 
but not in src will be deleted.  It uses FsShell to do delete, so that it will 
use trash if  the trash is enable.</td>
 </tr>
 <tr>
-<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3889";>HADOOP-3889</a></td><td>tools/distcp</td><td>Changed
 DistCp error messages when there is a RemoteException.  Changed the 
corresponding return value from -999 to -3.</td>
+<td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3873";>HADOOP-3873</a></td><td>tools/distcp</td><td>Added
 two new options -filelimit &lt;n&gt; and -sizelimit &lt;n&gt; to DistCp for 
limiting the total number of files and the total size in bytes, 
respectively.</td>
 </tr>
 <tr>
 <td><a 
href="https://issues.apache.org:443/jira/browse/HADOOP-3854";>HADOOP-3854</a></td><td>util</td><td>Added
 a configuration property hadoop.http.filter.initializers and a class 
org.apache.hadoop.http.FilterInitializer for supporting servlet filter.  
Cluster administrator could possibly configure customized filters for their web 
site.</td>


Reply via email to