Author: cdouglas
Date: Fri Feb 19 07:48:15 2010
New Revision: 911729

URL: http://svn.apache.org/viewvc?rev=911729&view=rev
Log:
Prepare for 0.20.2-rc4

Modified:
    hadoop/common/branches/branch-0.20/CHANGES.txt
    hadoop/common/branches/branch-0.20/src/docs/releasenotes.html

Modified: hadoop/common/branches/branch-0.20/CHANGES.txt
URL: 
http://svn.apache.org/viewvc/hadoop/common/branches/branch-0.20/CHANGES.txt?rev=911729&r1=911728&r2=911729&view=diff
==============================================================================
--- hadoop/common/branches/branch-0.20/CHANGES.txt (original)
+++ hadoop/common/branches/branch-0.20/CHANGES.txt Fri Feb 19 07:48:15 2010
@@ -2,7 +2,7 @@
 
 Release 0.20.3 - Unreleased
 
-Release 0.20.2 - 2010-2-18
+Release 0.20.2 - 2010-2-19
 
   NEW FEATURES
 

Modified: hadoop/common/branches/branch-0.20/src/docs/releasenotes.html
URL: 
http://svn.apache.org/viewvc/hadoop/common/branches/branch-0.20/src/docs/releasenotes.html?rev=911729&r1=911728&r2=911729&view=diff
==============================================================================
--- hadoop/common/branches/branch-0.20/src/docs/releasenotes.html (original)
+++ hadoop/common/branches/branch-0.20/src/docs/releasenotes.html Fri Feb 19 
07:48:15 2010
@@ -17,18 +17,20 @@
 <a name="changes"></a>
 <h2>Changes Since Hadoop 0.20.1</h2>
 
-<h3>Common</h3>        
+<h3>Common</h3>
 <h4>        Bug
 </h4>
 <ul>
 <li>[<a 
href='https://issues.apache.org/jira/browse/HADOOP-4802'>HADOOP-4802</a>] -     
    RPC Server send buffer retains size of largest response ever sent 
 </li>
+<li>[<a 
href='https://issues.apache.org/jira/browse/HADOOP-5611'>HADOOP-5611</a>] -     
    C++ libraries do not build on Debian Lenny
+</li>
+<li>[<a 
href='https://issues.apache.org/jira/browse/HADOOP-5612'>HADOOP-5612</a>] -     
    Some c++ scripts are not chmodded before ant execution
+</li>
 <li>[<a 
href='https://issues.apache.org/jira/browse/HADOOP-5623'>HADOOP-5623</a>] -     
    Streaming: process provided status messages are overwritten every 10 seoncds
 </li>
 <li>[<a 
href='https://issues.apache.org/jira/browse/HADOOP-5759'>HADOOP-5759</a>] -     
    IllegalArgumentException when CombineFileInputFormat is used as job 
InputFormat
 </li>
-<li>[<a 
href='https://issues.apache.org/jira/browse/HADOOP-5997'>HADOOP-5997</a>] -     
    Many test jobs write to HDFS under /
-</li>
 <li>[<a 
href='https://issues.apache.org/jira/browse/HADOOP-6097'>HADOOP-6097</a>] -     
    Multiple bugs w/ Hadoop archives
 </li>
 <li>[<a 
href='https://issues.apache.org/jira/browse/HADOOP-6231'>HADOOP-6231</a>] -     
    Allow caching of filesystem instances to be disabled on a per-instance basis
@@ -39,39 +41,22 @@
 </li>
 <li>[<a 
href='https://issues.apache.org/jira/browse/HADOOP-6386'>HADOOP-6386</a>] -     
    NameNode's HttpServer can't instantiate InetSocketAddress: 
IllegalArgumentException is thrown
 </li>
-<li>[<a 
href='https://issues.apache.org/jira/browse/HADOOP-6417'>HADOOP-6417</a>] -     
    Alternative Java Distributions in the Hadoop Documention
-</li>
 <li>[<a 
href='https://issues.apache.org/jira/browse/HADOOP-6428'>HADOOP-6428</a>] -     
    HttpServer sleeps with negative values
 </li>
-<li>[<a 
href='https://issues.apache.org/jira/browse/HADOOP-6453'>HADOOP-6453</a>] -     
    Hadoop wrapper script shouldn't ignore an existing JAVA_LIBRARY_PATH
-</li>
 <li>[<a 
href='https://issues.apache.org/jira/browse/HADOOP-6460'>HADOOP-6460</a>] -     
    Namenode runs of out of memory due to memory leak in ipc Server
 </li>
 <li>[<a 
href='https://issues.apache.org/jira/browse/HADOOP-6498'>HADOOP-6498</a>] -     
    IPC client  bug may cause rpc call hang
 </li>
-<li>[<a 
href='https://issues.apache.org/jira/browse/HADOOP-6502'>HADOOP-6502</a>] -     
    DistributedFileSystem#listStatus is very slow when listing a directory with 
a size of 1300
-</li>
 <li>[<a 
href='https://issues.apache.org/jira/browse/HADOOP-6506'>HADOOP-6506</a>] -     
    Failing tests prevent the rest of test targets from execution.
 </li>
 <li>[<a 
href='https://issues.apache.org/jira/browse/HADOOP-6524'>HADOOP-6524</a>] -     
    Contrib tests are failing Clover'ed build
 </li>
-</ul>
-    
-<h4>        Improvement
-</h4>
-<ul>
-<li>[<a 
href='https://issues.apache.org/jira/browse/HADOOP-3659'>HADOOP-3659</a>] -     
    Patch to allow hadoop native to compile on Mac OS X
-</li>
-<li>[<a 
href='https://issues.apache.org/jira/browse/HADOOP-6304'>HADOOP-6304</a>] -     
    Use java.io.File.set{Readable|Writable|Executable} where possible in 
RawLocalFileSystem 
-</li>
-<li>[<a 
href='https://issues.apache.org/jira/browse/HADOOP-6376'>HADOOP-6376</a>] -     
    slaves file to have a header specifying the format of conf/slaves file 
+<li>[<a 
href='https://issues.apache.org/jira/browse/HADOOP-6575'>HADOOP-6575</a>] -     
    Tests do not run on 0.20 branch
 </li>
-<li>[<a 
href='https://issues.apache.org/jira/browse/HADOOP-6475'>HADOOP-6475</a>] -     
    Improvements to the hadoop-config script
-</li>
-<li>[<a 
href='https://issues.apache.org/jira/browse/HADOOP-6542'>HADOOP-6542</a>] -     
    Add a -Dno-docs option to build.xml
+<li>[<a 
href='https://issues.apache.org/jira/browse/HADOOP-6576'>HADOOP-6576</a>] -     
    TestStreamingStatus is failing on 0.20 branch
 </li>
 </ul>
-                    
+
 <h4>        Task
 </h4>
 <ul>
@@ -79,6 +64,7 @@
 </li>
 </ul>
 
+
 <h3>HDFS</h3>
 <h4>        Bug
 </h4>
@@ -89,26 +75,16 @@
 </li>
 <li>[<a href='https://issues.apache.org/jira/browse/HDFS-187'>HDFS-187</a>] -  
       TestStartup fails if hdfs is running in the same machine
 </li>
-<li>[<a href='https://issues.apache.org/jira/browse/HDFS-442'>HDFS-442</a>] -  
       dfsthroughput in test.jar throws NPE
-</li>
 <li>[<a href='https://issues.apache.org/jira/browse/HDFS-495'>HDFS-495</a>] -  
       Hadoop FSNamesystem startFileInternal() getLease() has bug
 </li>
 <li>[<a href='https://issues.apache.org/jira/browse/HDFS-579'>HDFS-579</a>] -  
       HADOOP-3792 update of DfsTask incomplete
 </li>
 <li>[<a href='https://issues.apache.org/jira/browse/HDFS-596'>HDFS-596</a>] -  
       Memory leak in libhdfs: hdfsFreeFileInfo() in libhdfs does not free 
memory for mOwner and mGroup
 </li>
-<li>[<a href='https://issues.apache.org/jira/browse/HDFS-645'>HDFS-645</a>] -  
       Namenode does not leave safe mode even if all blocks are available
-</li>
-<li>[<a href='https://issues.apache.org/jira/browse/HDFS-667'>HDFS-667</a>] -  
       test-contrib target fails on hdfsproxy tests
-</li>
 <li>[<a href='https://issues.apache.org/jira/browse/HDFS-677'>HDFS-677</a>] -  
       Rename failure due to quota results in deletion of src directory
 </li>
-<li>[<a href='https://issues.apache.org/jira/browse/HDFS-686'>HDFS-686</a>] -  
       NullPointerException is thrown while merging edit log and image
-</li>
 <li>[<a href='https://issues.apache.org/jira/browse/HDFS-723'>HDFS-723</a>] -  
       Deadlock in DFSClient#DFSOutputStream
 </li>
-<li>[<a href='https://issues.apache.org/jira/browse/HDFS-727'>HDFS-727</a>] -  
       bug setting block size hdfsOpenFile 
-</li>
 <li>[<a href='https://issues.apache.org/jira/browse/HDFS-732'>HDFS-732</a>] -  
       HDFS files are ending up truncated
 </li>
 <li>[<a href='https://issues.apache.org/jira/browse/HDFS-734'>HDFS-734</a>] -  
       TestDatanodeBlockScanner times out in branch 0.20
@@ -123,29 +99,12 @@
 </li>
 <li>[<a href='https://issues.apache.org/jira/browse/HDFS-795'>HDFS-795</a>] -  
       DFS Write pipeline does not detect defective datanode correctly in some 
cases (HADOOP-3339)
 </li>
-<li>[<a href='https://issues.apache.org/jira/browse/HDFS-846'>HDFS-846</a>] -  
       SetSpaceQuota of value 9223372036854775807 does not apply quota.
-</li>
 <li>[<a href='https://issues.apache.org/jira/browse/HDFS-872'>HDFS-872</a>] -  
       DFSClient 0.20.1 is incompatible with HDFS 0.20.2
 </li>
-<li>[<a href='https://issues.apache.org/jira/browse/HDFS-886'>HDFS-886</a>] -  
       TestHDFSTrash fails on Windows
-</li>
-<li>[<a href='https://issues.apache.org/jira/browse/HDFS-920'>HDFS-920</a>] -  
       Incorrect metrics reporting of transcations metrics.  
-</li>
 <li>[<a href='https://issues.apache.org/jira/browse/HDFS-927'>HDFS-927</a>] -  
       DFSInputStream retries too many times for new block locations
 </li>
-<li>[<a href='https://issues.apache.org/jira/browse/HDFS-937'>HDFS-937</a>] -  
       Port HDFS-101 to branch 0.20
-</li>
-<li>[<a href='https://issues.apache.org/jira/browse/HDFS-961'>HDFS-961</a>] -  
       dfs_readdir incorrectly parses paths
-</li>
-</ul>
-    
-<h4>        Improvement
-</h4>
-<ul>
-<li>[<a href='https://issues.apache.org/jira/browse/HDFS-959'>HDFS-959</a>] -  
       Performance improvements to DFSClient and DataNode for faster DFS write 
at replication factor of 1
-</li>
 </ul>
-                                
+
 <h4>        Test
 </h4>
 <ul>
@@ -157,70 +116,46 @@
 </li>
 </ul>
 
-<h3>MapReduce</h3>
 
+<h3>MapReduce</h3>
 <h4>        Bug
 </h4>
 <ul>
 <li>[<a 
href='https://issues.apache.org/jira/browse/MAPREDUCE-112'>MAPREDUCE-112</a>] - 
        Reduce Input Records and Reduce Output Records counters are not being 
set when using the new Mapreduce reducer API
 </li>
-<li>[<a 
href='https://issues.apache.org/jira/browse/MAPREDUCE-118'>MAPREDUCE-118</a>] - 
        Job.getJobID() will always return null
-</li>
 <li>[<a 
href='https://issues.apache.org/jira/browse/MAPREDUCE-433'>MAPREDUCE-433</a>] - 
        TestReduceFetch failed.
 </li>
-<li>[<a 
href='https://issues.apache.org/jira/browse/MAPREDUCE-806'>MAPREDUCE-806</a>] - 
        WordCount example does not compile given the current instructions
-</li>
 <li>[<a 
href='https://issues.apache.org/jira/browse/MAPREDUCE-826'>MAPREDUCE-826</a>] - 
        harchive doesn't use ToolRunner / harchive returns 0 even if the job 
fails with exception
 </li>
 <li>[<a 
href='https://issues.apache.org/jira/browse/MAPREDUCE-979'>MAPREDUCE-979</a>] - 
        JobConf.getMemoryFor{Map|Reduce}Task doesn't fallback to newer config 
knobs when mapred.taskmaxvmem is set to DISABLED_MEMORY_LIMIT of -1
 </li>
 <li>[<a 
href='https://issues.apache.org/jira/browse/MAPREDUCE-1010'>MAPREDUCE-1010</a>] 
-         Adding tests for changes in archives.
 </li>
-<li>[<a 
href='https://issues.apache.org/jira/browse/MAPREDUCE-1057'>MAPREDUCE-1057</a>] 
-         java tasks are not honouring the value of mapred.userlog.limit.kb
-</li>
 <li>[<a 
href='https://issues.apache.org/jira/browse/MAPREDUCE-1068'>MAPREDUCE-1068</a>] 
-         In hadoop-0.20.0 streaming job do not throw proper verbose error 
message if file is not present
 </li>
 <li>[<a 
href='https://issues.apache.org/jira/browse/MAPREDUCE-1070'>MAPREDUCE-1070</a>] 
-         Deadlock in FairSchedulerServlet
 </li>
-<li>[<a 
href='https://issues.apache.org/jira/browse/MAPREDUCE-1088'>MAPREDUCE-1088</a>] 
-         JobHistory files should have narrower 0600 perms 
-</li>
 <li>[<a 
href='https://issues.apache.org/jira/browse/MAPREDUCE-1112'>MAPREDUCE-1112</a>] 
-         Fix CombineFileInputFormat for hadoop 0.20
 </li>
 <li>[<a 
href='https://issues.apache.org/jira/browse/MAPREDUCE-1147'>MAPREDUCE-1147</a>] 
-         Map output records counter missing for map-only jobs in new API
 </li>
-<li>[<a 
href='https://issues.apache.org/jira/browse/MAPREDUCE-1157'>MAPREDUCE-1157</a>] 
-         JT UI shows incorrect scheduling info for failed/killed retired jobs
-</li>
 <li>[<a 
href='https://issues.apache.org/jira/browse/MAPREDUCE-1163'>MAPREDUCE-1163</a>] 
-         hdfsJniHelper.h: Yahoo! specific paths are encoded
 </li>
 <li>[<a 
href='https://issues.apache.org/jira/browse/MAPREDUCE-1182'>MAPREDUCE-1182</a>] 
-         Reducers fail with OutOfMemoryError while copying Map outputs
 </li>
-<li>[<a 
href='https://issues.apache.org/jira/browse/MAPREDUCE-1262'>MAPREDUCE-1262</a>] 
-         Eclipse Plugin does not build for Hadoop 0.20.1
-</li>
-<li>[<a 
href='https://issues.apache.org/jira/browse/MAPREDUCE-1264'>MAPREDUCE-1264</a>] 
-         Error Recovery failed, task will continue but run forever as new data 
only comes in very very slowly
-</li>
-<li>[<a 
href='https://issues.apache.org/jira/browse/MAPREDUCE-1321'>MAPREDUCE-1321</a>] 
-         Spurios logs with 
org.apache.hadoop.util.DiskChecker$DiskErrorException in TaskTracker
+<li>[<a 
href='https://issues.apache.org/jira/browse/MAPREDUCE-1251'>MAPREDUCE-1251</a>] 
-         c++ utils doesn't compile
 </li>
 <li>[<a 
href='https://issues.apache.org/jira/browse/MAPREDUCE-1328'>MAPREDUCE-1328</a>] 
-         contrib/index  - modify build / ivy files as appropriate 
 </li>
-<li>[<a 
href='https://issues.apache.org/jira/browse/MAPREDUCE-1346'>MAPREDUCE-1346</a>] 
-         TestStreamingExitStatus / TestStreamingKeyValue - correct text 
fixtures in place 
-</li>
-<li>[<a 
href='https://issues.apache.org/jira/browse/MAPREDUCE-1381'>MAPREDUCE-1381</a>] 
-         Incorrect values being displayed for blacklisted_maps and 
blacklisted_reduces
-</li>
-<li>[<a 
href='https://issues.apache.org/jira/browse/MAPREDUCE-1389'>MAPREDUCE-1389</a>] 
-         TestDFSIO creates TestDFSIO_results.log file directly under 
hadoop.home
-</li>
-<li>[<a 
href='https://issues.apache.org/jira/browse/MAPREDUCE-1397'>MAPREDUCE-1397</a>] 
-         NullPointerException observed during task failures
-</li>
 </ul>
-    
+
 <h4>        Improvement
 </h4>
 <ul>
-<li>[<a 
href='https://issues.apache.org/jira/browse/MAPREDUCE-1315'>MAPREDUCE-1315</a>] 
-         taskdetails.jsp and jobfailures.jsp should have consistent convention 
for machine names in case of lost task tracker
-</li>
-<li>[<a 
href='https://issues.apache.org/jira/browse/MAPREDUCE-1361'>MAPREDUCE-1361</a>] 
-         In the pools with minimum slots, new job will always receive slots 
even if the minimum slots limit has been fulfilled
+<li>[<a 
href='https://issues.apache.org/jira/browse/MAPREDUCE-623'>MAPREDUCE-623</a>] - 
        Resolve javac warnings in mapred
 </li>
 </ul>
-    
+
 <h4>        New Feature
 </h4>
 <ul>
@@ -229,14 +164,8 @@
 <li>[<a 
href='https://issues.apache.org/jira/browse/MAPREDUCE-1170'>MAPREDUCE-1170</a>] 
-         MultipleInputs doesn't work with new API in 0.20 branch
 </li>
 </ul>
-                            
-<h4>        Test
-</h4>
-<ul>
-<li>[<a 
href='https://issues.apache.org/jira/browse/MAPREDUCE-1336'>MAPREDUCE-1336</a>] 
-         TestStreamingExitStatus - Fix deprecated use of StreamJob submission 
API
-</li>
-</ul>
-                
+
+
 <h2>Changes Since Hadoop 0.20.0</h2>
 
 <h3>Common</h3>


Reply via email to