I downloaded RC3 and encountered a repeatable unit test failure on my Mac laptop
Test set: org.apache.hadoop.hbase.util.TestMergeTool ------------------------------------------------------------------------------- Tests run: 1, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 6.087 sec <<< FAILURE! testMergeTool(org.apache.hadoop.hbase.util.TestMergeTool) Time elapsed: 6.06 sec <<< FAILURE! junit.framework.AssertionFailedError: 'merging regions 0 and 1' failed at junit.framework.Assert.fail(Assert.java:47) at junit.framework.Assert.assertTrue(Assert.java:20) at org.apache.hadoop.hbase.util.TestMergeTool.mergeAndVerify(TestMergeTool.java:182) at org.apache.hadoop.hbase.util.TestMergeTool.testMergeTool(TestMergeTool.java:257) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) at java.lang.reflect.Method.invoke(Method.java:597) My OS: Darwin tyumac.local 10.5.0 Darwin Kernel Version 10.5.0: Fri Nov 5 23:20:39 PDT 2010; root:xnu-1504.9.17~1/RELEASE_I386 i386 i386 >From org.apache.hadoop.hbase.util.TestMergeTool-output.txt: 2011-01-08 20:57:44,469 WARN [org.apache.hadoop.hdfs.server.datanode.dataxceiverser...@1afd1810] datanode.DataXceiverServer(137): DatanodeRegistration(127.0.0.1:60366, storageID=DS-788800082-192.168.2.107-60366-1294549062630, infoPort=60367, ipcPort=60368):DataXceiveServer: java.nio.channels.AsynchronousCloseException at java.nio.channels.spi.AbstractInterruptibleChannel.end(AbstractInterruptibleChannel.java:185) at sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:152) at sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:84) at org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:130) at java.lang.Thread.run(Thread.java:680) Shutting down DataNode 0 FYI On Fri, Jan 7, 2011 at 5:03 PM, Stack <st...@duboce.net> wrote: > The fourth hbase 0.90.0 release candidate is available for download: > > > http://people.apache.org/~stack/hbase-0.90.0-candidate-3/<http://people.apache.org/%7Estack/hbase-0.90.0-candidate-3/> > > This is going to be the one! > > Should we release this candidate as hbase 0.90.0? Take it for a spin. > Check out the doc., etc. Vote +1/-1 by next Friday, the 14th of January. > > HBase 0.90.0 is the major HBase release that follows 0.20.0 and the > fruit of the 0.89.x development release series we've been running of > late. > > Over 1k issues have been closed since 0.20.0. Release notes are > available here: http://su.pr/8LbgvK. > > HBase 0.90.0 runs on Hadoop 0.20.x. It does not currently run on > Hadoop 0.21.0 nor on Hadoop TRUNK. HBase will lose data unless it is > running on an Hadoop HDFS 0.20.x that has a durable sync. Currently > only the branch-0.20-append branch [1] has this attribute (See > CHANGES.txt [3] in branch-0.20-append to see the list of patches > involved adding an append). No official releases have been made from > this branch as yet so you will have to build your own Hadoop from the > tip of this branch, OR install Cloudera's CDH3 [2] (Its currently in > beta). CDH3b2 or CDHb3 have the 0.20-append patches needed to add a > durable sync. If using CDH, be sure to replace the hadoop jars that > are bundled with HBase with those from your CDH distribution. > > There is no migration necessary. Your data written with HBase 0.20.x > (or with HBase 0.89.x) is readable by HBase 0.90.0. A shutdown and > restart after putting in place the new HBase should be all thats > involved. That said, once done, there is no going back to 0.20.x once > the transition has been made. HBase 0.90.0 and HBase 0.89.x write > region names differently in the filesystem. Rolling restart from > 0.20.x or 0.89.x to 0.90.0RC1 will not work. > > Yours, > The HBasistas > P.S. For why the version 0.90 and whats new in HBase 0.90, see slides > 4-10 in this deck [4] > > 1. http://svn.apache.org/viewvc/hadoop/common/branches/branch-0.20-append > 2. http://archive.cloudera.com/docs/ > 3. > http://svn.apache.org/viewvc/hadoop/common/branches/branch-0.20-append/CHANGES.txt > 4. http://hbaseblog.com/2010/07/04/hug11-hbase-0-90-preview-wrap-up/ >