Modified: 
hbase/branches/0.89-fb/src/test/java/org/apache/hadoop/hbase/regionserver/TestStoreFile.java
URL: 
http://svn.apache.org/viewvc/hbase/branches/0.89-fb/src/test/java/org/apache/hadoop/hbase/regionserver/TestStoreFile.java?rev=1294395&r1=1294394&r2=1294395&view=diff
==============================================================================
--- 
hbase/branches/0.89-fb/src/test/java/org/apache/hadoop/hbase/regionserver/TestStoreFile.java
 (original)
+++ 
hbase/branches/0.89-fb/src/test/java/org/apache/hadoop/hbase/regionserver/TestStoreFile.java
 Mon Feb 27 23:37:53 2012
@@ -36,7 +36,6 @@ import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.Path;
 import org.apache.hadoop.hbase.HBaseTestCase;
 import org.apache.hadoop.hbase.HBaseTestingUtility;
-import org.apache.hadoop.hbase.HColumnDescriptor;
 import org.apache.hadoop.hbase.HConstants;
 import org.apache.hadoop.hbase.KeyValue;
 import org.apache.hadoop.hbase.client.Scan;
@@ -101,9 +100,12 @@ public class TestStoreFile extends HBase
    */
   public void testBasicHalfMapFile() throws Exception {
     // Make up a directory hierarchy that has a regiondir and familyname.
-    StoreFile.Writer writer = StoreFile.createWriter(this.fs,
-      new Path(new Path(this.testDir, "regionname"), "familyname"), 2 * 1024,
-      conf, cacheConf);
+    Path outputDir = new Path(new Path(this.testDir, "regionname"),
+        "familyname");
+    StoreFile.Writer writer = new StoreFile.WriterBuilder(conf, cacheConf,
+        this.fs, 2 * 1024)
+            .withOutputDir(outputDir)
+            .build();
     writeStoreFile(writer);
     checkHalfHFile(new StoreFile(this.fs, writer.getPath(), conf, cacheConf,
         StoreFile.BloomType.NONE, NoOpDataBlockEncoder.INSTANCE));
@@ -143,8 +145,10 @@ public class TestStoreFile extends HBase
     Path storedir = new Path(new Path(this.testDir, "regionname"), 
"familyname");
     Path dir = new Path(storedir, "1234567890");
     // Make a store file and write data to it.
-    StoreFile.Writer writer = StoreFile.createWriter(this.fs, dir, 8 * 1024,
-        conf, cacheConf);
+    StoreFile.Writer writer = new StoreFile.WriterBuilder(conf, cacheConf,
+        this.fs, 8 * 1024)
+            .withOutputDir(dir)
+            .build();
     writeStoreFile(writer);
     StoreFile hsf = new StoreFile(this.fs, writer.getPath(), conf, cacheConf,
         StoreFile.BloomType.NONE, NoOpDataBlockEncoder.INSTANCE);
@@ -404,10 +408,12 @@ public class TestStoreFile extends HBase
 
     // write the file
     Path f = new Path(ROOT_DIR, getName());
-    StoreFile.Writer writer =
-        new StoreFile.Writer(fs, f, StoreFile.DEFAULT_BLOCKSIZE_SMALL,
-            HFile.DEFAULT_COMPRESSION_ALGORITHM, conf, cacheConf,
-            KeyValue.COMPARATOR, StoreFile.BloomType.ROW, 2000);
+    StoreFile.Writer writer = new StoreFile.WriterBuilder(conf, cacheConf, fs,
+        StoreFile.DEFAULT_BLOCKSIZE_SMALL)
+            .withFilePath(f)
+            .withBloomType(StoreFile.BloomType.ROW)
+            .withMaxKeyCount(2000)
+            .build();
     bloomWriteRead(writer, fs);
   }
 
@@ -421,10 +427,11 @@ public class TestStoreFile extends HBase
 
     // write the file
     Path f = new Path(ROOT_DIR, getName());
-    StoreFile.Writer writer = new StoreFile.Writer(fs, f,
-        StoreFile.DEFAULT_BLOCKSIZE_SMALL, HFile.DEFAULT_COMPRESSION_ALGORITHM,
-        conf, cacheConf, KeyValue.COMPARATOR, StoreFile.BloomType.NONE,
-        2000);
+    StoreFile.Writer writer = new StoreFile.WriterBuilder(conf, cacheConf,
+        fs, StoreFile.DEFAULT_BLOCKSIZE_SMALL)
+            .withFilePath(f)
+            .withMaxKeyCount(2000)
+            .build();
 
     // add delete family
     long now = System.currentTimeMillis();
@@ -489,10 +496,12 @@ public class TestStoreFile extends HBase
     for (int x : new int[]{0,1}) {
       // write the file
       Path f = new Path(ROOT_DIR, getName() + x);
-      StoreFile.Writer writer = new StoreFile.Writer(fs, f,
-          StoreFile.DEFAULT_BLOCKSIZE_SMALL,
-          HFile.DEFAULT_COMPRESSION_ALGORITHM,
-          conf, cacheConf, KeyValue.COMPARATOR, bt[x], expKeys[x]);
+      StoreFile.Writer writer = new StoreFile.WriterBuilder(conf, cacheConf,
+          fs, StoreFile.DEFAULT_BLOCKSIZE_SMALL)
+              .withFilePath(f)
+              .withBloomType(bt[x])
+              .withMaxKeyCount(expKeys[x])
+              .build();
 
       long now = System.currentTimeMillis();
       for (int i = 0; i < rowCount*2; i += 2) { // rows
@@ -562,10 +571,12 @@ public class TestStoreFile extends HBase
     conf.setInt(HFile.FORMAT_VERSION_KEY, 1);
 
     // this should not create a bloom because the max keys is too small
-    StoreFile.Writer writer = new StoreFile.Writer(fs, f,
-        StoreFile.DEFAULT_BLOCKSIZE_SMALL, HFile.DEFAULT_COMPRESSION_ALGORITHM,
-        conf, cacheConf, KeyValue.COMPARATOR, StoreFile.BloomType.ROW,
-        2000);
+    StoreFile.Writer writer = new StoreFile.WriterBuilder(conf, cacheConf, fs,
+        StoreFile.DEFAULT_BLOCKSIZE_SMALL)
+            .withFilePath(f)
+            .withBloomType(StoreFile.BloomType.ROW)
+            .withMaxKeyCount(2000)
+            .build();
     assertFalse(writer.hasGeneralBloom());
     writer.close();
     fs.delete(f, true);
@@ -573,23 +584,23 @@ public class TestStoreFile extends HBase
     conf.setInt(BloomFilterFactory.IO_STOREFILE_BLOOM_MAX_KEYS,
         Integer.MAX_VALUE);
 
-    // TODO: commented out because we ran out of java heap space on trunk
-    /*
-    // the below config caused IllegalArgumentException in our production 
cluster
-    // however, the resulting byteSize is < MAX_INT, so this should work 
properly
-    writer = new StoreFile.Writer(fs, f,
-        StoreFile.DEFAULT_BLOCKSIZE_SMALL, HFile.DEFAULT_COMPRESSION_ALGORITHM,
-        conf, KeyValue.COMPARATOR, StoreFile.BloomType.ROW, 272446963);
-    assertTrue(writer.hasBloom());
+    writer = new StoreFile.WriterBuilder(conf, cacheConf, fs,
+        StoreFile.DEFAULT_BLOCKSIZE_SMALL)
+            .withFilePath(f)
+            .withBloomType(StoreFile.BloomType.ROW)
+            .withMaxKeyCount(27244696)
+            .build();
+    assertTrue(writer.hasGeneralBloom());
     bloomWriteRead(writer, fs);
-    */
 
     // this, however, is too large and should not create a bloom
     // because Java can't create a contiguous array > MAX_INT
-    writer = new StoreFile.Writer(fs, f,
-        StoreFile.DEFAULT_BLOCKSIZE_SMALL, HFile.DEFAULT_COMPRESSION_ALGORITHM,
-        conf, cacheConf, KeyValue.COMPARATOR, StoreFile.BloomType.ROW,
-        Integer.MAX_VALUE);
+    writer = new StoreFile.WriterBuilder(conf, cacheConf, fs,
+        StoreFile.DEFAULT_BLOCKSIZE_SMALL)
+            .withFilePath(f)
+            .withBloomType(StoreFile.BloomType.ROW)
+            .withMaxKeyCount(Integer.MAX_VALUE)
+            .build();
     assertFalse(writer.hasGeneralBloom());
     writer.close();
     fs.delete(f, true);
@@ -680,8 +691,10 @@ public class TestStoreFile extends HBase
     Path storedir = new Path(new Path(this.testDir, "regionname"),
     "familyname");
     Path dir = new Path(storedir, "1234567890");
-    StoreFile.Writer writer = StoreFile.createWriter(this.fs, dir, 8 * 1024,
-        conf, cacheConf);
+    StoreFile.Writer writer = new StoreFile.WriterBuilder(conf, cacheConf,
+        this.fs, 8 * 1024)
+            .withOutputDir(dir)
+            .build();
 
     List<KeyValue> kvList = getKeyValueSet(timestamps,numRows,
         family, qualifier);
@@ -853,10 +866,11 @@ public class TestStoreFile extends HBase
       totalSize += kv.getLength() + 1;
     }
     int blockSize = totalSize / numBlocks;
-    StoreFile.Writer writer = new StoreFile.Writer(fs, path, blockSize,
-        HFile.DEFAULT_COMPRESSION_ALGORITHM,
-        conf, cacheConf, KeyValue.COMPARATOR, StoreFile.BloomType.NONE,
-        2000);
+    StoreFile.Writer writer = new StoreFile.WriterBuilder(conf, cacheConf, fs,
+        blockSize)
+            .withFilePath(path)
+            .withMaxKeyCount(2000)
+            .build();
     // We'll write N-1 KVs to ensure we don't write an extra block
     kvs.remove(kvs.size()-1);
     for (KeyValue kv : kvs) {
@@ -882,15 +896,12 @@ public class TestStoreFile extends HBase
             dataBlockEncoderAlgo,
             dataBlockEncoderAlgo);
     cacheConf = new CacheConfig(conf);
-    StoreFile.Writer writer = new StoreFile.Writer(fs,
-        path, HFile.DEFAULT_BLOCKSIZE,
-        HFile.DEFAULT_COMPRESSION_ALGORITHM,
-        dataBlockEncoder,
-        conf,
-        cacheConf,
-        KeyValue.COMPARATOR,
-        StoreFile.BloomType.NONE,
-        HColumnDescriptor.DEFAULT_BLOOMFILTER_ERROR_RATE, 2000, null);
+    StoreFile.Writer writer = new StoreFile.WriterBuilder(conf, cacheConf, fs,
+        HFile.DEFAULT_BLOCKSIZE)
+            .withFilePath(path)
+            .withDataBlockEncoder(dataBlockEncoder)
+            .withMaxKeyCount(2000)
+            .build();
     writer.close();
 
     StoreFile storeFile = new StoreFile(fs, writer.getPath(), conf,
@@ -902,4 +913,4 @@ public class TestStoreFile extends HBase
 
     assertEquals(dataBlockEncoderAlgo.getNameInBytes(), value);
   }
-}
+}
\ No newline at end of file

Modified: 
hbase/branches/0.89-fb/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestWALReplay.java
URL: 
http://svn.apache.org/viewvc/hbase/branches/0.89-fb/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestWALReplay.java?rev=1294395&r1=1294394&r2=1294395&view=diff
==============================================================================
--- 
hbase/branches/0.89-fb/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestWALReplay.java
 (original)
+++ 
hbase/branches/0.89-fb/src/test/java/org/apache/hadoop/hbase/regionserver/wal/TestWALReplay.java
 Mon Feb 27 23:37:53 2012
@@ -41,7 +41,6 @@ import org.apache.hadoop.hbase.KeyValue;
 import org.apache.hadoop.hbase.client.Get;
 import org.apache.hadoop.hbase.client.Put;
 import org.apache.hadoop.hbase.client.Result;
-import org.apache.hadoop.hbase.io.hfile.CacheConfig;
 import org.apache.hadoop.hbase.io.hfile.HFile;
 import org.apache.hadoop.hbase.monitoring.MonitoredTask;
 import org.apache.hadoop.hbase.regionserver.HRegion;
@@ -188,7 +187,8 @@ public class TestWALReplay {
     HLog wal = createWAL(this.conf);
     HRegion region = HRegion.openHRegion(hri, basedir, wal, this.conf);
     Path f =  new Path(basedir, "hfile");
-    HFile.Writer writer = HFile.getWriterFactory(conf).createWriter(this.fs, 
f);
+    HFile.Writer writer =
+      HFile.getWriterFactoryNoCache(conf).withPath(fs, f).create();
     byte [] family = 
hri.getTableDesc().getFamilies().iterator().next().getName();
     byte [] row = Bytes.toBytes(tableNameStr);
     writer.append(new KeyValue(row, family, family, row));
@@ -550,4 +550,4 @@ public class TestWALReplay {
     HBaseTestingUtility.setMaxRecoveryErrorCount(wal.getOutputStream(), 1);
     return wal;
   }
-}
+}
\ No newline at end of file


Reply via email to