ACCUMULO-1563 committing Jonathan Hsieh's patch to prevent the test from writing to /
git-svn-id: https://svn.apache.org/repos/asf/accumulo/branches/1.5@1502583 13f79535-47bb-0310-9956-ffa450edef68 Project: http://git-wip-us.apache.org/repos/asf/accumulo/repo Commit: http://git-wip-us.apache.org/repos/asf/accumulo/commit/1fc73a9d Tree: http://git-wip-us.apache.org/repos/asf/accumulo/tree/1fc73a9d Diff: http://git-wip-us.apache.org/repos/asf/accumulo/diff/1fc73a9d Branch: refs/heads/ACCUMULO-1000 Commit: 1fc73a9d718b24350191b0d46c6c6012748dc76b Parents: bd6c426 Author: Eric C. Newton <e...@apache.org> Authored: Fri Jul 12 15:11:14 2013 +0000 Committer: Eric C. Newton <e...@apache.org> Committed: Fri Jul 12 15:11:14 2013 +0000 ---------------------------------------------------------------------- .../main/java/org/apache/accumulo/test/CreateRFiles.java | 2 +- .../test/functional/BulkSplitOptimizationTest.java | 8 ++++---- test/system/auto/simple/bulk.py | 10 +++++----- test/system/auto/simple/compaction.py | 4 ++-- 4 files changed, 12 insertions(+), 12 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/accumulo/blob/1fc73a9d/test/src/main/java/org/apache/accumulo/test/CreateRFiles.java ---------------------------------------------------------------------- diff --git a/test/src/main/java/org/apache/accumulo/test/CreateRFiles.java b/test/src/main/java/org/apache/accumulo/test/CreateRFiles.java index fc50ed9..82f9b6b 100644 --- a/test/src/main/java/org/apache/accumulo/test/CreateRFiles.java +++ b/test/src/main/java/org/apache/accumulo/test/CreateRFiles.java @@ -58,7 +58,7 @@ public class CreateRFiles { int count = 0; while (currEnd <= opts.end && currStart < currEnd) { - final String tia = String.format("--rfile /%s/mf%05d --timestamp 1 --size 50 --random 56 --rows %d --start %d --user root", opts.outputDirectory, count, currEnd - currStart, currStart); + final String tia = String.format("--rfile %s/mf%05d --timestamp 1 --size 50 --random 56 --rows %d --start %d --user root", opts.outputDirectory, count, currEnd - currStart, currStart); Runnable r = new Runnable() { http://git-wip-us.apache.org/repos/asf/accumulo/blob/1fc73a9d/test/src/main/java/org/apache/accumulo/test/functional/BulkSplitOptimizationTest.java ---------------------------------------------------------------------- diff --git a/test/src/main/java/org/apache/accumulo/test/functional/BulkSplitOptimizationTest.java b/test/src/main/java/org/apache/accumulo/test/functional/BulkSplitOptimizationTest.java index 5a62e5e..0cdd3b7 100644 --- a/test/src/main/java/org/apache/accumulo/test/functional/BulkSplitOptimizationTest.java +++ b/test/src/main/java/org/apache/accumulo/test/functional/BulkSplitOptimizationTest.java @@ -44,8 +44,8 @@ public class BulkSplitOptimizationTest extends FunctionalTest { @Override public void cleanup() throws Exception { FileSystem fs = FileSystem.get(CachedConfiguration.getInstance()); - fs.delete(new Path("/tmp/testmf"), true); - fs.delete(new Path("/tmp/testmf_failures"), true); + fs.delete(new Path("tmp/testmf"), true); + fs.delete(new Path("tmp/testmf_failures"), true); } @Override @@ -63,11 +63,11 @@ public class BulkSplitOptimizationTest extends FunctionalTest { public void run() throws Exception { FileSystem fs = FileSystem.get(CachedConfiguration.getInstance()); - fs.delete(new Path("/tmp/testmf"), true); + fs.delete(new Path("tmp/testmf"), true); AuthenticationToken token = this.getToken(); CreateRFiles.main(new String[] {"--output", "tmp/testmf", "--numThreads", "8", "--start", "0", "--end", "100000", "--splits", "99"}); - bulkImport(fs, TABLE_NAME, "/tmp/testmf"); + bulkImport(fs, TABLE_NAME, "tmp/testmf"); checkSplits(TABLE_NAME, 0, 0); checkRFiles(TABLE_NAME, 1, 1, 100, 100); http://git-wip-us.apache.org/repos/asf/accumulo/blob/1fc73a9d/test/system/auto/simple/bulk.py ---------------------------------------------------------------------- diff --git a/test/system/auto/simple/bulk.py b/test/system/auto/simple/bulk.py index 3899176..b0b73b4 100755 --- a/test/system/auto/simple/bulk.py +++ b/test/system/auto/simple/bulk.py @@ -42,13 +42,13 @@ class SimpleBulkTest(TestUtilsMixin, unittest.TestCase): self.masterHost(), 'org.apache.accumulo.test.BulkImportDirectory', ['-u', ROOT, '-p', ROOT_PASSWORD, - '-t', 'test_ingest', '--source', dir, '--failures', '/testBulkFail', '-i', INSTANCE_NAME]) + '-t', 'test_ingest', '--source', dir, '--failures', 'testBulkFail', '-i', INSTANCE_NAME]) self.wait(handle) self.assert_(handle.returncode == 0) def createRFiles(self): - args = '--rfile /testrf/rf%02d --timestamp 1 --size 50 --random 56 --rows %1d --start %ld --cols 1 -u root -i ' + INSTANCE_NAME + args = '--rfile testrf/rf%02d --timestamp 1 --size 50 --random 56 --rows %1d --start %ld --cols 1 -u root -i ' + INSTANCE_NAME log.info('creating rfiles') handles = [] for i in range(COUNT): @@ -73,9 +73,9 @@ class SimpleBulkTest(TestUtilsMixin, unittest.TestCase): # initialize the database self.createTable('test_ingest') - self.execute(self.masterHost(), 'hadoop dfs -rmr /testrf'.split()) - self.execute(self.masterHost(), 'hadoop dfs -rmr /testBulkFail'.split()) - self.execute(self.masterHost(), 'hadoop dfs -mkdir /testBulkFail'.split()) + self.execute(self.masterHost(), 'hadoop dfs -rmr testrf'.split()) + self.execute(self.masterHost(), 'hadoop dfs -rmr testBulkFail'.split()) + self.execute(self.masterHost(), 'hadoop dfs -mkdir testBulkFail'.split()) # insert some data self.createRFiles() http://git-wip-us.apache.org/repos/asf/accumulo/blob/1fc73a9d/test/system/auto/simple/compaction.py ---------------------------------------------------------------------- diff --git a/test/system/auto/simple/compaction.py b/test/system/auto/simple/compaction.py index f2ff90d..9bd0b56 100755 --- a/test/system/auto/simple/compaction.py +++ b/test/system/auto/simple/compaction.py @@ -52,8 +52,8 @@ class CompactionTest(SimpleBulkTest): # initialize the database self.createTable('test_ingest') - self.execute(self.masterHost(), 'hadoop dfs -rmr /testrf'.split()) - self.execute(self.masterHost(), 'hadoop dfs -rmr /testrfFail'.split()) + self.execute(self.masterHost(), 'hadoop dfs -rmr testrf'.split()) + self.execute(self.masterHost(), 'hadoop dfs -rmr testrfFail'.split()) # insert some data self.createRFiles(self.masterHost())