Modified: hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupInserters.java URL: http://svn.apache.org/viewvc/hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupInserters.java?rev=823411&r1=823410&r2=823411&view=diff ============================================================================== --- hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupInserters.java (original) +++ hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupInserters.java Fri Oct 9 05:04:10 2009 @@ -69,8 +69,8 @@ public void testInsertNullValues() throws IOException, ParseException { fs.delete(path, true); System.out.println("testInsertNullValues"); - writer = new ColumnGroup.Writer(path, "abc, def", false, "pig", "lzo2", - true, conf); + writer = new ColumnGroup.Writer(path, "abc, def", false, "pig", "gz", + null, null, (short) -1, true, conf); TableInserter ins = writer.getInserter("part1", true); // Tuple row = TypesUtils.createTuple(writer.getSchema()); // ins.insert(new BytesWritable("key".getBytes()), row); @@ -82,8 +82,8 @@ public void testFailureInvalidSchema() throws IOException, ParseException { fs.delete(path, true); System.out.println("testFailureInvalidSchema"); - writer = new ColumnGroup.Writer(path, "abc, def", false, "pig", "lzo2", - true, conf); + writer = new ColumnGroup.Writer(path, "abc, def", false, "pig", "gz", + null, null, (short) -1, true, conf); TableInserter ins = writer.getInserter("part1", true); Tuple row = TypesUtils.createTuple(Schema.parse("xyz, ijk, def")); try { @@ -102,8 +102,8 @@ ParseException { fs.delete(path, true); System.out.println("testFailureGetInserterAfterWriterClosed"); - writer = new ColumnGroup.Writer(path, "abc, def", false, "pig", "lzo2", - true, conf); + writer = new ColumnGroup.Writer(path, "abc, def", false, "pig", "gz", + null, null, (short) -1, true, conf); try { writer.close(); TableInserter ins = writer.getInserter("part1", true); @@ -120,8 +120,8 @@ ParseException { fs.delete(path, true); System.out.println("testFailureInsertAfterClose"); - writer = new ColumnGroup.Writer(path, "abc, def ", false, "pig", "lzo2", - true, conf); + writer = new ColumnGroup.Writer(path, "abc, def ", false, "pig", "gz", + null, null, (short) -1, true, conf); TableInserter ins = writer.getInserter("part1", true); Tuple row = TypesUtils.createTuple(writer.getSchema()); @@ -151,7 +151,7 @@ ParseException { fs.delete(path, true); System.out.println("testFailureInsertXtraColumn"); - writer = new ColumnGroup.Writer(path, "abc ", false, "pig", "lzo2", true, + writer = new ColumnGroup.Writer(path, "abc ", false, "pig", "gz", null, null, (short) -1, true, conf); TableInserter ins = writer.getInserter("part1", true); @@ -164,7 +164,7 @@ row.set(1, map); Assert .fail("Failed to catch insertion an extra column not defined in schema."); - } catch (ExecException e) { + } catch (IndexOutOfBoundsException e) { // noop, expecting exceptions } finally { ins.close(); @@ -177,8 +177,8 @@ ParseException { fs.delete(path, true); System.out.println("testInsertOneRow"); - writer = new ColumnGroup.Writer(path, "abc, def", false, "pig", "lzo2", - true, conf); + writer = new ColumnGroup.Writer(path, "abc, def", false, "pig", "gz", + null, null, (short) -1, true, conf); TableInserter ins = writer.getInserter("part1", true); Tuple row = TypesUtils.createTuple(writer.getSchema()); @@ -199,8 +199,8 @@ ParseException { fs.delete(path, true); System.out.println("testInsert2Rows"); - writer = new ColumnGroup.Writer(path, "abc, def", false, "pig", "lzo2", - true, conf); + writer = new ColumnGroup.Writer(path, "abc, def", false, "pig", "gz", + null, null, (short) -1, true, conf); TableInserter ins = writer.getInserter("part1", true); // row 1 @@ -233,8 +233,8 @@ ParseException { fs.delete(path, true); System.out.println("testInsert2Inserters"); - writer = new ColumnGroup.Writer(path, "abc, def", false, "pig", "lzo2", - true, conf); + writer = new ColumnGroup.Writer(path, "abc, def", false, "pig", "gz", + null, null, (short) -1, true, conf); TableInserter ins1 = writer.getInserter("part1", true); TableInserter ins2 = writer.getInserter("part2", true); @@ -270,8 +270,8 @@ ParseException { fs.delete(path, true); System.out.println("testFailureOverlappingKeys"); - writer = new ColumnGroup.Writer(path, "abc, def ", true, "pig", "lzo2", - true, conf); + writer = new ColumnGroup.Writer(path, "abc, def ", true, "pig", "gz", + null, null, (short) -1, true, conf); TableInserter ins1 = writer.getInserter("part1", false); TableInserter ins2 = writer.getInserter("part2", false);
Modified: hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupOpen.java URL: http://svn.apache.org/viewvc/hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupOpen.java?rev=823411&r1=823410&r2=823411&view=diff ============================================================================== --- hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupOpen.java (original) +++ hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupOpen.java Fri Oct 9 05:04:10 2009 @@ -62,8 +62,8 @@ @Test public void testNew() throws IOException, ParseException { System.out.println("testNew"); - writer = new ColumnGroup.Writer(path, "abc, def ", false, "pig", "lzo2", - true, conf); + writer = new ColumnGroup.Writer(path, "abc, def ", false, "pig", "gz", + null, null, (short) -1, true, conf); // NOTE: don't call writer.close() here close(); } @@ -73,11 +73,11 @@ ParseException { System.out.println("testFailureExistingSortedDiff"); try { - writer = new ColumnGroup.Writer(path, "abc, def ", false, "pig", "lzo2", - true, conf); + writer = new ColumnGroup.Writer(path, "abc, def ", false, "pig", "gz", + null, null, (short) -1, true, conf); finish(); - writer = new ColumnGroup.Writer(path, "abc, def", true, "pig", "lzo2", - false, conf); + writer = new ColumnGroup.Writer(path, "abc, def", true, "pig", "gz", + null, null, (short) -1, false, conf); Assert.fail("Failed to catch sorted flag alteration."); } catch (IOException e) { // noop, expecting exceptions @@ -89,8 +89,8 @@ @Test public void testExisting() throws IOException, ParseException { System.out.println("testExisting"); - writer = new ColumnGroup.Writer(path, " abc , def ", false, "pig", "lzo2", - false, conf); + writer = new ColumnGroup.Writer(path, " abc , def ", false, "pig", "gz", + null, null, (short) -1, false, conf); writer.close(); close(); } @@ -105,7 +105,7 @@ FSDataOutputStream in = fs.create(path); in.close(); writer = new ColumnGroup.Writer(path, " abc , def ", false, "pig", - "lzo2", false, conf); + "gz", null, null, (short) -1, false, conf); Assert.fail("Failed to catch path not a directory."); } catch (IOException e) { // noop, expecting exceptions @@ -121,7 +121,7 @@ fs.delete(path, true); FSDataOutputStream in = fs.create(new Path(path, ColumnGroup.META_FILE)); in.close(); - writer = new ColumnGroup.Writer(path, "abc", false, "pig", "lzo2", false, + writer = new ColumnGroup.Writer(path, "abc", false, "pig", "gz", null, null, (short) -1, false, conf); Assert.fail("Failed to catch meta file existence."); } catch (IOException e) { @@ -135,10 +135,10 @@ public void testFailureDiffSchema() throws IOException, ParseException { System.out.println("testFailureDiffSchema"); try { - writer = new ColumnGroup.Writer(path, "abc", false, "pig", "lzo2", false, + writer = new ColumnGroup.Writer(path, "abc", false, "pig", "gz", null, null, (short) -1, false, conf); writer.finish(); - writer = new ColumnGroup.Writer(path, "efg", false, "pig", "lzo2", false, + writer = new ColumnGroup.Writer(path, "efg", false, "pig", "gz", null, null, (short) -1, false, conf); Assert.fail("Failed to catch schema differences."); } catch (IOException e) { @@ -155,7 +155,7 @@ ColumnGroup.Writer writer2 = null; ColumnGroup.Writer writer3 = null; try { - writer1 = new ColumnGroup.Writer(path, "abc", false, "pig", "lzo2", true, + writer1 = new ColumnGroup.Writer(path, "abc", false, "pig", "gz", null, null, (short) -1, true, conf); writer2 = new ColumnGroup.Writer(path, conf); writer3 = new ColumnGroup.Writer(path, conf); Modified: hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupProjections.java URL: http://svn.apache.org/viewvc/hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupProjections.java?rev=823411&r1=823410&r2=823411&view=diff ============================================================================== --- hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupProjections.java (original) +++ hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupProjections.java Fri Oct 9 05:04:10 2009 @@ -63,7 +63,7 @@ schema = new Schema("a,b,c,d,e,f,g"); ColumnGroup.Writer writer = new ColumnGroup.Writer(path, schema, false, - "pig", "lzo2", true, conf); + "pig", "gz", null, null, (short) -1, true, conf); TableInserter ins = writer.getInserter("part0", true); // row 1 @@ -168,7 +168,7 @@ try { row.get(0); Assert.fail("Failed to catch out of boundary exceptions."); - } catch (ExecException e) { + } catch (IndexOutOfBoundsException e) { // no op, expecting out of bounds exceptions } @@ -202,7 +202,7 @@ try { row.get(1); Assert.fail("Failed to catch out of boundary exceptions."); - } catch (ExecException e) { + } catch (IndexOutOfBoundsException e) { // no op, expecting out of bounds exceptions } @@ -236,7 +236,7 @@ try { row.get(1); Assert.fail("Failed to catch 'out of boundary' exceptions."); - } catch (ExecException e) { + } catch (IndexOutOfBoundsException e) { // no op, expecting out of bounds exceptions } @@ -270,7 +270,7 @@ try { row.get(2); Assert.fail("Failed to catch 'out of boundary' exceptions."); - } catch (ExecException e) { + } catch (IndexOutOfBoundsException e) { // no op, expecting out of bounds exceptions } @@ -304,7 +304,7 @@ try { row.get(2); Assert.fail("Failed to catch 'out of boundary' exceptions."); - } catch (ExecException e) { + } catch (IndexOutOfBoundsException e) { // no op, expecting out of bounds exceptions } Modified: hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupReaders.java URL: http://svn.apache.org/viewvc/hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupReaders.java?rev=823411&r1=823410&r2=823411&view=diff ============================================================================== --- hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupReaders.java (original) +++ hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupReaders.java Fri Oct 9 05:04:10 2009 @@ -77,7 +77,7 @@ System.out.println("testInsert2Inserters"); boolean sorted = false; // true; writer = new ColumnGroup.Writer(path, "col1, colTWO", sorted, "pig", - "lzo2", true, conf); + "gz", null, null, (short) -1, true, conf); TableInserter ins1 = writer.getInserter("part1", false); TableInserter ins2 = writer.getInserter("part2", false); @@ -197,8 +197,8 @@ throws IOException, ExecException, ParseException { ColumnGroup.Writer writer = null; if (schemaString != null) { - writer = new ColumnGroup.Writer(path, schemaString, false, "pig", "lzo2", - true, conf); + writer = new ColumnGroup.Writer(path, schemaString, false, "pig", "gz", + null, null, (short) -1, true, conf); } else { writer = new ColumnGroup.Writer(path, conf); } Modified: hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupSplits.java URL: http://svn.apache.org/viewvc/hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupSplits.java?rev=823411&r1=823410&r2=823411&view=diff ============================================================================== --- hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupSplits.java (original) +++ hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/io/TestColumnGroupSplits.java Fri Oct 9 05:04:10 2009 @@ -73,7 +73,7 @@ schema = new Schema(STR_SCHEMA); ColumnGroup.Writer writer = new ColumnGroup.Writer(path, schema, false, - "pig", "lzo2", true, conf); + "pig", "gz", null, null, (short) -1, true, conf); TableInserter ins = writer.getInserter("part0", true); Tuple row = TypesUtils.createTuple(schema); @@ -273,7 +273,7 @@ try { row.get(0); Assert.fail("Failed to catch out of boundary exceptions."); - } catch (ExecException e) { + } catch (IndexOutOfBoundsException e) { // no op, expecting out of bounds exceptions } @@ -306,7 +306,7 @@ try { row.get(1); Assert.fail("Failed to catch out of boundary exceptions."); - } catch (ExecException e) { + } catch (IndexOutOfBoundsException e) { // no op, expecting out of bounds exceptions } @@ -340,7 +340,7 @@ try { row.get(1); Assert.fail("Failed to catch 'out of boundary' exceptions."); - } catch (ExecException e) { + } catch (IndexOutOfBoundsException e) { // no op, expecting out of bounds exceptions } @@ -374,7 +374,7 @@ try { row.get(2); Assert.fail("Failed to catch 'out of boundary' exceptions."); - } catch (ExecException e) { + } catch (IndexOutOfBoundsException e) { // no op, expecting out of bounds exceptions } @@ -415,7 +415,7 @@ try { Object obj = tupColl.get(1); Assert.fail("Failed to catch 'out of boundary' exceptions."); - } catch (ExecException e) { + } catch (IndexOutOfBoundsException e) { // no op, expecting out of bounds exceptions } tupColl = itorColl.next(); @@ -424,7 +424,7 @@ try { row.get(3); Assert.fail("Failed to catch 'out of boundary' exceptions."); - } catch (ExecException e) { + } catch (IndexOutOfBoundsException e) { // no op, expecting out of bounds exceptions } @@ -474,7 +474,7 @@ try { Object obj = tupColl.get(1); Assert.fail("Failed to catch 'out of boundary' exceptions."); - } catch (ExecException e) { + } catch (IndexOutOfBoundsException e) { // no op, expecting out of bounds exceptions } tupColl = itorColl.next(); @@ -482,7 +482,7 @@ try { Object obj = tupColl.get(1); Assert.fail("Failed to catch 'out of boundary' exceptions."); - } catch (ExecException e) { + } catch (IndexOutOfBoundsException e) { // no op, expecting out of bounds exceptions } Assert.assertFalse(itorColl.hasNext()); @@ -492,7 +492,7 @@ try { Object obj = row.get(4); Assert.fail("Failed to catch 'out of boundary' exceptions."); - } catch (ExecException e) { + } catch (IndexOutOfBoundsException e) { // no op, expecting out of bounds exceptions } @@ -545,7 +545,7 @@ try { row.get(3); Assert.fail("Failed to catch 'out of boundary' exceptions."); - } catch (ExecException e) { + } catch (IndexOutOfBoundsException e) { // no op, expecting out of bounds exceptions } Modified: hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/pig/TestCollection.java URL: http://svn.apache.org/viewvc/hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/pig/TestCollection.java?rev=823411&r1=823410&r2=823411&view=diff ============================================================================== --- hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/pig/TestCollection.java (original) +++ hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/pig/TestCollection.java Fri Oct 9 05:04:10 2009 @@ -210,7 +210,7 @@ @Test public void testRead1() throws IOException, ParseException { String query = "records = LOAD '" + path.toString() - + "' USING org.apache.pig.table.pig.TableLoader('c');"; + + "' USING org.apache.hadoop.zebra.pig.TableLoader('c');"; System.out.println(query); pigServer.registerQuery(query); Iterator<Tuple> it = pigServer.openIterator("records"); Modified: hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorageMap.java URL: http://svn.apache.org/viewvc/hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorageMap.java?rev=823411&r1=823410&r2=823411&view=diff ============================================================================== --- hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorageMap.java (original) +++ hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorageMap.java Fri Oct 9 05:04:10 2009 @@ -262,7 +262,7 @@ System.out.println(cgs1); } catch (Exception e) { String errMsg = e.getMessage(); - String str = "Encountered \" \"[\" \"[ \"\" at line 1, column 1."; + String str = "Encountered \" \"{\" \"{ \"\" at line 1, column 4."; System.out.println(errMsg); System.out.println(str); Assert.assertEquals(errMsg.startsWith(str), true); Modified: hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorageMisc1.java URL: http://svn.apache.org/viewvc/hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorageMisc1.java?rev=823411&r1=823410&r2=823411&view=diff ============================================================================== --- hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorageMisc1.java (original) +++ hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorageMisc1.java Fri Oct 9 05:04:10 2009 @@ -81,8 +81,15 @@ @Test public void testStorageValid1() { try { - String strStorage = "[r.r.f1,r.f2#{k1}] COMPRESS BY gzip ; [r.r.f2, r.f2#{k2}] COMPRESS BY lzo2 SERIALIZE BY avro"; - Partition p = new Partition(schema.toString(), strStorage); + String strStorage = "[r.r.f1,r.f2#{k1}] COMPRESS BY gzip SECURE BY user:gaurav; [r.r.f2, r.f2#{k2}] COMPRESS BY lzo SERIALIZE BY avro"; +// String strStorage = "[r.r.f1,r.f2#{k1}] COMPRESS BY gzip SECURE BY user:gaurav; [r.r.f2, r.f2#{k2}] COMPRESS BY lzo SERIALIZE BY avro"; +// String strStorage = "[r.r.f1,r.f2#{k1}] COMPRESS BY gzip SECURE BY user:gaurav group:data perm:0766; [r.r.f2, r.f2#{k2}] COMPRESS BY lzo SERIALIZE BY avro"; +// String strStorage = "[r.r.f1,r.f2#{k1}] COMPRESS BY gzip SECURE BY user:gaurav group:data perm:966; [r.r.f2, r.f2#{k2}] COMPRESS BY lzo SERIALIZE BY avro"; +// String strStorage = "[r.r.f1,r.f2#{k1}] COMPRESS BY gzip SECURE BY; [r.r.f2, r.f2#{k2}] COMPRESS BY lzo SERIALIZE BY avro"; +// String strStorage = "[r.r.f1,r.f2#{k1}] COMPRESS BY gzip SECURE BY user:ggg SECURE BY group:fff; [r.r.f2, r.f2#{k2}] COMPRESS BY lzo SERIALIZE BY avro"; +// String strStorage = "[r.r.f1,r.f2#{k1}] COMPRESS BY gzip SECURE BY user:gauravj user:gauravj; [r.r.f2, r.f2#{k2}] COMPRESS BY lzo SERIALIZE BY avro"; + + Partition p = new Partition(schema.toString(), strStorage); CGSchema[] cgschemas = p.getCGSchemas(); // 3 column group; @@ -157,4 +164,4 @@ Assert.assertTrue(false); } } -} \ No newline at end of file +} Modified: hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorageMisc3.java URL: http://svn.apache.org/viewvc/hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorageMisc3.java?rev=823411&r1=823410&r2=823411&view=diff ============================================================================== --- hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorageMisc3.java (original) +++ hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorageMisc3.java Fri Oct 9 05:04:10 2009 @@ -207,9 +207,9 @@ Assert.assertEquals(cgs1.getCompressor(), "gzip"); Assert.assertEquals(cgs1.getSerializer(), "pig"); - Assert.assertEquals(cgs2.getCompressor(), "lzo2"); + Assert.assertEquals(cgs2.getCompressor(), "lzo"); Assert.assertEquals(cgs2.getSerializer(), "avro"); - Assert.assertEquals(cgs3.getCompressor(), "lzo2"); + Assert.assertEquals(cgs3.getCompressor(), "lzo"); Assert.assertEquals(cgs3.getSerializer(), "pig"); System.out.println("*********** Column Map **********"); @@ -249,4 +249,4 @@ } catch (Exception e) { } } -} \ No newline at end of file +} Modified: hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorePrimitive.java URL: http://svn.apache.org/viewvc/hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorePrimitive.java?rev=823411&r1=823410&r2=823411&view=diff ============================================================================== --- hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorePrimitive.java (original) +++ hadoop/pig/trunk/contrib/zebra/src/test/org/apache/hadoop/zebra/types/TestStorePrimitive.java Fri Oct 9 05:04:10 2009 @@ -107,11 +107,11 @@ Assert.assertEquals("f6", f32.name); Assert.assertEquals(ColumnType.BYTES, f32.type); - Assert.assertEquals(cgs1.getCompressor(), "lzo2"); + Assert.assertEquals(cgs1.getCompressor(), "lzo"); Assert.assertEquals(cgs1.getSerializer(), "pig"); Assert.assertEquals(cgs2.getCompressor(), "gzip"); Assert.assertEquals(cgs2.getSerializer(), "avro"); - Assert.assertEquals(cgs3.getCompressor(), "lzo2"); + Assert.assertEquals(cgs3.getCompressor(), "lzo"); Assert.assertEquals(cgs3.getSerializer(), "pig"); System.out.println("*********** Column Map **********"); @@ -204,10 +204,10 @@ System.out.println(cgs1); } catch (Exception e) { String errMsg = e.getMessage(); - String str = "Encountered \" <IDENTIFIER> \"f1 \"\" at line 1, column 1.\nWas expecting one of:\n <EOF> \n \"compress by\" ...\n \"serialize by\" ...\n \"[\" ...\n "; + String str = "Encountered \" <IDENTIFIER> \"f1 \"\" at line 1, column 1.\n"; System.out.println(errMsg); System.out.println(str); - Assert.assertEquals(errMsg, str); + Assert.assertEquals(errMsg.startsWith(str), true); } } @@ -238,10 +238,10 @@ System.out.println(cgs1); } catch (Exception e) { String errMsg = e.getMessage(); - String str = "Encountered \" <IDENTIFIER> \"f1 \"\" at line 1, column 1.\nWas expecting one of:\n <EOF> \n \"compress by\" ...\n \"serialize by\" ...\n \"[\" ...\n "; + String str = "Encountered \" <IDENTIFIER> \"f1 \"\" at line 1, column 1.\nWas expecting one of:\n"; System.out.println(errMsg); System.out.println(str); - Assert.assertEquals(errMsg, str); + Assert.assertEquals(errMsg.startsWith(str), true); } } @@ -255,10 +255,10 @@ System.out.println(cgs1); } catch (Exception e) { String errMsg = e.getMessage(); - String str = "Encountered \" \"[\" \"[ \"\" at line 1, column 1.\nWas expecting one of:\n \"compress by\" ...\n \"serialize by\" ...\n "; + String str = "Encountered \"<EOF>\" at line 1, column 1.\nWas expecting one of:\n"; System.out.println(errMsg); System.out.println(str); - Assert.assertEquals(errMsg, str); + Assert.assertEquals(errMsg.startsWith(str), true); } } @@ -289,24 +289,24 @@ System.out.println(cgs1); } catch (Exception e) { String errMsg = e.getMessage(); - String str = "Lexical error at line 1, column 2. Encountered: <EOF> after : \"\""; + String str = "Encountered \" \":\" \": \"\" at line 1, column 1.\n"; System.out.println(errMsg); System.out.println(str); - Assert.assertEquals(errMsg, str); + Assert.assertEquals(errMsg.startsWith(str), true); } } @Test public void testStorageInvalid7() { try { - String strStorage = "[f1, f2] serialize by xyz compress by gzip; [f3, f4] SERIALIZE BY avro COMPRESS BY lzo2"; + String strStorage = "[f1, f2] serialize by xyz compress by gzip; [f3, f4] SERIALIZE BY avro COMPRESS BY lzo"; Partition p = new Partition(schema.toString(), strStorage); CGSchema[] cgschemas = p.getCGSchemas(); CGSchema cgs1 = cgschemas[0]; System.out.println(cgs1); } catch (Exception e) { String errMsg = e.getMessage(); - String str = "Encountered \" \"serialize by\" \"serialize by \"\" at line 1, column 10."; + String str = "Encountered \" <IDENTIFIER> \"xyz \"\" at line 1, column 23."; System.out.println(errMsg); System.out.println(str); Assert.assertEquals(errMsg.startsWith(str), true); @@ -316,17 +316,17 @@ @Test public void testStorageInvalid8() { try { - String strStorage = "[f1, f2] serialize by avro compress by xyz; [f3, f4] SERIALIZE BY avro COMPRESS BY lzo2"; + String strStorage = "[f1, f2] serialize by avro compress by xyz; [f3, f4] SERIALIZE BY avro COMPRESS BY lzo"; Partition p = new Partition(schema.toString(), strStorage); CGSchema[] cgschemas = p.getCGSchemas(); CGSchema cgs1 = cgschemas[0]; System.out.println(cgs1); } catch (Exception e) { String errMsg = e.getMessage(); - String str = "Encountered \" \"compress by\" \"compress by \"\" at line 1, column 28."; + String str = "Encountered \" <IDENTIFIER> \"xyz \"\" at line 1, column 40."; System.out.println(errMsg); System.out.println(str); Assert.assertEquals(errMsg.startsWith(str), true); } } -} \ No newline at end of file +}
