See <https://builds.apache.org/job/Tajo-master-nightly/803/changes>

Changes:

[jhkim] TAJO-1777: JsonLineDeserializer returns invalid unicode text, if 
contains control character

------------------------------------------
[...truncated 1896 lines...]
        at 
org.apache.hadoop.fs.RawLocalFileSystem$LocalFSFileOutputStream.<init>(RawLocalFileSystem.java:225)
        at 
org.apache.hadoop.fs.RawLocalFileSystem$LocalFSFileOutputStream.<init>(RawLocalFileSystem.java:209)
        at 
org.apache.hadoop.fs.RawLocalFileSystem.createOutputStreamWithMode(RawLocalFileSystem.java:305)
        at 
org.apache.hadoop.fs.RawLocalFileSystem.create(RawLocalFileSystem.java:294)
        at 
org.apache.hadoop.fs.RawLocalFileSystem.create(RawLocalFileSystem.java:326)
        at 
org.apache.hadoop.fs.ChecksumFileSystem$ChecksumFSOutputSummer.<init>(ChecksumFileSystem.java:393)
        at 
org.apache.hadoop.fs.ChecksumFileSystem.create(ChecksumFileSystem.java:456)
        at 
org.apache.hadoop.fs.ChecksumFileSystem.create(ChecksumFileSystem.java:435)
        at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:909)
        at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:890)
        at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:787)
        at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:776)
        at 
org.apache.tajo.storage.text.DelimitedTextFile$DelimitedTextFileAppender.init(DelimitedTextFile.java:161)
        at 
org.apache.tajo.storage.TestFileTablespace.testGetScannerAndAppender(TestFileTablespace.java:88)

testGetFileTablespace(org.apache.tajo.storage.TestFileTablespace)  Time 
elapsed: 0.17 sec  <<< ERROR!
java.lang.OutOfMemoryError: unable to create new native thread
        at java.lang.Thread.start0(Native Method)
        at java.lang.Thread.start(Thread.java:693)
        at 
org.apache.hadoop.hdfs.server.namenode.FSImage.saveFSImageInAllDirs(FSImage.java:1169)
        at 
org.apache.hadoop.hdfs.server.namenode.FSImage.saveFSImageInAllDirs(FSImage.java:1133)
        at 
org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:163)
        at 
org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:991)
        at 
org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:342)
        at 
org.apache.hadoop.hdfs.DFSTestUtil.formatNameNode(DFSTestUtil.java:176)
        at 
org.apache.hadoop.hdfs.MiniDFSCluster.createNameNodesAndSetConf(MiniDFSCluster.java:973)
        at 
org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:811)
        at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:471)
        at 
org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:430)
        at 
org.apache.tajo.storage.TestFileTablespace.testGetFileTablespace(TestFileTablespace.java:219)

Running org.apache.tajo.storage.TestDelimitedTextFile
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.119 sec - in 
org.apache.tajo.storage.TestDelimitedTextFile
Running org.apache.tajo.storage.TestByteBufLineReader
Formatting using clusterid: testClusterID
Tests run: 3, Failures: 0, Errors: 3, Skipped: 0, Time elapsed: 0.245 sec <<< 
FAILURE! - in org.apache.tajo.storage.TestByteBufLineReader
testReaderWithLocalFS(org.apache.tajo.storage.TestByteBufLineReader)  Time 
elapsed: 0.066 sec  <<< ERROR!
java.io.IOException: Cannot run program "chmod": error=11, Resource temporarily 
unavailable
        at java.lang.UNIXProcess.forkAndExec(Native Method)
        at java.lang.UNIXProcess.<init>(UNIXProcess.java:135)
        at java.lang.ProcessImpl.start(ProcessImpl.java:130)
        at java.lang.ProcessBuilder.start(ProcessBuilder.java:1022)
        at org.apache.hadoop.util.Shell.runCommand(Shell.java:486)
        at org.apache.hadoop.util.Shell.run(Shell.java:456)
        at 
org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:722)
        at org.apache.hadoop.util.Shell.execCommand(Shell.java:815)
        at org.apache.hadoop.util.Shell.execCommand(Shell.java:798)
        at 
org.apache.hadoop.fs.RawLocalFileSystem.setPermission(RawLocalFileSystem.java:728)
        at 
org.apache.hadoop.fs.RawLocalFileSystem$LocalFSFileOutputStream.<init>(RawLocalFileSystem.java:225)
        at 
org.apache.hadoop.fs.RawLocalFileSystem$LocalFSFileOutputStream.<init>(RawLocalFileSystem.java:209)
        at 
org.apache.hadoop.fs.RawLocalFileSystem.createOutputStreamWithMode(RawLocalFileSystem.java:305)
        at 
org.apache.hadoop.fs.RawLocalFileSystem.create(RawLocalFileSystem.java:294)
        at 
org.apache.hadoop.fs.RawLocalFileSystem.create(RawLocalFileSystem.java:326)
        at 
org.apache.hadoop.fs.ChecksumFileSystem$ChecksumFSOutputSummer.<init>(ChecksumFileSystem.java:393)
        at 
org.apache.hadoop.fs.ChecksumFileSystem.create(ChecksumFileSystem.java:456)
        at 
org.apache.hadoop.fs.ChecksumFileSystem.create(ChecksumFileSystem.java:435)
        at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:909)
        at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:890)
        at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:787)
        at 
org.apache.tajo.storage.TestByteBufLineReader.testReaderWithLocalFS(TestByteBufLineReader.java:68)

testReaderWithDFS(org.apache.tajo.storage.TestByteBufLineReader)  Time elapsed: 
0.111 sec  <<< ERROR!
java.lang.OutOfMemoryError: unable to create new native thread
        at java.lang.Thread.start0(Native Method)
        at java.lang.Thread.start(Thread.java:693)
        at 
org.apache.hadoop.hdfs.server.namenode.FSImage.saveFSImageInAllDirs(FSImage.java:1169)
        at 
org.apache.hadoop.hdfs.server.namenode.FSImage.saveFSImageInAllDirs(FSImage.java:1133)
        at 
org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:163)
        at 
org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:991)
        at 
org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:342)
        at 
org.apache.hadoop.hdfs.DFSTestUtil.formatNameNode(DFSTestUtil.java:176)
        at 
org.apache.hadoop.hdfs.MiniDFSCluster.createNameNodesAndSetConf(MiniDFSCluster.java:973)
        at 
org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:811)
        at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:471)
        at 
org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:430)
        at 
org.apache.tajo.storage.TestByteBufLineReader.testReaderWithDFS(TestByteBufLineReader.java:96)

testReaderWithNIO(org.apache.tajo.storage.TestByteBufLineReader)  Time elapsed: 
0.066 sec  <<< ERROR!
java.io.IOException: Cannot run program "chmod": error=11, Resource temporarily 
unavailable
        at java.lang.UNIXProcess.forkAndExec(Native Method)
        at java.lang.UNIXProcess.<init>(UNIXProcess.java:135)
        at java.lang.ProcessImpl.start(ProcessImpl.java:130)
        at java.lang.ProcessBuilder.start(ProcessBuilder.java:1022)
        at org.apache.hadoop.util.Shell.runCommand(Shell.java:486)
        at org.apache.hadoop.util.Shell.run(Shell.java:456)
        at 
org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:722)
        at org.apache.hadoop.util.Shell.execCommand(Shell.java:815)
        at org.apache.hadoop.util.Shell.execCommand(Shell.java:798)
        at 
org.apache.hadoop.fs.RawLocalFileSystem.setPermission(RawLocalFileSystem.java:728)
        at 
org.apache.hadoop.fs.RawLocalFileSystem$LocalFSFileOutputStream.<init>(RawLocalFileSystem.java:225)
        at 
org.apache.hadoop.fs.RawLocalFileSystem$LocalFSFileOutputStream.<init>(RawLocalFileSystem.java:209)
        at 
org.apache.hadoop.fs.RawLocalFileSystem.createOutputStreamWithMode(RawLocalFileSystem.java:305)
        at 
org.apache.hadoop.fs.RawLocalFileSystem.create(RawLocalFileSystem.java:294)
        at 
org.apache.hadoop.fs.RawLocalFileSystem.create(RawLocalFileSystem.java:326)
        at 
org.apache.hadoop.fs.ChecksumFileSystem$ChecksumFSOutputSummer.<init>(ChecksumFileSystem.java:393)
        at 
org.apache.hadoop.fs.ChecksumFileSystem.create(ChecksumFileSystem.java:456)
        at 
org.apache.hadoop.fs.ChecksumFileSystem.create(ChecksumFileSystem.java:435)
        at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:909)
        at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:890)
        at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:787)
        at 
org.apache.tajo.storage.TestByteBufLineReader.testReaderWithNIO(TestByteBufLineReader.java:137)

Running org.apache.tajo.storage.TestLineReader
Tests run: 6, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.436 sec <<< 
FAILURE! - in org.apache.tajo.storage.TestLineReader
testByteBufLineReader(org.apache.tajo.storage.TestLineReader)  Time elapsed: 
0.068 sec  <<< ERROR!
java.lang.OutOfMemoryError: unable to create new native thread
        at java.lang.Thread.start0(Native Method)
        at java.lang.Thread.start(Thread.java:693)
        at org.apache.hadoop.util.Shell.runCommand(Shell.java:522)
        at org.apache.hadoop.util.Shell.run(Shell.java:456)
        at 
org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:722)
        at org.apache.hadoop.util.Shell.execCommand(Shell.java:815)
        at org.apache.hadoop.util.Shell.execCommand(Shell.java:798)
        at 
org.apache.hadoop.fs.RawLocalFileSystem.setPermission(RawLocalFileSystem.java:728)
        at 
org.apache.hadoop.fs.RawLocalFileSystem$LocalFSFileOutputStream.<init>(RawLocalFileSystem.java:225)
        at 
org.apache.hadoop.fs.RawLocalFileSystem$LocalFSFileOutputStream.<init>(RawLocalFileSystem.java:209)
        at 
org.apache.hadoop.fs.RawLocalFileSystem.createOutputStreamWithMode(RawLocalFileSystem.java:305)
        at 
org.apache.hadoop.fs.RawLocalFileSystem.create(RawLocalFileSystem.java:294)
        at 
org.apache.hadoop.fs.RawLocalFileSystem.create(RawLocalFileSystem.java:326)
        at 
org.apache.hadoop.fs.ChecksumFileSystem$ChecksumFSOutputSummer.<init>(ChecksumFileSystem.java:393)
        at 
org.apache.hadoop.fs.ChecksumFileSystem.create(ChecksumFileSystem.java:456)
        at 
org.apache.hadoop.fs.ChecksumFileSystem.create(ChecksumFileSystem.java:435)
        at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:909)
        at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:890)
        at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:787)
        at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:776)
        at 
org.apache.tajo.storage.text.DelimitedTextFile$DelimitedTextFileAppender.init(DelimitedTextFile.java:161)
        at 
org.apache.tajo.storage.TestLineReader.testByteBufLineReader(TestLineReader.java:71)

rquet.hadoop.InternalParquetRecordReader: block read in memory in 1 ms. row 
count = 1
Aug 19, 2015 3:30:03 AM INFO: 
org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem columnStore 
to file. allocated memory: 280,000
Aug 19, 2015 3:30:03 AM INFO: 
org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 40,047B for [id] 
INT32: 10,000 values, 40,008B raw, 40,008B comp, 1 pages, encodings: [PLAIN, 
BIT_PACKED, RLE]
Aug 19, 2015 3:30:03 AM INFO: 
org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 80,055B for [age] 
INT64: 10,000 values, 80,008B raw, 80,008B comp, 1 pages, encodings: [PLAIN, 
BIT_PACKED, RLE]
Aug 19, 2015 3:30:03 AM INFO: 
org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 40,047B for 
[score] FLOAT: 10,000 values, 40,008B raw, 40,008B comp, 1 pages, encodings: 
[PLAIN, BIT_PACKED, RLE]
Aug 19, 2015 3:30:03 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: 
Initiating action with parallelism: 5
Aug 19, 2015 3:30:03 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: 
reading another 1 footers
Aug 19, 2015 3:30:03 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: 
Initiating action with parallelism: 5
Aug 19, 2015 3:30:03 AM INFO: 
org.apache.parquet.hadoop.InternalParquetRecordReader: RecordReader initialized 
will read a total of 10000 records.
Aug 19, 2015 3:30:03 AM INFO: 
org.apache.parquet.hadoop.InternalParquetRecordReader: at row 0. reading next 
block
Aug 19, 2015 3:30:03 AM INFO: 
org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 
1 ms. row count = 10000
Aug 19, 2015 3:30:03 AM INFO: 
org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem columnStore 
to file. allocated memory: 66,794
Aug 19, 2015 3:30:03 AM INFO: 
org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 36B for [col1] 
BOOLEAN: 12 values, 9B raw, 9B comp, 1 pages, encodings: [PLAIN, BIT_PACKED, 
RLE]
Aug 19, 2015 3:30:03 AM INFO: 
org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 48B for [col2] 
BINARY: 12 values, 9B raw, 9B comp, 1 pages, encodings: [PLAIN_DICTIONARY, 
BIT_PACKED, RLE], dic { 1 entries, 11B raw, 1B comp}
Aug 19, 2015 3:30:03 AM INFO: 
org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 42B for [col3] 
INT32: 12 values, 9B raw, 9B comp, 1 pages, encodings: [PLAIN_DICTIONARY, 
BIT_PACKED, RLE], dic { 1 entries, 4B raw, 1B comp}
Aug 19, 2015 3:30:03 AM INFO: 
org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 42B for [col4] 
INT32: 12 values, 9B raw, 9B comp, 1 pages, encodings: [PLAIN_DICTIONARY, 
BIT_PACKED, RLE], dic { 1 entries, 4B raw, 1B comp}
Aug 19, 2015 3:30:03 AM INFO: 
org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 50B for [col5] 
INT64: 12 values, 9B raw, 9B comp, 1 pages, encodings: [PLAIN_DICTIONARY, 
BIT_PACKED, RLE], dic { 1 entries, 8B raw, 1B comp}
Aug 19, 2015 3:30:03 AM INFO: 
org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 42B for [col6] 
FLOAT: 12 values, 9B raw, 9B comp, 1 pages, encodings: [PLAIN_DICTIONARY, 
BIT_PACKED, RLE], dic { 1 entries, 4B raw, 1B comp}
Aug 19, 2015 3:30:03 AM INFO: 
org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 50B for [col7] 
DOUBLE: 12 values, 9B raw, 9B comp, 1 pages, encodings: [PLAIN_DICTIONARY, 
BIT_PACKED, RLE], dic { 1 entries, 8B raw, 1B comp}
Aug 19, 2015 3:30:03 AM INFO: 
org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 48B for [col8] 
BINARY: 12 values, 9B raw, 9B comp, 1 pages, encodings: [PLAIN_DICTIONARY, 
BIT_PACKED, RLE], dic { 1 entries, 11B raw, 1B comp}
Aug 19, 2015 3:30:03 AM INFO: 
org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 49B for [col9] 
BINARY: 12 values, 10B raw, 10B comp, 1 pages, encodings: [PLAIN_DICTIONARY, 
BIT_PACKED, RLE], dic { 1 entries, 11B raw, 1B comp}
Aug 19, 2015 3:30:03 AM INFO: 
org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 43B for [col10] 
BINARY: 12 values, 10B raw, 10B comp, 1 pages, encodings: [PLAIN_DICTIONARY, 
BIT_PACKED, RLE], dic { 1 entries, 8B raw, 1B comp}
Aug 19, 2015 3:30:03 AM INFO: 
org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 53B for [col12] 
BINARY: 12 values, 10B raw, 10B comp, 1 pages, encodings: [PLAIN_DICTIONARY, 
BIT_PACKED, RLE], dic { 1 entries, 13B raw, 1B comp}
Aug 19, 2015 3:30:04 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: 
Initiating action with parallelism: 5
Aug 19, 2015 3:30:04 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: 
reading another 1 footers
Aug 19, 2015 3:30:04 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: 
Initiating action with parallelism: 5
Aug 19, 2015 3:30:04 AM INFO: 
org.apache.parquet.hadoop.InternalParquetRecordReader: RecordReader initialized 
will read a total of 12 records.
Aug 19, 2015 3:30:04 AM INFO: 
org.apache.parquet.hadoop.InternalParquetRecordReader: at row 0. reading next 
block
Aug 19, 2015 3:30:04 AM INFO: 
org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 
1 ms. row count = 12
Aug 19, 2015 3:30:13 AM INFO: 
org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem columnStore 
to file. allocated memory: 65,659
Aug 19, 2015 3:30:13 AM INFO: 
org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 34B for 
[myboolean] BOOLEAN: 1 values, 7B raw, 7B comp, 1 pages, encodings: [PLAIN, 
BIT_PACKED, RLE]
Aug 19, 2015 3:30:13 AM INFO: 
org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 43B for [mybit] 
INT32: 1 values, 10B raw, 10B comp, 1 pages, encodings: [PLAIN, BIT_PACKED, RLE]
Aug 19, 2015 3:30:13 AM INFO: 
org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 38B for [mychar] 
BINARY: 1 values, 11B raw, 11B comp, 1 pages, encodings: [PLAIN, BIT_PACKED, 
RLE]
Aug 19, 2015 3:30:13 AM INFO: 
org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 43B for [myint2] 
INT32: 1 values, 10B raw, 10B comp, 1 pages, encodings: [PLAIN, BIT_PACKED, RLE]
Aug 19, 2015 3:30:13 AM INFO: 
org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 43B for [myint4] 
INT32: 1 values, 10B raw, 10B comp, 1 pages, encodings: [PLAIN, BIT_PACKED, RLE]
Aug 19, 2015 3:30:13 AM INFO: 
org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 55B for [myint8] 
INT64: 1 values, 14B raw, 14B comp, 1 pages, encodings: [PLAIN, BIT_PACKED, RLE]
Aug 19, 2015 3:30:13 AM INFO: 
org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 43B for [myfloat4] 
FLOAT: 1 values, 10B raw, 10B comp, 1 pages, encodings: [PLAIN, BIT_PACKED, RLE]
Aug 19, 2015 3:30:13 AM INFO: 
org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 55B for [myfloat8] 
DOUBLE: 1 values, 14B raw, 14B comp, 1 pages, encodings: [PLAIN, BIT_PACKED, 
RLE]
Aug 19, 2015 3:30:13 AM INFO: 
org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 50B for [mytext] 
BINARY: 1 values, 15B raw, 15B comp, 1 pages, encodings: [PLAIN, BIT_PACKED, 
RLE]
Aug 19, 2015 3:30:13 AM INFO: 
org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 50B for [myblob] 
BINARY: 1 values, 15B raw, 15B comp, 1 pages, encodings: [PLAIN, BIT_PACKED, 
RLE]
Aug 19, 2015 3:30:13 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: 
Initiating action with parallelism: 5
Aug 19, 2015 3:30:13 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: 
reading another 1 footers
Aug 19, 2015 3:30:13 AM INFO: org.apache.parquet.hadoop.ParquetFileReader: 
Initiating action with parallelism: 5
Aug 19, 2015 3:30:13 AM INFO: 
org.apache.parquet.hadoop.InternalParquetRecordReader: RecordReader initialized 
will read a total of 1 records.
Aug 19, 2015 3:30:13 AM INFO: 
org.apache.parquet.hadoop.InternalParquetRecordReader: at row 0. reading next 
block
Aug 19, 2015 3:30:13 AM INFO: 
org.apache.parquet.hadoop.InternalParquetRecordReader: block read in memory in 
1 ms. row count = 1

Results :

Tests in error: 
  TestStorages.testMaxValue:923 » IO Cannot run program "chmod": error=11, 
Resou...
  TestStorages.testNullHandlingTypes:429 » OutOfMemory unable to create new 
nati...
  TestFileSystems.testBlockSplit:120 » IO Cannot run program "chmod": error=11, 
...
  TestAvroUtil.testGetSchema:80 » OutOfMemory unable to create new native thread
  TestFileTablespace.testGetSplitWithBlockStorageLocationsBatching:207 » 
OutOfMemory
  TestFileTablespace.testGetScannerAndAppender:88 » IO Cannot run program 
"chmod...
  TestFileTablespace.testGetFileTablespace:219 » OutOfMemory unable to create 
ne...
  TestByteBufLineReader.testReaderWithLocalFS:68 » IO Cannot run program 
"chmod"...
  TestByteBufLineReader.testReaderWithDFS:96 » OutOfMemory unable to create new 
...
  TestByteBufLineReader.testReaderWithNIO:137 » IO Cannot run program "chmod": 
e...
  TestLineReader.testByteBufLineReader:71 » OutOfMemory unable to create new 
nat...

Tests run: 178, Failures: 0, Errors: 11, Skipped: 0

[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO] 
[INFO] Tajo Main ......................................... SUCCESS [  2.133 s]
[INFO] Tajo Project POM .................................. SUCCESS [  2.797 s]
[INFO] Tajo Maven Plugins ................................ SUCCESS [  3.660 s]
[INFO] Tajo Common ....................................... SUCCESS [ 34.828 s]
[INFO] Tajo Algebra ...................................... SUCCESS [  3.505 s]
[INFO] Tajo Catalog Common ............................... SUCCESS [  5.638 s]
[INFO] Tajo Plan ......................................... SUCCESS [  8.209 s]
[INFO] Tajo Rpc Common ................................... SUCCESS [  1.591 s]
[INFO] Tajo Protocol Buffer Rpc .......................... SUCCESS [01:30 min]
[INFO] Tajo Catalog Client ............................... SUCCESS [  1.531 s]
[INFO] Tajo Catalog Server ............................... SUCCESS [ 12.691 s]
[INFO] Tajo Storage Common ............................... SUCCESS [ 12.845 s]
[INFO] Tajo HDFS Storage ................................. FAILURE [ 59.949 s]
[INFO] Tajo PullServer ................................... SKIPPED
[INFO] Tajo Client ....................................... SKIPPED
[INFO] Tajo CLI tools .................................... SKIPPED
[INFO] Tajo JDBC Driver .................................. SKIPPED
[INFO] ASM (thirdparty) .................................. SKIPPED
[INFO] Tajo RESTful Container ............................ SKIPPED
[INFO] Tajo Metrics ...................................... SKIPPED
[INFO] Tajo Core ......................................... SKIPPED
[INFO] Tajo RPC .......................................... SKIPPED
[INFO] Tajo Catalog Drivers Hive ......................... SKIPPED
[INFO] Tajo Catalog Drivers .............................. SKIPPED
[INFO] Tajo Catalog ...................................... SKIPPED
[INFO] Tajo HBase Storage ................................ SKIPPED
[INFO] Tajo Storage ...................................... SKIPPED
[INFO] Tajo Distribution ................................. SKIPPED
[INFO] Tajo Cluster Tests ................................ SKIPPED
[INFO] Tajo Core Tests ................................... SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 04:00 min
[INFO] Finished at: 2015-08-19T03:30:18+00:00
[INFO] Final Memory: 75M/406M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal 
org.apache.maven.plugins:maven-surefire-plugin:2.17:test (default-test) on 
project tajo-storage-hdfs: There are test failures.
[ERROR] 
[ERROR] Please refer to 
<https://builds.apache.org/job/Tajo-master-nightly/ws/tajo-storage/tajo-storage-hdfs/target/surefire-reports>
 for the individual test results.
[ERROR] -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e 
switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please 
read the following articles:
[ERROR] [Help 1] 
http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR] 
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR]   mvn <goals> -rf :tajo-storage-hdfs
Build step 'Execute shell' marked build as failure
Archiving artifacts
Recording test results

Reply via email to