This is an automated email from the ASF dual-hosted git repository.
vbalaji pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/incubator-hudi.git
The following commit(s) were added to refs/heads/master by this push:
new d0d2fa0 Reduce logging in unit-test runs
d0d2fa0 is described below
commit d0d2fa03379525e8156d5dc0adae9eb852603899
Author: Balaji Varadarajan <[email protected]>
AuthorDate: Fri May 24 22:20:10 2019 -0700
Reduce logging in unit-test runs
---
.../java/com/uber/hoodie/TestCompactionAdminClient.java | 2 +-
.../java/com/uber/hoodie/common/HoodieClientTestUtils.java | 5 ++++-
.../java/com/uber/hoodie/func/TestUpdateMapFunction.java | 1 -
.../java/com/uber/hoodie/table/TestCopyOnWriteTable.java | 6 +++++-
.../main/java/com/uber/hoodie/common/util/RocksDBDAO.java | 2 +-
.../common/table/view/HoodieTableFileSystemViewTest.java | 6 +++++-
.../common/table/view/IncrementalFSViewSyncTest.java | 14 +++++++++-----
.../com/uber/hoodie/common/util/TestRocksDBManager.java | 5 -----
8 files changed, 25 insertions(+), 16 deletions(-)
diff --git
a/hoodie-client/src/test/java/com/uber/hoodie/TestCompactionAdminClient.java
b/hoodie-client/src/test/java/com/uber/hoodie/TestCompactionAdminClient.java
index 0ed1f7c..220b2fb 100644
--- a/hoodie-client/src/test/java/com/uber/hoodie/TestCompactionAdminClient.java
+++ b/hoodie-client/src/test/java/com/uber/hoodie/TestCompactionAdminClient.java
@@ -158,7 +158,7 @@ public class TestCompactionAdminClient extends
TestHoodieClientBase {
Assert.assertTrue("Rename Files must be empty", renameFiles.isEmpty());
}
expRenameFiles.entrySet().stream().forEach(r -> {
- System.out.println("Key :" + r.getKey() + " renamed to " + r.getValue()
+ " rolled back to "
+ logger.info("Key :" + r.getKey() + " renamed to " + r.getValue() + "
rolled back to "
+ renameFilesFromUndo.get(r.getKey()));
});
diff --git
a/hoodie-client/src/test/java/com/uber/hoodie/common/HoodieClientTestUtils.java
b/hoodie-client/src/test/java/com/uber/hoodie/common/HoodieClientTestUtils.java
index e77653e..ebd05d3 100644
---
a/hoodie-client/src/test/java/com/uber/hoodie/common/HoodieClientTestUtils.java
+++
b/hoodie-client/src/test/java/com/uber/hoodie/common/HoodieClientTestUtils.java
@@ -50,6 +50,8 @@ import org.apache.avro.Schema;
import org.apache.avro.generic.GenericRecord;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
+import org.apache.log4j.LogManager;
+import org.apache.log4j.Logger;
import org.apache.parquet.avro.AvroSchemaConverter;
import org.apache.parquet.hadoop.ParquetWriter;
import org.apache.parquet.hadoop.metadata.CompressionCodecName;
@@ -64,6 +66,7 @@ import org.apache.spark.sql.SQLContext;
*/
public class HoodieClientTestUtils {
+ private static final transient Logger log =
LogManager.getLogger(HoodieClientTestUtils.class);
public static List<WriteStatus> collectStatuses(Iterator<List<WriteStatus>>
statusListItr) {
List<WriteStatus> statuses = new ArrayList<>();
@@ -137,7 +140,7 @@ public class HoodieClientTestUtils {
try {
HashMap<String, String> paths = getLatestFileIDsToFullPath(basePath,
commitTimeline,
Arrays.asList(commitInstant));
- System.out.println("Path :" + paths.values());
+ log.info("Path :" + paths.values());
return sqlContext.read().parquet(paths.values().toArray(new
String[paths.size()]))
.filter(String.format("%s ='%s'",
HoodieRecord.COMMIT_TIME_METADATA_FIELD, commitTime));
} catch (Exception e) {
diff --git
a/hoodie-client/src/test/java/com/uber/hoodie/func/TestUpdateMapFunction.java
b/hoodie-client/src/test/java/com/uber/hoodie/func/TestUpdateMapFunction.java
index d3c78a8..064021e 100644
---
a/hoodie-client/src/test/java/com/uber/hoodie/func/TestUpdateMapFunction.java
+++
b/hoodie-client/src/test/java/com/uber/hoodie/func/TestUpdateMapFunction.java
@@ -80,7 +80,6 @@ public class TestUpdateMapFunction implements Serializable {
public void testSchemaEvolutionOnUpdate() throws Exception {
// Create a bunch of records with a old version of schema
final HoodieWriteConfig config =
makeHoodieClientConfig("/exampleSchema.txt");
- System.out.println("JSC =" + jsc);
final HoodieCopyOnWriteTable table = new HoodieCopyOnWriteTable(config,
jsc);
final List<WriteStatus> statuses = jsc.parallelize(Arrays.asList(1)).map(x
-> {
diff --git
a/hoodie-client/src/test/java/com/uber/hoodie/table/TestCopyOnWriteTable.java
b/hoodie-client/src/test/java/com/uber/hoodie/table/TestCopyOnWriteTable.java
index 822ec93..76c12d8 100644
---
a/hoodie-client/src/test/java/com/uber/hoodie/table/TestCopyOnWriteTable.java
+++
b/hoodie-client/src/test/java/com/uber/hoodie/table/TestCopyOnWriteTable.java
@@ -50,6 +50,8 @@ import java.util.UUID;
import org.apache.avro.generic.GenericRecord;
import org.apache.commons.io.IOUtils;
import org.apache.hadoop.fs.Path;
+import org.apache.log4j.LogManager;
+import org.apache.log4j.Logger;
import org.apache.parquet.avro.AvroReadSupport;
import org.apache.parquet.hadoop.ParquetReader;
import org.apache.spark.TaskContext;
@@ -64,6 +66,8 @@ import scala.Tuple2;
public class TestCopyOnWriteTable {
+ protected static Logger log =
LogManager.getLogger(TestCopyOnWriteTable.class);
+
private String basePath = null;
private transient JavaSparkContext jsc = null;
@@ -378,7 +382,7 @@ public class TestCopyOnWriteTable {
int counts = 0;
for (File file : new File(basePath + "/2016/01/31").listFiles()) {
if (file.getName().endsWith(".parquet") &&
FSUtils.getCommitTime(file.getName()).equals(commitTime)) {
- System.out.println(file.getName() + "-" + file.length());
+ log.info(file.getName() + "-" + file.length());
counts++;
}
}
diff --git
a/hoodie-common/src/main/java/com/uber/hoodie/common/util/RocksDBDAO.java
b/hoodie-common/src/main/java/com/uber/hoodie/common/util/RocksDBDAO.java
index 29cd8d6..0322796 100644
--- a/hoodie-common/src/main/java/com/uber/hoodie/common/util/RocksDBDAO.java
+++ b/hoodie-common/src/main/java/com/uber/hoodie/common/util/RocksDBDAO.java
@@ -86,7 +86,7 @@ public class RocksDBDAO {
*/
private void init() throws HoodieException {
try {
- log.warn("DELETING RocksDB persisted at " + rocksDBBasePath);
+ log.info("DELETING RocksDB persisted at " + rocksDBBasePath);
FileUtils.deleteDirectory(new File(rocksDBBasePath));
managedHandlesMap = new ConcurrentHashMap<>();
diff --git
a/hoodie-common/src/test/java/com/uber/hoodie/common/table/view/HoodieTableFileSystemViewTest.java
b/hoodie-common/src/test/java/com/uber/hoodie/common/table/view/HoodieTableFileSystemViewTest.java
index cd26a2b..fbfa214 100644
---
a/hoodie-common/src/test/java/com/uber/hoodie/common/table/view/HoodieTableFileSystemViewTest.java
+++
b/hoodie-common/src/test/java/com/uber/hoodie/common/table/view/HoodieTableFileSystemViewTest.java
@@ -56,6 +56,8 @@ import java.util.stream.Collectors;
import java.util.stream.Stream;
import org.apache.hadoop.fs.FileStatus;
import org.apache.hadoop.fs.Path;
+import org.apache.log4j.LogManager;
+import org.apache.log4j.Logger;
import org.junit.Assert;
import org.junit.Before;
import org.junit.Rule;
@@ -65,6 +67,8 @@ import org.junit.rules.TemporaryFolder;
@SuppressWarnings("ResultOfMethodCallIgnored")
public class HoodieTableFileSystemViewTest {
+ private static final transient Logger log =
LogManager.getLogger(HoodieTableFileSystemViewTest.class);
+
private static String TEST_WRITE_TOKEN = "1-0-1";
protected HoodieTableMetaClient metaClient;
@@ -502,7 +506,7 @@ public class HoodieTableFileSystemViewTest {
roView.getAllDataFiles(partitionPath);
fileSliceList =
rtView.getLatestFileSlices(partitionPath).collect(Collectors.toList());
- System.out.println("FILESLICE LIST=" + fileSliceList);
+ log.info("FILESLICE LIST=" + fileSliceList);
dataFiles = fileSliceList.stream().map(FileSlice::getDataFile)
.filter(Optional::isPresent).map(Optional::get).collect(Collectors.toList());
assertEquals("Expect only one data-files in latest view as there is only
one file-group", 1, dataFiles.size());
diff --git
a/hoodie-common/src/test/java/com/uber/hoodie/common/table/view/IncrementalFSViewSyncTest.java
b/hoodie-common/src/test/java/com/uber/hoodie/common/table/view/IncrementalFSViewSyncTest.java
index a57ebad..7e73bce 100644
---
a/hoodie-common/src/test/java/com/uber/hoodie/common/table/view/IncrementalFSViewSyncTest.java
+++
b/hoodie-common/src/test/java/com/uber/hoodie/common/table/view/IncrementalFSViewSyncTest.java
@@ -61,6 +61,8 @@ import java.util.UUID;
import java.util.stream.Collectors;
import java.util.stream.IntStream;
import org.apache.hadoop.fs.Path;
+import org.apache.log4j.LogManager;
+import org.apache.log4j.Logger;
import org.junit.Assert;
import org.junit.Before;
import org.junit.Rule;
@@ -69,6 +71,8 @@ import org.junit.rules.TemporaryFolder;
public class IncrementalFSViewSyncTest {
+ private static final transient Logger log =
LogManager.getLogger(IncrementalFSViewSyncTest.class);
+
private static String TEST_WRITE_TOKEN = "1-0-1";
protected HoodieTableMetaClient metaClient;
@@ -344,7 +348,7 @@ public class IncrementalFSViewSyncTest {
Assert.assertEquals(newCleanerInstants.size(), cleanedInstants.size());
long initialFileSlices = partitions.stream().mapToLong(p ->
view.getAllFileSlices(p).count()).findAny().getAsLong();
long exp = initialFileSlices;
- System.out.println("Initial File Slices :" + exp);
+ log.info("Initial File Slices :" + exp);
for (int idx = 0; idx < newCleanerInstants.size(); idx++) {
String instant = cleanedInstants.get(idx);
try {
@@ -361,8 +365,8 @@ public class IncrementalFSViewSyncTest {
Assert.assertEquals(State.COMPLETED,
view.getLastInstant().get().getState());
Assert.assertEquals(HoodieTimeline.CLEAN_ACTION,
view.getLastInstant().get().getAction());
partitions.forEach(p -> {
- System.out.println("PARTTITION : " + p);
- System.out.println("\tFileSlices :" +
view.getAllFileSlices(p).collect(Collectors.toList()));
+ log.info("PARTTITION : " + p);
+ log.info("\tFileSlices :" +
view.getAllFileSlices(p).collect(Collectors.toList()));
});
partitions.forEach(p ->
Assert.assertEquals(fileIdsPerPartition.size(),
view.getLatestFileSlices(p).count()));
@@ -404,7 +408,7 @@ public class IncrementalFSViewSyncTest {
initialFileSlices - ((idx + 1) * fileIdsPerPartition.size());
view.sync();
Assert.assertTrue(view.getLastInstant().isPresent());
- System.out.println("Last Instant is :" + view.getLastInstant().get());
+ log.info("Last Instant is :" + view.getLastInstant().get());
if (isRestore) {
Assert.assertEquals(newRestoreInstants.get(idx),
view.getLastInstant().get().getTimestamp());
Assert.assertEquals(isRestore ? HoodieTimeline.RESTORE_ACTION :
HoodieTimeline.ROLLBACK_ACTION,
@@ -645,7 +649,7 @@ public class IncrementalFSViewSyncTest {
int multiple = begin;
for (int idx = 0; idx < instants.size(); idx++) {
String instant = instants.get(idx);
- System.out.println("Adding instant=" + instant);
+ log.info("Adding instant=" + instant);
HoodieInstant lastInstant = lastInstants.get(idx);
// Add a non-empty ingestion to COW table
List<String> filePaths = addInstant(metaClient, instant, deltaCommit,
diff --git
a/hoodie-common/src/test/java/com/uber/hoodie/common/util/TestRocksDBManager.java
b/hoodie-common/src/test/java/com/uber/hoodie/common/util/TestRocksDBManager.java
index 7bd3cb3..5b04fe3 100644
---
a/hoodie-common/src/test/java/com/uber/hoodie/common/util/TestRocksDBManager.java
+++
b/hoodie-common/src/test/java/com/uber/hoodie/common/util/TestRocksDBManager.java
@@ -29,17 +29,12 @@ import java.util.Objects;
import java.util.UUID;
import java.util.stream.Collectors;
import java.util.stream.IntStream;
-import org.apache.log4j.Level;
import org.junit.AfterClass;
import org.junit.Assert;
import org.junit.Test;
public class TestRocksDBManager {
- static {
- RocksDBDAO.log.setLevel(Level.INFO);
- }
-
private static RocksDBDAO dbManager;
@AfterClass