This is an automated email from the ASF dual-hosted git repository.

mcvsubbu pushed a commit to branch try-helix-0.9.1
in repository https://gitbox.apache.org/repos/asf/incubator-pinot.git


The following commit(s) were added to refs/heads/try-helix-0.9.1 by this push:
     new 7916922  Comment out a log message that is flooding 
HadoopSegmentBuildPushOfflineClusterIntegrationTest
7916922 is described below

commit 7916922c710f2048690e3001661708c78251045b
Author: Subbu Subramaniam <[email protected]>
AuthorDate: Thu Oct 24 09:46:22 2019 -0700

    Comment out a log message that is flooding 
HadoopSegmentBuildPushOfflineClusterIntegrationTest
---
 .../org/apache/pinot/hadoop/job/mappers/SegmentPreprocessingMapper.java | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)

diff --git 
a/pinot-hadoop/src/main/java/org/apache/pinot/hadoop/job/mappers/SegmentPreprocessingMapper.java
 
b/pinot-hadoop/src/main/java/org/apache/pinot/hadoop/job/mappers/SegmentPreprocessingMapper.java
index 5f2b7ac..e7aa0a5 100644
--- 
a/pinot-hadoop/src/main/java/org/apache/pinot/hadoop/job/mappers/SegmentPreprocessingMapper.java
+++ 
b/pinot-hadoop/src/main/java/org/apache/pinot/hadoop/job/mappers/SegmentPreprocessingMapper.java
@@ -94,7 +94,7 @@ public class SegmentPreprocessingMapper extends 
Mapper<AvroKey<GenericRecord>, N
       if (!normalizedTimeColumnValue.equals(_sampleNormalizedTimeColumnValue)) 
{
         // TODO: Create a custom exception and gracefully catch this exception 
outside, changing what the path to input
         // into segment creation should be
-        LOGGER.error("This segment contains multiple time units. Sample is {}, 
current is {}", _sampleNormalizedTimeColumnValue, normalizedTimeColumnValue);
+        //LOGGER.error("This segment contains multiple time units. Sample is 
{}, current is {}", _sampleNormalizedTimeColumnValue, 
normalizedTimeColumnValue);
       }
     }
 


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to