This is an automated email from the ASF dual-hosted git repository.

jonwei pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/incubator-druid.git


The following commit(s) were added to refs/heads/master by this push:
     new 5726692  Accept total rows over Integer.MAX_VALUE. (#6080)
5726692 is described below

commit 5726692f8fe3228c88be628defce44a281387568
Author: es1220 <[email protected]>
AuthorDate: Thu Aug 16 06:03:22 2018 +0900

    Accept total rows over Integer.MAX_VALUE. (#6080)
---
 .../java/io/druid/indexer/DeterminePartitionsJob.java    | 16 ++++++++--------
 .../io/druid/indexer/hadoop/DatasourceRecordReader.java  |  4 ++--
 2 files changed, 10 insertions(+), 10 deletions(-)

diff --git 
a/indexing-hadoop/src/main/java/io/druid/indexer/DeterminePartitionsJob.java 
b/indexing-hadoop/src/main/java/io/druid/indexer/DeterminePartitionsJob.java
index 9d61d98..fc4dde5 100644
--- a/indexing-hadoop/src/main/java/io/druid/indexer/DeterminePartitionsJob.java
+++ b/indexing-hadoop/src/main/java/io/druid/indexer/DeterminePartitionsJob.java
@@ -577,7 +577,7 @@ public class DeterminePartitionsJob implements Jobby
               }
 
               // Respect "poisoning" (negative values mean we can't use this 
dimension)
-              final int newNumRows = (arg1.numRows >= 0 && arg2.numRows >= 0 ? 
arg1.numRows + arg2.numRows : -1);
+              final long newNumRows = (arg1.numRows >= 0 && arg2.numRows >= 0 
? arg1.numRows + arg2.numRows : -1);
               return new DimValueCount(arg1.dim, arg1.value, newNumRows);
             }
           }
@@ -620,7 +620,7 @@ public class DeterminePartitionsJob implements Jobby
 
       // First DVC should be the total row count indicator
       final DimValueCount firstDvc = iterator.next();
-      final int totalRows = firstDvc.numRows;
+      final long totalRows = firstDvc.numRows;
 
       if (!"".equals(firstDvc.dim) || !"".equals(firstDvc.value)) {
         throw new IllegalStateException("WTF?! Expected total row indicator on 
first k/v pair!");
@@ -899,9 +899,9 @@ public class DeterminePartitionsJob implements Jobby
       return distance;
     }
 
-    public int getRows()
+    public long getRows()
     {
-      int sum = 0;
+      long sum = 0;
       for (final DimPartition dimPartition : partitions) {
         sum += dimPartition.rows;
       }
@@ -913,16 +913,16 @@ public class DeterminePartitionsJob implements Jobby
   {
     public ShardSpec shardSpec = null;
     public int cardinality = 0;
-    public int rows = 0;
+    public long rows = 0;
   }
 
   private static class DimValueCount
   {
     public final String dim;
     public final String value;
-    public final int numRows;
+    public final long numRows;
 
-    private DimValueCount(String dim, String value, int numRows)
+    private DimValueCount(String dim, String value, long numRows)
     {
       this.dim = dim;
       this.value = value;
@@ -938,7 +938,7 @@ public class DeterminePartitionsJob implements Jobby
     {
       final Iterator<String> splits = 
TAB_SPLITTER.limit(3).split(text.toString()).iterator();
       final String dim = splits.next();
-      final int numRows = Integer.parseInt(splits.next());
+      final long numRows = Long.parseLong(splits.next());
       final String value = splits.next();
 
       return new DimValueCount(dim, value, numRows);
diff --git 
a/indexing-hadoop/src/main/java/io/druid/indexer/hadoop/DatasourceRecordReader.java
 
b/indexing-hadoop/src/main/java/io/druid/indexer/hadoop/DatasourceRecordReader.java
index 739bd9e..b3dde78 100644
--- 
a/indexing-hadoop/src/main/java/io/druid/indexer/hadoop/DatasourceRecordReader.java
+++ 
b/indexing-hadoop/src/main/java/io/druid/indexer/hadoop/DatasourceRecordReader.java
@@ -52,12 +52,12 @@ public class DatasourceRecordReader extends 
RecordReader<NullWritable, InputRow>
   private DatasourceIngestionSpec spec;
   private IngestSegmentFirehose firehose;
 
-  private int rowNum;
+  private long rowNum;
   private Row currRow;
 
   private List<QueryableIndex> indexes = Lists.newArrayList();
   private List<File> tmpSegmentDirs = Lists.newArrayList();
-  private int numRows;
+  private long numRows;
 
   @Override
   public void initialize(InputSplit split, final TaskAttemptContext context) 
throws IOException


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to