[ 
https://issues.apache.org/jira/browse/PHOENIX-3784?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16025740#comment-16025740
 ] 

Hadoop QA commented on PHOENIX-3784:
------------------------------------

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12869972/PHOENIX-3784.patch
  against master branch at commit 5f9cf15e272fc9d92a3165753ac2157396851bd6.
  ATTACHMENT ID: 12869972

    {color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

    {color:green}+1 tests included{color}.  The patch appears to include 3 new 
or modified tests.

    {color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

    {color:red}-1 javadoc{color}.  The javadoc tool appears to have generated 
47 warning messages.

    {color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

    {color:red}-1 lineLengths{color}.  The patch introduces the following lines 
longer than 100:
    +        connection = (PhoenixConnection) 
DriverManager.getConnection(getUrl(), connectionProperties);
+                        if (readyToCommit(rowCount, mutations.heapSize(), 
maxBatchSize, maxBatchSizeBytes)) {
+                        if (readyToCommit(rowCount, indexMutations.heapSize(), 
maxBatchSize, maxBatchSizeBytes)) {
+    private boolean readyToCommit(int rowCount, long mutationSize, int 
maxBatchSize, long maxBatchSizeBytes) {
+            int maxBatchSize = config.getInt(MUTATE_BATCH_SIZE_ATTRIB, 
QueryServicesOptions.DEFAULT_MUTATE_BATCH_SIZE);
+                        if (readyToCommit(rowCount, mutations.heapSize(), 
maxBatchSize, maxBatchSizeBytes)) {
+                        List<List<Mutation>> mutationBatchList = 
getMutationBatchList(batchSize, batchSizeBytes, mutationList);
+    public static List<List<Mutation>> getMutationBatchList(long batchSize, 
long batchSizeBytes, List<Mutation> allMutationList) {
+            if (currentList.size() == batchSize || currentBatchSizeBytes + 
mutationSizeBytes > batchSizeBytes) {

     {color:red}-1 core tests{color}.  The patch failed these unit tests:
     

Test results: 
https://builds.apache.org/job/PreCommit-PHOENIX-Build/908//testReport/
Javadoc warnings: 
https://builds.apache.org/job/PreCommit-PHOENIX-Build/908//artifact/patchprocess/patchJavadocWarnings.txt
Console output: 
https://builds.apache.org/job/PreCommit-PHOENIX-Build/908//console

This message is automatically generated.

> Chunk commit data using lower of byte-based and row-count limits
> ----------------------------------------------------------------
>
>                 Key: PHOENIX-3784
>                 URL: https://issues.apache.org/jira/browse/PHOENIX-3784
>             Project: Phoenix
>          Issue Type: Bug
>            Reporter: James Taylor
>            Assignee: Thomas D'Silva
>             Fix For: 4.11.0
>
>         Attachments: PHOENIX-3784.patch
>
>
> We have a byte-based limit that determines how much data we send over at a 
> time when a commit occurs (PHOENIX-541), but we should also have a row-count 
> limit. We could check both the byte-based limit and the row-count limit and 
> ensure the batch size meets both constraints. This would help prevent too 
> many rows from being submitted to the server at one time and decrease the 
> likelihood of conflicting rows amongst batches. 



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

Reply via email to