[GitHub] carbondata issue #890: [CARBONDATA-1008] Make Caron table schema compatible ...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/890 Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2167/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #984: [CARBONDATA-1008] Make Caron table schema compatible ...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/984 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-2.1/62/Failed Tests: 1carbondata-pr-spark-2.1/org.apache.carbondata:carbondata-spark-common-test: 1org.apache.carbondata.spark.testsuite.allqueries.InsertIntoCarbonTableTestCase.insert into carbon table from carbon table union query --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #984: [CARBONDATA-1008] Make Caron table schema compatible ...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/984 Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2166/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #890: [CARBONDATA-1008] Make Caron table schema compatible ...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/890 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-2.1/63/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #890: [CARBONDATA-1008] Make Caron table schema compatible ...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/890 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/34/Build result: FAILURE[...truncated 101.51 KB...][ERROR] Re-run Maven using the -X switch to enable full debug logging.[ERROR] [ERROR] For more information about the errors and possible solutions, please read the following articles:[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException[ERROR] [ERROR] After correcting the problems, you can resume the build with the command[ERROR] mvn -rf :carbondata-spark[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/processing/pom.xml to org.apache.carbondata/carbondata-processing/1.2.0-SNAPSHOT/carbondata-processing-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/spark-common/pom.xml to org.apache.carbondata/carbondata-spark-common/1.2.0-SNAPSHOT/carbondata-spark-common-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave /workspace/carbondata-pr-spark-1.6/examples/spark/pom.xml to org.apache.carbondata/carbondata-examples-spark/1.2.0-SNAPSHOT/carbondata-examples-spark-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/hive/pom.xml to org.apache.carbondata/carbondata-hive/1.2.0-SNAPSHOT/carbondata-hive-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/common/pom.xml to org.apache.carbondata/carbondata-common/1.2.0-SNAPSHOT/carbondata-common-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/assembly/pom.xml to org.apache.carbondata/carbondata-assembly/1.2.0-SNAPSHOT/carbondata-assembly-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/pom.xml to org.apache.carbondata/carbondata-parent/1.2.0-SNAPSHOT/carbondata-parent-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/ca rbondata-pr-spark-1.6/examples/flink/pom.xml to org.apache.carbondata/carbondata-examples-flink/1.2.0-SNAPSHOT/carbondata-examples-flink-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/core/pom.xml to org.apache.carbondata/carbondata-core/1.2.0-SNAPSHOT/carbondata-core-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/hadoop/pom.xml to org.apache.carbondata/carbondata-hadoop/1.2.0-SNAPSHOT/carbondata-hadoop-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/presto/pom.xml to org.apache.carbondata/carbondata-presto/1.2.0-SNAPSHOT/carbondata-presto-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/spark-common-test/pom.xml to org.apache.carbondata/carbondata-spark-common-test/1.2.0-SNAPSHOT/carbondata-spark-common-test-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/j enkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/spark/pom.xml to org.apache.carbondata/carbondata-spark/1.2.0-SNAPSHOT/carbondata-spark-1.2.0-SNAPSHOT.pomchannel stoppedSetting status of 8aba3d4e9e7de376eb9036860cbe4e76b38bd5cf to FAILURE with url https://builds.apache.org/job/carbondata-pr-spark-1.6/34/ and message: 'Tests Failed for Spark1.6 'Using context: Jenkins(Spark1.6): mvn clean test -Pspark-1.6 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #984: [CARBONDATA-1008] Make Caron table schema compatible ...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/984 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/33/Build result: FAILURE[...truncated 102.30 KB...][ERROR] [ERROR] After correcting the problems, you can resume the build with the command[ERROR] mvn -rf :carbondata-spark[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/processing/pom.xml to org.apache.carbondata/carbondata-processing/1.2.0-SNAPSHOT/carbondata-processing-1.2.0-SNAPSHOT.pomNo artifacts from carbondata-pr-spark-1.6 » Apache CarbonData :: Processing #34 to compare, so performing full copy of artifacts[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/spark-common/pom.xml to org.apache.carbondata/carbondata-spark-common/1.2.0-SNAPSHOT/carbondata-spark-common-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/examples/spark/pom.xml to org.apache.carbondata/carbondata-examples-spark/1.2.0-SNAPSHOT/carbondata-example s-spark-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/hive/pom.xml to org.apache.carbondata/carbondata-hive/1.2.0-SNAPSHOT/carbondata-hive-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/common/pom.xml to org.apache.carbondata/carbondata-common/1.2.0-SNAPSHOT/carbondata-common-1.2.0-SNAPSHOT.pomNo artifacts from carbondata-pr-spark-1.6 » Apache CarbonData :: Common #34 to compare, so performing full copy of artifacts[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/assembly/pom.xml to org.apache.carbondata/carbondata-assembly/1.2.0-SNAPSHOT/carbondata-assembly-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/pom.xml to org.apache.carbondata/carbondata-parent/1.2.0-SNAPSHOT/carbondata-parent-1.2.0-SNAPSHOT.pomNo artifacts from carbondata-pr-spark-1.6 » Apache CarbonData :: Parent #34 to compare, so performing full copy of artifacts[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/examples/flink/pom.xml to org.apache.carbondata/carbondata-examples-flink/1.2.0-SNAPSHOT/carbondata-examples-flink-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/core/pom.xml to org.apache.carbondata/carbondata-core/1.2.0-SNAPSHOT/carbondata-core-1.2.0-SNAPSHOT.pomNo artifacts from carbondata-pr-spark-1.6 » Apache CarbonData :: Core #34 to compare, so performing full copy of artifacts[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/hadoop/pom.xml to org.apache.carbondata/carbondata-hadoop/1.2.0-SNAPSHOT/carbondata-hadoop-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/presto/pom.xml to org.apache.carbondata/carbondata-presto/1.2.0-SNAPSHOT/carbondata-presto-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkin s/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/spark-common-test/pom.xml to org.apache.carbondata/carbondata-spark-common-test/1.2.0-SNAPSHOT/carbondata-spark-common-test-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/spark/pom.xml to org.apache.carbondata/carbondata-spark/1.2.0-SNAPSHOT/carbondata-spark-1.2.0-SNAPSHOT.pomchannel stoppedSetting status of 8aba3d4e9e7de376eb9036860cbe4e76b38bd5cf to FAILURE with url https://builds.apache.org/job/carbondata-pr-spark-1.6/33/ and message: 'Tests Failed for Spark1.6 'Using context: Jenkins(Spark1.6): mvn clean test -Pspark-1.6 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #890: [CARBONDATA-1008] Make Caron table schema compatible ...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/890 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-2.1/61/Failed Tests: 1carbondata-pr-spark-2.1/org.apache.carbondata:carbondata-core: 1org.apache.carbondata.core.dictionary.client.DictionaryClientTest.testClient --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #890: [CARBONDATA-1008] Make Caron table schema compatible ...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/890 Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2165/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #984: [CARBONDATA-1008] Make Caron table schema compatible ...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/984 Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2164/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #984: [CARBONDATA-1008] Make Caron table schema compatible ...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/984 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-2.1/60/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #890: [CARBONDATA-1008] Make Caron table schema compatible ...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/890 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/32/Build result: FAILURE[...truncated 80.69 KB...][ERROR] Re-run Maven using the -X switch to enable full debug logging.[ERROR] [ERROR] For more information about the errors and possible solutions, please read the following articles:[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException[ERROR] [ERROR] After correcting the problems, you can resume the build with the command[ERROR] mvn -rf :carbondata-spark[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/processing/pom.xml to org.apache.carbondata/carbondata-processing/1.2.0-SNAPSHOT/carbondata-processing-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/spark-common/pom.xml to org.apache.carbondata/carbondata-spark-common/1.2.0-SNAPSHOT/carbondata-spark-common-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/ workspace/carbondata-pr-spark-1.6/examples/spark/pom.xml to org.apache.carbondata/carbondata-examples-spark/1.2.0-SNAPSHOT/carbondata-examples-spark-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/hive/pom.xml to org.apache.carbondata/carbondata-hive/1.2.0-SNAPSHOT/carbondata-hive-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/common/pom.xml to org.apache.carbondata/carbondata-common/1.2.0-SNAPSHOT/carbondata-common-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/assembly/pom.xml to org.apache.carbondata/carbondata-assembly/1.2.0-SNAPSHOT/carbondata-assembly-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/pom.xml to org.apache.carbondata/carbondata-parent/1.2.0-SNAPSHOT/carbondata-parent-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/car bondata-pr-spark-1.6/examples/flink/pom.xml to org.apache.carbondata/carbondata-examples-flink/1.2.0-SNAPSHOT/carbondata-examples-flink-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/core/pom.xml to org.apache.carbondata/carbondata-core/1.2.0-SNAPSHOT/carbondata-core-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/hadoop/pom.xml to org.apache.carbondata/carbondata-hadoop/1.2.0-SNAPSHOT/carbondata-hadoop-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/presto/pom.xml to org.apache.carbondata/carbondata-presto/1.2.0-SNAPSHOT/carbondata-presto-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/spark-common-test/pom.xml to org.apache.carbondata/carbondata-spark-common-test/1.2.0-SNAPSHOT/carbondata-spark-common-test-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/je nkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/spark/pom.xml to org.apache.carbondata/carbondata-spark/1.2.0-SNAPSHOT/carbondata-spark-1.2.0-SNAPSHOT.pomchannel stoppedSetting status of 011581a4a2b00e9440619d7de565a0443ffe14aa to FAILURE with url https://builds.apache.org/job/carbondata-pr-spark-1.6/32/ and message: 'Tests Failed for Spark1.6 'Using context: Jenkins(Spark1.6): mvn clean test -Pspark-1.6 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #984: [CARBONDATA-1008] Make Caron table schema compatible ...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/984 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/31/Build result: FAILURE[...truncated 81.20 KB...][ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException[ERROR] [ERROR] After correcting the problems, you can resume the build with the command[ERROR] mvn -rf :carbondata-spark[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/processing/pom.xml to org.apache.carbondata/carbondata-processing/1.2.0-SNAPSHOT/carbondata-processing-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/spark-common/pom.xml to org.apache.carbondata/carbondata-spark-common/1.2.0-SNAPSHOT/carbondata-spark-common-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/examples/spark/pom.xml to org.apache.carbondata/carbondata-examples-spark/1.2.0-SNAPSHOT/carbondata-examples-spark-1.2.0-SNAPSHOT.pom[JENKINS] Archiv ing /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/hive/pom.xml to org.apache.carbondata/carbondata-hive/1.2.0-SNAPSHOT/carbondata-hive-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/common/pom.xml to org.apache.carbondata/carbondata-common/1.2.0-SNAPSHOT/carbondata-common-1.2.0-SNAPSHOT.pomNo artifacts from carbondata-pr-spark-1.6 » Apache CarbonData :: Common #32 to compare, so performing full copy of artifacts[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/assembly/pom.xml to org.apache.carbondata/carbondata-assembly/1.2.0-SNAPSHOT/carbondata-assembly-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/pom.xml to org.apache.carbondata/carbondata-parent/1.2.0-SNAPSHOT/carbondata-parent-1.2.0-SNAPSHOT.pomNo artifacts from carbondata-pr-spark-1.6 » Apache CarbonData :: Parent #32 to compare, so performing full copy of art ifacts[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/examples/flink/pom.xml to org.apache.carbondata/carbondata-examples-flink/1.2.0-SNAPSHOT/carbondata-examples-flink-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/core/pom.xml to org.apache.carbondata/carbondata-core/1.2.0-SNAPSHOT/carbondata-core-1.2.0-SNAPSHOT.pomNo artifacts from carbondata-pr-spark-1.6 » Apache CarbonData :: Core #32 to compare, so performing full copy of artifacts[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/hadoop/pom.xml to org.apache.carbondata/carbondata-hadoop/1.2.0-SNAPSHOT/carbondata-hadoop-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/presto/pom.xml to org.apache.carbondata/carbondata-presto/1.2.0-SNAPSHOT/carbondata-presto-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-sp ark-1.6/integration/spark-common-test/pom.xml to org.apache.carbondata/carbondata-spark-common-test/1.2.0-SNAPSHOT/carbondata-spark-common-test-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/spark/pom.xml to org.apache.carbondata/carbondata-spark/1.2.0-SNAPSHOT/carbondata-spark-1.2.0-SNAPSHOT.pomchannel stoppedSetting status of 011581a4a2b00e9440619d7de565a0443ffe14aa to FAILURE with url https://builds.apache.org/job/carbondata-pr-spark-1.6/31/ and message: 'Tests Failed for Spark1.6 'Using context: Jenkins(Spark1.6): mvn clean test -Pspark-1.6 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #890: [CARBONDATA-1008] Make Caron table schema compatible ...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/890 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/30/Build result: FAILURE[...truncated 81.50 KB...][ERROR] mvn -rf :carbondata-spark[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/processing/pom.xml to org.apache.carbondata/carbondata-processing/1.2.0-SNAPSHOT/carbondata-processing-1.2.0-SNAPSHOT.pomNo artifacts from carbondata-pr-spark-1.6 » Apache CarbonData :: Processing #31 to compare, so performing full copy of artifacts[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/spark-common/pom.xml to org.apache.carbondata/carbondata-spark-common/1.2.0-SNAPSHOT/carbondata-spark-common-1.2.0-SNAPSHOT.pomNo artifacts from carbondata-pr-spark-1.6 » Apache CarbonData :: Spark Common #31 to compare, so performing full copy of artifacts[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/examples/spark/pom.xml to org.apache.carbondata/carbondata-exampl es-spark/1.2.0-SNAPSHOT/carbondata-examples-spark-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/hive/pom.xml to org.apache.carbondata/carbondata-hive/1.2.0-SNAPSHOT/carbondata-hive-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/common/pom.xml to org.apache.carbondata/carbondata-common/1.2.0-SNAPSHOT/carbondata-common-1.2.0-SNAPSHOT.pomNo artifacts from carbondata-pr-spark-1.6 » Apache CarbonData :: Common #32 to compare, so performing full copy of artifacts[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/assembly/pom.xml to org.apache.carbondata/carbondata-assembly/1.2.0-SNAPSHOT/carbondata-assembly-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/pom.xml to org.apache.carbondata/carbondata-parent/1.2.0-SNAPSHOT/carbondata-parent-1.2.0-SNAPSHOT.pomNo artifacts from carbondata-pr-spa rk-1.6 » Apache CarbonData :: Parent #32 to compare, so performing full copy of artifacts[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/examples/flink/pom.xml to org.apache.carbondata/carbondata-examples-flink/1.2.0-SNAPSHOT/carbondata-examples-flink-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/core/pom.xml to org.apache.carbondata/carbondata-core/1.2.0-SNAPSHOT/carbondata-core-1.2.0-SNAPSHOT.pomNo artifacts from carbondata-pr-spark-1.6 » Apache CarbonData :: Core #32 to compare, so performing full copy of artifacts[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/hadoop/pom.xml to org.apache.carbondata/carbondata-hadoop/1.2.0-SNAPSHOT/carbondata-hadoop-1.2.0-SNAPSHOT.pomNo artifacts from carbondata-pr-spark-1.6 » Apache CarbonData :: Hadoop #31 to compare, so performing full copy of artifacts[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata -pr-spark-1.6/integration/presto/pom.xml to org.apache.carbondata/carbondata-presto/1.2.0-SNAPSHOT/carbondata-presto-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/spark-common-test/pom.xml to org.apache.carbondata/carbondata-spark-common-test/1.2.0-SNAPSHOT/carbondata-spark-common-test-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/spark/pom.xml to org.apache.carbondata/carbondata-spark/1.2.0-SNAPSHOT/carbondata-spark-1.2.0-SNAPSHOT.pomchannel stoppedSetting status of dc11502cbdecc7c42af6d18026a73d8852889e20 to FAILURE with url https://builds.apache.org/job/carbondata-pr-spark-1.6/30/ and message: 'Tests Failed for Spark1.6 'Using context: Jenkins(Spark1.6): mvn clean test -Pspark-1.6 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #984: [CARBONDATA-1008] Make Caron table schema compatible ...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/984 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/29/Build result: FAILURE[...truncated 80.23 KB...][ERROR] [ERROR] After correcting the problems, you can resume the build with the command[ERROR] mvn -rf :carbondata-spark[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/processing/pom.xml to org.apache.carbondata/carbondata-processing/1.2.0-SNAPSHOT/carbondata-processing-1.2.0-SNAPSHOT.pomNo artifacts from carbondata-pr-spark-1.6 » Apache CarbonData :: Processing #31 to compare, so performing full copy of artifacts[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/spark-common/pom.xml to org.apache.carbondata/carbondata-spark-common/1.2.0-SNAPSHOT/carbondata-spark-common-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/examples/spark/pom.xml to org.apache.carbondata/carbondata-examples-spark/1.2.0-SNAPSHOT/carbondata-examples -spark-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/hive/pom.xml to org.apache.carbondata/carbondata-hive/1.2.0-SNAPSHOT/carbondata-hive-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/common/pom.xml to org.apache.carbondata/carbondata-common/1.2.0-SNAPSHOT/carbondata-common-1.2.0-SNAPSHOT.pomNo artifacts from carbondata-pr-spark-1.6 » Apache CarbonData :: Common #32 to compare, so performing full copy of artifacts[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/assembly/pom.xml to org.apache.carbondata/carbondata-assembly/1.2.0-SNAPSHOT/carbondata-assembly-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/pom.xml to org.apache.carbondata/carbondata-parent/1.2.0-SNAPSHOT/carbondata-parent-1.2.0-SNAPSHOT.pomNo artifacts from carbondata-pr-spark-1.6 » Apache CarbonData :: Parent #32 t o compare, so performing full copy of artifacts[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/examples/flink/pom.xml to org.apache.carbondata/carbondata-examples-flink/1.2.0-SNAPSHOT/carbondata-examples-flink-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/core/pom.xml to org.apache.carbondata/carbondata-core/1.2.0-SNAPSHOT/carbondata-core-1.2.0-SNAPSHOT.pomNo artifacts from carbondata-pr-spark-1.6 » Apache CarbonData :: Core #31 to compare, so performing full copy of artifacts[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/hadoop/pom.xml to org.apache.carbondata/carbondata-hadoop/1.2.0-SNAPSHOT/carbondata-hadoop-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/presto/pom.xml to org.apache.carbondata/carbondata-presto/1.2.0-SNAPSHOT/carbondata-presto-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins /jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/spark-common-test/pom.xml to org.apache.carbondata/carbondata-spark-common-test/1.2.0-SNAPSHOT/carbondata-spark-common-test-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/spark/pom.xml to org.apache.carbondata/carbondata-spark/1.2.0-SNAPSHOT/carbondata-spark-1.2.0-SNAPSHOT.pomchannel stoppedSetting status of dc11502cbdecc7c42af6d18026a73d8852889e20 to FAILURE with url https://builds.apache.org/job/carbondata-pr-spark-1.6/29/ and message: 'Tests Failed for Spark1.6 'Using context: Jenkins(Spark1.6): mvn clean test -Pspark-1.6 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #890: [CARBONDATA-1008] Make Caron table schema compatible ...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/890 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-2.1/57/Build result: FAILURE[...truncated 80.44 KB...][ERROR] Re-run Maven using the -X switch to enable full debug logging.[ERROR] [ERROR] For more information about the errors and possible solutions, please read the following articles:[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException[ERROR] [ERROR] After correcting the problems, you can resume the build with the command[ERROR] mvn -rf :carbondata-spark2[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-2.1@2/processing/pom.xml to org.apache.carbondata/carbondata-processing/1.2.0-SNAPSHOT/carbondata-processing-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-2.1@2/integration/spark-common/pom.xml to org.apache.carbondata/carbondata-spark-common/1.2.0-SNAPSHOT/carbondata-spark-common-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-s lave/workspace/carbondata-pr-spark-2.1@2/integration/hive/pom.xml to org.apache.carbondata/carbondata-hive/1.2.0-SNAPSHOT/carbondata-hive-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-2.1@2/common/pom.xml to org.apache.carbondata/carbondata-common/1.2.0-SNAPSHOT/carbondata-common-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-2.1@2/examples/spark2/pom.xml to org.apache.carbondata/carbondata-examples-spark2/1.2.0-SNAPSHOT/carbondata-examples-spark2-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-2.1@2/integration/spark2/pom.xml to org.apache.carbondata/carbondata-spark2/1.2.0-SNAPSHOT/carbondata-spark2-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-2.1@2/assembly/pom.xml to org.apache.carbondata/carbondata-assembly/1.2.0-SNAPSHOT/carbondata-assembly-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /hom e/jenkins/jenkins-slave/workspace/carbondata-pr-spark-2.1@2/pom.xml to org.apache.carbondata/carbondata-parent/1.2.0-SNAPSHOT/carbondata-parent-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-2.1@2/examples/flink/pom.xml to org.apache.carbondata/carbondata-examples-flink/1.2.0-SNAPSHOT/carbondata-examples-flink-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-2.1@2/core/pom.xml to org.apache.carbondata/carbondata-core/1.2.0-SNAPSHOT/carbondata-core-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-2.1@2/hadoop/pom.xml to org.apache.carbondata/carbondata-hadoop/1.2.0-SNAPSHOT/carbondata-hadoop-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-2.1@2/integration/presto/pom.xml to org.apache.carbondata/carbondata-presto/1.2.0-SNAPSHOT/carbondata-presto-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins /jenkins-slave/workspace/carbondata-pr-spark-2.1@2/integration/spark-common-test/pom.xml to org.apache.carbondata/carbondata-spark-common-test/1.2.0-SNAPSHOT/carbondata-spark-common-test-1.2.0-SNAPSHOT.pomchannel stoppedSetting status of 4d979eb039e3c125d599e3c71ff6d46028be6b27 to FAILURE with url https://builds.apache.org/job/carbondata-pr-spark-2.1/57/ and message: 'Tests Failed for Spark2.1 'Using context: Jenkins(Spark2.1): mvn clean test -Pspark-2.1 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #890: [CARBONDATA-1008] Make Caron table schema compatible ...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/890 Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2163/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #984: [CARBONDATA-1008] Make Caron table schema compatible ...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/984 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/27/Build result: FAILURE[...truncated 78.26 KB...][ERROR] mvn -rf :carbondata-spark[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/processing/pom.xml to org.apache.carbondata/carbondata-processing/1.2.0-SNAPSHOT/carbondata-processing-1.2.0-SNAPSHOT.pomNo artifacts from carbondata-pr-spark-1.6 » Apache CarbonData :: Processing #28 to compare, so performing full copy of artifacts[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/spark-common/pom.xml to org.apache.carbondata/carbondata-spark-common/1.2.0-SNAPSHOT/carbondata-spark-common-1.2.0-SNAPSHOT.pomNo artifacts from carbondata-pr-spark-1.6 » Apache CarbonData :: Spark Common #28 to compare, so performing full copy of artifacts[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/examples/spark/pom.xml to org.apache.carbondata/carbondata-exampl es-spark/1.2.0-SNAPSHOT/carbondata-examples-spark-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/hive/pom.xml to org.apache.carbondata/carbondata-hive/1.2.0-SNAPSHOT/carbondata-hive-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/common/pom.xml to org.apache.carbondata/carbondata-common/1.2.0-SNAPSHOT/carbondata-common-1.2.0-SNAPSHOT.pomNo artifacts from carbondata-pr-spark-1.6 » Apache CarbonData :: Common #28 to compare, so performing full copy of artifacts[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/assembly/pom.xml to org.apache.carbondata/carbondata-assembly/1.2.0-SNAPSHOT/carbondata-assembly-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/pom.xml to org.apache.carbondata/carbondata-parent/1.2.0-SNAPSHOT/carbondata-parent-1.2.0-SNAPSHOT.pomNo artifacts from carbondata-pr-spa rk-1.6 » Apache CarbonData :: Parent #28 to compare, so performing full copy of artifacts[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/examples/flink/pom.xml to org.apache.carbondata/carbondata-examples-flink/1.2.0-SNAPSHOT/carbondata-examples-flink-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/core/pom.xml to org.apache.carbondata/carbondata-core/1.2.0-SNAPSHOT/carbondata-core-1.2.0-SNAPSHOT.pomNo artifacts from carbondata-pr-spark-1.6 » Apache CarbonData :: Core #28 to compare, so performing full copy of artifacts[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/hadoop/pom.xml to org.apache.carbondata/carbondata-hadoop/1.2.0-SNAPSHOT/carbondata-hadoop-1.2.0-SNAPSHOT.pomNo artifacts from carbondata-pr-spark-1.6 » Apache CarbonData :: Hadoop #28 to compare, so performing full copy of artifacts[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata -pr-spark-1.6/integration/presto/pom.xml to org.apache.carbondata/carbondata-presto/1.2.0-SNAPSHOT/carbondata-presto-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/spark-common-test/pom.xml to org.apache.carbondata/carbondata-spark-common-test/1.2.0-SNAPSHOT/carbondata-spark-common-test-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/spark/pom.xml to org.apache.carbondata/carbondata-spark/1.2.0-SNAPSHOT/carbondata-spark-1.2.0-SNAPSHOT.pomchannel stoppedSetting status of 4d979eb039e3c125d599e3c71ff6d46028be6b27 to FAILURE with url https://builds.apache.org/job/carbondata-pr-spark-1.6/27/ and message: 'Tests Failed for Spark1.6 'Using context: Jenkins(Spark1.6): mvn clean test -Pspark-1.6 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #890: [CARBONDATA-1008] Make Caron table schema compatible ...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/890 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/28/Build result: FAILURE[...truncated 76.77 KB...][ERROR] Re-run Maven using the -X switch to enable full debug logging.[ERROR] [ERROR] For more information about the errors and possible solutions, please read the following articles:[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException[ERROR] [ERROR] After correcting the problems, you can resume the build with the command[ERROR] mvn -rf :carbondata-spark[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/processing/pom.xml to org.apache.carbondata/carbondata-processing/1.2.0-SNAPSHOT/carbondata-processing-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/spark-common/pom.xml to org.apache.carbondata/carbondata-spark-common/1.2.0-SNAPSHOT/carbondata-spark-common-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/ workspace/carbondata-pr-spark-1.6/examples/spark/pom.xml to org.apache.carbondata/carbondata-examples-spark/1.2.0-SNAPSHOT/carbondata-examples-spark-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/hive/pom.xml to org.apache.carbondata/carbondata-hive/1.2.0-SNAPSHOT/carbondata-hive-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/common/pom.xml to org.apache.carbondata/carbondata-common/1.2.0-SNAPSHOT/carbondata-common-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/assembly/pom.xml to org.apache.carbondata/carbondata-assembly/1.2.0-SNAPSHOT/carbondata-assembly-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/pom.xml to org.apache.carbondata/carbondata-parent/1.2.0-SNAPSHOT/carbondata-parent-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/car bondata-pr-spark-1.6/examples/flink/pom.xml to org.apache.carbondata/carbondata-examples-flink/1.2.0-SNAPSHOT/carbondata-examples-flink-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/core/pom.xml to org.apache.carbondata/carbondata-core/1.2.0-SNAPSHOT/carbondata-core-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/hadoop/pom.xml to org.apache.carbondata/carbondata-hadoop/1.2.0-SNAPSHOT/carbondata-hadoop-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/presto/pom.xml to org.apache.carbondata/carbondata-presto/1.2.0-SNAPSHOT/carbondata-presto-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/spark-common-test/pom.xml to org.apache.carbondata/carbondata-spark-common-test/1.2.0-SNAPSHOT/carbondata-spark-common-test-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/je nkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/spark/pom.xml to org.apache.carbondata/carbondata-spark/1.2.0-SNAPSHOT/carbondata-spark-1.2.0-SNAPSHOT.pomchannel stoppedSetting status of 4d979eb039e3c125d599e3c71ff6d46028be6b27 to FAILURE with url https://builds.apache.org/job/carbondata-pr-spark-1.6/28/ and message: 'Tests Failed for Spark1.6 'Using context: Jenkins(Spark1.6): mvn clean test -Pspark-1.6 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #984: [CARBONDATA-1008] Make Caron table schema compatible ...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/984 Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2162/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #984: [CARBONDATA-1008] Make Caron table schema compatible ...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/984 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-2.1/56/Build result: FAILURE[...truncated 80.98 KB...][ERROR] After correcting the problems, you can resume the build with the command[ERROR] mvn -rf :carbondata-spark2[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-2.1/processing/pom.xml to org.apache.carbondata/carbondata-processing/1.2.0-SNAPSHOT/carbondata-processing-1.2.0-SNAPSHOT.pomNo artifacts from carbondata-pr-spark-2.1 » Apache CarbonData :: Processing #57 to compare, so performing full copy of artifacts[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-2.1/integration/spark-common/pom.xml to org.apache.carbondata/carbondata-spark-common/1.2.0-SNAPSHOT/carbondata-spark-common-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-2.1/integration/hive/pom.xml to org.apache.carbondata/carbondata-hive/1.2.0-SNAPSHOT/carbondata-hive-1.2.0-SNAPSHOT.pom [JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-2.1/common/pom.xml to org.apache.carbondata/carbondata-common/1.2.0-SNAPSHOT/carbondata-common-1.2.0-SNAPSHOT.pomNo artifacts from carbondata-pr-spark-2.1 » Apache CarbonData :: Common #57 to compare, so performing full copy of artifacts[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-2.1/examples/spark2/pom.xml to org.apache.carbondata/carbondata-examples-spark2/1.2.0-SNAPSHOT/carbondata-examples-spark2-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-2.1/integration/spark2/pom.xml to org.apache.carbondata/carbondata-spark2/1.2.0-SNAPSHOT/carbondata-spark2-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-2.1/assembly/pom.xml to org.apache.carbondata/carbondata-assembly/1.2.0-SNAPSHOT/carbondata-assembly-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbo ndata-pr-spark-2.1/pom.xml to org.apache.carbondata/carbondata-parent/1.2.0-SNAPSHOT/carbondata-parent-1.2.0-SNAPSHOT.pomNo artifacts from carbondata-pr-spark-2.1 » Apache CarbonData :: Parent #57 to compare, so performing full copy of artifacts[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-2.1/examples/flink/pom.xml to org.apache.carbondata/carbondata-examples-flink/1.2.0-SNAPSHOT/carbondata-examples-flink-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-2.1/core/pom.xml to org.apache.carbondata/carbondata-core/1.2.0-SNAPSHOT/carbondata-core-1.2.0-SNAPSHOT.pomNo artifacts from carbondata-pr-spark-2.1 » Apache CarbonData :: Core #57 to compare, so performing full copy of artifacts[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-2.1/hadoop/pom.xml to org.apache.carbondata/carbondata-hadoop/1.2.0-SNAPSHOT/carbondata-hadoop-1.2.0-SNAPSHOT.pomNo artifacts from carbondata-pr-spark -2.1 » Apache CarbonData :: Hadoop #57 to compare, so performing full copy of artifacts[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-2.1/integration/presto/pom.xml to org.apache.carbondata/carbondata-presto/1.2.0-SNAPSHOT/carbondata-presto-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-2.1/integration/spark-common-test/pom.xml to org.apache.carbondata/carbondata-spark-common-test/1.2.0-SNAPSHOT/carbondata-spark-common-test-1.2.0-SNAPSHOT.pomchannel stoppedSetting status of 4d979eb039e3c125d599e3c71ff6d46028be6b27 to FAILURE with url https://builds.apache.org/job/carbondata-pr-spark-2.1/56/ and message: 'Tests Failed for Spark2.1 'Using context: Jenkins(Spark2.1): mvn clean test -Pspark-2.1 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #988: [CARBONDATA-1110] put if clause out of the for clause
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/988 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/26/Failed Tests: 4carbondata-pr-spark-1.6/org.apache.carbondata:carbondata-spark-common-test: 4org.apache.carbondata.spark.testsuite.dataload.MultiFilesDataLoagdingTestCase.test data loading for multi files and nested folderorg.apache.carbondata.spark.testsuite.iud.HorizontalCompactionTestCase.test IUD Horizontal Compaction Check Column Cardinalityorg.apache.carbondata.spark.testsuite.iud.HorizontalCompactionTestCase.test IUD Horizontal Compaction Segment Delete Test Caseorg.apache.carbondata.spark.testsuite.iud. HorizontalCompactionTestCase.test case full table delete --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #988: [CARBONDATA-1110] put if clause out of the for clause
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/988 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-2.1/55/Failed Tests: 9carbondata-pr-spark-2.1/org.apache.carbondata:carbondata-spark-common-test: 9org.apache.carbondata.integration.spark.testsuite.dataload.TestLoadDataWithAutoLoadMerge.test data loading with auto load mergeorg.apache.carbondata.spark.testsuite.partition.TestDataLoadingForPartitionTable.data loading for partition table: range partitionorg.apache.carbondata.spark.testsuite.partition.TestDataLoadingForPartitionTable.data loading for partition table: list partitionor g.apache.carbondata.spark.testsuite.partition.TestDataLoadingForPartitionTable.Insert into for partition table: hash partitionorg.apache.carbondata.spark.testsuite.partition.TestDataLoadingForPartitionTable.Insert into for partition table: range partitionorg.apache.carbondata.spark.testsuite.partition.TestDataLoadingForPartitionTable.Insert into partition table: list partitionorg.apache.carbondata.spark.testsuite.partition.TestDataLoadingForPartitionTable.multiple data loading for partition tableorg.apache.carbondata.spark.testsuite.partition.TestDataLoadingForPartitionTable.multiple insertInto for partition tableorg.apache.carbondata.spark.testsuite.partition.TestDataLoadingForPartitionTable.multiple data loading and insertInto for partition table --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #988: [CARBONDATA-1110] put if clause out of the for clause
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/988 Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2161/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #971: [CARBONDATA-1015] Extract interface in data load writ...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/971 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-2.1/54/Failed Tests: 3carbondata-pr-spark-2.1/org.apache.carbondata:carbondata-spark-common-test: 3org.apache.carbondata.integration.spark.testsuite.dataload.TestLoadDataWithSinglePass.test data loading use one passorg.apache.carbondata.integration.spark.testsuite.dataload.TestLoadDataWithSinglePass.test data loading use one pass when offer column dictionary fileorg.apache.carbondata.integration.spark.testsuite.dataload.TestLoadDataWithSinglePass.test data loading use one pass when do incremental load --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #971: [CARBONDATA-1015] Extract interface in data load writ...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/971 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/25/Failed Tests: 5carbondata-pr-spark-1.6/org.apache.carbondata:carbondata-spark-common-test: 5org.apache.carbondata.integration.spark.testsuite.dataload.TestLoadDataWithSinglePass.test data loading use one passorg.apache.carbondata.integration.spark.testsuite.dataload.TestLoadDataWithSinglePass.test data loading use one pass when offer column dictionary fileorg.apache.carbondata.integration.spark.testsuite.dataload.TestLoadDataWithSinglePass.test data loading use one pass when do incremental loadorg.apache.carbondata.spark.testsuite.allqueries.InsertIntoCarbonTableTestCase.insert from hive-sum expressionorg.apache.carbondata.spark.testsuite.allqueries.InsertIntoCarbonTableTestCase.insert into carbon table from carbon table union query --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #987: [WIP] Add EncodingStrategy and use ColumnPage for mea...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/987 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-2.1/53/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #987: [WIP] Add EncodingStrategy and use ColumnPage for mea...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/987 Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2159/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #971: [CARBONDATA-1015] Extract interface in data load writ...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/971 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-2.1/52/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #971: [CARBONDATA-1015] Extract interface in data load writ...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/971 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/23/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #971: [CARBONDATA-1015] Extract interface in data load writ...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/971 Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2158/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata pull request #971: [CARBONDATA-1015] Extract interface in data lo...
Github user jackylk commented on a diff in the pull request: https://github.com/apache/carbondata/pull/971#discussion_r119909916 --- Diff: core/src/main/java/org/apache/carbondata/core/util/CarbonUtil.java --- @@ -841,17 +837,14 @@ public static WriterCompressModel getValueCompressionModel( */ for (int i = 0; i < dataTypeSelected.length; i++) { // always 1 ValueEncoderMeta valueEncoderMeta = encodeMetaList.get(i); - maxValue[i] = valueEncoderMeta.getMaxValue(); - minValue[i] = valueEncoderMeta.getMinValue(); - uniqueValue[i] = valueEncoderMeta.getUniqueValue(); + max[i] = valueEncoderMeta.getMaxValue(); + min[i] = valueEncoderMeta.getMinValue(); + nonExistValue[i] = valueEncoderMeta.getUniqueValue(); decimal[i] = valueEncoderMeta.getDecimal(); type[i] = valueEncoderMeta.getType(); dataTypeSelected[i] = valueEncoderMeta.getDataTypeSelected(); } -MeasureMetaDataModel measureMetadataModel = -new MeasureMetaDataModel(minValue, maxValue, decimal, dataTypeSelected.length, uniqueValue, -type, dataTypeSelected); -return ValueCompressionUtil.getWriterCompressModel(measureMetadataModel); +return MeasurePageStatistics.build(min, max, nonExistValue, decimal, type, dataTypeSelected); --- End diff -- ok --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata pull request #971: [CARBONDATA-1015] Extract interface in data lo...
Github user jackylk commented on a diff in the pull request: https://github.com/apache/carbondata/pull/971#discussion_r119909710 --- Diff: core/src/main/java/org/apache/carbondata/core/util/CarbonMetadataUtil.java --- @@ -549,24 +562,23 @@ private static ValueEncoderMeta deserializeValueEncoderMeta(ByteBuffer byteBuffe } - private static WriterCompressModel getValueCompressionModel(ValueEncoderMeta[] encoderMetas) { -Object[] maxValue = new Object[encoderMetas.length]; -Object[] minValue = new Object[encoderMetas.length]; -int[] decimalLength = new int[encoderMetas.length]; -Object[] uniqueValue = new Object[encoderMetas.length]; -DataType[] aggType = new DataType[encoderMetas.length]; + private static MeasurePageStatistics getMeasurePageStats(ValueEncoderMeta[] encoderMetas) { +Object[] max = new Object[encoderMetas.length]; +Object[] min = new Object[encoderMetas.length]; +int[] decimal = new int[encoderMetas.length]; +Object[] nonExistValue = new Object[encoderMetas.length]; +DataType[] types = new DataType[encoderMetas.length]; byte[] dataTypeSelected = new byte[encoderMetas.length]; for (int i = 0; i < encoderMetas.length; i++) { - maxValue[i] = encoderMetas[i].getMaxValue(); - minValue[i] = encoderMetas[i].getMinValue(); - decimalLength[i] = encoderMetas[i].getDecimal(); - uniqueValue[i] = encoderMetas[i].getUniqueValue(); - aggType[i] = encoderMetas[i].getType(); + max[i] = encoderMetas[i].getMaxValue(); + min[i] = encoderMetas[i].getMinValue(); + decimal[i] = encoderMetas[i].getDecimal(); + nonExistValue[i] = encoderMetas[i].getUniqueValue(); + types[i] = encoderMetas[i].getType(); dataTypeSelected[i] = encoderMetas[i].getDataTypeSelected(); } -return ValueCompressionUtil -.getWriterCompressModel(maxValue, minValue, decimalLength, uniqueValue, aggType, -dataTypeSelected); + +return MeasurePageStatistics.build(min, max, nonExistValue, decimal, types, dataTypeSelected); --- End diff -- ok --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata pull request #971: [CARBONDATA-1015] Extract interface in data lo...
Github user jackylk commented on a diff in the pull request: https://github.com/apache/carbondata/pull/971#discussion_r119905806 --- Diff: core/src/main/java/org/apache/carbondata/core/datastore/page/statistics/MeasurePageStatistics.java --- @@ -0,0 +1,88 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one or more + * contributor license agreements. See the NOTICE file distributed with + * this work for additional information regarding copyright ownership. + * The ASF licenses this file to You under the Apache License, Version 2.0 + * (the "License"); you may not use this file except in compliance with + * the License. You may obtain a copy of the License at + * + *http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.carbondata.core.datastore.page.statistics; + +import org.apache.carbondata.core.datastore.page.ColumnPage; +import org.apache.carbondata.core.metadata.datatype.DataType; + +public class MeasurePageStatistics { --- End diff -- ok, change to `MeasurePageStatsVO` --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata pull request #971: [CARBONDATA-1015] Extract interface in data lo...
Github user jackylk commented on a diff in the pull request: https://github.com/apache/carbondata/pull/971#discussion_r119895320 --- Diff: core/src/main/java/org/apache/carbondata/core/datastore/chunk/reader/measure/v3/CompressedMeasureChunkFileBasedReaderV3.java --- @@ -220,13 +223,25 @@ public CompressedMeasureChunkFileBasedReaderV3(BlockletInfo blockletInfo, String valueEncodeMeta.add(CarbonUtil .deserializeEncoderMetaNew(measureColumnChunk.getEncoder_meta().get(i).array())); } -WriterCompressModel compressionModel = CarbonUtil.getValueCompressionModel(valueEncodeMeta); -ValueCompressionHolder values = compressionModel.getValueCompressionHolder()[0]; + +MeasurePageStatistics stats = CarbonUtil.getMeasurePageStats(valueEncodeMeta); --- End diff -- This code is repeated in CompressedMeasureChunkFileBasedReaderV2 and CompressedMeasureChunkFileBasedReaderV3, if we move it to a shared function, a result object need to be created, since multiple results are needed (stats, convertedType, values). Actually in #987 , usage of `CompressionFinder` is removed in write path, I think another PR is needed to remove it for read path also, after this is done, `CompressionFinder` related class and the whole `org.apache.carbondata.core.datastore.compression` package. --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #962: Only for test ci(vacuous PR-don't merge)
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/962 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/22/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata pull request #971: [CARBONDATA-1015] Extract interface in data lo...
Github user jackylk commented on a diff in the pull request: https://github.com/apache/carbondata/pull/971#discussion_r119889985 --- Diff: core/src/test/java/org/apache/carbondata/core/datastore/chunk/reader/measure/CompressedMeasureChunkFileBasedReaderTest.java --- @@ -1,90 +0,0 @@ -/* --- End diff -- ok. And I will remove `CompressedDimensionChunkFileBasedReaderTest` also --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #962: Only for test ci(vacuous PR-don't merge)
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/962 Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2157/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #962: Only for test ci(vacuous PR-don't merge)
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/962 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-2.1/51/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata pull request #971: [CARBONDATA-1015] Extract interface in data lo...
Github user jackylk commented on a diff in the pull request: https://github.com/apache/carbondata/pull/971#discussion_r119888798 --- Diff: core/src/main/java/org/apache/carbondata/core/util/ByteUtil.java --- @@ -670,4 +671,23 @@ public static int putBytes(byte[] tgtBytes, int tgtOffset, byte[] srcBytes, int System.arraycopy(srcBytes, srcOffset, tgtBytes, tgtOffset, srcLength); return tgtOffset + srcLength; } + + /** + * flatten the byte[][] to byte[] and return data after applying compression by compressor + * @param compressor compressor to use + * @return compressed data + */ + public static byte[] flattenAndCompress(Compressor compressor, byte[][] byteArrayData) { --- End diff -- ok --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata pull request #971: [CARBONDATA-1015] Extract interface in data lo...
Github user jackylk commented on a diff in the pull request: https://github.com/apache/carbondata/pull/971#discussion_r119887960 --- Diff: docs/useful-tips-on-carbondata.md --- @@ -127,7 +127,7 @@ query performance. The create table command can be modified as below : TBLPROPERTIES ( 'DICTIONARY_EXCLUDE'='MSISDN,HOST,IMSI', 'DICTIONARY_INCLUDE'='Dime_1,END_TIME,BEGIN_TIME'); ``` - The result of performance analysis of test-case shows reduction in query execution time from 15 to 3 seconds, thereby improving performance by nearly 5 times. + The encodedData of performance analysis of test-case shows reduction in query execution time from 15 to 3 seconds, thereby improving performance by nearly 5 times. --- End diff -- ok --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata pull request #971: [CARBONDATA-1015] Extract interface in data lo...
Github user jackylk commented on a diff in the pull request: https://github.com/apache/carbondata/pull/971#discussion_r119887844 --- Diff: docs/release-guide.md --- @@ -109,7 +109,7 @@ staging repository and promote the artifacts to Maven Central. 4. Choose `User Token` from the dropdown, then click `Access User Token`. Copy a snippet of the Maven XML configuration block. 5. Insert this snippet twice into your global Maven `settings.xml` file, typically `${HOME]/ -.m2/settings.xml`. The end result should look like this, where `TOKEN_NAME` and `TOKEN_PASSWORD` +.m2/settings.xml`. The end encodedData should look like this, where `TOKEN_NAME` and `TOKEN_PASSWORD` --- End diff -- ok --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #962: Only for test ci(vacuous PR-don't merge)
Github user chenliang613 commented on the issue: https://github.com/apache/carbondata/pull/962 retest this please --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata pull request #971: [CARBONDATA-1015] Extract interface in data lo...
Github user jackylk commented on a diff in the pull request: https://github.com/apache/carbondata/pull/971#discussion_r119887721 --- Diff: docs/faq.md --- @@ -123,7 +123,7 @@ id cityname 3 davishenzhen ``` -As result shows, the second column is city in carbon table, but what inside is name, such as jack. This phenomenon is same with insert data into hive table. +As encodedData shows, the second column is city in carbon table, but what inside is name, such as jack. This phenomenon is same with insert data into hive table. --- End diff -- ok --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata pull request #971: [CARBONDATA-1015] Extract interface in data lo...
Github user jackylk commented on a diff in the pull request: https://github.com/apache/carbondata/pull/971#discussion_r119887697 --- Diff: docs/dml-operation-on-carbondata.md --- @@ -211,7 +211,7 @@ By default the above configuration will be false. ### Examples ``` -INSERT INTO table1 SELECT item1 ,sum(item2 + 1000) as result FROM +INSERT INTO table1 SELECT item1 ,sum(item2 + 1000) as encodedData FROM --- End diff -- ok --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #962: Only for test ci(vacuous PR-don't merge)
Github user chenliang613 commented on the issue: https://github.com/apache/carbondata/pull/962 text --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata pull request #971: [CARBONDATA-1015] Extract interface in data lo...
Github user jackylk commented on a diff in the pull request: https://github.com/apache/carbondata/pull/971#discussion_r119887373 --- Diff: dev/scalastyle-config.xml --- @@ -193,12 +193,12 @@ This file is divided into 3 sections: - Await\.result + Await\.encodedData --- End diff -- ok --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata pull request #971: [CARBONDATA-1015] Extract interface in data lo...
Github user jackylk commented on a diff in the pull request: https://github.com/apache/carbondata/pull/971#discussion_r119887399 --- Diff: LICENSE --- @@ -157,7 +157,7 @@ negligent acts) or agreed to in writing, shall any Contributor be liable to You for damages, including any direct, indirect, special, incidental, or consequential damages of any character arising as a - result of this License or out of the use or inability to use the + encodedData of this License or out of the use or inability to use the --- End diff -- ok --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata pull request #971: [CARBONDATA-1015] Extract interface in data lo...
Github user jackylk commented on a diff in the pull request: https://github.com/apache/carbondata/pull/971#discussion_r119886685 --- Diff: core/src/main/java/org/apache/carbondata/core/datastore/page/statistics/ColumnPageStatistics.java --- @@ -114,6 +115,46 @@ private int getDecimalCount(double value) { return decimalPlaces; } + /** + * return min value as byte array + */ + public byte[] minBytes() { +return getValueAsBytes(getMin()); + } + + /** + * return max value as byte array + */ + public byte[] maxBytes() { +return getValueAsBytes(getMax()); + } + + /** + * convert value to byte array + */ + private byte[] getValueAsBytes(Object value) { +ByteBuffer b = null; +Object max = getMax(); --- End diff -- ok --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata pull request #971: [CARBONDATA-1015] Extract interface in data lo...
Github user jackylk commented on a diff in the pull request: https://github.com/apache/carbondata/pull/971#discussion_r119886349 --- Diff: core/src/main/java/org/apache/carbondata/core/datastore/page/statistics/ColumnPageStatistics.java --- @@ -33,30 +34,30 @@ * the unique value is the non-exist value in the row, * and will be used as storage key for null values of measures */ - private Object uniqueValue; + private Object nonExistValue; --- End diff -- But it will be written into file by `CarbonMetadataUtil.serializeEncodeMetaUsingByteBuffer`, you mean read path is not using it? --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata pull request #971: [CARBONDATA-1015] Extract interface in data lo...
Github user jackylk commented on a diff in the pull request: https://github.com/apache/carbondata/pull/971#discussion_r119885662 --- Diff: core/src/main/java/org/apache/carbondata/core/datastore/page/ComplexColumnPage.java --- @@ -74,4 +75,70 @@ public void putComplexData(int rowId, int depth, Listvalue) { public int getDepth() { return depth; } + + @Override + public void putInt(int rowId, int value) { --- End diff -- How about add default implementation in ColumnPage and override the one that is needed in ComplexColumnPage --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata pull request #971: [CARBONDATA-1015] Extract interface in data lo...
Github user jackylk commented on a diff in the pull request: https://github.com/apache/carbondata/pull/971#discussion_r119885006 --- Diff: processing/src/main/java/org/apache/carbondata/processing/store/TablePage.java --- @@ -65,20 +68,20 @@ this.pageSize = pageSize; keyColumnPage = new KeyColumnPage(pageSize, model.getSegmentProperties().getDimensionPartitions().length); -noDictDimensionPage = new VarLengthColumnPage[model.getNoDictionaryCount()]; +noDictDimensionPage = new PrimitiveColumnPage[model.getNoDictionaryCount()]; for (int i = 0; i < noDictDimensionPage.length; i++) { - noDictDimensionPage[i] = new VarLengthColumnPage(pageSize); + noDictDimensionPage[i] = new PrimitiveColumnPage(DataType.STRING, pageSize); --- End diff -- If create another class for string and decimal, then it will be more complex for measure handling. And actually in future I think all simple column handling should be similar, so better to keep them in one class. I think a better class name can solve your comment? How about SimpleColumnPage and ComplexColumnPage? --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #987: [WIP] Add EncoderStrategy and use ColumnPage for meas...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/987 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-2.1/50/Failed Tests: 7carbondata-pr-spark-2.1/org.apache.carbondata:carbondata-spark-common-test: 7org.apache.carbondata.integration.spark.testsuite.dataload.TestLoadDataWithMaxMinInteger.test carbon table data loading when the int column contains min integer valueorg.apache.carbondata.integration.spark.testsuite.dataload.TestLoadDataWithSinglePass.test data loading use one passorg.apache.carbondata.integration.spark.testsuite.dataload.TestLoadDataWithSinglePass.test data loading use one pass when offer column dictionary fileorg.apache.carbondata.integration.spark.testsuite.dataload.TestLoadDataWithSinglePass.test data loading use one pass when do incremental loadorg.apache.carbondata.spark.testsuite.bigdecimal.TestBigInt.test big int data type storage for boundary valuesorg.apache.carbondata.spark.testsuite.detailquery.ValueCompressionDataType TestCase.When the values of Double datatype are negative valuesorg.apache.carbondata.spark.testsuite.detailquery.ValueCompressionDataTypeTestCase.When the values of Double datatype have both postive and negative values --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #987: [WIP] Add EncoderStrategy and use ColumnPage for meas...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/987 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/21/Failed Tests: 7carbondata-pr-spark-1.6/org.apache.carbondata:carbondata-spark-common-test: 7org.apache.carbondata.integration.spark.testsuite.dataload.TestLoadDataWithMaxMinInteger.test carbon table data loading when the int column contains min integer valueorg.apache.carbondata.integration.spark.testsuite.dataload.TestLoadDataWithSinglePass.test data loading use one passorg.apache.carbondata.integration.spark.testsuite.dataload.TestLoadDataWithSinglePass.test data loading use one pass when offer column dictionary fileorg.apache.carbondata.integration.spark.testsuite.dataload.TestLoadDataWithSinglePass.test data loading use one pass when do incremental loadorg.apache.carbondata.spark.testsuite.bigdecimal.TestBigInt.test big int data type storage for boundary valuesorg.apache.carbondata.spark.testsuite.detailquery.ValueCompressionDataType TestCase.When the values of Double datatype are negative valuesorg.apache.carbondata.spark.testsuite.detailquery.ValueCompressionDataTypeTestCase.When the values of Double datatype have both postive and negative values --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #987: [WIP] Add EncoderStrategy and use ColumnPage for meas...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/987 Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2156/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #955: [CARBONDATA-1118] Inset Filter Pushdown in 2.1
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/955 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/20/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #955: [CARBONDATA-1118] Inset Filter Pushdown in 2.1
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/955 Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2155/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #983: [CARBONDATA-1113] Added validation for partition colu...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/983 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/17/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #966: [CARBONDATA-1101] Avoid widening between wrapper clas...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/966 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/18/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #981: [CARBONDATA-1111]Improve No dictionary column Include...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/981 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-2.1/48/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #966: [CARBONDATA-1101] Avoid widening between wrapper clas...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/966 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-2.1/47/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #985: [CARBONDATA-1090] added integration test cases for al...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/985 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/16/Build result: FAILURE[...truncated 990.16 KB...][JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6@2/processing/pom.xml to org.apache.carbondata/carbondata-processing/1.2.0-SNAPSHOT/carbondata-processing-1.2.0-SNAPSHOT.pomNo artifacts from carbondata-pr-spark-1.6 » Apache CarbonData :: Processing #19 to compare, so performing full copy of artifacts[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6@2/integration/spark-common/pom.xml to org.apache.carbondata/carbondata-spark-common/1.2.0-SNAPSHOT/carbondata-spark-common-1.2.0-SNAPSHOT.pomNo artifacts from carbondata-pr-spark-1.6 » Apache CarbonData :: Spark Common #19 to compare, so performing full copy of artifacts[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6@2/examples/spark/pom.xml to org.apache.carbondata/carbondata-examples-spark/1.2.0-SNAPSHOT/carbondata-e xamples-spark-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6@2/integration/hive/pom.xml to org.apache.carbondata/carbondata-hive/1.2.0-SNAPSHOT/carbondata-hive-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6@2/common/pom.xml to org.apache.carbondata/carbondata-common/1.2.0-SNAPSHOT/carbondata-common-1.2.0-SNAPSHOT.pomNo artifacts from carbondata-pr-spark-1.6 » Apache CarbonData :: Common #20 to compare, so performing full copy of artifacts[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6@2/assembly/pom.xml to org.apache.carbondata/carbondata-assembly/1.2.0-SNAPSHOT/carbondata-assembly-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6@2/pom.xml to org.apache.carbondata/carbondata-parent/1.2.0-SNAPSHOT/carbondata-parent-1.2.0-SNAPSHOT.pomNo artifacts from carbondata-pr-spark-1.6 » Apache CarbonData :: Parent #20 to compare, so performing full copy of artifacts[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6@2/examples/flink/pom.xml to org.apache.carbondata/carbondata-examples-flink/1.2.0-SNAPSHOT/carbondata-examples-flink-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6@2/core/pom.xml to org.apache.carbondata/carbondata-core/1.2.0-SNAPSHOT/carbondata-core-1.2.0-SNAPSHOT.pomNo artifacts from carbondata-pr-spark-1.6 » Apache CarbonData :: Core #19 to compare, so performing full copy of artifacts[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6@2/hadoop/pom.xml to org.apache.carbondata/carbondata-hadoop/1.2.0-SNAPSHOT/carbondata-hadoop-1.2.0-SNAPSHOT.pomNo artifacts from carbondata-pr-spark-1.6 » Apache CarbonData :: Hadoop #19 to compare, so performing full copy of artifacts[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6@2/integr ation/presto/pom.xml to org.apache.carbondata/carbondata-presto/1.2.0-SNAPSHOT/carbondata-presto-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6@2/integration/spark-common-test/pom.xml to org.apache.carbondata/carbondata-spark-common-test/1.2.0-SNAPSHOT/carbondata-spark-common-test-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6@2/integration/spark/pom.xml to org.apache.carbondata/carbondata-spark/1.2.0-SNAPSHOT/carbondata-spark-1.2.0-SNAPSHOT.pomNo artifacts from carbondata-pr-spark-1.6 » Apache CarbonData :: Spark #19 to compare, so performing full copy of artifactschannel stoppedSetting status of 6dc4c994bf0f2baf6c078f8d237ee955e861ae8d to FAILURE with url https://builds.apache.org/job/carbondata-pr-spark-1.6/16/ and message: 'Tests Failed for Spark1.6 'Using context: Jenkins(Spark1.6): mvn clean test -Pspark-1.6 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #955: [CARBONDATA-1118] Inset Filter Pushdown in 2.1
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/955 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-2.1/49/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #982: [CARBONDATA-1112] fix failing IUD testcase
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/982 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-2.1/44/Failed Tests: 2carbondata-pr-spark-2.1/org.apache.carbondata:carbondata-core: 2org.apache.carbondata.core.writer.CarbonFooterWriterTest.testWriteFactMetadataorg.apache.carbondata.core.writer.CarbonFooterWriterTest.te stReadFactMetadata --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #982: [CARBONDATA-1112] fix failing IUD testcase
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/982 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/15/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #982: [CARBONDATA-1112] fix failing IUD testcase
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/982 Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2151/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[jira] [Created] (CARBONDATA-1118) Inset Pushdown in Carbondata.
sounak chakraborty created CARBONDATA-1118: -- Summary: Inset Pushdown in Carbondata. Key: CARBONDATA-1118 URL: https://issues.apache.org/jira/browse/CARBONDATA-1118 Project: CarbonData Issue Type: Bug Components: core Affects Versions: 1.1.0 Reporter: sounak chakraborty PushDown Inset Filter in carbon Layer. -- This message was sent by Atlassian JIRA (v6.3.15#6346)
[GitHub] carbondata issue #983: [CARBONDATA-1113] Added validation for partition colu...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/983 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-2.1/46/Build result: FAILURE[...truncated 81.97 KB...][JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-2.1/processing/pom.xml to org.apache.carbondata/carbondata-processing/1.2.0-SNAPSHOT/carbondata-processing-1.2.0-SNAPSHOT.pomNo artifacts from carbondata-pr-spark-2.1 » Apache CarbonData :: Processing #47 to compare, so performing full copy of artifacts[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-2.1/integration/spark-common/pom.xml to org.apache.carbondata/carbondata-spark-common/1.2.0-SNAPSHOT/carbondata-spark-common-1.2.0-SNAPSHOT.pomNo artifacts from carbondata-pr-spark-2.1 » Apache CarbonData :: Spark Common #45 to compare, so performing full copy of artifacts[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-2.1/integration/hive/pom.xml to org.apache.carbondata/carbondata-hive/1.2.0-SNAPSHOT/carbondata-hive-1.2.0-SNAPS HOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-2.1/common/pom.xml to org.apache.carbondata/carbondata-common/1.2.0-SNAPSHOT/carbondata-common-1.2.0-SNAPSHOT.pomNo artifacts from carbondata-pr-spark-2.1 » Apache CarbonData :: Common #48 to compare, so performing full copy of artifacts[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-2.1/examples/spark2/pom.xml to org.apache.carbondata/carbondata-examples-spark2/1.2.0-SNAPSHOT/carbondata-examples-spark2-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-2.1/integration/spark2/pom.xml to org.apache.carbondata/carbondata-spark2/1.2.0-SNAPSHOT/carbondata-spark2-1.2.0-SNAPSHOT.pomNo artifacts from carbondata-pr-spark-2.1 » Apache CarbonData :: Spark2 #45 to compare, so performing full copy of artifacts[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-2.1/assembly/pom.xml to org.apache.carbondata/carb ondata-assembly/1.2.0-SNAPSHOT/carbondata-assembly-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-2.1/pom.xml to org.apache.carbondata/carbondata-parent/1.2.0-SNAPSHOT/carbondata-parent-1.2.0-SNAPSHOT.pomNo artifacts from carbondata-pr-spark-2.1 » Apache CarbonData :: Parent #48 to compare, so performing full copy of artifacts[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-2.1/examples/flink/pom.xml to org.apache.carbondata/carbondata-examples-flink/1.2.0-SNAPSHOT/carbondata-examples-flink-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-2.1/core/pom.xml to org.apache.carbondata/carbondata-core/1.2.0-SNAPSHOT/carbondata-core-1.2.0-SNAPSHOT.pomNo artifacts from carbondata-pr-spark-2.1 » Apache CarbonData :: Core #47 to compare, so performing full copy of artifacts[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-2.1/hadoop/pom.xml to org.apache.carbondata/carbondata-hadoop/1.2.0-SNAPSHOT/carbondata-hadoop-1.2.0-SNAPSHOT.pomNo artifacts from carbondata-pr-spark-2.1 » Apache CarbonData :: Hadoop #47 to compare, so performing full copy of artifacts[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-2.1/integration/presto/pom.xml to org.apache.carbondata/carbondata-presto/1.2.0-SNAPSHOT/carbondata-presto-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-2.1/integration/spark-common-test/pom.xml to org.apache.carbondata/carbondata-spark-common-test/1.2.0-SNAPSHOT/carbondata-spark-common-test-1.2.0-SNAPSHOT.pomchannel stoppedSetting status of da36939a9f8718e7d8604c0333216b541c5869d5 to FAILURE with url https://builds.apache.org/job/carbondata-pr-spark-2.1/46/ and message: 'Tests Failed for Spark2.1 'Using context: Jenkins(Spark2.1): mvn clean test -Pspark-2.1 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #985: [CARBONDATA-1090] added integration test cases for al...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/985 Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2150/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #955: [WIP] Inset Filter Pushdown in 2.1
Github user ravipesala commented on the issue: https://github.com/apache/carbondata/pull/955 retest this please --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #988: [CARBONDATA-1110] put if clause out of the for clause
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/988 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-2.1/42/Failed Tests: 1carbondata-pr-spark-2.1/org.apache.carbondata:carbondata-spark-common-test: 1org.apache.carbondata.spark.testsuite.allqueries.InsertIntoCarbonTableTestCase.insert into carbon table from carbon table union query --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #988: [CARBONDATA-1110] put if clause out of the for clause
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/988 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/13/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #981: [CARBONDATA-1111]Improve No dictionary column Include...
Github user chenliang613 commented on the issue: https://github.com/apache/carbondata/pull/981 retest this please --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #966: [CARBONDATA-1101] Avoid widening between wrapper clas...
Github user chenliang613 commented on the issue: https://github.com/apache/carbondata/pull/966 retest this please --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #983: [CARBONDATA-1113] Added validation for partition colu...
Github user chenliang613 commented on the issue: https://github.com/apache/carbondata/pull/983 retest this please --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #985: [CARBONDATA-1090] added integration test cases for al...
Github user chenliang613 commented on the issue: https://github.com/apache/carbondata/pull/985 retest this please --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #987: [WIP] Add EncoderStrategy and use ColumnPage for meas...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/987 Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2149/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #987: [WIP] Add EncoderStrategy and use ColumnPage for meas...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/987 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/14/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #987: [WIP] Add EncoderStrategy and use ColumnPage for meas...
Github user chenliang613 commented on the issue: https://github.com/apache/carbondata/pull/987 retest this please --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #988: [CARBONDATA-1110] put if clause out of the for clause
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/988 Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2148/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #988: [CARBONDATA-1110] put if clause out of the for clause
Github user jbonofre commented on the issue: https://github.com/apache/carbondata/pull/988 retest this please --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[jira] [Assigned] (CARBONDATA-1052) Complete result does not display in Sort_column table.
[ https://issues.apache.org/jira/browse/CARBONDATA-1052?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Pallavi Singh reassigned CARBONDATA-1052: - Assignee: Pallavi Singh > Complete result does not display in Sort_column table. > -- > > Key: CARBONDATA-1052 > URL: https://issues.apache.org/jira/browse/CARBONDATA-1052 > Project: CarbonData > Issue Type: Bug > Components: data-load > Environment: Spark 2.1 >Reporter: Vinod Rohilla >Assignee: Pallavi Singh >Priority: Minor > Attachments: newdata.csv > > > Incomplete result displays to the user while select query. > Steps to reproduces: > 1: Create table: > CREATE TABLE sorttable4_offheap_safe (empno int, empname String, designation > String, doj Timestamp, workgroupcategory int, workgroupcategoryname String, > deptno int, deptname String, projectcode int, projectjoindate Timestamp, > projectenddate Timestamp,attendance int,utilization int,salary int) STORED BY > 'org.apache.carbondata.format' > tblproperties('sort_columns'='workgroupcategory, empname'); > 2: Load Data > LOAD DATA local inpath 'hdfs://localhost:54310/newdata.csv' INTO TABLE > sorttable4_offheap_safe OPTIONS('DELIMITER'= ',', 'QUOTECHAR'= '\"'); > 3: Select Query & result set: > 0: jdbc:hive2://localhost:1> Select * from sorttable4_offheap_safe; > ++---+---++++-+---+--+++-+--+-+--+ > | empno | empname | designation >| doj | workgroupcategory | workgroupcategoryname > | deptno | deptname| > projectcode |projectjoindate | projectenddate | attendance > | utilization | salary | > ++---+---++++-+---+--+++-+--+-+--+ > | 15012 | Robert| Component Engineer >| 1980-06-29 01:31:41.0 | 2067 | NM > | 10963 | Component Engineer| > 121 | 2015-07-13 17:34:48.0 | 2016-09-01 09:13:53.0 | 33691 > | 6341805 | 66463 | > | 15013 | Richard | Conference Coordinator >| 2010-05-27 13:04:06.0 | 10408 | AL > | 76341 | Conference Coordinator| > 131 | 2015-07-15 07:56:36.0 | 2016-11-25 13:14:17.0 | 61992 > | 58731| 59651 | > | 15027 | Jonathon | Flight Nurse >| 1999-05-14 09:17:19.0 | 10635 | MA > | 14845 | Flight Nurse | > 271 | 2015-09-29 23:19:06.0 | 2016-04-28 19:55:40.0 | 57908 > | 161869 | 78633 | > | 15010 | Sue | Clinical Reviewer >| 2004-01-02 20:04:47.0 | 11623 | IA > | 34148 | Clinical Reviewer | > 101 | 2015-09-16 09:49:18.0 | 2016-03-24 06:31:12.0 | 46362 > | 6696935 | 32713 | > | 15011 | Carlos| Commercial Print Management Consultant >| 1993-07-03 04:17:06.0 | 20091 | RI > | 32071 | Commercial Print Management Consultant| > 111 | 2015-07-09 04:04:15.0 | 2016-03-13 14:42:45.0 | 19847 > | 6654411 | 18747 | > | 15028 | Daniell | Front Desk Agent >| 1981-04-19 13:14:14.0 | 20192 | PR > | 31469 | Front Desk Agent | > 281 | 2015-03-08 22:12:39.0 | 2016-12-20 11:59:57.0 | 98291 > | 4566010 | 61481 | > | 15020 | Kathryn | Dispatcher >| 1981-08-21 04:34:26.0 | 21670 | ME > | 2543| Dispatcher| > 201 | 2015-11-28 11:17:36.0 | 2016-02-11 23:25:53.0 | 69271 > | 5991103
[GitHub] carbondata pull request #966: [CARBONDATA-1101] Avoid widening between wrapp...
Github user ravipesala commented on a diff in the pull request: https://github.com/apache/carbondata/pull/966#discussion_r119805770 --- Diff: processing/src/main/java/org/apache/carbondata/processing/newflow/sort/unsafe/merger/UnsafeIntermediateFileMerger.java --- @@ -314,7 +314,7 @@ private void writeDataTofile(Object[] row) throws CarbonSortKeyAndGroupByExcepti case SHORT: case INT: case LONG: -rowData.putLong(size, (Long) value); +rowData.putLong(size, Long.valueOf(value.toString())); --- End diff -- Should be kept in the respective datatypes, it is already fixed. --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata pull request #971: [CARBONDATA-1015] Extract interface in data lo...
Github user ravipesala commented on a diff in the pull request: https://github.com/apache/carbondata/pull/971#discussion_r119804061 --- Diff: dev/findbugs-exclude.xml --- @@ -143,7 +143,7 @@ This method returns a value that is not checked. The return value should be checked since it can indicate an unusual or unexpected function execution. For example, the File.delete() method returns false if the file could not be successfully deleted - (rather than throwing an Exception). If you don't check the result, you won't notice + (rather than throwing an Exception). If you don't check the encodedData, you won't notice --- End diff -- I think added by mistake --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata pull request #971: [CARBONDATA-1015] Extract interface in data lo...
Github user ravipesala commented on a diff in the pull request: https://github.com/apache/carbondata/pull/971#discussion_r119803990 --- Diff: core/src/test/java/org/apache/carbondata/core/datastore/chunk/reader/measure/CompressedMeasureChunkFileBasedReaderTest.java --- @@ -1,90 +0,0 @@ -/* --- End diff -- Better remove the file --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata pull request #971: [CARBONDATA-1015] Extract interface in data lo...
Github user ravipesala commented on a diff in the pull request: https://github.com/apache/carbondata/pull/971#discussion_r119802768 --- Diff: core/src/main/java/org/apache/carbondata/core/util/CarbonUtil.java --- @@ -841,17 +837,14 @@ public static WriterCompressModel getValueCompressionModel( */ for (int i = 0; i < dataTypeSelected.length; i++) { // always 1 ValueEncoderMeta valueEncoderMeta = encodeMetaList.get(i); - maxValue[i] = valueEncoderMeta.getMaxValue(); - minValue[i] = valueEncoderMeta.getMinValue(); - uniqueValue[i] = valueEncoderMeta.getUniqueValue(); + max[i] = valueEncoderMeta.getMaxValue(); + min[i] = valueEncoderMeta.getMinValue(); + nonExistValue[i] = valueEncoderMeta.getUniqueValue(); decimal[i] = valueEncoderMeta.getDecimal(); type[i] = valueEncoderMeta.getType(); dataTypeSelected[i] = valueEncoderMeta.getDataTypeSelected(); } -MeasureMetaDataModel measureMetadataModel = -new MeasureMetaDataModel(minValue, maxValue, decimal, dataTypeSelected.length, uniqueValue, -type, dataTypeSelected); -return ValueCompressionUtil.getWriterCompressModel(measureMetadataModel); +return MeasurePageStatistics.build(min, max, nonExistValue, decimal, type, dataTypeSelected); --- End diff -- Better pass `ValueEncoderMeta` list to `build` method to avoid duplicate code in other places --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata pull request #971: [CARBONDATA-1015] Extract interface in data lo...
Github user ravipesala commented on a diff in the pull request: https://github.com/apache/carbondata/pull/971#discussion_r119802447 --- Diff: core/src/main/java/org/apache/carbondata/core/util/CarbonMetadataUtil.java --- @@ -549,24 +562,23 @@ private static ValueEncoderMeta deserializeValueEncoderMeta(ByteBuffer byteBuffe } - private static WriterCompressModel getValueCompressionModel(ValueEncoderMeta[] encoderMetas) { -Object[] maxValue = new Object[encoderMetas.length]; -Object[] minValue = new Object[encoderMetas.length]; -int[] decimalLength = new int[encoderMetas.length]; -Object[] uniqueValue = new Object[encoderMetas.length]; -DataType[] aggType = new DataType[encoderMetas.length]; + private static MeasurePageStatistics getMeasurePageStats(ValueEncoderMeta[] encoderMetas) { +Object[] max = new Object[encoderMetas.length]; +Object[] min = new Object[encoderMetas.length]; +int[] decimal = new int[encoderMetas.length]; +Object[] nonExistValue = new Object[encoderMetas.length]; +DataType[] types = new DataType[encoderMetas.length]; byte[] dataTypeSelected = new byte[encoderMetas.length]; for (int i = 0; i < encoderMetas.length; i++) { - maxValue[i] = encoderMetas[i].getMaxValue(); - minValue[i] = encoderMetas[i].getMinValue(); - decimalLength[i] = encoderMetas[i].getDecimal(); - uniqueValue[i] = encoderMetas[i].getUniqueValue(); - aggType[i] = encoderMetas[i].getType(); + max[i] = encoderMetas[i].getMaxValue(); + min[i] = encoderMetas[i].getMinValue(); + decimal[i] = encoderMetas[i].getDecimal(); + nonExistValue[i] = encoderMetas[i].getUniqueValue(); + types[i] = encoderMetas[i].getType(); dataTypeSelected[i] = encoderMetas[i].getDataTypeSelected(); } -return ValueCompressionUtil -.getWriterCompressModel(maxValue, minValue, decimalLength, uniqueValue, aggType, -dataTypeSelected); + +return MeasurePageStatistics.build(min, max, nonExistValue, decimal, types, dataTypeSelected); --- End diff -- Better pass `ValueEncoderMeta` list to `build` method to avoid duplicate code in other places --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata pull request #971: [CARBONDATA-1015] Extract interface in data lo...
Github user ravipesala commented on a diff in the pull request: https://github.com/apache/carbondata/pull/971#discussion_r119801431 --- Diff: core/src/main/java/org/apache/carbondata/core/util/ByteUtil.java --- @@ -670,4 +671,23 @@ public static int putBytes(byte[] tgtBytes, int tgtOffset, byte[] srcBytes, int System.arraycopy(srcBytes, srcOffset, tgtBytes, tgtOffset, srcLength); return tgtOffset + srcLength; } + + /** + * flatten the byte[][] to byte[] and return data after applying compression by compressor + * @param compressor compressor to use + * @return compressed data + */ + public static byte[] flattenAndCompress(Compressor compressor, byte[][] byteArrayData) { --- End diff -- Since it is ByteUtil, better use only flatten not compress, compress can be done in caller method --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[jira] [Resolved] (CARBONDATA-1109) Page lost in load process when last page is not be consumed at the end
[ https://issues.apache.org/jira/browse/CARBONDATA-1109?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Ravindra Pesala resolved CARBONDATA-1109. - Resolution: Fixed Assignee: Yadong Qi Fix Version/s: 1.1.1 1.2.0 > Page lost in load process when last page is not be consumed at the end > -- > > Key: CARBONDATA-1109 > URL: https://issues.apache.org/jira/browse/CARBONDATA-1109 > Project: CarbonData > Issue Type: Bug > Components: data-load >Affects Versions: 1.2.0 >Reporter: Yadong Qi >Assignee: Yadong Qi > Fix For: 1.2.0, 1.1.1 > > Time Spent: 3.5h > Remaining Estimate: 0h > > First, we use Producer-Consumer model in the write step, we have n(default > value is 2 and can be configured) producers and one consumer. The task of > generate last page(less than 32000) is added to thread pool at the end, but > can't be guaranteed to be finished and add to BlockletDataHolder at the end. > Because we have n tasks running concurrently. > Second, we have 2 ways to invoke `writeDataToFile`, one is the size of > `DataWriterHolder` reach the size of blocklet and two is the page is the last > page. > So if the last page is not be consumed at the end, we lost the page which be > consumed after last page. -- This message was sent by Atlassian JIRA (v6.3.15#6346)
[GitHub] carbondata pull request #978: [CARBONDATA-1109] Acquire semaphore before sub...
Github user asfgit closed the pull request at: https://github.com/apache/carbondata/pull/978 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #978: [CARBONDATA-1109] Acquire semaphore before submit a p...
Github user ravipesala commented on the issue: https://github.com/apache/carbondata/pull/978 @manishgupta88 It was the case in V1 and V2 formats, but in V3 format we should launch producer at finish to flush the older pages it was holding. --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #978: [CARBONDATA-1109] Acquire semaphore before submit a p...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/978 Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2147/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #978: [CARBONDATA-1109] Acquire semaphore before submit a p...
Github user watermen commented on the issue: https://github.com/apache/carbondata/pull/978 retest this please --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #978: [CARBONDATA-1109] Acquire semaphore before submit a p...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/978 Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2146/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #978: [CARBONDATA-1109] Acquire semaphore before submit a p...
Github user ravipesala commented on the issue: https://github.com/apache/carbondata/pull/978 retest this please --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---