[GitHub] carbondata issue #1004: [CARBONDATA-1138] restrict duplicate column name in ...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/1004 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/305/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #1015: [CARBONDATA-1151] Update useful-tips-on-carbondata.m...
Github user sgururajshetty commented on the issue: https://github.com/apache/carbondata/pull/1015 LGTM --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #1004: [CARBONDATA-1138] restrict duplicate column name in ...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1004 Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2426/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #1025: [CARBONDATA-1165] solved bug for class cast exceptio...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/1025 Can one of the admins verify this patch? --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata pull request #1025: [CARBONDATA-1165] solved bug for class cast e...
GitHub user anubhav100 opened a pull request: https://github.com/apache/carbondata/pull/1025 [CARBONDATA-1165] solved bug for class cast exception in intermediate file merger You can merge this pull request into a Git repository by running: $ git pull https://github.com/anubhav100/incubator-carbondata CARBONDATA-1165 Alternatively you can review and apply these changes as the patch at: https://github.com/apache/carbondata/pull/1025.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #1025 commit 9ca5e37bcf006abde8db5041a436697fd8ca8aa0 Author: anubhav100Date: 2017-06-13T05:46:57Z solved bug for class cast exception in intermediate file merger --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #1024: [CARBONDATA-1064] Make column group feature deprecat...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/1024 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/304/Failed Tests: 1carbondata-pr-spark-1.6/org.apache.carbondata:carbondata-spark: 1org.apache.carbondata.spark.util.AutoHighCardinalityIdentifyTestCase.skip auto identify high cardinality column for column group --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #1024: [CARBONDATA-1064] Make column group feature deprecat...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1024 Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2425/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[jira] [Updated] (CARBONDATA-1165) Class Cast exception in intermediate file merger when loading data
[ https://issues.apache.org/jira/browse/CARBONDATA-1165?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] anubhav tarar updated CARBONDATA-1165: -- Summary: Class Cast exception in intermediate file merger when loading data (was: Class Cast excpeyion in intermediate file merger when loading data) > Class Cast exception in intermediate file merger when loading data > -- > > Key: CARBONDATA-1165 > URL: https://issues.apache.org/jira/browse/CARBONDATA-1165 > Project: CarbonData > Issue Type: Bug > Components: data-load >Affects Versions: 1.2.0 > Environment: spark2.1 >Reporter: anubhav tarar >Assignee: anubhav tarar >Priority: Trivial > > query: > spark.sql("CREATE TABLE ORDERS ( O_ORDERKEY INT ,\n O_CUSTKEY INT ,\n > O_ORDERSTATUS STRING ," + > "\n O_TOTALPRICE DECIMAL(15,2) ,\n O_ORDERDATE TIMESTAMP ,\n > O_ORDERPRIORITY STRING" + > " , \n O_CLERK STRING , \n O_SHIPPRIORITY INT ,\n O_COMMENT > STRING ) STORED BY " + > "'carbondata'") > spark.sql("LOAD DATA INPATH \"hdfs://localhost:54310/user1/orders.csv\" > INTO TABLE orders " + > "OPTIONS('DELIMITER'='|' , > 'QUOTECHAR'='\"','FILEHEADER'='O_ORDERKEY,O_CUSTKEY," + > > "O_ORDERSTATUS,O_TOTALPRICE,O_ORDERDATE,O_ORDERPRIORITY,O_CLERK,O_SHIPPRIORITY," > + > "O_COMMENT')") > logs: > java.lang.ClassCastException: java.lang.Integer cannot be cast to > java.lang.Long > at > org.apache.carbondata.processing.sortandgroupby.sortdata.IntermediateFileMerger.writeDataTofile(IntermediateFileMerger.java:347) > at > org.apache.carbondata.processing.sortandgroupby.sortdata.IntermediateFileMerger.call(IntermediateFileMerger.java:112) > at > org.apache.carbondata.processing.sortandgroupby.sortdata.IntermediateFileMerger.call(IntermediateFileMerger.java:37) > at java.util.concurrent.FutureTask.run(FutureTask.java:266) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) > at java.lang.Thread.run(Thread.java:745) -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[jira] [Created] (CARBONDATA-1165) Class Cast excpeyion in intermediate file merger when loading data
anubhav tarar created CARBONDATA-1165: - Summary: Class Cast excpeyion in intermediate file merger when loading data Key: CARBONDATA-1165 URL: https://issues.apache.org/jira/browse/CARBONDATA-1165 Project: CarbonData Issue Type: Bug Components: data-load Affects Versions: 1.2.0 Environment: spark2.1 Reporter: anubhav tarar Assignee: anubhav tarar Priority: Trivial query: spark.sql("CREATE TABLE ORDERS ( O_ORDERKEY INT ,\n O_CUSTKEY INT ,\n O_ORDERSTATUS STRING ," + "\n O_TOTALPRICE DECIMAL(15,2) ,\n O_ORDERDATE TIMESTAMP ,\n O_ORDERPRIORITY STRING" + " , \n O_CLERK STRING , \n O_SHIPPRIORITY INT ,\n O_COMMENT STRING ) STORED BY " + "'carbondata'") spark.sql("LOAD DATA INPATH \"hdfs://localhost:54310/user1/orders.csv\" INTO TABLE orders " + "OPTIONS('DELIMITER'='|' , 'QUOTECHAR'='\"','FILEHEADER'='O_ORDERKEY,O_CUSTKEY," + "O_ORDERSTATUS,O_TOTALPRICE,O_ORDERDATE,O_ORDERPRIORITY,O_CLERK,O_SHIPPRIORITY," + "O_COMMENT')") logs: java.lang.ClassCastException: java.lang.Integer cannot be cast to java.lang.Long at org.apache.carbondata.processing.sortandgroupby.sortdata.IntermediateFileMerger.writeDataTofile(IntermediateFileMerger.java:347) at org.apache.carbondata.processing.sortandgroupby.sortdata.IntermediateFileMerger.call(IntermediateFileMerger.java:112) at org.apache.carbondata.processing.sortandgroupby.sortdata.IntermediateFileMerger.call(IntermediateFileMerger.java:37) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745) -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[GitHub] carbondata pull request #1024: [CARBONDATA-1064] Make column group feature d...
GitHub user jackylk opened a pull request: https://github.com/apache/carbondata/pull/1024 [CARBONDATA-1064] Make column group feature deprecated After discussion in community (http://apache-carbondata-dev-mailing-list-archive.1130556.n5.nabble.com/About-ColumnGroup-feature-td14436.html), we conclude that column group feature will be deprecated. In this PR, column group feature is removed by throwing exception in parser You can merge this pull request into a Git repository by running: $ git pull https://github.com/jackylk/incubator-carbondata removecg Alternatively you can review and apply these changes as the patch at: https://github.com/apache/carbondata/pull/1024.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #1024 commit be81a8453f6052c61ece06f8cc670b8eea866ad0 Author: jackylkDate: 2017-06-13T05:22:01Z remove column group --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[jira] [Created] (CARBONDATA-1164) Make Column Group feature deprecated
Jacky Li created CARBONDATA-1164: Summary: Make Column Group feature deprecated Key: CARBONDATA-1164 URL: https://issues.apache.org/jira/browse/CARBONDATA-1164 Project: CarbonData Issue Type: Improvement Reporter: Jacky Li After discussion in community (http://apache-carbondata-dev-mailing-list-archive.1130556.n5.nabble.com/About-ColumnGroup-feature-td14436.html), we conclude that column group feature will be deprecated. -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[GitHub] carbondata pull request #1011: [carbondata-1149] Fix issue of mismatch type ...
Github user chenerlu closed the pull request at: https://github.com/apache/carbondata/pull/1011 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #1011: [carbondata-1149] Fix issue of mismatch type of part...
Github user chenerlu commented on the issue: https://github.com/apache/carbondata/pull/1011 Same modify to PR1023, so close it --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata pull request #987: [CARBONDATA-1017] Add EncodingStrategy and use...
Github user ravipesala commented on a diff in the pull request: https://github.com/apache/carbondata/pull/987#discussion_r121580802 --- Diff: core/src/main/java/org/apache/carbondata/core/datastore/page/encoding/ColumnPageCodec.java --- @@ -21,15 +21,28 @@ /** * Codec for a column page data, implementation should not keep state across pages, - * caller will use the same object to encode multiple pages. + * caller may use the same object to apply multiple pages. */ -public interface ColumnCodec { +public interface ColumnPageCodec { --- End diff -- We better bring RLE of mdk also to similar interfaces, may we can work on future PR. --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata pull request #987: [CARBONDATA-1017] Add EncodingStrategy and use...
Github user ravipesala commented on a diff in the pull request: https://github.com/apache/carbondata/pull/987#discussion_r121580534 --- Diff: core/src/main/java/org/apache/carbondata/core/datastore/page/encoding/AdaptiveIntegerCodec.java --- @@ -0,0 +1,213 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one or more + * contributor license agreements. See the NOTICE file distributed with + * this work for additional information regarding copyright ownership. + * The ASF licenses this file to You under the Apache License, Version 2.0 + * (the "License"); you may not use this file except in compliance with + * the License. You may obtain a copy of the License at + * + *http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.carbondata.core.datastore.page.encoding; + +import org.apache.carbondata.core.datastore.compression.Compressor; +import org.apache.carbondata.core.datastore.page.ColumnPage; +import org.apache.carbondata.core.datastore.page.LazyColumnPage; +import org.apache.carbondata.core.datastore.page.PrimitiveCodec; +import org.apache.carbondata.core.datastore.page.statistics.ColumnPageStatsVO; +import org.apache.carbondata.core.metadata.datatype.DataType; + +/** + * Codec for integer (byte, short, int, long) data type page. + * This codec will do type casting on page data to make storage minimum. + */ +class AdaptiveIntegerCodec extends AdaptiveCompressionCodec { + + private ColumnPage encodedPage; + + public static ColumnPageCodec newInstance(DataType srcDataType, DataType targetDataType, + ColumnPageStatsVO stats, Compressor compressor) { +return new AdaptiveIntegerCodec(srcDataType, targetDataType, stats, compressor); + } + + private AdaptiveIntegerCodec(DataType srcDataType, DataType targetDataType, + ColumnPageStatsVO stats, Compressor compressor) { +super(srcDataType, targetDataType, stats, compressor); + } + + @Override + public String getName() { +return "AdaptiveIntegerCodec"; + } + + @Override + public byte[] encode(ColumnPage input) { +if (srcDataType.equals(targetDataType)) { + return input.compress(compressor); +} else { + encodedPage = ColumnPage.newPage(targetDataType, input.getPageSize()); + input.encode(codec); + return encodedPage.compress(compressor); +} + } + + @Override + public ColumnPage decode(byte[] input, int offset, int length) { +if (srcDataType.equals(targetDataType)) { + return ColumnPage.decompress(compressor, targetDataType, input, offset, length); +} else { + ColumnPage page = ColumnPage.decompress(compressor, targetDataType, input, offset, length); + return LazyColumnPage.newPage(page, codec); +} + } + + private PrimitiveCodec codec = new PrimitiveCodec() { +@Override +public void encode(int rowId, byte value) { + switch (targetDataType) { +default: + assert (false); + throw new RuntimeException("internal error: " + debugInfo()); + } +} + +@Override +public void encode(int rowId, short value) { + switch (targetDataType) { +case BYTE: + encodedPage.putByte(rowId, (byte) value); + break; +default: + assert (false); --- End diff -- This assertion is not required as we are already throwing exception --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata pull request #987: [CARBONDATA-1017] Add EncodingStrategy and use...
Github user ravipesala commented on a diff in the pull request: https://github.com/apache/carbondata/pull/987#discussion_r121580390 --- Diff: core/src/main/java/org/apache/carbondata/core/datastore/page/PrimitiveCodec.java --- @@ -0,0 +1,50 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one or more + * contributor license agreements. See the NOTICE file distributed with + * this work for additional information regarding copyright ownership. + * The ASF licenses this file to You under the Apache License, Version 2.0 + * (the "License"); you may not use this file except in compliance with + * the License. You may obtain a copy of the License at + * + *http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.carbondata.core.datastore.page; + +// no operation, new value = page value +// NO_OP, +// +// // new value = (max value of page) - (page value) +// MAX_DELTA, +// +// // new value = (10 power of decimal) * (page value) +// UPSCALE, +// +// // new value = (10 power of decimal) * ((max value of page) - (page value)) +// UPSCALE_MAX_DELTA + +// Transformation type that can be applied to ColumnPage +public interface PrimitiveCodec { --- End diff -- We better bring RLE of mdk also to similar interfaces, may we can work on future PR. --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata pull request #987: [CARBONDATA-1017] Add EncodingStrategy and use...
Github user ravipesala commented on a diff in the pull request: https://github.com/apache/carbondata/pull/987#discussion_r121579950 --- Diff: core/src/main/java/org/apache/carbondata/core/datastore/page/ColumnPage.java --- @@ -46,51 +62,157 @@ // The index of the rowId whose value is null, will be set to 1 private BitSet nullBitSet; - public ColumnPage(DataType dataType, int pageSize) { -this.dataType = dataType; + protected ColumnPage(DataType dataType, int pageSize) { this.pageSize = pageSize; -this.stats = new ColumnPageStatistics(dataType); -this.nullBitSet = new BitSet(pageSize); +this.dataType = dataType; + } + + // create a new page + public static ColumnPage newPage(DataType dataType, int pageSize) { +ColumnPage instance; switch (dataType) { + case BYTE: +instance = newBytePage(new byte[pageSize]); +break; case SHORT: +instance = newShortPage(new short[pageSize]); +break; case INT: +instance = newIntPage(new int[pageSize]); +break; case LONG: -longData = new long[pageSize]; +instance = newLongPage(new long[pageSize]); +break; + case FLOAT: +instance = newFloatPage(new float[pageSize]); break; case DOUBLE: -doubleData = new double[pageSize]; +instance = newDoublePage(new double[pageSize]); break; case DECIMAL: -byteArrayData = new byte[pageSize][]; +instance = newDecimalPage(new byte[pageSize][]); break; case STRING: -byteArrayData = new byte[pageSize][]; +instance = newStringPage(new byte[pageSize][]); break; default: throw new RuntimeException("Unsupported data dataType: " + dataType); } +instance.stats = new ColumnPageStatsVO(dataType); +instance.nullBitSet = new BitSet(pageSize); +return instance; + } + + // create a new page and set data with input `pageData` + public static ColumnPage newPage(DataType dataType, Object pageData) { +// This is used in read path, since statistics and nullBitSet is not required, not creating +// them to make object minimum +switch (dataType) { + case BYTE: +byte[] byteData = (byte[]) pageData; +return newBytePage(byteData); + case SHORT: +short[] shortData = (short[]) pageData; +return newShortPage(shortData); + case INT: +int[] intData = (int[]) pageData; +return newIntPage(intData); + case LONG: +long[] longData = (long[]) pageData; +return newLongPage(longData); + case FLOAT: +float[] floatData = (float[]) pageData; +return newFloatPage(floatData); + case DOUBLE: +double[] doubleData = (double[]) pageData; +return newDoublePage(doubleData); + case DECIMAL: +byte[][] decimalData = (byte[][]) pageData; +return newDecimalPage(decimalData); + case STRING: --- End diff -- Better change this to BYTE_ARRAY and bring KeyColumnPage also here --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata pull request #987: [CARBONDATA-1017] Add EncodingStrategy and use...
Github user ravipesala commented on a diff in the pull request: https://github.com/apache/carbondata/pull/987#discussion_r121579752 --- Diff: core/src/main/java/org/apache/carbondata/core/datastore/page/ColumnPage.java --- @@ -22,22 +22,38 @@ import java.util.BitSet; import org.apache.carbondata.core.constants.CarbonCommonConstants; -import org.apache.carbondata.core.datastore.page.statistics.ColumnPageStatistics; --- End diff -- I think we should unify KeyColumnPage and ColumnPage or atleast one is derived from another. --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #1023: [carbondata-1149] Fix issue of mismatch type of part...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1023 Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2424/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #1023: [carbondata-1149] Fix issue of mismatch type of part...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/1023 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/303/Failed Tests: 3carbondata-pr-spark-1.6/org.apache.carbondata:carbondata-spark-common-test: 3org.apache.carbondata.spark.testsuite.partition.TestDDLForPartitionTable.create partition table: range partitionorg.apache.carbondata.spark.testsuite.partition.TestDDLForPartitionTable.test exception when values in list_info can not match partition column typeorg.apache.carbondata.spark.testsuite.partition.TestDDLForPartitionTable.test exception when values in range_info can not match partition column type --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #951: [CARBONDATA-1069] Fixed Compaction Failure issue on C...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/951 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/302/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #951: [CARBONDATA-1069] Fixed Compaction Failure issue on C...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/951 Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2423/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #1023: [carbondata-1149] Fix issue of mismatch type of part...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/1023 Can one of the admins verify this patch? --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #1016: [CARBONDATA-1152] List partition not working properl...
Github user QiangCai commented on the issue: https://github.com/apache/carbondata/pull/1016 @BJangir Spark 1.6 has same issue. can you check CarbonDDLSqlParser and fix it? --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata pull request #1016: [CARBONDATA-1152] List partition not working ...
Github user QiangCai commented on a diff in the pull request: https://github.com/apache/carbondata/pull/1016#discussion_r121578200 --- Diff: integration/spark2/src/main/scala/org/apache/spark/sql/parser/CarbonSparkSqlParser.scala --- @@ -173,9 +173,20 @@ class CarbonSqlAstBuilder(conf: SQLConf) extends SparkSqlAstBuilder(conf) { s"Values must be specified for key(s): ${ badKeys.mkString("[", ",", "]") }", ctx) } props.map { case (key, value) => - (key.toLowerCase, value.toLowerCase) + if (needToConvertToLowerCase(key)) { + (key.toLowerCase, value.toLowerCase) +} +else --- End diff -- code style --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #1011: [carbondata-1149] Fix issue of mismatch type of part...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/1011 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/301/Build result: FAILURE[...truncated 162.99 KB...][ERROR] Re-run Maven using the -X switch to enable full debug logging.[ERROR] [ERROR] For more information about the errors and possible solutions, please read the following articles:[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException[ERROR] [ERROR] After correcting the problems, you can resume the build with the command[ERROR] mvn -rf :carbondata-spark-common-test[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/processing/pom.xml to org.apache.carbondata/carbondata-processing/1.2.0-SNAPSHOT/carbondata-processing-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/spark-common/pom.xml to org.apache.carbondata/carbondata-spark-common/1.2.0-SNAPSHOT/carbondata-spark-common-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/ jenkins-slave/workspace/carbondata-pr-spark-1.6/examples/spark/pom.xml to org.apache.carbondata/carbondata-examples-spark/1.2.0-SNAPSHOT/carbondata-examples-spark-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/hive/pom.xml to org.apache.carbondata/carbondata-hive/1.2.0-SNAPSHOT/carbondata-hive-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/common/pom.xml to org.apache.carbondata/carbondata-common/1.2.0-SNAPSHOT/carbondata-common-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/assembly/pom.xml to org.apache.carbondata/carbondata-assembly/1.2.0-SNAPSHOT/carbondata-assembly-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/pom.xml to org.apache.carbondata/carbondata-parent/1.2.0-SNAPSHOT/carbondata-parent-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave /workspace/carbondata-pr-spark-1.6/examples/flink/pom.xml to org.apache.carbondata/carbondata-examples-flink/1.2.0-SNAPSHOT/carbondata-examples-flink-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/core/pom.xml to org.apache.carbondata/carbondata-core/1.2.0-SNAPSHOT/carbondata-core-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/hadoop/pom.xml to org.apache.carbondata/carbondata-hadoop/1.2.0-SNAPSHOT/carbondata-hadoop-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/presto/pom.xml to org.apache.carbondata/carbondata-presto/1.2.0-SNAPSHOT/carbondata-presto-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/spark-common-test/pom.xml to org.apache.carbondata/carbondata-spark-common-test/1.2.0-SNAPSHOT/carbondata-spark-common-test-1.2.0-SNAPSHOT.pom[JENKINS] Arch iving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/spark/pom.xml to org.apache.carbondata/carbondata-spark/1.2.0-SNAPSHOT/carbondata-spark-1.2.0-SNAPSHOT.pomchannel stoppedSetting status of 98284f023ce5e89671aa1df267b139f2e6213ac4 to FAILURE with url https://builds.apache.org/job/carbondata-pr-spark-1.6/301/ and message: 'Tests Failed for Spark1.6 'Using context: Jenkins(Spark1.6): mvn clean test -Pspark-1.6 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #1011: [carbondata-1149] Fix issue of mismatch type of part...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1011 Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2421/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #951: [CARBONDATA-1069] Fixed Compaction Failure issue on C...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/951 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/300/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata pull request #987: [CARBONDATA-1017] Add EncodingStrategy and use...
Github user ravipesala commented on a diff in the pull request: https://github.com/apache/carbondata/pull/987#discussion_r121577179 --- Diff: core/src/main/java/org/apache/carbondata/core/datastore/page/ColumnPage.java --- @@ -22,22 +22,38 @@ import java.util.BitSet; import org.apache.carbondata.core.constants.CarbonCommonConstants; -import org.apache.carbondata.core.datastore.page.statistics.ColumnPageStatistics; +import org.apache.carbondata.core.constants.CarbonV3DataFormatConstants; +import org.apache.carbondata.core.datastore.compression.Compressor; +import org.apache.carbondata.core.datastore.page.statistics.ColumnPageStatsVO; import org.apache.carbondata.core.metadata.datatype.DataType; +import org.apache.carbondata.core.util.ByteUtil; +import org.apache.carbondata.core.util.CarbonProperties; import org.apache.carbondata.core.util.DataTypeUtil; -// Represent a columnar data in one page for one column. +import static org.apache.carbondata.core.metadata.datatype.DataType.BYTE; +import static org.apache.carbondata.core.metadata.datatype.DataType.DECIMAL; +import static org.apache.carbondata.core.metadata.datatype.DataType.DOUBLE; +import static org.apache.carbondata.core.metadata.datatype.DataType.FLOAT; +import static org.apache.carbondata.core.metadata.datatype.DataType.INT; +import static org.apache.carbondata.core.metadata.datatype.DataType.LONG; +import static org.apache.carbondata.core.metadata.datatype.DataType.SHORT; +import static org.apache.carbondata.core.metadata.datatype.DataType.STRING; + +/** + * Represent a columnar data in one page for one column. + */ public class ColumnPage { - protected final DataType dataType; - protected final int pageSize; - protected ColumnPageStatistics stats; + private final int pageSize; + private DataType dataType; + private ColumnPageStatsVO stats; --- End diff -- I think ColumnPageStats supposed to be interface so that user can implement to have his own stats --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata pull request #1013: [CARBONDATA-1154] IUD Performance Changes
Github user asfgit closed the pull request at: https://github.com/apache/carbondata/pull/1013 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #1013: [CARBONDATA-1154] IUD Performance Changes
Github user ravipesala commented on the issue: https://github.com/apache/carbondata/pull/1013 LGTM --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #1016: [CARBONDATA-1152] List partition not working properl...
Github user ravipesala commented on the issue: https://github.com/apache/carbondata/pull/1016 @BJangir Can you add testcase for it. --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #1013: [CARBONDATA-1154] IUD Performance Changes
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1013 Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2420/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #1011: [carbondata-1149] Fix issue of mismatch type of part...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/1011 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/298/Failed Tests: 6carbondata-pr-spark-1.6/org.apache.carbondata:carbondata-spark-common-test: 6org.apache.carbondata.spark.testsuite.partition.TestDDLForPartitionTable.create partition table: range partitionorg.apache.carbondata.spark.testsuite.partition.TestDDLForPartitionTable.test exception when values in list_info can not match partition column typeorg.apache.carbondata.spark.testsuite.partition.TestDDLForPartitionTable.test exception when values in range_info can not match partition column typeorg.apache.carbondata.spark.testsuite.partition.TestDataLoadingForPartitionTable.data loading for partition table: range partitionorg.apache.carbondata.spark.testsuite.partition.TestDataLoadingForPartitionTable.Insert into for partition table: range partitionorg.apache.carbondata.spark.testsuite.partition.TestQueryForPartitionTable.detail query on partition table: ra nge partition --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #1013: [CARBONDATA-1154] IUD Performance Changes
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/1013 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/299/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #1011: [carbondata-1149] Fix issue of mismatch type of part...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1011 Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2419/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[jira] [Created] (CARBONDATA-1162) After compact segments, data sorted in global level.
Yadong Qi created CARBONDATA-1162: - Summary: After compact segments, data sorted in global level. Key: CARBONDATA-1162 URL: https://issues.apache.org/jira/browse/CARBONDATA-1162 Project: CarbonData Issue Type: Sub-task Reporter: Yadong Qi -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[GitHub] carbondata pull request #1013: [CARBONDATA-1154] IUD Performance Changes
Github user sounakr commented on a diff in the pull request: https://github.com/apache/carbondata/pull/1013#discussion_r121572395 --- Diff: hadoop/src/main/java/org/apache/carbondata/hadoop/CarbonInputFormat.java --- @@ -340,23 +340,37 @@ private static AbsoluteTableIdentifier getAbsoluteTableIdentifier(Configuration List result = new LinkedList(); FilterExpressionProcessor filterExpressionProcessor = new FilterExpressionProcessor(); +UpdateVO invalidBlockVOForSegmentId = null; +Boolean IUDTable = false; --- End diff -- Done --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[jira] [Created] (CARBONDATA-1161) Generate one index file per segment to improve the performance of first query
Yadong Qi created CARBONDATA-1161: - Summary: Generate one index file per segment to improve the performance of first query Key: CARBONDATA-1161 URL: https://issues.apache.org/jira/browse/CARBONDATA-1161 Project: CarbonData Issue Type: Sub-task Reporter: Yadong Qi -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[GitHub] carbondata issue #910: [WIP] Global sort by spark in load process
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/910 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/297/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata pull request #1013: [CARBONDATA-1154] IUD Performance Changes
Github user sounakr commented on a diff in the pull request: https://github.com/apache/carbondata/pull/1013#discussion_r121572267 --- Diff: core/src/main/java/org/apache/carbondata/core/datastore/SegmentTaskIndexStore.java --- @@ -184,22 +184,29 @@ private SegmentTaskIndexWrapper loadAndGetTaskIdToSegmentsMap( SegmentUpdateStatusManager updateStatusManager = new SegmentUpdateStatusManager(absoluteTableIdentifier); String segmentId = null; +UpdateVO updateVO = null; TaskBucketHolder taskBucketHolder = null; try { while (iteratorOverSegmentBlocksInfos.hasNext()) { +// Initialize the UpdateVO to Null for each segment. +updateVO = null; --- End diff -- Done. --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[jira] [Created] (CARBONDATA-1160) Use spark multi-threads model to load data
Yadong Qi created CARBONDATA-1160: - Summary: Use spark multi-threads model to load data Key: CARBONDATA-1160 URL: https://issues.apache.org/jira/browse/CARBONDATA-1160 Project: CarbonData Issue Type: New Feature Components: data-load Affects Versions: 1.1.0 Reporter: Yadong Qi -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[GitHub] carbondata issue #910: [WIP] Global sort by spark in load process
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/910 Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2418/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #910: [WIP] Global sort by spark in load process
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/910 Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2417/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #910: [WIP] Global sort by spark in load process
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/910 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/296/Build result: FAILURE[...truncated 52.60 KB...][ERROR] Re-run Maven using the -X switch to enable full debug logging.[ERROR] [ERROR] For more information about the errors and possible solutions, please read the following articles:[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException[ERROR] [ERROR] After correcting the problems, you can resume the build with the command[ERROR] mvn -rf :carbondata-processing[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/processing/pom.xml to org.apache.carbondata/carbondata-processing/1.2.0-SNAPSHOT/carbondata-processing-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/spark-common/pom.xml to org.apache.carbondata/carbondata-spark-common/1.2.0-SNAPSHOT/carbondata-spark-common-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins- slave/workspace/carbondata-pr-spark-1.6/examples/spark/pom.xml to org.apache.carbondata/carbondata-examples-spark/1.2.0-SNAPSHOT/carbondata-examples-spark-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/hive/pom.xml to org.apache.carbondata/carbondata-hive/1.2.0-SNAPSHOT/carbondata-hive-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/common/pom.xml to org.apache.carbondata/carbondata-common/1.2.0-SNAPSHOT/carbondata-common-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/assembly/pom.xml to org.apache.carbondata/carbondata-assembly/1.2.0-SNAPSHOT/carbondata-assembly-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/pom.xml to org.apache.carbondata/carbondata-parent/1.2.0-SNAPSHOT/carbondata-parent-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspa ce/carbondata-pr-spark-1.6/examples/flink/pom.xml to org.apache.carbondata/carbondata-examples-flink/1.2.0-SNAPSHOT/carbondata-examples-flink-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/core/pom.xml to org.apache.carbondata/carbondata-core/1.2.0-SNAPSHOT/carbondata-core-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/hadoop/pom.xml to org.apache.carbondata/carbondata-hadoop/1.2.0-SNAPSHOT/carbondata-hadoop-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/presto/pom.xml to org.apache.carbondata/carbondata-presto/1.2.0-SNAPSHOT/carbondata-presto-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/spark-common-test/pom.xml to org.apache.carbondata/carbondata-spark-common-test/1.2.0-SNAPSHOT/carbondata-spark-common-test-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /h ome/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/spark/pom.xml to org.apache.carbondata/carbondata-spark/1.2.0-SNAPSHOT/carbondata-spark-1.2.0-SNAPSHOT.pomchannel stoppedSetting status of e7da989faab09b08b82416f036d3b8a2474e978e to FAILURE with url https://builds.apache.org/job/carbondata-pr-spark-1.6/296/ and message: 'Tests Failed for Spark1.6 'Using context: Jenkins(Spark1.6): mvn clean test -Pspark-1.6 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #910: [WIP] Global sort by spark in load process
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/910 Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2416/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #910: [WIP] Global sort by spark in load process
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/910 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/295/Build result: FAILURE[...truncated 78.94 KB...][ERROR] Re-run Maven using the -X switch to enable full debug logging.[ERROR] [ERROR] For more information about the errors and possible solutions, please read the following articles:[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException[ERROR] [ERROR] After correcting the problems, you can resume the build with the command[ERROR] mvn -rf :carbondata-spark[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/processing/pom.xml to org.apache.carbondata/carbondata-processing/1.2.0-SNAPSHOT/carbondata-processing-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/spark-common/pom.xml to org.apache.carbondata/carbondata-spark-common/1.2.0-SNAPSHOT/carbondata-spark-common-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave /workspace/carbondata-pr-spark-1.6/examples/spark/pom.xml to org.apache.carbondata/carbondata-examples-spark/1.2.0-SNAPSHOT/carbondata-examples-spark-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/hive/pom.xml to org.apache.carbondata/carbondata-hive/1.2.0-SNAPSHOT/carbondata-hive-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/common/pom.xml to org.apache.carbondata/carbondata-common/1.2.0-SNAPSHOT/carbondata-common-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/assembly/pom.xml to org.apache.carbondata/carbondata-assembly/1.2.0-SNAPSHOT/carbondata-assembly-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/pom.xml to org.apache.carbondata/carbondata-parent/1.2.0-SNAPSHOT/carbondata-parent-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/ca rbondata-pr-spark-1.6/examples/flink/pom.xml to org.apache.carbondata/carbondata-examples-flink/1.2.0-SNAPSHOT/carbondata-examples-flink-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/core/pom.xml to org.apache.carbondata/carbondata-core/1.2.0-SNAPSHOT/carbondata-core-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/hadoop/pom.xml to org.apache.carbondata/carbondata-hadoop/1.2.0-SNAPSHOT/carbondata-hadoop-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/presto/pom.xml to org.apache.carbondata/carbondata-presto/1.2.0-SNAPSHOT/carbondata-presto-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/spark-common-test/pom.xml to org.apache.carbondata/carbondata-spark-common-test/1.2.0-SNAPSHOT/carbondata-spark-common-test-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/j enkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/spark/pom.xml to org.apache.carbondata/carbondata-spark/1.2.0-SNAPSHOT/carbondata-spark-1.2.0-SNAPSHOT.pomchannel stoppedSetting status of 7d0b8801f2917d9dd5cbe9debd68322a80d43131 to FAILURE with url https://builds.apache.org/job/carbondata-pr-spark-1.6/295/ and message: 'Tests Failed for Spark1.6 'Using context: Jenkins(Spark1.6): mvn clean test -Pspark-1.6 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #910: [WIP] Global sort by spark in load process
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/910 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/294/Build result: FAILURE[...truncated 65.64 KB...][ERROR] Re-run Maven using the -X switch to enable full debug logging.[ERROR] [ERROR] For more information about the errors and possible solutions, please read the following articles:[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException[ERROR] [ERROR] After correcting the problems, you can resume the build with the command[ERROR] mvn -rf :carbondata-spark-common[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/processing/pom.xml to org.apache.carbondata/carbondata-processing/1.2.0-SNAPSHOT/carbondata-processing-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/spark-common/pom.xml to org.apache.carbondata/carbondata-spark-common/1.2.0-SNAPSHOT/carbondata-spark-common-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkin s-slave/workspace/carbondata-pr-spark-1.6/examples/spark/pom.xml to org.apache.carbondata/carbondata-examples-spark/1.2.0-SNAPSHOT/carbondata-examples-spark-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/hive/pom.xml to org.apache.carbondata/carbondata-hive/1.2.0-SNAPSHOT/carbondata-hive-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/common/pom.xml to org.apache.carbondata/carbondata-common/1.2.0-SNAPSHOT/carbondata-common-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/assembly/pom.xml to org.apache.carbondata/carbondata-assembly/1.2.0-SNAPSHOT/carbondata-assembly-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/pom.xml to org.apache.carbondata/carbondata-parent/1.2.0-SNAPSHOT/carbondata-parent-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/works pace/carbondata-pr-spark-1.6/examples/flink/pom.xml to org.apache.carbondata/carbondata-examples-flink/1.2.0-SNAPSHOT/carbondata-examples-flink-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/core/pom.xml to org.apache.carbondata/carbondata-core/1.2.0-SNAPSHOT/carbondata-core-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/hadoop/pom.xml to org.apache.carbondata/carbondata-hadoop/1.2.0-SNAPSHOT/carbondata-hadoop-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/presto/pom.xml to org.apache.carbondata/carbondata-presto/1.2.0-SNAPSHOT/carbondata-presto-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/spark-common-test/pom.xml to org.apache.carbondata/carbondata-spark-common-test/1.2.0-SNAPSHOT/carbondata-spark-common-test-1.2.0-SNAPSHOT.pom[JENKINS] Archiving /home/jenkins/jenkins-slave/workspace/carbondata-pr-spark-1.6/integration/spark/pom.xml to org.apache.carbondata/carbondata-spark/1.2.0-SNAPSHOT/carbondata-spark-1.2.0-SNAPSHOT.pomchannel stoppedSetting status of 763d586c806579d5df482639b357914d247faf75 to FAILURE with url https://builds.apache.org/job/carbondata-pr-spark-1.6/294/ and message: 'Tests Failed for Spark1.6 'Using context: Jenkins(Spark1.6): mvn clean test -Pspark-1.6 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #910: [WIP] Global sort by spark in load process
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/910 Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2415/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #1020: [CARBONDATA-1155] Load query were failing if sort_co...
Github user QiangCai commented on the issue: https://github.com/apache/carbondata/pull/1020 please rebase to resolve confilct --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata pull request #987: [CARBONDATA-1017] Add EncodingStrategy and use...
Github user ravipesala commented on a diff in the pull request: https://github.com/apache/carbondata/pull/987#discussion_r121509773 --- Diff: core/src/main/java/org/apache/carbondata/core/cache/dictionary/DoubleArrayTrieDictionary.java --- @@ -43,7 +43,7 @@ private static final int UUSD_ROOM_VALUE = -2; // unused position, only for zero private static final int EPTY_BACK_VALUE = 0;// value of empty position - private static final int ENCODE_BASE_VALUE = 10; // encode start number + private static final int ENCODE_BASE_VALUE = 10; // apply start number --- End diff -- I think no need to change this file --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[jira] [Resolved] (CARBONDATA-1159) Batch sort loading is not proper without synchronization
[ https://issues.apache.org/jira/browse/CARBONDATA-1159?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Ravindra Pesala resolved CARBONDATA-1159. - Resolution: Fixed Fix Version/s: 1.1.1 1.2.0 > Batch sort loading is not proper without synchronization > > > Key: CARBONDATA-1159 > URL: https://issues.apache.org/jira/browse/CARBONDATA-1159 > Project: CarbonData > Issue Type: Bug >Reporter: dhatchayani >Assignee: dhatchayani >Priority: Minor > Fix For: 1.2.0, 1.1.1 > > -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[GitHub] carbondata pull request #1022: [CARBONDATA - 1159] Batch sort loading is not...
Github user asfgit closed the pull request at: https://github.com/apache/carbondata/pull/1022 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #1022: [CARBONDATA - 1159] Batch sort loading is not proper...
Github user ravipesala commented on the issue: https://github.com/apache/carbondata/pull/1022 LGTM --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #1022: [CARBONDATA - 1159] Batch sort loading is not proper...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1022 Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2414/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[jira] [Resolved] (CARBONDATA-1092) alter table add column query should support no_inverted_index
[ https://issues.apache.org/jira/browse/CARBONDATA-1092?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Ravindra Pesala resolved CARBONDATA-1092. - Resolution: Fixed Fix Version/s: 1.2.0 > alter table add column query should support no_inverted_index > - > > Key: CARBONDATA-1092 > URL: https://issues.apache.org/jira/browse/CARBONDATA-1092 > Project: CarbonData > Issue Type: Bug >Reporter: Rahul Kumar >Assignee: Rahul Kumar >Priority: Minor > Fix For: 1.2.0 > > Time Spent: 4h 10m > Remaining Estimate: 0h > -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[GitHub] carbondata pull request #950: [CARBONDATA-1092] alter table should support n...
Github user asfgit closed the pull request at: https://github.com/apache/carbondata/pull/950 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #950: [CARBONDATA-1092] alter table should support no_inver...
Github user ravipesala commented on the issue: https://github.com/apache/carbondata/pull/950 LGTM --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata pull request #997: [CARBONDATA-1132] describe formatted query sho...
Github user asfgit closed the pull request at: https://github.com/apache/carbondata/pull/997 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #997: [CARBONDATA-1132] describe formatted query should dis...
Github user ravipesala commented on the issue: https://github.com/apache/carbondata/pull/997 LGTM --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #1022: [CARBONDATA - 1159] Batch sort loading is not proper...
Github user ravipesala commented on the issue: https://github.com/apache/carbondata/pull/1022 add to whitelist --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata pull request #1013: [CARBONDATA-1154] IUD Performance Changes
Github user ravipesala commented on a diff in the pull request: https://github.com/apache/carbondata/pull/1013#discussion_r121486113 --- Diff: hadoop/src/main/java/org/apache/carbondata/hadoop/CarbonInputFormat.java --- @@ -340,23 +340,37 @@ private static AbsoluteTableIdentifier getAbsoluteTableIdentifier(Configuration List result = new LinkedList(); FilterExpressionProcessor filterExpressionProcessor = new FilterExpressionProcessor(); +UpdateVO invalidBlockVOForSegmentId = null; +Boolean IUDTable = false; --- End diff -- Use `isIUDTable` --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #1022: [CARBONDATA - 1159] Batch sort loading is not proper...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/1022 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/292/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata pull request #1022: [CARBONDATA - 1159] Batch sort loading is not...
GitHub user dhatchayani opened a pull request: https://github.com/apache/carbondata/pull/1022 [CARBONDATA - 1159] Batch sort loading is not proper without synchronization You can merge this pull request into a Git repository by running: $ git pull https://github.com/dhatchayani/incubator-carbondata batch_sort_sync Alternatively you can review and apply these changes as the patch at: https://github.com/apache/carbondata/pull/1022.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #1022 commit a4c79ba0f16f7e22add560c3ec408dafb1032d61 Author: dhatchayaniDate: 2017-06-12T16:26:47Z [CARBONDATA - 1159] Batch sort loading is not proper without synchronization --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #1022: [CARBONDATA - 1159] Batch sort loading is not proper...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/1022 Can one of the admins verify this patch? --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata pull request #1019: [CARBONDATA-1156]Improve IUD performance and ...
Github user asfgit closed the pull request at: https://github.com/apache/carbondata/pull/1019 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #1019: [CARBONDATA-1156]Improve IUD performance and fixed s...
Github user ravipesala commented on the issue: https://github.com/apache/carbondata/pull/1019 LGTM --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #877: [CARBONDATA-951] create table ddl can specify a bitma...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/877 Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2411/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #1013: [CARBONDATA-1154] IUD Performance Changes
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1013 Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2410/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata pull request #1020: [CARBONDATA-1155] Load query were failing if ...
Github user rahulforallp commented on a diff in the pull request: https://github.com/apache/carbondata/pull/1020#discussion_r121440033 --- Diff: integration/spark-common-test/src/test/scala/org/apache/carbondata/spark/testsuite/sortcolumns/TestSortColumns.scala --- @@ -37,6 +37,12 @@ class TestSortColumns extends QueryTest with BeforeAndAfterAll { checkAnswer(sql("select empno from sorttable1"), sql("select empno from sorttable1 order by empno")) } + test("create table with no dictionary sort_columns with dictionary exclude") { --- End diff -- @QiangCai i have added test case for NumberOfNoDictSortColumns < NoDictionaryCount --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #910: [WIP] Global sort by spark in load process
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/910 Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2409/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #1020: [CARBONDATA-1155] Load query were failing if sort_co...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/1020 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/291/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #1015: [CARBONDATA-1151] Update useful-tips-on-carbondata.m...
Github user chenliang613 commented on the issue: https://github.com/apache/carbondata/pull/1015 @chenerlu please squash into one commit, and update the commit name "Create useful-tips-on-carbondata.md" --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #1019: [CARBONDATA-1156]Improve IUD performance and fixed s...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1019 Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2408/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #1015: [CARBONDATA-1151] Update useful-tips-on-carbondata.m...
Github user chenliang613 commented on the issue: https://github.com/apache/carbondata/pull/1015 LGTM --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #1013: [CARBONDATA-1154] IUD Performance Changes
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1013 Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2407/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #877: [CARBONDATA-951] create table ddl can specify a bitma...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/877 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/290/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #1003: [CARBONDATA-988] Added Presto benchmarking
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1003 Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2406/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #987: [CARBONDATA-1017] Add EncodingStrategy and use Column...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/987 Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2405/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #1013: [CARBONDATA-1154] IUD Performance Changes
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/1013 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/289/Failed Tests: 1carbondata-pr-spark-1.6/org.apache.carbondata:carbondata-spark-common-test: 1org.apache.carbondata.spark.testsuite.allqueries.InsertIntoCarbonTableTestCase.insert into carbon table from carbon table union query --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #910: [WIP] Global sort by spark in load process
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/910 Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2404/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #910: [WIP] Global sort by spark in load process
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/910 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/288/Failed Tests: 5carbondata-pr-spark-1.6/org.apache.carbondata:carbondata-spark-common-test: 5org.apache.carbondata.spark.testsuite.dataload.TestGlobalSortDataLoad.Make sure the result is right and sorted in global levelorg.apache.carbondata.spark.testsuite.dataload.TestGlobalSortDataLoad.Don't support GLOBAL_SORT on partitioned tableorg.apache.carbondata.spark.testsuite.dataload.TestGlobalSortDataLoad.Number of partitions should be greater than 0org.apache.carbondata.spark.testsuite.dataload.TestGlobalSortDataLoad.DDL > carbon.proper tiesorg.apache.carbondata.spark.testsuite.dataload.TestGlobalSortDataLoad.LOAD with DELETE --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #910: [WIP] Global sort by spark in load process
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/910 Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2403/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #1013: [CARBONDATA-1154] IUD Performance Changes
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/1013 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/286/Failed Tests: 2carbondata-pr-spark-1.6/org.apache.carbondata:carbondata-spark-common-test: 2org.apache.carbondata.spark.testsuite.dataload.TestLoadDataFrame.test load dataframe with single pass enabledorg.apache.carbondata.spark.testsuite.dataload.TestLoadDataFrame.test load dataframe with single pass disabled --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #1019: [CARBONDATA-1156]Improve IUD performance and fixed s...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/1019 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/287/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #1020: [CARBONDATA-1155] Load query were failing if sort_co...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1020 Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2402/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #1003: [CARBONDATA-988] Added Presto benchmarking
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/1003 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/285/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata pull request #1019: [CARBONDATA-1156]Improve IUD performance and ...
Github user kumarvishal09 commented on a diff in the pull request: https://github.com/apache/carbondata/pull/1019#discussion_r121399194 --- Diff: core/src/main/java/org/apache/carbondata/core/scan/result/iterator/AbstractDetailQueryResultIterator.java --- @@ -126,6 +144,82 @@ private void intialiseInfos() { } } + /** + * Below method will be used to get the delete delta rows for a block + * + * @param dataBlock data block + * @param deleteDeltaInfo delete delta info + * @return blockid+pageid to deleted row mapping + */ + private MapgetDeleteDeltaDetails(AbstractIndex dataBlock, + DeleteDeltaInfo deleteDeltaInfo) { +// if datablock deleted delta timestamp is more then the current delete delta files timestamp +// then return the current deleted rows +if (dataBlock.getDeleteDeltaTimestamp() >= deleteDeltaInfo +.getLatestDeleteDeltaFileTimestamp()) { + return dataBlock.getDeletedRowsMap(); +} +CarbonDeleteFilesDataReader carbonDeleteDeltaFileReader = null; +// get the lock object so in case of concurrent query only one task will read the delete delta +// files other tasks will wait +Object lockObject = deleteDeltaToLockObjectMap.get(deleteDeltaInfo); +// if lock object is null then add a lock object +if (null == lockObject) { + synchronized (deleteDeltaToLockObjectMap) { +// double checking --- End diff -- ok. I missed it:) --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #987: [CARBONDATA-1017] Add EncodingStrategy and use Column...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/987 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/284/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata pull request #1001: [CARBONDATA-1135] Added partition column info...
Github user asfgit closed the pull request at: https://github.com/apache/carbondata/pull/1001 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata pull request #184: [CARBONDATA-264]Fixed limit query scan time st...
Github user kumarvishal09 closed the pull request at: https://github.com/apache/carbondata/pull/184 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #910: [WIP] Global sort by spark in load process
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/910 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/281/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #910: [WIP] Global sort by spark in load process
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/910 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/283/Failed Tests: 10carbondata-pr-spark-1.6/org.apache.carbondata:carbondata-spark-common-test: 10org.apache.carbondata.spark.testsuite.dataload.TestGlobalSortDataLoad.Make sure the result is right and sorted in global levelorg.apache.carbondata.spark.testsuite.dataload.TestGlobalSortDataLoad.Test GLOBAL_SORT with BAD_RECORDS_ACTION = 'REDIRECT'org.apache.carbondata.spark.testsuite.dataload.TestGlobalSortDataLoad.Don't support GLOBAL_SORT on partitioned tableorg.apache.carbondata.spark.testsuite.dataload.Te stGlobalSortDataLoad.Number of partitions should be greater than 0org.apache.carbondata.spark.testsuite.dataload.TestGlobalSortDataLoad.Compaction GLOBAL_SORT * 2org.apache.carbondata.spark.testsuite.dataload.TestGlobalSortDataLoad.Compaction GLOBAL_SORT + LOCAL_SORT + BATCH_SORTorg.apache.carbondata.spark.testsuite.dataload.TestGlobalSortDataLoad.DDL > carbon.propertiesorg.apache.carbondata.spark.testsuite.dataload.TestGlobalSortDataLoad.carbon.propertiesorg.apache.carbondata.spark.testsuite.dataload.TestGlobalSortDataLoad.LOAD with DELETEorg.apache.carbondata.spark.testsuite.dataload.TestGlobalSortDataLoad.INSERT INTO --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #910: [WIP] Global sort by spark in load process
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/910 Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2401/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata pull request #1019: [CARBONDATA-1156]Improve IUD performance and ...
Github user ravipesala commented on a diff in the pull request: https://github.com/apache/carbondata/pull/1019#discussion_r121395234 --- Diff: core/src/main/java/org/apache/carbondata/core/scan/result/iterator/AbstractDetailQueryResultIterator.java --- @@ -126,6 +144,82 @@ private void intialiseInfos() { } } + /** + * Below method will be used to get the delete delta rows for a block + * + * @param dataBlock data block + * @param deleteDeltaInfo delete delta info + * @return blockid+pageid to deleted row mapping + */ + private MapgetDeleteDeltaDetails(AbstractIndex dataBlock, + DeleteDeltaInfo deleteDeltaInfo) { +// if datablock deleted delta timestamp is more then the current delete delta files timestamp +// then return the current deleted rows +if (dataBlock.getDeleteDeltaTimestamp() >= deleteDeltaInfo +.getLatestDeleteDeltaFileTimestamp()) { + return dataBlock.getDeletedRowsMap(); +} +CarbonDeleteFilesDataReader carbonDeleteDeltaFileReader = null; +// get the lock object so in case of concurrent query only one task will read the delete delta +// files other tasks will wait +Object lockObject = deleteDeltaToLockObjectMap.get(deleteDeltaInfo); +// if lock object is null then add a lock object +if (null == lockObject) { + synchronized (deleteDeltaToLockObjectMap) { +// double checking --- End diff -- Again do `deleteDeltaToLockObjectMap.get(deleteDeltaInfo);` to avoid null pointer exception --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #992: [CARBONDATA-808] Added Example for Presto
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/992 Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2400/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #972: [CARBONDATA-1065] Added set command in carbon to upda...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/972 Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2399/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata pull request #1019: [CARBONDATA-1156]Improve IUD performance and ...
Github user ravipesala commented on a diff in the pull request: https://github.com/apache/carbondata/pull/1019#discussion_r121390830 --- Diff: core/src/main/java/org/apache/carbondata/core/reader/CarbonDeleteFilesDataReader.java --- @@ -120,7 +122,53 @@ private void initThreadPoolSize() { } } return pageIdDeleteRowsMap; + } + /** + * Below method will be used to read the delete delta files + * and get the map of blockletid and page id mapping to deleted + * rows + * + * @param deltaFiles delete delta files array + * @return map of blockletid_pageid to deleted rows + */ + public MapgetDeletedRowsDataVo(String[] deltaFiles) { +List taskSubmitList = new ArrayList<>(); +ExecutorService executorService = Executors.newFixedThreadPool(thread_pool_size); +for (final String deltaFile : deltaFiles) { + taskSubmitList.add(executorService.submit(new Callable() { +@Override public DeleteDeltaBlockDetails call() throws IOException { + CarbonDeleteDeltaFileReaderImpl deltaFileReader = + new CarbonDeleteDeltaFileReaderImpl(deltaFile, FileFactory.getFileType(deltaFile)); + return deltaFileReader.readJson(); +} + })); +} +try { + executorService.shutdown(); + executorService.awaitTermination(30, TimeUnit.MINUTES); +} catch (InterruptedException e) { + LOGGER.error("Error while reading the delete delta files : " + e.getMessage()); +} +Map pageIdToBlockLetVo = new HashMap<>(); +List blockletDetails = null; +for (int i = 0; i < taskSubmitList.size(); i++) { + try { +blockletDetails = taskSubmitList.get(i).get().getBlockletDetails(); + } catch (InterruptedException | ExecutionException e) { +throw new RuntimeException(e); + } + for (DeleteDeltaBlockletDetails blockletDetail : blockletDetails) { +DeleteDeltaVo deleteDeltaVo = pageIdToBlockLetVo.get(blockletDetail.getBlockletKey()); +if (null == deleteDeltaVo) { + deleteDeltaVo = new DeleteDeltaVo(); + pageIdToBlockLetVo.put(blockletDetail.getBlockletKey(), deleteDeltaVo); +} +deleteDeltaVo.insertData(blockletDetail.getDeletedRows()); +; --- End diff -- remove semicolon --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #1011: [carbondata-1149] Fix issue of mismatch type of part...
Github user asfgit commented on the issue: https://github.com/apache/carbondata/pull/1011 Refer to this link for build results (access rights to CI server needed): https://builds.apache.org/job/carbondata-pr-spark-1.6/277/Failed Tests: 11carbondata-pr-spark-1.6/org.apache.carbondata:carbondata-spark-common-test: 11org.apache.carbondata.spark.testsuite.partition.TestDDLForPartitionTable.create partition table: range partitionorg.apache.carbondata.spark.testsuite.partition.TestDDLForPartitionTable.create partition table: list partitionorg.apache.carbondata.spark.testsuite.partition.TestDDLForPartitionTable.test exception if partition column is droppedorg.apache.carbondata.spark.testsuite.partition.TestDDLForPartit ionTable.test exception when values in list_info can not match partition column typeorg.apache.carbondata.spark.testsuite.partition.TestDDLForPartitionTable.test exception when values in range_info can not match partition column typeorg.apache.carbondata.spark.testsuite.partition.TestDataLoadingForPartitionTable.data loading for partition table: range partitionorg.apache.carbondata.spark.testsuite.partition.TestDataLoadingForPartitionTable.data loading for partition table: list partitionorg.apache.carbondata.spark.testsuite.partition.TestDataLoadingForPartitionTable.Insert into for partition table: range partitionorg.apache.carbondata.spark.testsuite.partition.TestDataLoadingForPartitionTable.Insert into partition table: list partitionorg.apache.carbondata.spark.testsuite.partition.TestQueryForPartitionTable.detail query on partition table: range partitionorg.apache.carbondata.spark.testsuite.partition.TestQueryForPartitionTable.detail q uery on partition table: list partition --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] carbondata issue #972: [CARBONDATA-1065] Added set command in carbon to upda...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/972 Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/2397/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---