carbondata git commit: [CARBONDATA-2715][LuceneDataMap] Fix bug in search mode with lucene datamap in windows

2018-07-10 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/master d8562e5bd -> a5039baa3 [CARBONDATA-2715][LuceneDataMap] Fix bug in search mode with lucene datamap in windows While comparing two pathes, the file separator is different in windows, thus causing empty pruned blocklets. This PR will ig

carbondata git commit: [CARBONDATA-2703][Tests] Clear up env after tests

2018-07-10 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/master a5039baa3 -> 75126c6ca [CARBONDATA-2703][Tests] Clear up env after tests 1.reset session parameters after test 2.clean up output after test This closes #2458 Project: http://git-wip-us.apache.org/repos/asf/carbondata/repo Commit: ht

carbondata git commit: [CARBONDATA-2702][BloomDatamap] Fix bugs in clear bloom datamap concurrently

2018-07-11 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/master 438b4421e -> 641ec098f [CARBONDATA-2702][BloomDatamap] Fix bugs in clear bloom datamap concurrently add synchronization for clearing bloom datamap, this implementation refers to BlockletDataMapFactory in PR2324 for CARBONDATA-2496 Thi

carbondata git commit: [CARBONDATA-2716][DataMap] Add validation for datamap writer listener during data loading

2018-07-11 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/master 641ec098f -> 57b457153 [CARBONDATA-2716][DataMap] Add validation for datamap writer listener during data loading In some scenarios, while doing data loading, the loading will use the datamap writer listener that does not belong to th

carbondata git commit: [CARBONDATA-2655][BloomDataMap] BloomFilter datamap support in operator

2018-07-12 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/master f91140367 -> 202d099d6 [CARBONDATA-2655][BloomDataMap] BloomFilter datamap support in operator Now queries with in expression on bloom index column can leverage the BloomFilter datamap. This closes #2445 Project: http://git-wip-us.a

carbondata git commit: [CARBONDATA-2637][DataMap] Fix bugs in rebuild datamap

2018-07-12 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/master 202d099d6 -> 9d7a9a2a9 [CARBONDATA-2637][DataMap] Fix bugs in rebuild datamap In cluster mode, readCommitScope is null while rebuilding datamap for segments, this will cause NPE. Here we use the origin segment object whose readCommitSc

carbondata git commit: [CARBONDATA-2708][BloomDataMap] clear index file in case of data load failure

2018-07-13 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/master 18381e3db -> 1fd370399 [CARBONDATA-2708][BloomDataMap] clear index file in case of data load failure When data loading failed, clean the index DataMap files that generated This closes #2463 Project: http://git-wip-us.apache.org/repo

[1/4] carbondata git commit: Revert "[CARBONDATA-2532][Integration] Carbon to support spark 2.3 version, compatability issues"

2018-07-13 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/carbonstore 7306b59dd -> 0aab4e7c6 http://git-wip-us.apache.org/repos/asf/carbondata/blob/0aab4e7c/pom.xml -- diff --git a/pom.xml b/pom.xml index d6436d7..74e5c01 100644 ---

[2/4] carbondata git commit: Revert "[CARBONDATA-2532][Integration] Carbon to support spark 2.3 version, compatability issues"

2018-07-13 Thread jackylk
http://git-wip-us.apache.org/repos/asf/carbondata/blob/0aab4e7c/integration/spark2/src/main/spark2.2/org/apache/spark/sql/execution/BatchedDataSourceScanExec.scala -- diff --git a/integration/spark2/src/main/spark2.2/org/apache/sp

[3/4] carbondata git commit: Revert "[CARBONDATA-2532][Integration] Carbon to support spark 2.3 version, compatability issues"

2018-07-13 Thread jackylk
http://git-wip-us.apache.org/repos/asf/carbondata/blob/0aab4e7c/integration/spark2/src/main/commonTo2.2And2.3/org/apache/spark/sql/hive/CreateCarbonSourceTableAsSelectCommand.scala -- diff --git a/integration/spark2/src/main/commo

[4/4] carbondata git commit: Revert "[CARBONDATA-2532][Integration] Carbon to support spark 2.3 version, compatability issues"

2018-07-13 Thread jackylk
Revert "[CARBONDATA-2532][Integration] Carbon to support spark 2.3 version, compatability issues" This reverts commit d0fa52396687ccc1a5d029006e7204771c04a9eb. Project: http://git-wip-us.apache.org/repos/asf/carbondata/repo Commit: http://git-wip-us.apache.org/repos/asf/carbondata/commit/0aab4e

carbondata git commit: [HOTFIX] Removed BatchedDataSourceScanExec class and extended directly from FileSourceScanExec

2018-07-13 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/master 653efee02 -> 3df2fd030 [HOTFIX] Removed BatchedDataSourceScanExec class and extended directly from FileSourceScanExec Problem: Since some of the code of BatchedDataSourceScanExec is copied from spark, it is difficult to maintain from

carbondata git commit: [CARBONDATA-2729][file-format] Schema Compatibility problem between version 1.3.0 and 1.4.0

2018-07-13 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/master 3df2fd030 -> bc12de004 [CARBONDATA-2729][file-format] Schema Compatibility problem between version 1.3.0 and 1.4.0 Problem: In TableSchema the field Name schemaEvaluation is changed to schemaEvoluation and in DataMapSchema field name

carbondata git commit: [CARBONDATA-2528][MV] Fixed order by in mv and aggregation functions inside projection expressions are fixed

2018-07-14 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/master bc12de004 -> 0c33857fd [CARBONDATA-2528][MV] Fixed order by in mv and aggregation functions inside projection expressions are fixed Problem: Order by queries and the queries with functions like sum(a)+sum(b) are not working in MV. Pl

carbondata git commit: [CARBONDATA-2693][BloomDataMap]Fix bug for alter rename is renaming the existing table on which bloomfilter datamp exists

2018-07-15 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/master 75a602d01 -> cdee81d4d [CARBONDATA-2693][BloomDataMap]Fix bug for alter rename is renaming the existing table on which bloomfilter datamp exists Fix bug for alter rename is renaming the existing table on which bloom filter datamap ex

[2/2] carbondata git commit: Revert "[CARBONDATA-2532][Integration] Carbon to support spark 2.3 version, ColumnVector Interface"

2018-07-17 Thread jackylk
Revert "[CARBONDATA-2532][Integration] Carbon to support spark 2.3 version, ColumnVector Interface" This reverts commit 2b8ae2628d50efcd095696b5bf614eab2fcdb8d2. Project: http://git-wip-us.apache.org/repos/asf/carbondata/repo Commit: http://git-wip-us.apache.org/repos/asf/carbondata/commit/96fe

[1/2] carbondata git commit: Revert "[CARBONDATA-2532][Integration] Carbon to support spark 2.3 version, ColumnVector Interface"

2018-07-17 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/carbonstore 0aab4e7c6 -> 96fe233a2 http://git-wip-us.apache.org/repos/asf/carbondata/blob/96fe233a/streaming/src/main/java/org/apache/carbondata/streaming/CarbonStreamRecordReader.java --

carbondata git commit: [CARBONDATA-2736][CARBONSTORE] Kafka integration with Carbon StreamSQL

2018-07-17 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/carbonstore 96fe233a2 -> 4b96ed8ca [CARBONDATA-2736][CARBONSTORE] Kafka integration with Carbon StreamSQL Modification in this PR: 1.Pass source table properties to streamReader.load() 2.Do not pass schema when sparkSession.readStream 3.Remov

[12/50] [abbrv] carbondata git commit: [CARBONDATA-2637][DataMap] Fix bugs in rebuild datamap

2018-07-17 Thread jackylk
[CARBONDATA-2637][DataMap] Fix bugs in rebuild datamap In cluster mode, readCommitScope is null while rebuilding datamap for segments, this will cause NPE. Here we use the origin segment object whose readCommitScope is not null and will work fine. This closes #2493 Project: http://git-wip-us.ap

[33/50] [abbrv] carbondata git commit: [CARBONDATA-2727][BloomDataMap] Support create bloom datamap on newly added column

2018-07-17 Thread jackylk
[CARBONDATA-2727][BloomDataMap] Support create bloom datamap on newly added column Add a result collector with rowId infomation for datamap rebuild if table schema is changed; Use keygenerator to retrieve surrogate value of dictIndexColumn from query result; This closes #2490 Project: http:/

[06/50] [abbrv] carbondata git commit: [CARBONDATA-2720] Remove dead code

2018-07-17 Thread jackylk
http://git-wip-us.apache.org/repos/asf/carbondata/blob/f9114036/core/src/test/java/org/apache/carbondata/core/scan/expression/conditional/GreaterThanExpressionUnitTest.java -- diff --git a/core/src/test/java/org/apache/carbondata/

[43/50] [abbrv] carbondata git commit: [CARBONDATA-2613] Support csv based carbon table

2018-07-17 Thread jackylk
[CARBONDATA-2613] Support csv based carbon table 1. create csv based carbon table using CREATE TABLE fact_table (col1 bigint, col2 string, ..., col100 string) STORED BY 'CarbonData' TBLPROPERTIES( 'foramt'='csv', 'csv.delimiter'=',', 'csv.header'='col1,col2,col100') 2. Load data to this tab

[03/50] [abbrv] carbondata git commit: [CARBONDATA-2720] Remove dead code

2018-07-17 Thread jackylk
http://git-wip-us.apache.org/repos/asf/carbondata/blob/f9114036/processing/src/main/java/org/apache/carbondata/processing/loading/steps/DataWriterProcessorStepImpl.java -- diff --git a/processing/src/main/java/org/apache/carbondat

[48/50] [abbrv] carbondata git commit: [CARBONDATA-2705][CarbonStore] CarbonStore Java API and Implementation

2018-07-17 Thread jackylk
[CARBONDATA-2705][CarbonStore] CarbonStore Java API and Implementation Support two implementations: 1.LocalCarbonStore for usage in local mode 2.DistributedCarbonStore leveraging multiple server (Master and Workers) via RPC This closes #2473 Project: http://git-wip-us.apache.org/repos/asf/carbo

[27/50] [abbrv] carbondata git commit: [CARBONDATA-2482] Pass uuid while writing segment file if possible

2018-07-17 Thread jackylk
[CARBONDATA-2482] Pass uuid while writing segment file if possible Pass the uuid (segmentFileName) to the writeSegmentFile method file if possible. Problem: When the supporting tables depends on the segmentFileName of the main table, query is failing, as it is expected to be the same name as tha

[46/50] [abbrv] carbondata git commit: [CARBONDATA-2705][CarbonStore] CarbonStore Java API and Implementation

2018-07-17 Thread jackylk
http://git-wip-us.apache.org/repos/asf/carbondata/blob/85cdc404/store/core/src/main/java/org/apache/carbondata/store/impl/rpc/model/QueryResponse.java -- diff --git a/store/core/src/main/java/org/apache/carbondata/store/impl/rpc/m

[11/50] [abbrv] carbondata git commit: [CARBONDATA-2655][BloomDataMap] BloomFilter datamap support in operator

2018-07-17 Thread jackylk
[CARBONDATA-2655][BloomDataMap] BloomFilter datamap support in operator Now queries with in expression on bloom index column can leverage the BloomFilter datamap. This closes #2445 Project: http://git-wip-us.apache.org/repos/asf/carbondata/repo Commit: http://git-wip-us.apache.org/repos/asf/car

[05/50] [abbrv] carbondata git commit: [CARBONDATA-2720] Remove dead code

2018-07-17 Thread jackylk
http://git-wip-us.apache.org/repos/asf/carbondata/blob/f9114036/datamap/lucene/src/main/java/org/apache/carbondata/datamap/lucene/LuceneCoarseGrainDataMap.java -- diff --git a/datamap/lucene/src/main/java/org/apache/carbondata/dat

[02/50] [abbrv] carbondata git commit: [CARBONDATA-2712] Added fix for Local Dictionary Exclude for multi level complex columns

2018-07-17 Thread jackylk
[CARBONDATA-2712] Added fix for Local Dictionary Exclude for multi level complex columns What was the problem? When Local Dictionary Exclude was defined for multi level complex columns, the columns were still considered for Local Dictionary Include What has been changed? The index value was not

[18/50] [abbrv] carbondata git commit: [CARBONDATA-2723][DataMap] Fix bugs in recreate datamap on table

2018-07-17 Thread jackylk
[CARBONDATA-2723][DataMap] Fix bugs in recreate datamap on table While we drop datamap/table, the executor side cache for datamap is stale. So if we recreate the datamap with different index columns, when we are doing data loading, the cache should be cleaned, otherwise the DataMapWriterListener w

[49/50] [abbrv] carbondata git commit: [CARBONDATA-2688][CarbonStore] Support SQL in REST API

2018-07-17 Thread jackylk
[CARBONDATA-2688][CarbonStore] Support SQL in REST API Support SQL interface in Horizon service. Support REST client for SQL This closes #2481 Project: http://git-wip-us.apache.org/repos/asf/carbondata/repo Commit: http://git-wip-us.apache.org/repos/asf/carbondata/commit/d4a28a25 Tree: http://g

[08/50] [abbrv] carbondata git commit: [CARBONDATA-2720] Remove dead code

2018-07-17 Thread jackylk
http://git-wip-us.apache.org/repos/asf/carbondata/blob/f9114036/core/src/main/java/org/apache/carbondata/core/scan/executor/util/QueryUtil.java -- diff --git a/core/src/main/java/org/apache/carbondata/core/scan/executor/util/Query

[21/50] [abbrv] carbondata git commit: [HOTFIX] Removed BatchedDataSourceScanExec class and extended directly from FileSourceScanExec

2018-07-17 Thread jackylk
[HOTFIX] Removed BatchedDataSourceScanExec class and extended directly from FileSourceScanExec Problem: Since some of the code of BatchedDataSourceScanExec is copied from spark, it is difficult to maintain from version upgrades of spark. Currently we face issues during spark 2.3 upgrade so bett

[09/50] [abbrv] carbondata git commit: [CARBONDATA-2720] Remove dead code

2018-07-17 Thread jackylk
http://git-wip-us.apache.org/repos/asf/carbondata/blob/f9114036/core/src/main/java/org/apache/carbondata/core/datastore/block/SegmentTaskIndexWrapper.java -- diff --git a/core/src/main/java/org/apache/carbondata/core/datastore/blo

[24/50] [abbrv] carbondata git commit: [CARBONDATA-2606][Complex DataType Enhancements]Fix Null result if First two Projection column have same parent and third column has different Parent Struct

2018-07-17 Thread jackylk
[CARBONDATA-2606][Complex DataType Enhancements]Fix Null result if First two Projection column have same parent and third column has different Parent Struct Problem: When multiple columns are there,then the first child elements is only going to make parent Object Array. For all other cases it sho

[07/50] [abbrv] carbondata git commit: [CARBONDATA-2720] Remove dead code

2018-07-17 Thread jackylk
http://git-wip-us.apache.org/repos/asf/carbondata/blob/f9114036/core/src/test/java/org/apache/carbondata/core/datastore/SegmentTaskIndexStoreTest.java -- diff --git a/core/src/test/java/org/apache/carbondata/core/datastore/Segment

[47/50] [abbrv] carbondata git commit: [CARBONDATA-2705][CarbonStore] CarbonStore Java API and Implementation

2018-07-17 Thread jackylk
http://git-wip-us.apache.org/repos/asf/carbondata/blob/85cdc404/store/core/src/main/java/org/apache/carbondata/store/conf/StoreConf.java -- diff --git a/store/core/src/main/java/org/apache/carbondata/store/conf/StoreConf.java b/s

[37/50] [abbrv] carbondata git commit: [CARBONDATA-2609] Change RPC implementation to Hadoop RPC framework

2018-07-17 Thread jackylk
[CARBONDATA-2609] Change RPC implementation to Hadoop RPC framework This closes #2372 Project: http://git-wip-us.apache.org/repos/asf/carbondata/repo Commit: http://git-wip-us.apache.org/repos/asf/carbondata/commit/d9b40bf9 Tree: http://git-wip-us.apache.org/repos/asf/carbondata/tree/d9b40bf9 Di

[30/50] [abbrv] carbondata git commit: [CARBONDATA-2649] Fixed arrayIndexOutOfBoundException while loading Blocklet DataMap after alter add column operation

2018-07-17 Thread jackylk
[CARBONDATA-2649] Fixed arrayIndexOutOfBoundException while loading Blocklet DataMap after alter add column operation Things done as part of this PR Fixed arrayIndexOutOfBoundException while loading Blocklet DataMap after alter add column operation Problem: Array Index out of bound exception wa

[04/50] [abbrv] carbondata git commit: [CARBONDATA-2720] Remove dead code

2018-07-17 Thread jackylk
http://git-wip-us.apache.org/repos/asf/carbondata/blob/f9114036/integration/spark-common/src/main/scala/org/apache/carbondata/events/AlterTableEvents.scala -- diff --git a/integration/spark-common/src/main/scala/org/apache/carbond

[01/50] [abbrv] carbondata git commit: [CARBONDATA-2712] Added fix for Local Dictionary Exclude for multi level complex columns [Forced Update!]

2018-07-17 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/carbonstore 4b96ed8ca -> 239a6cadb (forced update) http://git-wip-us.apache.org/repos/asf/carbondata/blob/d267c40b/integration/spark-common-test/src/test/scala/org/apache/carbondata/spark/testsuite/localdictionary/LocalDictionarySupportCreateT

[45/50] [abbrv] carbondata git commit: [CARBONDATA-2705][CarbonStore] CarbonStore Java API and Implementation

2018-07-17 Thread jackylk
http://git-wip-us.apache.org/repos/asf/carbondata/blob/85cdc404/store/core/src/main/java/org/apache/carbondata/store/rpc/model/BaseResponse.java -- diff --git a/store/core/src/main/java/org/apache/carbondata/store/rpc/model/BaseRe

[23/50] [abbrv] carbondata git commit: [CARBONDATA-2528][MV] Fixed order by in mv and aggregation functions inside projection expressions are fixed

2018-07-17 Thread jackylk
[CARBONDATA-2528][MV] Fixed order by in mv and aggregation functions inside projection expressions are fixed Problem: Order by queries and the queries with functions like sum(a)+sum(b) are not working in MV. Please check jira for more details. Solution: The queries which have projection function

[13/50] [abbrv] carbondata git commit: [HOTFIX][CARBONDATA-2716][DataMap] fix bug for loading datamap

2018-07-17 Thread jackylk
[HOTFIX][CARBONDATA-2716][DataMap] fix bug for loading datamap In some scenarios, input parameter of getCarbonFactDataHandlerModel called carbonTable may be different from the one in loadmodel. This close #2497 Project: http://git-wip-us.apache.org/repos/asf/carbondata/repo Commit: http://git-

[16/50] [abbrv] carbondata git commit: [CARBONDATA-2708][BloomDataMap] clear index file in case of data load failure

2018-07-17 Thread jackylk
[CARBONDATA-2708][BloomDataMap] clear index file in case of data load failure When data loading failed, clean the index DataMap files that generated This closes #2463 Project: http://git-wip-us.apache.org/repos/asf/carbondata/repo Commit: http://git-wip-us.apache.org/repos/asf/carbondata/commit

[31/50] [abbrv] carbondata git commit: [CARBONDATA-2698][CARBONDATA-2700][CARBONDATA-2732][BloomDataMap] block some operations of bloomfilter datamap

2018-07-17 Thread jackylk
[CARBONDATA-2698][CARBONDATA-2700][CARBONDATA-2732][BloomDataMap] block some operations of bloomfilter datamap 1.Block create bloomfilter datamap index on column which its datatype is complex type; 2.Block change datatype for bloomfilter index datamap; 3.Block dropping index columns for bloomfil

[40/50] [abbrv] carbondata git commit: [CARBONDATA-2690][CarbonStore] implement RESTful API: create table, load data and select

2018-07-17 Thread jackylk
http://git-wip-us.apache.org/repos/asf/carbondata/blob/4437920a/store/core/src/main/java/org/apache/carbondata/store/rpc/impl/StoreServiceImpl.java -- diff --git a/store/core/src/main/java/org/apache/carbondata/store/rpc/impl/Stor

[32/50] [abbrv] carbondata git commit: [HotFix] Getting carbon table identifier to datamap events

2018-07-17 Thread jackylk
[HotFix] Getting carbon table identifier to datamap events Passing the table identifier to keep track of table in event in case preload and postload of datamap event. This closes #2448 Project: http://git-wip-us.apache.org/repos/asf/carbondata/repo Commit: http://git-wip-us.apache.org/repos/as

[39/50] [abbrv] carbondata git commit: [CARBONDATA-2690][CarbonStore] implement RESTful API: create table, load data and select

2018-07-17 Thread jackylk
http://git-wip-us.apache.org/repos/asf/carbondata/blob/4437920a/store/horizon/src/main/java/org/apache/carbondata/horizon/antlr/gen/ExpressionLexer.java -- diff --git a/store/horizon/src/main/java/org/apache/carbondata/horizon/ant

[10/50] [abbrv] carbondata git commit: [CARBONDATA-2720] Remove dead code

2018-07-17 Thread jackylk
[CARBONDATA-2720] Remove dead code For acturate coverage results and easy maintainance This closes #2354 Project: http://git-wip-us.apache.org/repos/asf/carbondata/repo Commit: http://git-wip-us.apache.org/repos/asf/carbondata/commit/f9114036 Tree: http://git-wip-us.apache.org/repos/asf/carbond

[19/50] [abbrv] carbondata git commit: [CARBONDATA-2717] fixed table id empty problem while taking drop lock

2018-07-17 Thread jackylk
[CARBONDATA-2717] fixed table id empty problem while taking drop lock This closes #2472 Project: http://git-wip-us.apache.org/repos/asf/carbondata/repo Commit: http://git-wip-us.apache.org/repos/asf/carbondata/commit/637a9746 Tree: http://git-wip-us.apache.org/repos/asf/carbondata/tree/637a9746

[38/50] [abbrv] carbondata git commit: [CARBONDATA-2690][CarbonStore] implement RESTful API: create table, load data and select

2018-07-17 Thread jackylk
http://git-wip-us.apache.org/repos/asf/carbondata/blob/4437920a/store/horizon/src/main/java/org/apache/carbondata/horizon/rest/model/view/LoadRequest.java -- diff --git a/store/horizon/src/main/java/org/apache/carbondata/horizon/r

[28/50] [abbrv] carbondata git commit: [CARBONDATA-2714] Support merge index files for the segment

2018-07-17 Thread jackylk
http://git-wip-us.apache.org/repos/asf/carbondata/blob/73419071/integration/spark2/src/test/scala/org/apache/spark/carbondata/TestStreamingTableWithRowParser.scala -- diff --git a/integration/spark2/src/test/scala/org/apache/spark

[50/50] [abbrv] carbondata git commit: [REBASE] Rebasing with master branch and Fixing rebase conflict

2018-07-17 Thread jackylk
[REBASE] Rebasing with master branch and Fixing rebase conflict Project: http://git-wip-us.apache.org/repos/asf/carbondata/repo Commit: http://git-wip-us.apache.org/repos/asf/carbondata/commit/239a6cad Tree: http://git-wip-us.apache.org/repos/asf/carbondata/tree/239a6cad Diff: http://git-wip-us.a

[17/50] [abbrv] carbondata git commit: [CARBONDATA-2648] Fixed NPE issue with legacy store when CACHE_LEVEL is Blocklet

2018-07-17 Thread jackylk
[CARBONDATA-2648] Fixed NPE issue with legacy store when CACHE_LEVEL is Blocklet Things done as part of this PR: Fixed Null pointer exception when store is of <= 1.1 version and DataMap is of type BlockletDataMap. Added clearing of SegmentProperties cache holder from executor Problem 1: Null poi

[34/50] [abbrv] carbondata git commit: [CARBONDATA-2724][DataMap]Unsupported create datamap on table with V1 or V2 format data

2018-07-17 Thread jackylk
[CARBONDATA-2724][DataMap]Unsupported create datamap on table with V1 or V2 format data block creating datamap on carbon table with V1 or V2 format Currently the version info is read from carbon data file This closes #2488 Project: http://git-wip-us.apache.org/repos/asf/carbondata/repo Commit:

[29/50] [abbrv] carbondata git commit: [CARBONDATA-2714] Support merge index files for the segment

2018-07-17 Thread jackylk
[CARBONDATA-2714] Support merge index files for the segment Problem : The first-time query of carbon becomes very slow. It is because of reading many small carbonindex files and cache to the driver at the first time. Many carbonindex files are created in below case Loading data in large cluster F

[41/50] [abbrv] carbondata git commit: [CARBONDATA-2690][CarbonStore] implement RESTful API: create table, load data and select

2018-07-17 Thread jackylk
[CARBONDATA-2690][CarbonStore] implement RESTful API: create table, load data and select This PR adds: 1.basic framework rewrite the carbon store's Master, Worker and Scheduler code in Java 2.RESTful API support create a table by using file meta store support load data to a table in single work

[14/50] [abbrv] carbondata git commit: [CARBONDATA-2719] Block update and delete on table having datamaps

2018-07-17 Thread jackylk
[CARBONDATA-2719] Block update and delete on table having datamaps Table update/delete is needed to block on table which has datamaps. This close #2483 Project: http://git-wip-us.apache.org/repos/asf/carbondata/repo Commit: http://git-wip-us.apache.org/repos/asf/carbondata/commit/56e7dad7 Tree:

[22/50] [abbrv] carbondata git commit: [CARBONDATA-2729][file-format] Schema Compatibility problem between version 1.3.0 and 1.4.0

2018-07-17 Thread jackylk
[CARBONDATA-2729][file-format] Schema Compatibility problem between version 1.3.0 and 1.4.0 Problem: In TableSchema the field Name schemaEvaluation is changed to schemaEvoluation and in DataMapSchema field name className is changed to providerName. Due to this current Carbon version & Version 1.

[42/50] [abbrv] carbondata git commit: [CARBONDATA-2613] Support csv based carbon table

2018-07-17 Thread jackylk
http://git-wip-us.apache.org/repos/asf/carbondata/blob/2009009a/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/management/CarbonAddSegmentCommand.scala -- diff --git a/integration/spark2/src/main/scala/or

[15/50] [abbrv] carbondata git commit: [CARBONDATA-2734] Fix struct of date issue in create table

2018-07-17 Thread jackylk
[CARBONDATA-2734] Fix struct of date issue in create table problem: Struct of date is not supported currently in create table flow as date datatype check is missing during parsing. Hence child date column was not appended with parent name, leading to StringOutOfIndex exception. solution: Handle t

[25/50] [abbrv] carbondata git commit: [CARBONDATA-2693][BloomDataMap]Fix bug for alter rename is renaming the existing table on which bloomfilter datamp exists

2018-07-17 Thread jackylk
[CARBONDATA-2693][BloomDataMap]Fix bug for alter rename is renaming the existing table on which bloomfilter datamp exists Fix bug for alter rename is renaming the existing table on which bloom filter datamap exists This closes #2452 Project: http://git-wip-us.apache.org/repos/asf/carbondata/r

[44/50] [abbrv] carbondata git commit: [CARBONDATA-2705][CarbonStore] CarbonStore Java API and Implementation

2018-07-17 Thread jackylk
http://git-wip-us.apache.org/repos/asf/carbondata/blob/85cdc404/store/horizon/src/main/java/org/apache/carbondata/horizon/rest/model/descriptor/SelectDescriptor.java -- diff --git a/store/horizon/src/main/java/org/apache/carbondat

[36/50] [abbrv] carbondata git commit: [CARBONDATA-2609] Change RPC implementation to Hadoop RPC framework

2018-07-17 Thread jackylk
http://git-wip-us.apache.org/repos/asf/carbondata/blob/d9b40bf9/store/core/src/main/scala/org/apache/carbondata/store/Master.scala -- diff --git a/store/core/src/main/scala/org/apache/carbondata/store/Master.scala b/store/core/src

[20/50] [abbrv] carbondata git commit: [CARBONDATA-2722] [CARBONDATA-2721] JsonWriter issue fixes

2018-07-17 Thread jackylk
[CARBONDATA-2722] [CARBONDATA-2721] JsonWriter issue fixes [CARBONDATA-2722][SDK] [JsonWriter] NPE when schema and data are not of same length or Data is null. problem: Null data is not handled in the json object to carbon row conversion. solution: add a null check when object is fetched from j

[35/50] [abbrv] carbondata git commit: [CARBONDATA-2609] Change RPC implementation to Hadoop RPC framework

2018-07-17 Thread jackylk
http://git-wip-us.apache.org/repos/asf/carbondata/blob/d9b40bf9/store/search/src/main/scala/org/apache/spark/rpc/Scheduler.scala -- diff --git a/store/search/src/main/scala/org/apache/spark/rpc/Scheduler.scala b/store/search/src/m

[26/50] [abbrv] carbondata git commit: [CARBONDATA-2704] Index file size in describe formatted command is not updated correctly with the segment file

2018-07-17 Thread jackylk
[CARBONDATA-2704] Index file size in describe formatted command is not updated correctly with the segment file Problem: Describe formatted command is not showing correct index files size after index files merge. Solution: Segment file should be updated with the actual index files size of that se

carbondata git commit: [CARBONDATA-2736][CARBONSTORE] Kafka integration with Carbon StreamSQL

2018-07-18 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/carbonstore 239a6cadb -> 9ac55a5a6 [CARBONDATA-2736][CARBONSTORE] Kafka integration with Carbon StreamSQL Modification in this PR: 1.Pass source table properties to streamReader.load() 2.Do not pass schema when sparkSession.readStream 3.Remov

carbondata git commit: [CARBONDATA-2530][MV] Disable the MV datamaps after main table load

2018-07-23 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/master 7ab670652 -> 0ab03f21f [CARBONDATA-2530][MV] Disable the MV datamaps after main table load Problem: MV datamaps are not disabled after the main table load is done. So the wrong data is displaying. Solution: Disable the MV datamaps aft

carbondata git commit: [CARBONDATA-2534][MV] Fix substring expression not working in MV creation

2018-07-23 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/master 4014b0f54 -> 2c291d620 [CARBONDATA-2534][MV] Fix substring expression not working in MV creation Problem: The column generated when subquery expression column present is wrong while creating of MV table. Solution: Corrected the column

carbondata git commit: [CARBONDATA-2542][MV] Fix the mv query from table with different database

2018-07-23 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/master 2c291d620 -> d820e3d51 [CARBONDATA-2542][MV] Fix the mv query from table with different database Problem: database name is not added to the table name while generating mv query. Solution: Add the database name to the table name while c

carbondata git commit: [CARBONDATA-2550][CARBONDATA-2576][MV] Fix limit and average function issue in MV query

2018-07-23 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/master d820e3d51 -> 83562ae75 [CARBONDATA-2550][CARBONDATA-2576][MV] Fix limit and average function issue in MV query Problem: Limit is not working on mv queries and the average is also not working. Solution: Correct the limit queries and av

carbondata git commit: [CARBONDATA-2770][BloomDataMap] Optimize code to get blocklet id when rebuilding datamap

2018-07-23 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/master 83562ae75 -> ce53b48a1 [CARBONDATA-2770][BloomDataMap] Optimize code to get blocklet id when rebuilding datamap we should get exactly number of blocklet id from blocklet scanned result instead of building it ourselves. This closes #

carbondata git commit: [CARBONDATA-2512][32k] Support writing longstring through SDK

2018-07-23 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/master ce53b48a1 -> 1345dc6a3 [CARBONDATA-2512][32k] Support writing longstring through SDK Support writing longstring through SDK. User can specify the datatype as 'varchar' for longstring columns. Please note that, the 'varchar' column cann

carbondata git commit: [CARBONDATA-2769] Fix bug when getting shard name from data before version 1.4

2018-07-23 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/master 1345dc6a3 -> fb2f9d33b [CARBONDATA-2769] Fix bug when getting shard name from data before version 1.4 datamap creation needs shardname. carbon imported segment id in carbondata file since version 1.4. we should return proper shardName

carbondata git commit: [CARBONDATA-2694][32k] Show longstring table property in descformatted

2018-07-23 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/master e5c1568de -> 42a80564c [CARBONDATA-2694][32k] Show longstring table property in descformatted add longstring table property in the output of desc formantted command This closes #2456 Project: http://git-wip-us.apache.org/repos/asf/c

carbondata git commit: [CARBONDATA-2540][CARBONDATA-2560][CARBONDATA-2568][MV] Add validations for unsupported MV queries

2018-07-23 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/master 6f1767b5a -> 498502d2b [CARBONDATA-2540][CARBONDATA-2560][CARBONDATA-2568][MV] Add validations for unsupported MV queries Problem: Validations are missing on the unsupported MV queries while creating MV datamap. Solution: Added valid

carbondata git commit: [CARBONDATA-2531][MV] Fix alias not working on MV query

2018-07-24 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/master bea277f83 -> a75b9db6a [CARBONDATA-2531][MV] Fix alias not working on MV query Problem: when alias present on actual query then MV match not happening because alias is not ignored. Solution : Do semantic check while doing match This

carbondata git commit: [HOTFIX] Fix a spelling mistake after PR2511 merged

2018-07-24 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/master b0aee53f5 -> 68e5b52c4 [HOTFIX] Fix a spelling mistake after PR2511 merged spelling mistakes: AtomicFileOperationsF modfity to: AtomicFileOperationFactory.getAtomicFileOperations This closes #2551 Project: http://git-wip-us.apache.

carbondata git commit: [CARBONDATA-2539][MV] Fix predicate subquery which uses leftsemi join

2018-07-24 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/master 68e5b52c4 -> 9a75ce53b [CARBONDATA-2539][MV] Fix predicate subquery which uses leftsemi join Problem: References to the top plan is not getting right when predicate subquery is present. Solution: Correct the refrences. This closes #2

[carbondata] Git Push Summary

2018-07-25 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/external-format [created] a37a2ff7f

[2/2] carbondata git commit: [CARBONDATA-2613] Support csv based carbon table

2018-07-25 Thread jackylk
[CARBONDATA-2613] Support csv based carbon table 1. create csv based carbon table using CREATE TABLE fact_table (col1 bigint, col2 string, ..., col100 string) STORED BY 'CarbonData' TBLPROPERTIES( 'foramt'='csv', 'csv.delimiter'=',', 'csv.header'='col1,col2,col100') 2. Load data to this tab

[1/2] carbondata git commit: [CARBONDATA-2613] Support csv based carbon table

2018-07-25 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/external-format a37a2ff7f -> 0d2769f75 http://git-wip-us.apache.org/repos/asf/carbondata/blob/0d2769f7/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/management/CarbonAddSegmentCommand.scala --

carbondata git commit: [CARBONDATA-2767][CarbonStore] Fix task locality issue

2018-07-25 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/carbonstore 7ad2fd951 -> 2d4628868 [CARBONDATA-2767][CarbonStore] Fix task locality issue If the Spark cluster and the Hadoop cluster are two different machine cluster, the Spark tasks will run in RACK_LOCAL mode. This closes #2528 Projec

carbondata git commit: [CARBONDATA-2782]delete dead code in class 'CarbonCleanFilesCommand'

2018-07-26 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/master d62fe9e65 -> c79fc90d5 [CARBONDATA-2782]delete dead code in class 'CarbonCleanFilesCommand' The variables(dms、indexDms) in function processMetadata are nerver used. This closes #2557 Project: http://git-wip-us.apache.org/repos/asf

carbondata git commit: [CARBONDATA-2776][CarbonStore] Support ingesting data from Kafka service

2018-07-31 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/carbonstore 2d4628868 -> a6027ae11 [CARBONDATA-2776][CarbonStore] Support ingesting data from Kafka service This closes #2544 Project: http://git-wip-us.apache.org/repos/asf/carbondata/repo Commit: http://git-wip-us.apache.org/repos/asf/car

carbondata git commit: [CARBONDATA-2790][BloomDataMap]Optimize default parameter for bloomfilter datamap

2018-08-01 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/master c29aef880 -> 6351c3a07 [CARBONDATA-2790][BloomDataMap]Optimize default parameter for bloomfilter datamap To provide better query performance for bloomfilter datamap by default, we optimize bloom_size from 32000 to 64 and optimize

carbondata git commit: [CARBONDATA-2793][32k][Doc] Add 32k support in document

2018-08-01 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/master de9246066 -> f9b02a5c1 [CARBONDATA-2793][32k][Doc] Add 32k support in document This closes #2572 Project: http://git-wip-us.apache.org/repos/asf/carbondata/repo Commit: http://git-wip-us.apache.org/repos/asf/carbondata/commit/f9b02a5

carbondata git commit: [CARBONDATA-2800][Doc] Add useful tips about bloomfilter datamap

2018-08-01 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/master f9b02a5c1 -> a302cd1ce [CARBONDATA-2800][Doc] Add useful tips about bloomfilter datamap add useful tips about bloomfilter datamap This closes #2581 Project: http://git-wip-us.apache.org/repos/asf/carbondata/repo Commit: http://git-w

carbondata git commit: [CARBONDATA-2806] Delete delete delta files upon clean files for flat folder

2018-08-01 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/master a302cd1ce -> af984101e [CARBONDATA-2806] Delete delete delta files upon clean files for flat folder Problem: Delete delta files are not removed after clean files operation. Solution: Get the delta files using Segment Status Manager and

carbondata git commit: [CARBONDATA-2796][32K]Fix data loading problem when table has complex column and long string column

2018-08-01 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/master af984101e -> 11fb422d9 [CARBONDATA-2796][32K]Fix data loading problem when table has complex column and long string column currently both varchar column and complex column believes itself is the last one member in noDictionary group

carbondata git commit: [CARBONDATA-2802][BloomDataMap] Remove clearing cache after rebuiding index datamap

2018-08-02 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/master 38384cb9f -> 26d9f3d8e [CARBONDATA-2802][BloomDataMap] Remove clearing cache after rebuiding index datamap This is no need to clear cache after rebuilding index datamap due to the following reasons: 1.currently it will clear all the

carbondata git commit: [CARBONDATA-2809][DataMap] Block rebuilding for bloom/lucene and preagg datamap

2018-08-07 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/master b702a1b01 -> abcd4f6e2 [CARBONDATA-2809][DataMap] Block rebuilding for bloom/lucene and preagg datamap As manual refresh currently only works fine for MV, it has some bugs with other types of datamap such as preaggregate, timeserials,

carbondata git commit: [CARBONDATA-2823] Support streaming property with datamap

2018-08-07 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/master abcd4f6e2 -> b9e510640 [CARBONDATA-2823] Support streaming property with datamap Since during query, carbondata get splits from streaming segment and columnar segments repectively, we can support streaming with index datamap. For prea

carbondata git commit: [CARBONDATA-2585] Fix local dictionary for both table level and system level property based on priority

2018-08-07 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/master f27efb3e3 -> 78438451b [CARBONDATA-2585] Fix local dictionary for both table level and system level property based on priority Added a System level Property for local dictionary Support. Property 'carbon.local.dictionary.enable' can b

carbondata git commit: [CARBONDATA-2539]Fix mv classcast exception issue

2018-08-07 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/master 78438451b -> 3d7fa1276 [CARBONDATA-2539]Fix mv classcast exception issue Class cast exception happens during min type aggregate function happening. It is corrected in this PR This closes #2602 Project: http://git-wip-us.apache.org/

carbondata git commit: [CARBONDATA-2768][CarbonStore] Fix error in tests for external csv format

2018-08-07 Thread jackylk
Repository: carbondata Updated Branches: refs/heads/external-format 0d2769f75 -> ccf64ce5a [CARBONDATA-2768][CarbonStore] Fix error in tests for external csv format In previous implementation earlier than PR2495, we only supportted csv as external format for carbondata. And we validated the re

[06/50] [abbrv] carbondata git commit: [CARBONDATA-2784][CARBONDATA-2786][SDK writer] Fixed:Forever blocking wait with more than 21 batch of data

2018-08-07 Thread jackylk
[CARBONDATA-2784][CARBONDATA-2786][SDK writer] Fixed:Forever blocking wait with more than 21 batch of data problem: [CARBONDATA-2784] [SDK writer] Forever blocking wait with more than 21 batch of data, when consumer is dead due to data loading exception (bad record / out of memory) root cause:

[30/50] [abbrv] carbondata git commit: [HOTFIX][PR 2575] Fixed modular plan creation only if valid datamaps are available

2018-08-07 Thread jackylk
[HOTFIX][PR 2575] Fixed modular plan creation only if valid datamaps are available update query is failing in spark-2.2 cluster if mv jars are available because catalogs are not empty if datamap are created for other table also and returns true from isValidPlan() inside MVAnalyzerRule. This cl

<    2   3   4   5   6   7   8   9   10   11   >