Repository: carbondata
Updated Branches:
refs/heads/master d8562e5bd -> a5039baa3
[CARBONDATA-2715][LuceneDataMap] Fix bug in search mode with lucene datamap in
windows
While comparing two pathes, the file separator is different in windows,
thus causing empty pruned blocklets. This PR will ig
Repository: carbondata
Updated Branches:
refs/heads/master a5039baa3 -> 75126c6ca
[CARBONDATA-2703][Tests] Clear up env after tests
1.reset session parameters after test
2.clean up output after test
This closes #2458
Project: http://git-wip-us.apache.org/repos/asf/carbondata/repo
Commit: ht
Repository: carbondata
Updated Branches:
refs/heads/master 438b4421e -> 641ec098f
[CARBONDATA-2702][BloomDatamap] Fix bugs in clear bloom datamap concurrently
add synchronization for clearing bloom datamap, this implementation
refers to BlockletDataMapFactory in PR2324 for CARBONDATA-2496
Thi
Repository: carbondata
Updated Branches:
refs/heads/master 641ec098f -> 57b457153
[CARBONDATA-2716][DataMap] Add validation for datamap writer listener during
data loading
In some scenarios, while doing data loading, the loading will use the datamap
writer listener that does not belong to th
Repository: carbondata
Updated Branches:
refs/heads/master f91140367 -> 202d099d6
[CARBONDATA-2655][BloomDataMap] BloomFilter datamap support in operator
Now queries with in expression on bloom index column can leverage the
BloomFilter datamap.
This closes #2445
Project: http://git-wip-us.a
Repository: carbondata
Updated Branches:
refs/heads/master 202d099d6 -> 9d7a9a2a9
[CARBONDATA-2637][DataMap] Fix bugs in rebuild datamap
In cluster mode, readCommitScope is null while rebuilding datamap for
segments, this will cause NPE. Here we use the origin segment object
whose readCommitSc
Repository: carbondata
Updated Branches:
refs/heads/master 18381e3db -> 1fd370399
[CARBONDATA-2708][BloomDataMap] clear index file in case of data load failure
When data loading failed, clean the index DataMap files that generated
This closes #2463
Project: http://git-wip-us.apache.org/repo
Repository: carbondata
Updated Branches:
refs/heads/carbonstore 7306b59dd -> 0aab4e7c6
http://git-wip-us.apache.org/repos/asf/carbondata/blob/0aab4e7c/pom.xml
--
diff --git a/pom.xml b/pom.xml
index d6436d7..74e5c01 100644
---
http://git-wip-us.apache.org/repos/asf/carbondata/blob/0aab4e7c/integration/spark2/src/main/spark2.2/org/apache/spark/sql/execution/BatchedDataSourceScanExec.scala
--
diff --git
a/integration/spark2/src/main/spark2.2/org/apache/sp
http://git-wip-us.apache.org/repos/asf/carbondata/blob/0aab4e7c/integration/spark2/src/main/commonTo2.2And2.3/org/apache/spark/sql/hive/CreateCarbonSourceTableAsSelectCommand.scala
--
diff --git
a/integration/spark2/src/main/commo
Revert "[CARBONDATA-2532][Integration] Carbon to support spark 2.3 version,
compatability issues"
This reverts commit d0fa52396687ccc1a5d029006e7204771c04a9eb.
Project: http://git-wip-us.apache.org/repos/asf/carbondata/repo
Commit: http://git-wip-us.apache.org/repos/asf/carbondata/commit/0aab4e
Repository: carbondata
Updated Branches:
refs/heads/master 653efee02 -> 3df2fd030
[HOTFIX] Removed BatchedDataSourceScanExec class and extended directly from
FileSourceScanExec
Problem:
Since some of the code of BatchedDataSourceScanExec is copied from spark, it is
difficult to maintain from
Repository: carbondata
Updated Branches:
refs/heads/master 3df2fd030 -> bc12de004
[CARBONDATA-2729][file-format] Schema Compatibility problem between version
1.3.0 and 1.4.0
Problem:
In TableSchema the field Name schemaEvaluation is changed to schemaEvoluation
and in DataMapSchema field name
Repository: carbondata
Updated Branches:
refs/heads/master bc12de004 -> 0c33857fd
[CARBONDATA-2528][MV] Fixed order by in mv and aggregation functions inside
projection expressions are fixed
Problem:
Order by queries and the queries with functions like sum(a)+sum(b) are not
working in MV.
Pl
Repository: carbondata
Updated Branches:
refs/heads/master 75a602d01 -> cdee81d4d
[CARBONDATA-2693][BloomDataMap]Fix bug for alter rename is renaming the
existing table on which bloomfilter datamp exists
Fix bug for alter rename is renaming the existing table on which bloom filter
datamap ex
Revert "[CARBONDATA-2532][Integration] Carbon to support spark 2.3 version,
ColumnVector Interface"
This reverts commit 2b8ae2628d50efcd095696b5bf614eab2fcdb8d2.
Project: http://git-wip-us.apache.org/repos/asf/carbondata/repo
Commit: http://git-wip-us.apache.org/repos/asf/carbondata/commit/96fe
Repository: carbondata
Updated Branches:
refs/heads/carbonstore 0aab4e7c6 -> 96fe233a2
http://git-wip-us.apache.org/repos/asf/carbondata/blob/96fe233a/streaming/src/main/java/org/apache/carbondata/streaming/CarbonStreamRecordReader.java
--
Repository: carbondata
Updated Branches:
refs/heads/carbonstore 96fe233a2 -> 4b96ed8ca
[CARBONDATA-2736][CARBONSTORE] Kafka integration with Carbon StreamSQL
Modification in this PR:
1.Pass source table properties to streamReader.load()
2.Do not pass schema when sparkSession.readStream
3.Remov
[CARBONDATA-2637][DataMap] Fix bugs in rebuild datamap
In cluster mode, readCommitScope is null while rebuilding datamap for
segments, this will cause NPE. Here we use the origin segment object
whose readCommitScope is not null and will work fine.
This closes #2493
Project: http://git-wip-us.ap
[CARBONDATA-2727][BloomDataMap] Support create bloom datamap on newly added
column
Add a result collector with rowId infomation for datamap rebuild if table
schema is changed;
Use keygenerator to retrieve surrogate value of dictIndexColumn from query
result;
This closes #2490
Project: http:/
http://git-wip-us.apache.org/repos/asf/carbondata/blob/f9114036/core/src/test/java/org/apache/carbondata/core/scan/expression/conditional/GreaterThanExpressionUnitTest.java
--
diff --git
a/core/src/test/java/org/apache/carbondata/
[CARBONDATA-2613] Support csv based carbon table
1. create csv based carbon table using
CREATE TABLE fact_table (col1 bigint, col2 string, ..., col100 string)
STORED BY 'CarbonData'
TBLPROPERTIES(
'foramt'='csv',
'csv.delimiter'=',',
'csv.header'='col1,col2,col100')
2. Load data to this tab
http://git-wip-us.apache.org/repos/asf/carbondata/blob/f9114036/processing/src/main/java/org/apache/carbondata/processing/loading/steps/DataWriterProcessorStepImpl.java
--
diff --git
a/processing/src/main/java/org/apache/carbondat
[CARBONDATA-2705][CarbonStore] CarbonStore Java API and Implementation
Support two implementations:
1.LocalCarbonStore for usage in local mode
2.DistributedCarbonStore leveraging multiple server (Master and Workers) via RPC
This closes #2473
Project: http://git-wip-us.apache.org/repos/asf/carbo
[CARBONDATA-2482] Pass uuid while writing segment file if possible
Pass the uuid (segmentFileName) to the writeSegmentFile method file if possible.
Problem:
When the supporting tables depends on the segmentFileName of the main table,
query is failing, as it is expected to be the same name as tha
http://git-wip-us.apache.org/repos/asf/carbondata/blob/85cdc404/store/core/src/main/java/org/apache/carbondata/store/impl/rpc/model/QueryResponse.java
--
diff --git
a/store/core/src/main/java/org/apache/carbondata/store/impl/rpc/m
[CARBONDATA-2655][BloomDataMap] BloomFilter datamap support in operator
Now queries with in expression on bloom index column can leverage the
BloomFilter datamap.
This closes #2445
Project: http://git-wip-us.apache.org/repos/asf/carbondata/repo
Commit: http://git-wip-us.apache.org/repos/asf/car
http://git-wip-us.apache.org/repos/asf/carbondata/blob/f9114036/datamap/lucene/src/main/java/org/apache/carbondata/datamap/lucene/LuceneCoarseGrainDataMap.java
--
diff --git
a/datamap/lucene/src/main/java/org/apache/carbondata/dat
[CARBONDATA-2712] Added fix for Local Dictionary Exclude for multi level
complex columns
What was the problem?
When Local Dictionary Exclude was defined for multi level complex columns, the
columns were still considered for Local Dictionary Include
What has been changed?
The index value was not
[CARBONDATA-2723][DataMap] Fix bugs in recreate datamap on table
While we drop datamap/table, the executor side cache for datamap is
stale. So if we recreate the datamap with different index columns, when
we are doing data loading, the cache should be cleaned, otherwise the
DataMapWriterListener w
[CARBONDATA-2688][CarbonStore] Support SQL in REST API
Support SQL interface in Horizon service.
Support REST client for SQL
This closes #2481
Project: http://git-wip-us.apache.org/repos/asf/carbondata/repo
Commit: http://git-wip-us.apache.org/repos/asf/carbondata/commit/d4a28a25
Tree: http://g
http://git-wip-us.apache.org/repos/asf/carbondata/blob/f9114036/core/src/main/java/org/apache/carbondata/core/scan/executor/util/QueryUtil.java
--
diff --git
a/core/src/main/java/org/apache/carbondata/core/scan/executor/util/Query
[HOTFIX] Removed BatchedDataSourceScanExec class and extended directly from
FileSourceScanExec
Problem:
Since some of the code of BatchedDataSourceScanExec is copied from spark, it is
difficult to maintain from version upgrades of spark. Currently we face issues
during spark 2.3 upgrade so bett
http://git-wip-us.apache.org/repos/asf/carbondata/blob/f9114036/core/src/main/java/org/apache/carbondata/core/datastore/block/SegmentTaskIndexWrapper.java
--
diff --git
a/core/src/main/java/org/apache/carbondata/core/datastore/blo
[CARBONDATA-2606][Complex DataType Enhancements]Fix Null result if First two
Projection column have same parent and third column has different Parent Struct
Problem:
When multiple columns are there,then the first child elements is only going
to make parent Object Array. For all other cases it sho
http://git-wip-us.apache.org/repos/asf/carbondata/blob/f9114036/core/src/test/java/org/apache/carbondata/core/datastore/SegmentTaskIndexStoreTest.java
--
diff --git
a/core/src/test/java/org/apache/carbondata/core/datastore/Segment
http://git-wip-us.apache.org/repos/asf/carbondata/blob/85cdc404/store/core/src/main/java/org/apache/carbondata/store/conf/StoreConf.java
--
diff --git
a/store/core/src/main/java/org/apache/carbondata/store/conf/StoreConf.java
b/s
[CARBONDATA-2609] Change RPC implementation to Hadoop RPC framework
This closes #2372
Project: http://git-wip-us.apache.org/repos/asf/carbondata/repo
Commit: http://git-wip-us.apache.org/repos/asf/carbondata/commit/d9b40bf9
Tree: http://git-wip-us.apache.org/repos/asf/carbondata/tree/d9b40bf9
Di
[CARBONDATA-2649] Fixed arrayIndexOutOfBoundException while loading Blocklet
DataMap after alter add column operation
Things done as part of this PR
Fixed arrayIndexOutOfBoundException while loading Blocklet DataMap after alter
add column operation
Problem:
Array Index out of bound exception wa
http://git-wip-us.apache.org/repos/asf/carbondata/blob/f9114036/integration/spark-common/src/main/scala/org/apache/carbondata/events/AlterTableEvents.scala
--
diff --git
a/integration/spark-common/src/main/scala/org/apache/carbond
Repository: carbondata
Updated Branches:
refs/heads/carbonstore 4b96ed8ca -> 239a6cadb (forced update)
http://git-wip-us.apache.org/repos/asf/carbondata/blob/d267c40b/integration/spark-common-test/src/test/scala/org/apache/carbondata/spark/testsuite/localdictionary/LocalDictionarySupportCreateT
http://git-wip-us.apache.org/repos/asf/carbondata/blob/85cdc404/store/core/src/main/java/org/apache/carbondata/store/rpc/model/BaseResponse.java
--
diff --git
a/store/core/src/main/java/org/apache/carbondata/store/rpc/model/BaseRe
[CARBONDATA-2528][MV] Fixed order by in mv and aggregation functions inside
projection expressions are fixed
Problem:
Order by queries and the queries with functions like sum(a)+sum(b) are not
working in MV.
Please check jira for more details.
Solution:
The queries which have projection function
[HOTFIX][CARBONDATA-2716][DataMap] fix bug for loading datamap
In some scenarios, input parameter of getCarbonFactDataHandlerModel called
carbonTable may be different from the one in loadmodel.
This close #2497
Project: http://git-wip-us.apache.org/repos/asf/carbondata/repo
Commit: http://git-
[CARBONDATA-2708][BloomDataMap] clear index file in case of data load failure
When data loading failed, clean the index DataMap files that generated
This closes #2463
Project: http://git-wip-us.apache.org/repos/asf/carbondata/repo
Commit: http://git-wip-us.apache.org/repos/asf/carbondata/commit
[CARBONDATA-2698][CARBONDATA-2700][CARBONDATA-2732][BloomDataMap] block some
operations of bloomfilter datamap
1.Block create bloomfilter datamap index on column which its datatype is
complex type;
2.Block change datatype for bloomfilter index datamap;
3.Block dropping index columns for bloomfil
http://git-wip-us.apache.org/repos/asf/carbondata/blob/4437920a/store/core/src/main/java/org/apache/carbondata/store/rpc/impl/StoreServiceImpl.java
--
diff --git
a/store/core/src/main/java/org/apache/carbondata/store/rpc/impl/Stor
[HotFix] Getting carbon table identifier to datamap events
Passing the table identifier to keep track of table in event in case preload
and postload of datamap event.
This closes #2448
Project: http://git-wip-us.apache.org/repos/asf/carbondata/repo
Commit: http://git-wip-us.apache.org/repos/as
http://git-wip-us.apache.org/repos/asf/carbondata/blob/4437920a/store/horizon/src/main/java/org/apache/carbondata/horizon/antlr/gen/ExpressionLexer.java
--
diff --git
a/store/horizon/src/main/java/org/apache/carbondata/horizon/ant
[CARBONDATA-2720] Remove dead code
For acturate coverage results and easy maintainance
This closes #2354
Project: http://git-wip-us.apache.org/repos/asf/carbondata/repo
Commit: http://git-wip-us.apache.org/repos/asf/carbondata/commit/f9114036
Tree: http://git-wip-us.apache.org/repos/asf/carbond
[CARBONDATA-2717] fixed table id empty problem while taking drop lock
This closes #2472
Project: http://git-wip-us.apache.org/repos/asf/carbondata/repo
Commit: http://git-wip-us.apache.org/repos/asf/carbondata/commit/637a9746
Tree: http://git-wip-us.apache.org/repos/asf/carbondata/tree/637a9746
http://git-wip-us.apache.org/repos/asf/carbondata/blob/4437920a/store/horizon/src/main/java/org/apache/carbondata/horizon/rest/model/view/LoadRequest.java
--
diff --git
a/store/horizon/src/main/java/org/apache/carbondata/horizon/r
http://git-wip-us.apache.org/repos/asf/carbondata/blob/73419071/integration/spark2/src/test/scala/org/apache/spark/carbondata/TestStreamingTableWithRowParser.scala
--
diff --git
a/integration/spark2/src/test/scala/org/apache/spark
[REBASE] Rebasing with master branch and Fixing rebase conflict
Project: http://git-wip-us.apache.org/repos/asf/carbondata/repo
Commit: http://git-wip-us.apache.org/repos/asf/carbondata/commit/239a6cad
Tree: http://git-wip-us.apache.org/repos/asf/carbondata/tree/239a6cad
Diff: http://git-wip-us.a
[CARBONDATA-2648] Fixed NPE issue with legacy store when CACHE_LEVEL is Blocklet
Things done as part of this PR:
Fixed Null pointer exception when store is of <= 1.1 version and DataMap is of
type BlockletDataMap.
Added clearing of SegmentProperties cache holder from executor
Problem 1:
Null poi
[CARBONDATA-2724][DataMap]Unsupported create datamap on table with V1 or V2
format data
block creating datamap on carbon table with V1 or V2 format
Currently the version info is read from carbon data file
This closes #2488
Project: http://git-wip-us.apache.org/repos/asf/carbondata/repo
Commit:
[CARBONDATA-2714] Support merge index files for the segment
Problem :
The first-time query of carbon becomes very slow. It is because of reading many
small carbonindex files and cache to the driver at the first time.
Many carbonindex files are created in below case
Loading data in large cluster
F
[CARBONDATA-2690][CarbonStore] implement RESTful API: create table, load data
and select
This PR adds:
1.basic framework
rewrite the carbon store's Master, Worker and Scheduler code in Java
2.RESTful API
support create a table by using file meta store
support load data to a table in single work
[CARBONDATA-2719] Block update and delete on table having datamaps
Table update/delete is needed to block on table which has datamaps.
This close #2483
Project: http://git-wip-us.apache.org/repos/asf/carbondata/repo
Commit: http://git-wip-us.apache.org/repos/asf/carbondata/commit/56e7dad7
Tree:
[CARBONDATA-2729][file-format] Schema Compatibility problem between version
1.3.0 and 1.4.0
Problem:
In TableSchema the field Name schemaEvaluation is changed to schemaEvoluation
and in DataMapSchema field name className is changed to providerName.
Due to this current Carbon version & Version 1.
http://git-wip-us.apache.org/repos/asf/carbondata/blob/2009009a/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/management/CarbonAddSegmentCommand.scala
--
diff --git
a/integration/spark2/src/main/scala/or
[CARBONDATA-2734] Fix struct of date issue in create table
problem: Struct of date is not supported currently in create table flow
as date datatype check is missing during parsing.
Hence child date column was not appended with parent name, leading to
StringOutOfIndex exception.
solution: Handle t
[CARBONDATA-2693][BloomDataMap]Fix bug for alter rename is renaming the
existing table on which bloomfilter datamp exists
Fix bug for alter rename is renaming the existing table on which bloom filter
datamap exists
This closes #2452
Project: http://git-wip-us.apache.org/repos/asf/carbondata/r
http://git-wip-us.apache.org/repos/asf/carbondata/blob/85cdc404/store/horizon/src/main/java/org/apache/carbondata/horizon/rest/model/descriptor/SelectDescriptor.java
--
diff --git
a/store/horizon/src/main/java/org/apache/carbondat
http://git-wip-us.apache.org/repos/asf/carbondata/blob/d9b40bf9/store/core/src/main/scala/org/apache/carbondata/store/Master.scala
--
diff --git a/store/core/src/main/scala/org/apache/carbondata/store/Master.scala
b/store/core/src
[CARBONDATA-2722] [CARBONDATA-2721] JsonWriter issue fixes
[CARBONDATA-2722][SDK] [JsonWriter] NPE when schema and data are not of same
length or Data is null.
problem: Null data is not handled in the json object to carbon row conversion.
solution: add a null check when object is fetched from j
http://git-wip-us.apache.org/repos/asf/carbondata/blob/d9b40bf9/store/search/src/main/scala/org/apache/spark/rpc/Scheduler.scala
--
diff --git a/store/search/src/main/scala/org/apache/spark/rpc/Scheduler.scala
b/store/search/src/m
[CARBONDATA-2704] Index file size in describe formatted command is not updated
correctly with the segment file
Problem:
Describe formatted command is not showing correct index files size after index
files merge.
Solution:
Segment file should be updated with the actual index files size of that se
Repository: carbondata
Updated Branches:
refs/heads/carbonstore 239a6cadb -> 9ac55a5a6
[CARBONDATA-2736][CARBONSTORE] Kafka integration with Carbon StreamSQL
Modification in this PR:
1.Pass source table properties to streamReader.load()
2.Do not pass schema when sparkSession.readStream
3.Remov
Repository: carbondata
Updated Branches:
refs/heads/master 7ab670652 -> 0ab03f21f
[CARBONDATA-2530][MV] Disable the MV datamaps after main table load
Problem:
MV datamaps are not disabled after the main table load is done. So the wrong
data is displaying.
Solution:
Disable the MV datamaps aft
Repository: carbondata
Updated Branches:
refs/heads/master 4014b0f54 -> 2c291d620
[CARBONDATA-2534][MV] Fix substring expression not working in MV creation
Problem: The column generated when subquery expression column present is wrong
while creating of MV table.
Solution: Corrected the column
Repository: carbondata
Updated Branches:
refs/heads/master 2c291d620 -> d820e3d51
[CARBONDATA-2542][MV] Fix the mv query from table with different database
Problem: database name is not added to the table name while generating mv query.
Solution: Add the database name to the table name while c
Repository: carbondata
Updated Branches:
refs/heads/master d820e3d51 -> 83562ae75
[CARBONDATA-2550][CARBONDATA-2576][MV] Fix limit and average function issue in
MV query
Problem: Limit is not working on mv queries and the average is also not working.
Solution: Correct the limit queries and av
Repository: carbondata
Updated Branches:
refs/heads/master 83562ae75 -> ce53b48a1
[CARBONDATA-2770][BloomDataMap] Optimize code to get blocklet id when
rebuilding datamap
we should get exactly number of blocklet id from blocklet scanned result
instead of building it ourselves.
This closes #
Repository: carbondata
Updated Branches:
refs/heads/master ce53b48a1 -> 1345dc6a3
[CARBONDATA-2512][32k] Support writing longstring through SDK
Support writing longstring through SDK.
User can specify the datatype as 'varchar' for longstring columns.
Please note that, the 'varchar' column cann
Repository: carbondata
Updated Branches:
refs/heads/master 1345dc6a3 -> fb2f9d33b
[CARBONDATA-2769] Fix bug when getting shard name from data before version 1.4
datamap creation needs shardname. carbon imported segment id in carbondata file
since version 1.4.
we should return proper shardName
Repository: carbondata
Updated Branches:
refs/heads/master e5c1568de -> 42a80564c
[CARBONDATA-2694][32k] Show longstring table property in descformatted
add longstring table property in the output of desc formantted command
This closes #2456
Project: http://git-wip-us.apache.org/repos/asf/c
Repository: carbondata
Updated Branches:
refs/heads/master 6f1767b5a -> 498502d2b
[CARBONDATA-2540][CARBONDATA-2560][CARBONDATA-2568][MV] Add validations for
unsupported MV queries
Problem: Validations are missing on the unsupported MV queries while creating
MV datamap.
Solution: Added valid
Repository: carbondata
Updated Branches:
refs/heads/master bea277f83 -> a75b9db6a
[CARBONDATA-2531][MV] Fix alias not working on MV query
Problem: when alias present on actual query then MV match not happening because
alias is not ignored.
Solution : Do semantic check while doing match
This
Repository: carbondata
Updated Branches:
refs/heads/master b0aee53f5 -> 68e5b52c4
[HOTFIX] Fix a spelling mistake after PR2511 merged
spelling mistakes: AtomicFileOperationsF modfity to:
AtomicFileOperationFactory.getAtomicFileOperations
This closes #2551
Project: http://git-wip-us.apache.
Repository: carbondata
Updated Branches:
refs/heads/master 68e5b52c4 -> 9a75ce53b
[CARBONDATA-2539][MV] Fix predicate subquery which uses leftsemi join
Problem: References to the top plan is not getting right when predicate
subquery is present.
Solution: Correct the refrences.
This closes #2
Repository: carbondata
Updated Branches:
refs/heads/external-format [created] a37a2ff7f
[CARBONDATA-2613] Support csv based carbon table
1. create csv based carbon table using
CREATE TABLE fact_table (col1 bigint, col2 string, ..., col100 string)
STORED BY 'CarbonData'
TBLPROPERTIES(
'foramt'='csv',
'csv.delimiter'=',',
'csv.header'='col1,col2,col100')
2. Load data to this tab
Repository: carbondata
Updated Branches:
refs/heads/external-format a37a2ff7f -> 0d2769f75
http://git-wip-us.apache.org/repos/asf/carbondata/blob/0d2769f7/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/management/CarbonAddSegmentCommand.scala
--
Repository: carbondata
Updated Branches:
refs/heads/carbonstore 7ad2fd951 -> 2d4628868
[CARBONDATA-2767][CarbonStore] Fix task locality issue
If the Spark cluster and the Hadoop cluster are two different machine cluster,
the Spark tasks will run in RACK_LOCAL mode.
This closes #2528
Projec
Repository: carbondata
Updated Branches:
refs/heads/master d62fe9e65 -> c79fc90d5
[CARBONDATA-2782]delete dead code in class 'CarbonCleanFilesCommand'
The variables(dmsãindexDms) in function processMetadata are nerver used.
This closes #2557
Project: http://git-wip-us.apache.org/repos/asf
Repository: carbondata
Updated Branches:
refs/heads/carbonstore 2d4628868 -> a6027ae11
[CARBONDATA-2776][CarbonStore] Support ingesting data from Kafka service
This closes #2544
Project: http://git-wip-us.apache.org/repos/asf/carbondata/repo
Commit: http://git-wip-us.apache.org/repos/asf/car
Repository: carbondata
Updated Branches:
refs/heads/master c29aef880 -> 6351c3a07
[CARBONDATA-2790][BloomDataMap]Optimize default parameter for bloomfilter
datamap
To provide better query performance for bloomfilter datamap by default,
we optimize bloom_size from 32000 to 64 and optimize
Repository: carbondata
Updated Branches:
refs/heads/master de9246066 -> f9b02a5c1
[CARBONDATA-2793][32k][Doc] Add 32k support in document
This closes #2572
Project: http://git-wip-us.apache.org/repos/asf/carbondata/repo
Commit: http://git-wip-us.apache.org/repos/asf/carbondata/commit/f9b02a5
Repository: carbondata
Updated Branches:
refs/heads/master f9b02a5c1 -> a302cd1ce
[CARBONDATA-2800][Doc] Add useful tips about bloomfilter datamap
add useful tips about bloomfilter datamap
This closes #2581
Project: http://git-wip-us.apache.org/repos/asf/carbondata/repo
Commit: http://git-w
Repository: carbondata
Updated Branches:
refs/heads/master a302cd1ce -> af984101e
[CARBONDATA-2806] Delete delete delta files upon clean files for flat folder
Problem:
Delete delta files are not removed after clean files operation.
Solution:
Get the delta files using Segment Status Manager and
Repository: carbondata
Updated Branches:
refs/heads/master af984101e -> 11fb422d9
[CARBONDATA-2796][32K]Fix data loading problem when table has complex column
and long string column
currently both varchar column and complex column believes itself is the last
one member in noDictionary group
Repository: carbondata
Updated Branches:
refs/heads/master 38384cb9f -> 26d9f3d8e
[CARBONDATA-2802][BloomDataMap] Remove clearing cache after rebuiding index
datamap
This is no need to clear cache after rebuilding index datamap due to the
following reasons:
1.currently it will clear all the
Repository: carbondata
Updated Branches:
refs/heads/master b702a1b01 -> abcd4f6e2
[CARBONDATA-2809][DataMap] Block rebuilding for bloom/lucene and preagg datamap
As manual refresh currently only works fine for MV, it has some bugs
with other types of datamap such as preaggregate, timeserials,
Repository: carbondata
Updated Branches:
refs/heads/master abcd4f6e2 -> b9e510640
[CARBONDATA-2823] Support streaming property with datamap
Since during query, carbondata get splits from streaming segment and
columnar segments repectively, we can support streaming with index
datamap.
For prea
Repository: carbondata
Updated Branches:
refs/heads/master f27efb3e3 -> 78438451b
[CARBONDATA-2585] Fix local dictionary for both table level and system level
property based on priority
Added a System level Property for local dictionary Support.
Property 'carbon.local.dictionary.enable' can b
Repository: carbondata
Updated Branches:
refs/heads/master 78438451b -> 3d7fa1276
[CARBONDATA-2539]Fix mv classcast exception issue
Class cast exception happens during min type aggregate function happening. It
is corrected in this PR
This closes #2602
Project: http://git-wip-us.apache.org/
Repository: carbondata
Updated Branches:
refs/heads/external-format 0d2769f75 -> ccf64ce5a
[CARBONDATA-2768][CarbonStore] Fix error in tests for external csv format
In previous implementation earlier than PR2495, we only supportted csv as
external format for carbondata. And we validated the re
[CARBONDATA-2784][CARBONDATA-2786][SDK writer] Fixed:Forever blocking wait with
more than 21 batch of data
problem: [CARBONDATA-2784]
[SDK writer] Forever blocking wait with more than 21 batch of data, when
consumer is dead due to data loading exception (bad record / out of memory)
root cause:
[HOTFIX][PR 2575] Fixed modular plan creation only if valid datamaps are
available
update query is failing in spark-2.2 cluster if mv jars are available because
catalogs
are not empty if datamap are created for other table also and returns true from
isValidPlan() inside MVAnalyzerRule.
This cl
601 - 700 of 2391 matches
Mail list logo