[carbondata] branch master updated: [CARBONDATA-4329] Fix multiple issues with External table

2022-04-01 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new 46b62cf  [CARBONDATA-4329] Fix multiple issues with External table
46b62cf is described below

commit 46b62cf6f79d1d826b498609435337b2ed342bbe
Author: Indhumathi27 
AuthorDate: Wed Mar 23 11:16:01 2022 +0530

[CARBONDATA-4329] Fix multiple issues with External table

Why is this PR needed?
Issue 1:
When we create external table on transactional table location,
schema file will be present. While creating external table,
which is also transactional, the schema file is overwritten

Issue 2:
If external table is created on a location, where the source table
already exists, on drop external table, it is deleting the table data.
Query on the source table fails

What changes were proposed in this PR?
Avoid writing schema file if table type is external and transactional
Dont drop external table location data, if table_type is external

This closes #4255
---
 .../core/metadata/schema/table/CarbonTable.java|   3 +-
 .../core/metadata/schema/table/TableInfo.java  |   4 +
 .../org/apache/spark/sql/CarbonCountStar.scala |  35 ---
 .../scala/org/apache/spark/sql/CarbonSource.scala  |   4 +-
 .../command/table/CarbonDropTableCommand.scala |   2 +-
 .../spark/sql/hive/CarbonFileMetastore.scala   |   6 ++
 .../sql/parser/CarbonSparkSqlParserUtil.scala  |   2 +-
 .../createTable/TestCreateExternalTable.scala  | 107 +
 8 files changed, 147 insertions(+), 16 deletions(-)

diff --git 
a/core/src/main/java/org/apache/carbondata/core/metadata/schema/table/CarbonTable.java
 
b/core/src/main/java/org/apache/carbondata/core/metadata/schema/table/CarbonTable.java
index 2c22142..1e7dfee 100644
--- 
a/core/src/main/java/org/apache/carbondata/core/metadata/schema/table/CarbonTable.java
+++ 
b/core/src/main/java/org/apache/carbondata/core/metadata/schema/table/CarbonTable.java
@@ -880,8 +880,7 @@ public class CarbonTable implements Serializable, Writable {
* an internal table property set during table creation)
*/
   public boolean isExternalTable() {
-String external = 
tableInfo.getFactTable().getTableProperties().get("_external");
-return external != null && external.equalsIgnoreCase("true");
+return tableInfo.isExternal();
   }
 
   public boolean isFileLevelFormat() {
diff --git 
a/core/src/main/java/org/apache/carbondata/core/metadata/schema/table/TableInfo.java
 
b/core/src/main/java/org/apache/carbondata/core/metadata/schema/table/TableInfo.java
index 1ce7230..84fea45 100644
--- 
a/core/src/main/java/org/apache/carbondata/core/metadata/schema/table/TableInfo.java
+++ 
b/core/src/main/java/org/apache/carbondata/core/metadata/schema/table/TableInfo.java
@@ -289,6 +289,10 @@ public class TableInfo implements Serializable, Writable {
 return isTransactionalTable;
   }
 
+  public boolean isExternal() {
+return 
Boolean.parseBoolean(factTable.getTableProperties().getOrDefault("_external", 
"false"));
+  }
+
   public void setTransactionalTable(boolean transactionalTable) {
 isTransactionalTable = transactionalTable;
   }
diff --git 
a/integration/spark/src/main/scala/org/apache/spark/sql/CarbonCountStar.scala 
b/integration/spark/src/main/scala/org/apache/spark/sql/CarbonCountStar.scala
index 678a0be..4d1ed11 100644
--- 
a/integration/spark/src/main/scala/org/apache/spark/sql/CarbonCountStar.scala
+++ 
b/integration/spark/src/main/scala/org/apache/spark/sql/CarbonCountStar.scala
@@ -24,6 +24,7 @@ import org.apache.hadoop.mapreduce.Job
 import org.apache.hadoop.mapreduce.lib.input.FileInputFormat
 import org.apache.spark.rdd.RDD
 import org.apache.spark.sql.catalyst.{InternalRow, TableIdentifier}
+import org.apache.spark.sql.catalyst.analysis.NoSuchTableException
 import org.apache.spark.sql.catalyst.expressions._
 import org.apache.spark.sql.execution.LeafExecNode
 import org.apache.spark.sql.optimizer.CarbonFilters
@@ -54,17 +55,29 @@ case class CarbonCountStar(
 CarbonInputFormat.setQuerySegment(job.getConfiguration, carbonTable)
 
 // get row count
-var rowCount = CarbonUpdateUtil.getRowCount(
-  tableInputFormat.getBlockRowCount(
-job,
-carbonTable,
-CarbonFilters.getPartitions(
-  Seq.empty,
-  sparkSession,
-  TableIdentifier(
-carbonTable.getTableName,
-Some(carbonTable.getDatabaseName))).map(_.toList.asJava).orNull, 
false),
-  carbonTable)
+var rowCount = try {
+  CarbonUpdateUtil.getRowCount(
+tableInputFormat.getBlockRowCount(
+  job,
+  carbonTable,
+  CarbonFilters.getPartitions(
+Seq.empty,
+  

[carbondata-site] 02/02: removed 1.6.0 download links

2022-03-21 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch asf-site
in repository https://gitbox.apache.org/repos/asf/carbondata-site.git

commit 098e57f2e445bb5904df18932c29544c107a2cb7
Author: kunal642 
AuthorDate: Mon Mar 21 22:34:46 2022 +0530

removed 1.6.0 download links
---
 content/js/custom.js | 1 -
 src/main/webapp/js/custom.js | 1 -
 2 files changed, 2 deletions(-)

diff --git a/content/js/custom.js b/content/js/custom.js
index e4ebb01..e5e8708 100644
--- a/content/js/custom.js
+++ b/content/js/custom.js
@@ -25,7 +25,6 @@ releaseMap.set("2.3.0", ["3.1.1", "2.4.5", "2.3.4"])
 releaseMap.set("2.2.0", ["3.1.1", "2.4.5", "2.3.4"])
 releaseMap.set("2.1.1", ["2.4.5", "2.3.4"])
 releaseMap.set("2.0.1", ["2.4.5", "2.3.4"])
-releaseMap.set("1.6.0", ["2.1.0", "2.2.1", "2.3.2"])
 
 var selectedRelease = "2.3.0"
 var selectedPackage = ""
diff --git a/src/main/webapp/js/custom.js b/src/main/webapp/js/custom.js
index e4ebb01..e5e8708 100644
--- a/src/main/webapp/js/custom.js
+++ b/src/main/webapp/js/custom.js
@@ -25,7 +25,6 @@ releaseMap.set("2.3.0", ["3.1.1", "2.4.5", "2.3.4"])
 releaseMap.set("2.2.0", ["3.1.1", "2.4.5", "2.3.4"])
 releaseMap.set("2.1.1", ["2.4.5", "2.3.4"])
 releaseMap.set("2.0.1", ["2.4.5", "2.3.4"])
-releaseMap.set("1.6.0", ["2.1.0", "2.2.1", "2.3.2"])
 
 var selectedRelease = "2.3.0"
 var selectedPackage = ""


[carbondata-site] branch asf-site updated (0c81575 -> 098e57f)

2022-03-21 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a change to branch asf-site
in repository https://gitbox.apache.org/repos/asf/carbondata-site.git.


from 0c81575  removed download link from index.html
 new 58b4f2e  removed 1.6.0 download links
 new 098e57f  removed 1.6.0 download links

The 2 revisions listed above as "new" are entirely new to this
repository and will be described in separate emails.  The revisions
listed as "add" were already present in the repository and have only
been added to this reference.


Summary of changes:
 content/js/custom.js | 1 -
 src/main/webapp/js/custom.js | 1 -
 2 files changed, 2 deletions(-)


[carbondata-site] 01/02: removed 1.6.0 download links

2022-03-21 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch asf-site
in repository https://gitbox.apache.org/repos/asf/carbondata-site.git

commit 58b4f2e105ea0ed180311970e82e0ab14829a4cb
Author: kunal642 
AuthorDate: Mon Mar 21 22:34:10 2022 +0530

removed 1.6.0 download links
---
 content/js/custom.js | 2 +-
 src/main/webapp/js/custom.js | 2 +-
 2 files changed, 2 insertions(+), 2 deletions(-)

diff --git a/content/js/custom.js b/content/js/custom.js
index 5b165bf..e4ebb01 100644
--- a/content/js/custom.js
+++ b/content/js/custom.js
@@ -25,7 +25,7 @@ releaseMap.set("2.3.0", ["3.1.1", "2.4.5", "2.3.4"])
 releaseMap.set("2.2.0", ["3.1.1", "2.4.5", "2.3.4"])
 releaseMap.set("2.1.1", ["2.4.5", "2.3.4"])
 releaseMap.set("2.0.1", ["2.4.5", "2.3.4"])
-releaseMap.set("1.6.0", ["2.4.5", "2.3.4"])
+releaseMap.set("1.6.0", ["2.1.0", "2.2.1", "2.3.2"])
 
 var selectedRelease = "2.3.0"
 var selectedPackage = ""
diff --git a/src/main/webapp/js/custom.js b/src/main/webapp/js/custom.js
index 5b165bf..e4ebb01 100644
--- a/src/main/webapp/js/custom.js
+++ b/src/main/webapp/js/custom.js
@@ -25,7 +25,7 @@ releaseMap.set("2.3.0", ["3.1.1", "2.4.5", "2.3.4"])
 releaseMap.set("2.2.0", ["3.1.1", "2.4.5", "2.3.4"])
 releaseMap.set("2.1.1", ["2.4.5", "2.3.4"])
 releaseMap.set("2.0.1", ["2.4.5", "2.3.4"])
-releaseMap.set("1.6.0", ["2.4.5", "2.3.4"])
+releaseMap.set("1.6.0", ["2.1.0", "2.2.1", "2.3.2"])
 
 var selectedRelease = "2.3.0"
 var selectedPackage = ""


[carbondata-site] branch asf-site updated: removed download link from index.html

2022-03-21 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch asf-site
in repository https://gitbox.apache.org/repos/asf/carbondata-site.git


The following commit(s) were added to refs/heads/asf-site by this push:
 new 0c81575  removed download link from index.html
0c81575 is described below

commit 0c815753b51c7af2c52ebf1e0d3f716ba0c5ac69
Author: kunal642 
AuthorDate: Mon Mar 21 22:05:13 2022 +0530

removed download link from index.html
---
 content/index.html | 18 ++
 src/main/webapp/index.html | 18 ++
 2 files changed, 12 insertions(+), 24 deletions(-)

diff --git a/content/index.html b/content/index.html
index fefeae8..92f123b 100644
--- a/content/index.html
+++ b/content/index.html
@@ -292,43 +292,37 @@
 
 
 
-https://dlcdn.apache.org/carbondata/2.3.0/;
-   target="_blank">Apache CarbonData 
2.3.0
+Apache CarbonData 2.3.0 
 Mar 
2022
 https://cwiki.apache.org/confluence/display/CARBONDATA/Apache+CarbonData+2.3.0+Release;
class="whatsnew" 
target="_blank">what's new
 
 
-https://dlcdn.apache.org/carbondata/2.2.0/;
-   target="_blank">Apache CarbonData 
2.2.0
+Apache CarbonData 2.2.0
 Aug 
2021
 https://cwiki.apache.org/confluence/display/CARBONDATA/Apache+CarbonData+2.2.0+Release;
class="whatsnew" 
target="_blank">what's new
 
 
-https://dlcdn.apache.org/carbondata/2.1.1/;
-   target="_blank">Apache CarbonData 
2.1.1
+Apache CarbonData 2.1.1
 May 
2021
 https://cwiki.apache.org/confluence/display/CARBONDATA/Apache+CarbonData+2.1.1+Release;
class="whatsnew" 
target="_blank">what's new
 
 
-https://dlcdn.apache.org/carbondata/2.0.1/;
-   target="_blank">Apache CarbonData 
2.0.1
+Apache CarbonData 2.0.1
 Jun 
2020
 https://cwiki.apache.org/confluence/display/CARBONDATA/Apache+CarbonData+2.0.1+Release;
class="whatsnew" 
target="_blank">what's new
 
 
-https://dlcdn.apache.org/carbondata/1.6.1/;
-   target="_blank">Apache CarbonData 
1.6.1
+Apache CarbonData 1.6.1
 Oct 
2019
 https://cwiki.apache.org/confluence/display/CARBONDATA/Apache+CarbonData+1.6.1+Release;
class="whatsnew" 
target="_blank">what's new
 
 
-https://archive.apache.org/dist/carbondata/1.5.4/;
-   target="_blank">Apache CarbonData 
1.5.4
+Apache CarbonData 1.5.4
 May 
2019
 https://cwiki.apache.org/confluence/display/CARBONDATA/Apache+CarbonData+1.5.4+Release;
class="whatsnew" 
target="_blank">what's new
diff --git a/src/main/webapp/index.html b/src/main/webapp/index.html
index fefeae8..92f123b 100644
--- a/src/main/webapp/index.html
+++ b/src/main/webapp/index.html
@@ -292,43 +292,37 @@
 
 
 
-https://dlcdn.apache.org/carbondata/2.3.0/;
-   target="_blank"&

[carbondata-site] branch asf-site updated (f7470ac -> a1d71d0)

2022-03-21 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a change to branch asf-site
in repository https://gitbox.apache.org/repos/asf/carbondata-site.git.


 discard f7470ac  Added a sperate page to download latest releases

This update removed existing revisions from the reference, leaving the
reference pointing at a previous point in the repository history.

 * -- * -- N   refs/heads/asf-site (a1d71d0)
\
 O -- O -- O   (f7470ac)

Any revisions marked "omit" are not gone; other references still
refer to them.  Any revisions marked "discard" are gone forever.

No new revisions were added by this update.

Summary of changes:
 content/alluxio-guide.html |  44 -
 content/bloomfilter-datamap-guide.html |  26 ++-
 content/bloomfilter-index-guide.html   |  44 -
 content/carbon-as-spark-datasource-guide.html  |  44 -
 content/clean-files.html   |  44 -
 content/configuration-parameters.html  |  44 -
 content/csdk-guide.html|  44 -
 content/datamap-developer-guide.html   |  26 ++-
 content/datamap-management.html|  26 ++-
 content/ddl-of-carbondata.html |  44 -
 content/dml-of-carbondata.html |  44 -
 content/documentation.html |  44 -
 content/downloads.html | 203 -
 content/faq.html   |  44 -
 content/file-structure-of-carbondata.html  |  44 -
 content/flink-integration-guide.html   |  44 -
 content/hive-guide.html|  44 -
 .../how-to-contribute-to-apache-carbondata.html|  44 -
 content/index-developer-guide.html |  44 -
 content/index-management.html  |  44 -
 content/index-server.html  |  44 -
 content/index.html |  30 ++-
 content/introduction.html  |  44 -
 content/js/custom.js   |  70 ---
 content/language-manual.html   |  44 -
 content/lucene-datamap-guide.html  |  26 ++-
 content/lucene-index-guide.html|  44 -
 content/mv-guide.html  |  44 -
 content/performance-tuning.html|  44 -
 content/preaggregate-datamap-guide.html|  26 ++-
 content/prestodb-guide.html|  44 -
 content/prestosql-guide.html   |  44 -
 content/quick-start-guide.html |  44 -
 content/release-guide.html |  44 -
 content/s3-guide.html  |  44 -
 content/scd-and-cdc-guide.html |  44 -
 content/sdk-guide.html |  44 -
 content/secondary-index-guide.html |  44 -
 content/security.html  |  44 -
 content/segment-management-on-carbondata.html  |  44 -
 content/spatial-index-guide.html   |  44 -
 content/streaming-guide.html   |  44 -
 content/supported-data-types-in-carbondata.html|  44 -
 content/timeseries-datamap-guide.html  |  26 ++-
 content/usecases.html  |  44 -
 content/videogallery.html  |  44 -
 src/main/webapp/alluxio-guide.html |  44 -
 src/main/webapp/bloomfilter-datamap-guide.html |  26 ++-
 src/main/webapp/bloomfilter-index-guide.html   |  44 -
 .../webapp/carbon-as-spark-datasource-guide.html   |  44 -
 src/main/webapp/clean-files.html   |  44 -
 src/main/webapp/configuration-parameters.html  |  44 -
 src/main/webapp/csdk-guide.html|  44 -
 src/main/webapp/datamap-developer-guide.html   |  26 ++-
 src/main/webapp/datamap-management.html|  26 ++-
 src/main/webapp/ddl-of-carbondata.html |  44 -
 src/main/webapp/dml-of-carbondata.html |  44 -
 src/main/webapp/documentation.html |  44 -
 src/main/webapp/downloads.html | 203 -
 src/main/webapp/faq.html   |  44 -
 src/main/webapp/file-structure-of-carbondata.html  |  44 -
 src/main/webapp/flink-integration-guide.html   |  44 -
 src/main/webapp/hive-guide.html|  44 -
 .../how-to-contribute-to-apache-carbondata.html|  44 -
 src/main/webapp/index-developer-guide.html |  44 -
 src/main/webapp/index-management.html  |  44 -
 src/main/webapp/index-server.html  

[carbondata-site] branch asf-site updated (98d8e76 -> a1d71d0)

2022-03-21 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a change to branch asf-site
in repository https://gitbox.apache.org/repos/asf/carbondata-site.git.


 discard 98d8e76  Added a sperate page to download latest releases

This update removed existing revisions from the reference, leaving the
reference pointing at a previous point in the repository history.

 * -- * -- N   refs/heads/asf-site (a1d71d0)
\
 O -- O -- O   (98d8e76)

Any revisions marked "omit" are not gone; other references still
refer to them.  Any revisions marked "discard" are gone forever.

No new revisions were added by this update.

Summary of changes:
 content/alluxio-guide.html |  44 -
 content/bloomfilter-datamap-guide.html |  26 ++-
 content/bloomfilter-index-guide.html   |  44 -
 content/carbon-as-spark-datasource-guide.html  |  44 -
 content/clean-files.html   |  44 -
 content/configuration-parameters.html  |  44 -
 content/csdk-guide.html|  44 -
 content/datamap-developer-guide.html   |  26 ++-
 content/datamap-management.html|  26 ++-
 content/ddl-of-carbondata.html |  44 -
 content/dml-of-carbondata.html |  44 -
 content/documentation.html |  44 -
 content/downloads.html | 203 -
 content/faq.html   |  44 -
 content/file-structure-of-carbondata.html  |  44 -
 content/flink-integration-guide.html   |  44 -
 content/hive-guide.html|  44 -
 .../how-to-contribute-to-apache-carbondata.html|  44 -
 content/index-developer-guide.html |  44 -
 content/index-management.html  |  44 -
 content/index-server.html  |  44 -
 content/index.html |  30 ++-
 content/introduction.html  |  44 -
 content/js/custom.js   |  70 ---
 content/language-manual.html   |  44 -
 content/lucene-datamap-guide.html  |  26 ++-
 content/lucene-index-guide.html|  44 -
 content/mv-guide.html  |  44 -
 content/performance-tuning.html|  44 -
 content/preaggregate-datamap-guide.html|  26 ++-
 content/prestodb-guide.html|  44 -
 content/prestosql-guide.html   |  44 -
 content/quick-start-guide.html |  44 -
 content/release-guide.html |  44 -
 content/s3-guide.html  |  44 -
 content/scd-and-cdc-guide.html |  44 -
 content/sdk-guide.html |  44 -
 content/secondary-index-guide.html |  44 -
 content/security.html  |  44 -
 content/segment-management-on-carbondata.html  |  44 -
 content/spatial-index-guide.html   |  44 -
 content/streaming-guide.html   |  44 -
 content/supported-data-types-in-carbondata.html|  44 -
 content/timeseries-datamap-guide.html  |  26 ++-
 content/usecases.html  |  44 -
 content/videogallery.html  |  44 -
 src/main/webapp/alluxio-guide.html |  44 -
 src/main/webapp/bloomfilter-datamap-guide.html |  26 ++-
 src/main/webapp/bloomfilter-index-guide.html   |  44 -
 .../webapp/carbon-as-spark-datasource-guide.html   |  44 -
 src/main/webapp/clean-files.html   |  44 -
 src/main/webapp/configuration-parameters.html  |  44 -
 src/main/webapp/csdk-guide.html|  44 -
 src/main/webapp/datamap-developer-guide.html   |  26 ++-
 src/main/webapp/datamap-management.html|  26 ++-
 src/main/webapp/ddl-of-carbondata.html |  44 -
 src/main/webapp/dml-of-carbondata.html |  44 -
 src/main/webapp/documentation.html |  44 -
 src/main/webapp/downloads.html | 203 -
 src/main/webapp/faq.html   |  44 -
 src/main/webapp/file-structure-of-carbondata.html  |  44 -
 src/main/webapp/flink-integration-guide.html   |  44 -
 src/main/webapp/hive-guide.html|  44 -
 .../how-to-contribute-to-apache-carbondata.html|  44 -
 src/main/webapp/index-developer-guide.html |  44 -
 src/main/webapp/index-management.html  |  44 -
 src/main/webapp/index-server.html  

svn commit: r53231 - in /release/carbondata: 2.0.1/ 2.1.1/ 2.2.0/ 2.3.0/

2022-03-21 Thread kunalkapoor
Author: kunalkapoor
Date: Mon Mar 21 10:32:29 2022
New Revision: 53231

Log:
deleted md5 files

Removed:
release/carbondata/2.0.1/apache-carbondata-2.0.1-source-release.zip.md5
release/carbondata/2.1.1/apache-carbondata-2.1.1-source-release.zip.md5
release/carbondata/2.2.0/apache-carbondata-2.2.0-source-release.zip.md5
release/carbondata/2.3.0/apache-carbondata-2.3.0-source-release.zip.md5



[carbondata] branch master updated: [CARBONDATA-4327] Update documentation related to partition

2022-03-17 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new 41831ce  [CARBONDATA-4327] Update documentation related to partition
41831ce is described below

commit 41831cefa60d9df8f33f1603531b18b8def1b8fa
Author: ShreelekhyaG 
AuthorDate: Thu Mar 17 19:41:05 2022 +0530

[CARBONDATA-4327] Update documentation related to partition

Why is this PR needed?
Drop partition with data is not supported and a few of the links are not 
working.

What changes were proposed in this PR?
Removed unsupported syntax , duplicate headings and updated the header with 
proper linkage.

This closes #4254
---
 docs/ddl-of-carbondata.md | 63 ---
 1 file changed, 16 insertions(+), 47 deletions(-)

diff --git a/docs/ddl-of-carbondata.md b/docs/ddl-of-carbondata.md
index dbf616b..3d04684 100644
--- a/docs/ddl-of-carbondata.md
+++ b/docs/ddl-of-carbondata.md
@@ -60,12 +60,11 @@ CarbonData DDL statements are documented here,which 
includes:
   * [REFRESH TABLE](#refresh-table)
   * [COMMENTS](#table-and-column-comment)
 * [PARTITION](#partition)
-  * [STANDARD PARTITION(HIVE)](#standard-partition)
-* [INSERT OVERWRITE PARTITION](#insert-overwrite)
+  * [CREATE PARTITION](#create-partition-table)
   * [SHOW PARTITIONS](#show-partitions)
-  * [ADD PARTITION](#add-a-new-partition)
-  * [SPLIT PARTITION](#split-a-partition)
-  * [DROP PARTITION](#drop-a-partition)
+  * [ADD PARTITION](#add-partition)
+  * [DROP PARTITION](#drop-partition)
+  * [INSERT OVERWRITE PARTITION](#insert-overwrite)
 * [BUCKETING](#bucketing)
 * [CACHE](#cache)
 
@@ -1085,9 +1084,20 @@ Users can specify which columns to include and exclude 
for local dictionary gene
   SHOW PARTITIONS [db_name.]table_name
   ```
 
+### Add Partition
+
+  This command adds the specified Hive partition.
+  ```
+  ALTER TABLE [db_name].table_name ADD PARTITION(part_spec, ...)
+  ```
+  Example:
+  ```
+  ALTER TABLE locationTable ADD PARTITION (country = 'US');
+  ```
+
  Drop Partition
 
-  This command drops the specified Hive partition only.
+  This command drops the specified Hive partition.
   ```
   ALTER TABLE table_name DROP [IF EXISTS] PARTITION (part_spec, ...)
   ```
@@ -1115,47 +1125,6 @@ Users can specify which columns to include and exclude 
for local dictionary gene
   WHERE au.country = 'US';
   ```
 
-### Show Partitions
-
-  The following command is executed to get the partition information of the 
table
-
-  ```
-  SHOW PARTITIONS [db_name.]table_name
-  ```
-
-### Add a new partition
-
-  ```
-  ALTER TABLE [db_name].table_name ADD PARTITION('new_partition')
-  ```
-
-### Drop a partition
-
-  Only drop partition definition, but keep data
-  ```
-  ALTER TABLE [db_name].table_name DROP PARTITION(partition_id)
-  ```
-
-  Drop both partition definition and data
-  ```
-  ALTER TABLE [db_name].table_name DROP PARTITION(partition_id) WITH DATA
-  ```
-
-  **NOTE:**
-  * Hash partition table is not supported for ADD, SPLIT and DROP commands.
-  * Partition Id: in CarbonData like the hive, folders are not used to divide 
partitions instead partition id is used to replace the task id. It could make 
use of the characteristic and meanwhile reduce some metadata.
-
-  ```
-  SegmentDir/0_batchno0-0-1502703086921.carbonindex
-^
-  SegmentDir/part-0-0_batchno0-0-1502703086921.carbondata
- ^
-  ```
-
-  Here are some useful tips to improve query performance of carbonData 
partition table:
-  * The partitioned column can be excluded from SORT_COLUMNS, this will let 
other columns to do the efficient sorting.
-  * When writing SQL on a partition table, try to use filters on the partition 
column.
-
 ## BUCKETING
 
   Bucketing feature can be used to distribute/organize the table/partition 
data into multiple files such


svn commit: r52992 - /release/carbondata/2.3.0/

2022-03-11 Thread kunalkapoor
Author: kunalkapoor
Date: Sat Mar 12 05:17:27 2022
New Revision: 52992

Log:
Uploading 2.3.0 signature files

Added:

release/carbondata/2.3.0/apache-carbondata-2.3.0-bin-spark2.3.4-hadoop2.7.2.jar.asc

release/carbondata/2.3.0/apache-carbondata-2.3.0-bin-spark2.3.4-hadoop2.7.2.jar.sha512

release/carbondata/2.3.0/apache-carbondata-2.3.0-bin-spark2.4.5-hadoop2.7.2.jar.asc

release/carbondata/2.3.0/apache-carbondata-2.3.0-bin-spark2.4.5-hadoop2.7.2.jar.sha512

release/carbondata/2.3.0/apache-carbondata-2.3.0-bin-spark3.1.1-hadoop2.7.2.jar.asc

release/carbondata/2.3.0/apache-carbondata-2.3.0-bin-spark3.1.1-hadoop2.7.2.jar.sha512
release/carbondata/2.3.0/apache-carbondata-2.3.0-source-release.zip   (with 
props)
release/carbondata/2.3.0/apache-carbondata-2.3.0-source-release.zip.asc
release/carbondata/2.3.0/apache-carbondata-2.3.0-source-release.zip.md5
release/carbondata/2.3.0/apache-carbondata-2.3.0-source-release.zip.sha512

Added: 
release/carbondata/2.3.0/apache-carbondata-2.3.0-bin-spark2.3.4-hadoop2.7.2.jar.asc
==
--- 
release/carbondata/2.3.0/apache-carbondata-2.3.0-bin-spark2.3.4-hadoop2.7.2.jar.asc
 (added)
+++ 
release/carbondata/2.3.0/apache-carbondata-2.3.0-bin-spark2.3.4-hadoop2.7.2.jar.asc
 Sat Mar 12 05:17:27 2022
@@ -0,0 +1,14 @@
+-BEGIN PGP SIGNATURE-
+
+iQGzBAABCgAdFiEER4oGoEjiZ19CtKBtzP3MO+Fg6doFAmIsK9UACgkQzP3MO+Fg
+6dppUgv8D3OXKxzXOXfZ2NBhRxtZHIbfoCC5SAaHopsa6IDPI80MKxaKoriqhTQO
+SF/PWFFSiNPMejjXXk4Vcg4dgO+SLc/PxwPc4w71eFsAvdqsmqHFhty+/xUtZyjt
+rJx6J3cepF1ChNeRK/NqNMp00+1slHOZ5cG3vYMrGwfmfCCMic3zCIsljsJYI6uw
+YeHV5yjLNdWi0m4OFHvkSWCQ/8OeDeot4lVz/2xTmLTWD+Xvz8pbv0MRjy6fO72D
+TjwuzCYraZInp2lAOVbNcREdxyge/MML5Cwzuk6Hz5vwRolLRp6g91HvI4oQSm8J
+LluYmHgiTvdzq5GjuG84V2ABfWOMVl/oICJQUk1YfRS7YguSYw052fZb3q4z9NBQ
+1jCY+u1jpv2VfHwRIrJ77V3RJ88zUTOd3fLjsHNRYOyN6XmcdVBiz9gQ6LNYuLuI
+Agw6YUd/b1bAAc5sLVBSZjJ2U50M3Xm7160Yap0FtAvp04jvUdh6qOzhiUi+ev0r
+og+eRhMG
+=XiRF
+-END PGP SIGNATURE-

Added: 
release/carbondata/2.3.0/apache-carbondata-2.3.0-bin-spark2.3.4-hadoop2.7.2.jar.sha512
==
--- 
release/carbondata/2.3.0/apache-carbondata-2.3.0-bin-spark2.3.4-hadoop2.7.2.jar.sha512
 (added)
+++ 
release/carbondata/2.3.0/apache-carbondata-2.3.0-bin-spark2.3.4-hadoop2.7.2.jar.sha512
 Sat Mar 12 05:17:27 2022
@@ -0,0 +1 @@
+470c12affdb547c79ad12e4848fce7bea3ceae2ecbd6dc1a2efbd03cce58cbe3cdaab973e583421385d39c3ba3116c12dc4ceb92c1e501c60d4dd3653ef2026a
  apache-carbondata-2.3.0-bin-spark2.3.4-hadoop2.7.2.jar

Added: 
release/carbondata/2.3.0/apache-carbondata-2.3.0-bin-spark2.4.5-hadoop2.7.2.jar.asc
==
--- 
release/carbondata/2.3.0/apache-carbondata-2.3.0-bin-spark2.4.5-hadoop2.7.2.jar.asc
 (added)
+++ 
release/carbondata/2.3.0/apache-carbondata-2.3.0-bin-spark2.4.5-hadoop2.7.2.jar.asc
 Sat Mar 12 05:17:27 2022
@@ -0,0 +1,14 @@
+-BEGIN PGP SIGNATURE-
+
+iQGzBAABCgAdFiEER4oGoEjiZ19CtKBtzP3MO+Fg6doFAmIsK+QACgkQzP3MO+Fg
+6dq5vgwA12CPcKYVfHqReujLkV9FkxZtFT4MHgtUHfTf+JT7P43twIATM8uPVoeb
+icvGyUYAPtOw718v4DEUrH/CfMZ0/FHF5p2/6sj0Ma/c9JTtyO3fzWO0NXo2IbAb
+NwRNmJ/sDZJhNNgo2wdSvG2fLAmEBb/b05TCFnXMlCz9rI1A8yqV45Ir3RjFgsB8
+tmxQSctTE+LSpMjNFPMfIydx/WrxbuBriwuZ7wiNJne8uIzSP2kpftrXF4vTsttw
+Qkt9k+ppHkR0VU5uz7+vvhzFLpGBA/uhv3KsT6D8AwECxrsozYyP9k/W7nZDSe4D
+mlr69SARprRFVq37Px9gi1awGXBlEagdr3vW6agLCzcuZuLZ+Se4VNcTy8wj9kNS
+S7oXcfEvkoxp5lZdW5LItsPbW/XDYKbww5tsY1W+yk6I0MLuI/6JQiX4d9bw47qw
+RLUyFg+GkZoWfRiNWNfg1teFC42fzUQVz7GFeMdeumh/XhM5ffmtiaY1qJ/TJiPI
+Vxdr0osH
+=vG4V
+-END PGP SIGNATURE-

Added: 
release/carbondata/2.3.0/apache-carbondata-2.3.0-bin-spark2.4.5-hadoop2.7.2.jar.sha512
==
--- 
release/carbondata/2.3.0/apache-carbondata-2.3.0-bin-spark2.4.5-hadoop2.7.2.jar.sha512
 (added)
+++ 
release/carbondata/2.3.0/apache-carbondata-2.3.0-bin-spark2.4.5-hadoop2.7.2.jar.sha512
 Sat Mar 12 05:17:27 2022
@@ -0,0 +1 @@
+7abc0a30d641746388ddac13258003c4d486956d8955cd5389dde40e423f259efea337c270bccdf6f4ae469d17269ff6e9e771e28a36afe18cf0b73730690f7f
  apache-carbondata-2.3.0-bin-spark2.4.5-hadoop2.7.2.jar

Added: 
release/carbondata/2.3.0/apache-carbondata-2.3.0-bin-spark3.1.1-hadoop2.7.2.jar.asc
==
--- 
release/carbondata/2.3.0/apache-carbondata-2.3.0-bin-spark3.1.1-hadoop2.7.2.jar.asc
 (added)
+++ 
release/carbondata/2.3.0/apache-carbondata-2.3.0-bin-spark3.1.1-hadoop2.7.2.jar.asc
 Sat Mar 12 05:17:27 2022
@@ -0,0 +1,14 @@
+-BEGIN PGP SIGNATURE-
+
+iQGzBAABCgAdFiEER4oGoEjiZ19CtKBtzP3MO+Fg6doFAmIsK+kACgkQzP3MO+Fg
+6dqPXQwAs2HkgrNgUBJ6p76uAlrUFUpzfEbE0Vc1NlSEe7LsPiWVC7MPpAWI3zhI
+H0rgf0wQPSVohxe2nckmLjCn4p1/vsVdHnO4GGxD/cgwo9Me1979r9lJ6bcegQuQ
+ALmoxHNP7wfEyWI74c2D

[carbondata] 02/03: [maven-release-plugin] prepare for next development iteration

2022-03-07 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git

commit e25d5b6fa292ba00cc0ed50b8f4610c2b968c0c5
Author: kunal642 
AuthorDate: Wed Jan 19 16:50:11 2022 +0530

[maven-release-plugin] prepare for next development iteration
---
 assembly/pom.xml| 2 +-
 common/pom.xml  | 2 +-
 core/pom.xml| 2 +-
 examples/flink/pom.xml  | 2 +-
 examples/spark/pom.xml  | 2 +-
 format/pom.xml  | 2 +-
 geo/pom.xml | 2 +-
 hadoop/pom.xml  | 2 +-
 index/bloom/pom.xml | 2 +-
 index/examples/pom.xml  | 2 +-
 index/lucene/pom.xml| 2 +-
 index/secondary-index/pom.xml   | 2 +-
 integration/flink-build/pom.xml | 2 +-
 integration/flink-proxy/pom.xml | 2 +-
 integration/flink/pom.xml   | 2 +-
 integration/hive/pom.xml| 2 +-
 integration/presto/pom.xml  | 2 +-
 integration/spark/pom.xml   | 2 +-
 mv/plan/pom.xml | 2 +-
 pom.xml | 4 ++--
 processing/pom.xml  | 2 +-
 sdk/sdk/pom.xml | 2 +-
 streaming/pom.xml   | 2 +-
 tools/cli/pom.xml   | 2 +-
 24 files changed, 25 insertions(+), 25 deletions(-)

diff --git a/assembly/pom.xml b/assembly/pom.xml
index 1107569..d8d03b7 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/common/pom.xml b/common/pom.xml
index 88dc1b8..6205781 100644
--- a/common/pom.xml
+++ b/common/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/core/pom.xml b/core/pom.xml
index 0e74e7a..a5f2e75 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/examples/flink/pom.xml b/examples/flink/pom.xml
index da34787..360fcff 100644
--- a/examples/flink/pom.xml
+++ b/examples/flink/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/examples/spark/pom.xml b/examples/spark/pom.xml
index ca666db..64f0d9f 100644
--- a/examples/spark/pom.xml
+++ b/examples/spark/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/format/pom.xml b/format/pom.xml
index a8c5b60..e992b2e 100644
--- a/format/pom.xml
+++ b/format/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/geo/pom.xml b/geo/pom.xml
index 83bbbd9..46f6695 100644
--- a/geo/pom.xml
+++ b/geo/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/hadoop/pom.xml b/hadoop/pom.xml
index 298337b..c38b0f9 100644
--- a/hadoop/pom.xml
+++ b/hadoop/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/index/bloom/pom.xml b/index/bloom/pom.xml
index 1096951..7ec45b4 100644
--- a/index/bloom/pom.xml
+++ b/index/bloom/pom.xml
@@ -4,7 +4,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/index/examples/pom.xml b/index/examples/pom.xml
index 3522097..bccbae1 100644
--- a/index/examples/pom.xml
+++ b/index/examples/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/index/lucene/pom.xml b/index/lucene/pom.xml
index 88d8afb..e23b94d 100644
--- a/index/lucene/pom.xml
+++ b/index/lucene/pom.xml
@@ -4,7 +4,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/index/secondary-index/pom.xml b/index/secondary-index/pom.xml
index f4ce62e..b1a2414 100644
--- a/index/secondary-index/pom.xml
+++ b/index/secondary-index/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/integration/flink-build/pom.xml b/integration/flink-build/pom.xml
index ba37c2e..a4b320e 100644
--- a/integration/flink-build/pom.xml
+++ b/integration/flink-build/pom.xml
@@ -4,7 +4,7 @@
 
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
 
 
diff --git a/integration/flink-proxy/pom.xml b/integration/flink-proxy/pom.xml
index 832d804..b7aee50 100644
--- a/integration/flink-proxy/pom.xml
+++ b/integration/flink-proxy

[carbondata] 03/03: [CARBONDATA-4306] Fix Query Performance issue for Spark 3.1

2022-03-07 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git

commit a838531f243f4b826778b51d1f238461d3414fe4
Author: Indhumathi27 
AuthorDate: Fri Mar 4 13:25:36 2022 +0530

[CARBONDATA-4306] Fix Query Performance issue for Spark 3.1

Why is this PR needed?
Some non-partition filters, which cannot be handled by carbon, is not 
pushed down to spark.

What changes were proposed in this PR?
If partition filters is non empty, then the filter column is not partition 
column, then push the filter to spark

This closes #4252
---
 .../apache/spark/sql/execution/strategy/CarbonSourceStrategy.scala  | 6 +++---
 1 file changed, 3 insertions(+), 3 deletions(-)

diff --git 
a/integration/spark/src/main/scala/org/apache/spark/sql/execution/strategy/CarbonSourceStrategy.scala
 
b/integration/spark/src/main/scala/org/apache/spark/sql/execution/strategy/CarbonSourceStrategy.scala
index 824c7fb..85044a2 100644
--- 
a/integration/spark/src/main/scala/org/apache/spark/sql/execution/strategy/CarbonSourceStrategy.scala
+++ 
b/integration/spark/src/main/scala/org/apache/spark/sql/execution/strategy/CarbonSourceStrategy.scala
@@ -158,10 +158,10 @@ private[sql] object CarbonSourceStrategy extends 
SparkStrategy {
 SparkSession.getActiveSession.get,
 relation.catalogTable.get.identifier
   )
+  // remove dynamic partition filter from predicates
+  filterPredicates = CarbonToSparkAdapter.getDataFilter(partitionSet,
+allPredicates, partitionsFilter)
 }
-// remove dynamic partition filter from predicates
-filterPredicates = CarbonToSparkAdapter.getDataFilter(partitionSet,
-  allPredicates, partitionsFilter)
 val table = relation.relation.asInstanceOf[CarbonDatasourceHadoopRelation]
 val projects = rawProjects.map {p =>
   p.transform {


[carbondata] 01/03: [maven-release-plugin] prepare release apache-carbondata-2.3.0-rc2

2022-03-07 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git

commit 9b749511153c653e15bfcfbd2757dd4d8f337530
Author: kunal642 
AuthorDate: Wed Jan 19 16:48:56 2022 +0530

[maven-release-plugin] prepare release apache-carbondata-2.3.0-rc2
---
 assembly/pom.xml| 2 +-
 common/pom.xml  | 2 +-
 core/pom.xml| 2 +-
 examples/flink/pom.xml  | 2 +-
 examples/spark/pom.xml  | 2 +-
 format/pom.xml  | 2 +-
 geo/pom.xml | 2 +-
 hadoop/pom.xml  | 2 +-
 index/bloom/pom.xml | 2 +-
 index/examples/pom.xml  | 2 +-
 index/lucene/pom.xml| 2 +-
 index/secondary-index/pom.xml   | 2 +-
 integration/flink-build/pom.xml | 2 +-
 integration/flink-proxy/pom.xml | 2 +-
 integration/flink/pom.xml   | 2 +-
 integration/hive/pom.xml| 2 +-
 integration/presto/pom.xml  | 2 +-
 integration/spark/pom.xml   | 2 +-
 mv/plan/pom.xml | 2 +-
 pom.xml | 4 ++--
 processing/pom.xml  | 2 +-
 sdk/sdk/pom.xml | 2 +-
 streaming/pom.xml   | 2 +-
 tools/cli/pom.xml   | 2 +-
 24 files changed, 25 insertions(+), 25 deletions(-)

diff --git a/assembly/pom.xml b/assembly/pom.xml
index d4c7aba..1107569 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/common/pom.xml b/common/pom.xml
index 68fea9c..88dc1b8 100644
--- a/common/pom.xml
+++ b/common/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/core/pom.xml b/core/pom.xml
index f4d2c16..0e74e7a 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/examples/flink/pom.xml b/examples/flink/pom.xml
index cd4b691..da34787 100644
--- a/examples/flink/pom.xml
+++ b/examples/flink/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/examples/spark/pom.xml b/examples/spark/pom.xml
index 0f804f4..ca666db 100644
--- a/examples/spark/pom.xml
+++ b/examples/spark/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/format/pom.xml b/format/pom.xml
index c547d37..a8c5b60 100644
--- a/format/pom.xml
+++ b/format/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/geo/pom.xml b/geo/pom.xml
index fcbd80c..83bbbd9 100644
--- a/geo/pom.xml
+++ b/geo/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/hadoop/pom.xml b/hadoop/pom.xml
index e8be41e..298337b 100644
--- a/hadoop/pom.xml
+++ b/hadoop/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/index/bloom/pom.xml b/index/bloom/pom.xml
index c4071ba..1096951 100644
--- a/index/bloom/pom.xml
+++ b/index/bloom/pom.xml
@@ -4,7 +4,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/index/examples/pom.xml b/index/examples/pom.xml
index 3873996..3522097 100644
--- a/index/examples/pom.xml
+++ b/index/examples/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/index/lucene/pom.xml b/index/lucene/pom.xml
index b10583d..88d8afb 100644
--- a/index/lucene/pom.xml
+++ b/index/lucene/pom.xml
@@ -4,7 +4,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/index/secondary-index/pom.xml b/index/secondary-index/pom.xml
index b998143..f4ce62e 100644
--- a/index/secondary-index/pom.xml
+++ b/index/secondary-index/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/integration/flink-build/pom.xml b/integration/flink-build/pom.xml
index c443497..ba37c2e 100644
--- a/integration/flink-build/pom.xml
+++ b/integration/flink-build/pom.xml
@@ -4,7 +4,7 @@
 
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
 
 
diff --git a/integration/flink-proxy/pom.xml b/integration/flink-proxy/pom.xml
index 23f5169..832d804 100644
--- a/integration/flink-proxy/pom.xml
+++ b/integration/flink-proxy

[carbondata] branch master updated (19343a7 -> a838531)

2022-03-07 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git.


from 19343a7  [CARBONDATA-4326] MV not hitting with multiple sessions issue 
fix
 new 9b74951  [maven-release-plugin] prepare release 
apache-carbondata-2.3.0-rc2
 new e25d5b6  [maven-release-plugin] prepare for next development iteration
 new a838531  [CARBONDATA-4306] Fix Query Performance issue for Spark 3.1

The 3 revisions listed above as "new" are entirely new to this
repository and will be described in separate emails.  The revisions
listed as "add" were already present in the repository and have only
been added to this reference.


Summary of changes:
 assembly/pom.xml| 2 +-
 common/pom.xml  | 2 +-
 core/pom.xml| 2 +-
 examples/flink/pom.xml  | 2 +-
 examples/spark/pom.xml  | 2 +-
 format/pom.xml  | 2 +-
 geo/pom.xml | 2 +-
 hadoop/pom.xml  | 2 +-
 index/bloom/pom.xml | 2 +-
 index/examples/pom.xml  | 2 +-
 index/lucene/pom.xml| 2 +-
 index/secondary-index/pom.xml   | 2 +-
 integration/flink-build/pom.xml | 2 +-
 integration/flink-proxy/pom.xml | 2 +-
 integration/flink/pom.xml   | 2 +-
 integration/hive/pom.xml| 2 +-
 integration/presto/pom.xml  | 2 +-
 integration/spark/pom.xml   | 2 +-
 .../apache/spark/sql/execution/strategy/CarbonSourceStrategy.scala  | 6 +++---
 mv/plan/pom.xml | 2 +-
 pom.xml | 2 +-
 processing/pom.xml  | 2 +-
 sdk/sdk/pom.xml | 2 +-
 streaming/pom.xml   | 2 +-
 tools/cli/pom.xml   | 2 +-
 25 files changed, 27 insertions(+), 27 deletions(-)


svn commit: r52235 - in /release/carbondata/2.3.0: ./ apache-carbondata-2.3.0-bin-spark2.3.4-hadoop2.7.2.jar apache-carbondata-2.3.0-bin-spark2.4.5-hadoop2.7.2.jar apache-carbondata-2.3.0-bin-spark3.1

2022-01-23 Thread kunalkapoor
Author: kunalkapoor
Date: Mon Jan 24 06:05:09 2022
New Revision: 52235

Log:
uploading 2.2.0 release

Added:
release/carbondata/2.3.0/

release/carbondata/2.3.0/apache-carbondata-2.3.0-bin-spark2.3.4-hadoop2.7.2.jar 
  (with props)

release/carbondata/2.3.0/apache-carbondata-2.3.0-bin-spark2.4.5-hadoop2.7.2.jar 
  (with props)

release/carbondata/2.3.0/apache-carbondata-2.3.0-bin-spark3.1.1-hadoop2.7.2.jar 
  (with props)

Added: 
release/carbondata/2.3.0/apache-carbondata-2.3.0-bin-spark2.3.4-hadoop2.7.2.jar
==
Binary file - no diff available.

Propchange: 
release/carbondata/2.3.0/apache-carbondata-2.3.0-bin-spark2.3.4-hadoop2.7.2.jar
--
svn:mime-type = application/octet-stream

Added: 
release/carbondata/2.3.0/apache-carbondata-2.3.0-bin-spark2.4.5-hadoop2.7.2.jar
==
Binary file - no diff available.

Propchange: 
release/carbondata/2.3.0/apache-carbondata-2.3.0-bin-spark2.4.5-hadoop2.7.2.jar
--
svn:mime-type = application/octet-stream

Added: 
release/carbondata/2.3.0/apache-carbondata-2.3.0-bin-spark3.1.1-hadoop2.7.2.jar
==
Binary file - no diff available.

Propchange: 
release/carbondata/2.3.0/apache-carbondata-2.3.0-bin-spark3.1.1-hadoop2.7.2.jar
--
svn:mime-type = application/octet-stream




svn commit: r52151 - in /dev/carbondata/2.3.0-rc2: ./ apache-carbondata-2.3.0-source-release.zip apache-carbondata-2.3.0-source-release.zip.asc apache-carbondata-2.3.0-source-release.zip.md5 apache-ca

2022-01-19 Thread kunalkapoor
Author: kunalkapoor
Date: Wed Jan 19 15:52:29 2022
New Revision: 52151

Log:
Upload 2.3.0-rc2

Added:
dev/carbondata/2.3.0-rc2/
dev/carbondata/2.3.0-rc2/apache-carbondata-2.3.0-source-release.zip   (with 
props)
dev/carbondata/2.3.0-rc2/apache-carbondata-2.3.0-source-release.zip.asc
dev/carbondata/2.3.0-rc2/apache-carbondata-2.3.0-source-release.zip.md5
dev/carbondata/2.3.0-rc2/apache-carbondata-2.3.0-source-release.zip.sha512

Added: dev/carbondata/2.3.0-rc2/apache-carbondata-2.3.0-source-release.zip
==
Binary file - no diff available.

Propchange: dev/carbondata/2.3.0-rc2/apache-carbondata-2.3.0-source-release.zip
--
svn:mime-type = application/octet-stream

Added: dev/carbondata/2.3.0-rc2/apache-carbondata-2.3.0-source-release.zip.asc
==
--- dev/carbondata/2.3.0-rc2/apache-carbondata-2.3.0-source-release.zip.asc 
(added)
+++ dev/carbondata/2.3.0-rc2/apache-carbondata-2.3.0-source-release.zip.asc Wed 
Jan 19 15:52:29 2022
@@ -0,0 +1,14 @@
+-BEGIN PGP SIGNATURE-
+
+iQGzBAABCgAdFiEER4oGoEjiZ19CtKBtzP3MO+Fg6doFAmHmUKQACgkQzP3MO+Fg
+6dp1QAv/R1fo7WKqxOqxN7Nw2Efk0VZ/ylNjwW0Cw0aQpcMw4SDMWkGYY6iZPXPm
+Y2aUfolFB7QRtFVPpl6EjT2O4zHBgYOy+4PXbyKa663YuBetiFp510wFPK6Wz/It
+p0EEbMM2Ec1xvWJaX/dZEhPX1GTeoxNDNaN+UAjkMCLCA0i9d4pLtZo0A22GV5bx
+4C0HPXUZQpUx6gavcM6vRNwAtJo6Hnw/rhayXSZmNmdADZEUhT/Zry+k1MAIvWV4
+IRHPpX0qZ65CrHbaK8QUWRzrfzIRZ4qv+ZEJ6S7KNPIT+J6fIEgdqtZ6S/DD5UO5
+mldGz4/69/TJyQPDDqTcl8EF62gVMHNemIzJjbpleYRHmfXtS1jd5bf9X9+SjWS4
+cDY15jHgKFPRrUie+MRsD1MuaaV4vXQaaFK6BIcv1NW2i1hJ2bhbVBhRVLXm+Hq2
+xX072nTY/v9vGQxRnw5Q93KTD+m00cgEPfdh56sm1MZKZ4teNwcBpKnILSrzwLld
+xT1c4XI8
+=Pi5J
+-END PGP SIGNATURE-

Added: dev/carbondata/2.3.0-rc2/apache-carbondata-2.3.0-source-release.zip.md5
==
--- dev/carbondata/2.3.0-rc2/apache-carbondata-2.3.0-source-release.zip.md5 
(added)
+++ dev/carbondata/2.3.0-rc2/apache-carbondata-2.3.0-source-release.zip.md5 Wed 
Jan 19 15:52:29 2022
@@ -0,0 +1 @@
+5240f4ff0a32a845a76038d7dd335ab3  apache-carbondata-2.3.0-source-release.zip

Added: 
dev/carbondata/2.3.0-rc2/apache-carbondata-2.3.0-source-release.zip.sha512
==
--- dev/carbondata/2.3.0-rc2/apache-carbondata-2.3.0-source-release.zip.sha512 
(added)
+++ dev/carbondata/2.3.0-rc2/apache-carbondata-2.3.0-source-release.zip.sha512 
Wed Jan 19 15:52:29 2022
@@ -0,0 +1 @@
+720addf7c3926812533102349c86d8030581aa5bdc7a93ee7d8e0ea4cbc42b7128762af816c60a44c43a983c988370496578fccc4889f830096db79e93dba6b1
  apache-carbondata-2.3.0-source-release.zip




[carbondata] branch branch-2.3 created (now 6db604a)

2022-01-19 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a change to branch branch-2.3
in repository https://gitbox.apache.org/repos/asf/carbondata.git.


  at 6db604a  [maven-release-plugin] prepare for next development iteration

This branch includes the following new commits:

 new 6db604a  [maven-release-plugin] prepare for next development iteration

The 1 revisions listed above as "new" are entirely new to this
repository and will be described in separate emails.  The revisions
listed as "add" were already present in the repository and have only
been added to this reference.



[carbondata] 01/01: [maven-release-plugin] prepare for next development iteration

2022-01-19 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch branch-2.3
in repository https://gitbox.apache.org/repos/asf/carbondata.git

commit 6db604a6389673194b30e3c45e7252af6400d54b
Author: kunal642 
AuthorDate: Wed Jan 19 16:50:11 2022 +0530

[maven-release-plugin] prepare for next development iteration
---
 assembly/pom.xml| 2 +-
 common/pom.xml  | 2 +-
 core/pom.xml| 2 +-
 examples/flink/pom.xml  | 2 +-
 examples/spark/pom.xml  | 2 +-
 format/pom.xml  | 2 +-
 geo/pom.xml | 2 +-
 hadoop/pom.xml  | 2 +-
 index/bloom/pom.xml | 2 +-
 index/examples/pom.xml  | 2 +-
 index/lucene/pom.xml| 2 +-
 index/secondary-index/pom.xml   | 2 +-
 integration/flink-build/pom.xml | 2 +-
 integration/flink-proxy/pom.xml | 2 +-
 integration/flink/pom.xml   | 2 +-
 integration/hive/pom.xml| 2 +-
 integration/presto/pom.xml  | 2 +-
 integration/spark/pom.xml   | 2 +-
 mv/plan/pom.xml | 2 +-
 pom.xml | 4 ++--
 processing/pom.xml  | 2 +-
 sdk/sdk/pom.xml | 2 +-
 streaming/pom.xml   | 2 +-
 tools/cli/pom.xml   | 2 +-
 24 files changed, 25 insertions(+), 25 deletions(-)

diff --git a/assembly/pom.xml b/assembly/pom.xml
index 1107569..d8d03b7 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/common/pom.xml b/common/pom.xml
index 88dc1b8..6205781 100644
--- a/common/pom.xml
+++ b/common/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/core/pom.xml b/core/pom.xml
index 0e74e7a..a5f2e75 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/examples/flink/pom.xml b/examples/flink/pom.xml
index da34787..360fcff 100644
--- a/examples/flink/pom.xml
+++ b/examples/flink/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/examples/spark/pom.xml b/examples/spark/pom.xml
index ca666db..64f0d9f 100644
--- a/examples/spark/pom.xml
+++ b/examples/spark/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/format/pom.xml b/format/pom.xml
index a8c5b60..e992b2e 100644
--- a/format/pom.xml
+++ b/format/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/geo/pom.xml b/geo/pom.xml
index 83bbbd9..46f6695 100644
--- a/geo/pom.xml
+++ b/geo/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/hadoop/pom.xml b/hadoop/pom.xml
index 298337b..c38b0f9 100644
--- a/hadoop/pom.xml
+++ b/hadoop/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/index/bloom/pom.xml b/index/bloom/pom.xml
index 1096951..7ec45b4 100644
--- a/index/bloom/pom.xml
+++ b/index/bloom/pom.xml
@@ -4,7 +4,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/index/examples/pom.xml b/index/examples/pom.xml
index 3522097..bccbae1 100644
--- a/index/examples/pom.xml
+++ b/index/examples/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/index/lucene/pom.xml b/index/lucene/pom.xml
index 88d8afb..e23b94d 100644
--- a/index/lucene/pom.xml
+++ b/index/lucene/pom.xml
@@ -4,7 +4,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/index/secondary-index/pom.xml b/index/secondary-index/pom.xml
index f4ce62e..b1a2414 100644
--- a/index/secondary-index/pom.xml
+++ b/index/secondary-index/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/integration/flink-build/pom.xml b/integration/flink-build/pom.xml
index ba37c2e..a4b320e 100644
--- a/integration/flink-build/pom.xml
+++ b/integration/flink-build/pom.xml
@@ -4,7 +4,7 @@
 
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
 
 
diff --git a/integration/flink-proxy/pom.xml b/integration/flink-proxy/pom.xml
index 832d804..b7aee50 100644
--- a/integration/flink-proxy/pom.xml
+++ b/integration/flink-proxy

[carbondata] 01/01: [maven-release-plugin] prepare release apache-carbondata-2.3.0-rc2

2022-01-19 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to annotated tag apache-carbondata-2.3.0-rc2
in repository https://gitbox.apache.org/repos/asf/carbondata.git

commit fee16db32bd29007d0de06e60fa74a1ee929cd90
Author: kunal642 
AuthorDate: Wed Jan 19 16:48:56 2022 +0530

[maven-release-plugin] prepare release apache-carbondata-2.3.0-rc2
---
 assembly/pom.xml| 2 +-
 common/pom.xml  | 2 +-
 core/pom.xml| 2 +-
 examples/flink/pom.xml  | 2 +-
 examples/spark/pom.xml  | 2 +-
 format/pom.xml  | 2 +-
 geo/pom.xml | 2 +-
 hadoop/pom.xml  | 2 +-
 index/bloom/pom.xml | 2 +-
 index/examples/pom.xml  | 2 +-
 index/lucene/pom.xml| 2 +-
 index/secondary-index/pom.xml   | 2 +-
 integration/flink-build/pom.xml | 2 +-
 integration/flink-proxy/pom.xml | 2 +-
 integration/flink/pom.xml   | 2 +-
 integration/hive/pom.xml| 2 +-
 integration/presto/pom.xml  | 2 +-
 integration/spark/pom.xml   | 2 +-
 mv/plan/pom.xml | 2 +-
 pom.xml | 4 ++--
 processing/pom.xml  | 2 +-
 sdk/sdk/pom.xml | 2 +-
 streaming/pom.xml   | 2 +-
 tools/cli/pom.xml   | 2 +-
 24 files changed, 25 insertions(+), 25 deletions(-)

diff --git a/assembly/pom.xml b/assembly/pom.xml
index d4c7aba..1107569 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/common/pom.xml b/common/pom.xml
index 68fea9c..88dc1b8 100644
--- a/common/pom.xml
+++ b/common/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/core/pom.xml b/core/pom.xml
index f4d2c16..0e74e7a 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/examples/flink/pom.xml b/examples/flink/pom.xml
index cd4b691..da34787 100644
--- a/examples/flink/pom.xml
+++ b/examples/flink/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/examples/spark/pom.xml b/examples/spark/pom.xml
index 0f804f4..ca666db 100644
--- a/examples/spark/pom.xml
+++ b/examples/spark/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/format/pom.xml b/format/pom.xml
index c547d37..a8c5b60 100644
--- a/format/pom.xml
+++ b/format/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/geo/pom.xml b/geo/pom.xml
index fcbd80c..83bbbd9 100644
--- a/geo/pom.xml
+++ b/geo/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/hadoop/pom.xml b/hadoop/pom.xml
index e8be41e..298337b 100644
--- a/hadoop/pom.xml
+++ b/hadoop/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/index/bloom/pom.xml b/index/bloom/pom.xml
index c4071ba..1096951 100644
--- a/index/bloom/pom.xml
+++ b/index/bloom/pom.xml
@@ -4,7 +4,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/index/examples/pom.xml b/index/examples/pom.xml
index 3873996..3522097 100644
--- a/index/examples/pom.xml
+++ b/index/examples/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/index/lucene/pom.xml b/index/lucene/pom.xml
index b10583d..88d8afb 100644
--- a/index/lucene/pom.xml
+++ b/index/lucene/pom.xml
@@ -4,7 +4,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/index/secondary-index/pom.xml b/index/secondary-index/pom.xml
index b998143..f4ce62e 100644
--- a/index/secondary-index/pom.xml
+++ b/index/secondary-index/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/integration/flink-build/pom.xml b/integration/flink-build/pom.xml
index c443497..ba37c2e 100644
--- a/integration/flink-build/pom.xml
+++ b/integration/flink-build/pom.xml
@@ -4,7 +4,7 @@
 
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
 
 
diff --git a/integration/flink-proxy/pom.xml b/integration/flink-proxy/pom.xml
index 23f5169..832d804 100644
--- a/integration/flink-proxy/pom.xml

[carbondata] annotated tag apache-carbondata-2.3.0-rc2 created (now eca13c5)

2022-01-19 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a change to annotated tag apache-carbondata-2.3.0-rc2
in repository https://gitbox.apache.org/repos/asf/carbondata.git.


  at eca13c5  (tag)
 tagging fee16db32bd29007d0de06e60fa74a1ee929cd90 (commit)
  by kunal642
  on Wed Jan 19 16:49:38 2022 +0530

- Log -
[maven-release-plugin] copy for tag apache-carbondata-2.3.0-rc2
---

This annotated tag includes the following new commits:

 new fee16db  [maven-release-plugin] prepare release 
apache-carbondata-2.3.0-rc2

The 1 revisions listed above as "new" are entirely new to this
repository and will be described in separate emails.  The revisions
listed as "add" were already present in the repository and have only
been added to this reference.



[carbondata] branch branch-2.3 deleted (was db0896a)

2022-01-19 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a change to branch branch-2.3
in repository https://gitbox.apache.org/repos/asf/carbondata.git.


 was db0896a  [maven-release-plugin] prepare for next development iteration

This change permanently discards the following revisions:

 discard db0896a  [maven-release-plugin] prepare for next development iteration
 discard 253e9cc  [maven-release-plugin] prepare release 
apache-carbondata-2.3.0-rc2


[carbondata] annotated tag apache-carbondata-2.3.0-rc2 deleted (was 65454a1)

2022-01-19 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a change to annotated tag apache-carbondata-2.3.0-rc2
in repository https://gitbox.apache.org/repos/asf/carbondata.git.


*** WARNING: tag apache-carbondata-2.3.0-rc2 was deleted! ***

   tag was  65454a1

The revisions that were on this annotated tag are still contained in
other references; therefore, this change does not discard any commits
from the repository.


[carbondata] 01/01: [maven-release-plugin] prepare for next development iteration

2022-01-19 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch branch-2.3
in repository https://gitbox.apache.org/repos/asf/carbondata.git

commit db0896a8bdd195ba6b8623d3e939e75d3b6433f0
Author: kunal642 
AuthorDate: Wed Jan 19 16:14:55 2022 +0530

[maven-release-plugin] prepare for next development iteration
---
 assembly/pom.xml| 2 +-
 common/pom.xml  | 2 +-
 core/pom.xml| 2 +-
 examples/flink/pom.xml  | 2 +-
 examples/spark/pom.xml  | 2 +-
 format/pom.xml  | 2 +-
 geo/pom.xml | 2 +-
 hadoop/pom.xml  | 2 +-
 index/bloom/pom.xml | 2 +-
 index/examples/pom.xml  | 2 +-
 index/lucene/pom.xml| 2 +-
 index/secondary-index/pom.xml   | 2 +-
 integration/flink-build/pom.xml | 2 +-
 integration/flink-proxy/pom.xml | 2 +-
 integration/flink/pom.xml   | 2 +-
 integration/hive/pom.xml| 2 +-
 integration/presto/pom.xml  | 2 +-
 integration/spark/pom.xml   | 2 +-
 mv/plan/pom.xml | 2 +-
 pom.xml | 4 ++--
 processing/pom.xml  | 2 +-
 sdk/sdk/pom.xml | 2 +-
 streaming/pom.xml   | 2 +-
 tools/cli/pom.xml   | 2 +-
 24 files changed, 25 insertions(+), 25 deletions(-)

diff --git a/assembly/pom.xml b/assembly/pom.xml
index 1107569..d8d03b7 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/common/pom.xml b/common/pom.xml
index 88dc1b8..6205781 100644
--- a/common/pom.xml
+++ b/common/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/core/pom.xml b/core/pom.xml
index 0e74e7a..a5f2e75 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/examples/flink/pom.xml b/examples/flink/pom.xml
index da34787..360fcff 100644
--- a/examples/flink/pom.xml
+++ b/examples/flink/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/examples/spark/pom.xml b/examples/spark/pom.xml
index ca666db..64f0d9f 100644
--- a/examples/spark/pom.xml
+++ b/examples/spark/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/format/pom.xml b/format/pom.xml
index a8c5b60..e992b2e 100644
--- a/format/pom.xml
+++ b/format/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/geo/pom.xml b/geo/pom.xml
index 83bbbd9..46f6695 100644
--- a/geo/pom.xml
+++ b/geo/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/hadoop/pom.xml b/hadoop/pom.xml
index 298337b..c38b0f9 100644
--- a/hadoop/pom.xml
+++ b/hadoop/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/index/bloom/pom.xml b/index/bloom/pom.xml
index 1096951..7ec45b4 100644
--- a/index/bloom/pom.xml
+++ b/index/bloom/pom.xml
@@ -4,7 +4,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/index/examples/pom.xml b/index/examples/pom.xml
index 3522097..bccbae1 100644
--- a/index/examples/pom.xml
+++ b/index/examples/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/index/lucene/pom.xml b/index/lucene/pom.xml
index 88d8afb..e23b94d 100644
--- a/index/lucene/pom.xml
+++ b/index/lucene/pom.xml
@@ -4,7 +4,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/index/secondary-index/pom.xml b/index/secondary-index/pom.xml
index f4ce62e..b1a2414 100644
--- a/index/secondary-index/pom.xml
+++ b/index/secondary-index/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/integration/flink-build/pom.xml b/integration/flink-build/pom.xml
index ba37c2e..a4b320e 100644
--- a/integration/flink-build/pom.xml
+++ b/integration/flink-build/pom.xml
@@ -4,7 +4,7 @@
 
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
 
 
diff --git a/integration/flink-proxy/pom.xml b/integration/flink-proxy/pom.xml
index 832d804..b7aee50 100644
--- a/integration/flink-proxy/pom.xml
+++ b/integration/flink-proxy

[carbondata] branch branch-2.3 created (now db0896a)

2022-01-19 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a change to branch branch-2.3
in repository https://gitbox.apache.org/repos/asf/carbondata.git.


  at db0896a  [maven-release-plugin] prepare for next development iteration

This branch includes the following new commits:

 new db0896a  [maven-release-plugin] prepare for next development iteration

The 1 revisions listed above as "new" are entirely new to this
repository and will be described in separate emails.  The revisions
listed as "add" were already present in the repository and have only
been added to this reference.



[carbondata] annotated tag apache-carbondata-2.3.0-rc2 created (now 65454a1)

2022-01-19 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a change to annotated tag apache-carbondata-2.3.0-rc2
in repository https://gitbox.apache.org/repos/asf/carbondata.git.


  at 65454a1  (tag)
 tagging 253e9cc7ce6a5dbc7787880738fec6582c34b685 (commit)
  by kunal642
  on Wed Jan 19 16:14:40 2022 +0530

- Log -
[maven-release-plugin] copy for tag apache-carbondata-2.3.0-rc2
---

This annotated tag includes the following new commits:

 new 253e9cc  [maven-release-plugin] prepare release 
apache-carbondata-2.3.0-rc2

The 1 revisions listed above as "new" are entirely new to this
repository and will be described in separate emails.  The revisions
listed as "add" were already present in the repository and have only
been added to this reference.



[carbondata] 01/01: [maven-release-plugin] prepare release apache-carbondata-2.3.0-rc2

2022-01-19 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to annotated tag apache-carbondata-2.3.0-rc2
in repository https://gitbox.apache.org/repos/asf/carbondata.git

commit 253e9cc7ce6a5dbc7787880738fec6582c34b685
Author: kunal642 
AuthorDate: Wed Jan 19 11:36:34 2022 +0530

[maven-release-plugin] prepare release apache-carbondata-2.3.0-rc2
---
 assembly/pom.xml| 2 +-
 common/pom.xml  | 2 +-
 core/pom.xml| 2 +-
 examples/flink/pom.xml  | 2 +-
 examples/spark/pom.xml  | 2 +-
 format/pom.xml  | 2 +-
 geo/pom.xml | 2 +-
 hadoop/pom.xml  | 2 +-
 index/bloom/pom.xml | 2 +-
 index/examples/pom.xml  | 2 +-
 index/lucene/pom.xml| 2 +-
 index/secondary-index/pom.xml   | 2 +-
 integration/flink-build/pom.xml | 2 +-
 integration/flink-proxy/pom.xml | 2 +-
 integration/flink/pom.xml   | 2 +-
 integration/hive/pom.xml| 2 +-
 integration/presto/pom.xml  | 2 +-
 integration/spark/pom.xml   | 2 +-
 mv/plan/pom.xml | 2 +-
 pom.xml | 4 ++--
 processing/pom.xml  | 2 +-
 sdk/sdk/pom.xml | 2 +-
 streaming/pom.xml   | 2 +-
 tools/cli/pom.xml   | 2 +-
 24 files changed, 25 insertions(+), 25 deletions(-)

diff --git a/assembly/pom.xml b/assembly/pom.xml
index d4c7aba..1107569 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/common/pom.xml b/common/pom.xml
index 68fea9c..88dc1b8 100644
--- a/common/pom.xml
+++ b/common/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/core/pom.xml b/core/pom.xml
index f4d2c16..0e74e7a 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/examples/flink/pom.xml b/examples/flink/pom.xml
index cd4b691..da34787 100644
--- a/examples/flink/pom.xml
+++ b/examples/flink/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/examples/spark/pom.xml b/examples/spark/pom.xml
index 0f804f4..ca666db 100644
--- a/examples/spark/pom.xml
+++ b/examples/spark/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/format/pom.xml b/format/pom.xml
index c547d37..a8c5b60 100644
--- a/format/pom.xml
+++ b/format/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/geo/pom.xml b/geo/pom.xml
index fcbd80c..83bbbd9 100644
--- a/geo/pom.xml
+++ b/geo/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/hadoop/pom.xml b/hadoop/pom.xml
index e8be41e..298337b 100644
--- a/hadoop/pom.xml
+++ b/hadoop/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/index/bloom/pom.xml b/index/bloom/pom.xml
index c4071ba..1096951 100644
--- a/index/bloom/pom.xml
+++ b/index/bloom/pom.xml
@@ -4,7 +4,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/index/examples/pom.xml b/index/examples/pom.xml
index 3873996..3522097 100644
--- a/index/examples/pom.xml
+++ b/index/examples/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/index/lucene/pom.xml b/index/lucene/pom.xml
index b10583d..88d8afb 100644
--- a/index/lucene/pom.xml
+++ b/index/lucene/pom.xml
@@ -4,7 +4,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/index/secondary-index/pom.xml b/index/secondary-index/pom.xml
index b998143..f4ce62e 100644
--- a/index/secondary-index/pom.xml
+++ b/index/secondary-index/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/integration/flink-build/pom.xml b/integration/flink-build/pom.xml
index c443497..ba37c2e 100644
--- a/integration/flink-build/pom.xml
+++ b/integration/flink-build/pom.xml
@@ -4,7 +4,7 @@
 
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
 
 
diff --git a/integration/flink-proxy/pom.xml b/integration/flink-proxy/pom.xml
index 23f5169..832d804 100644
--- a/integration/flink-proxy/pom.xml

[carbondata] annotated tag apache-carbondata-2.3.0-rc2 deleted (was d53c1e7)

2022-01-18 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a change to annotated tag apache-carbondata-2.3.0-rc2
in repository https://gitbox.apache.org/repos/asf/carbondata.git.


*** WARNING: tag apache-carbondata-2.3.0-rc2 was deleted! ***

   tag was  d53c1e7

This change permanently discards the following revisions:

 discard 56a71ca  [maven-release-plugin] prepare release 
apache-carbondata-2.3.0-rc2


[carbondata] branch branch-2.3 deleted (was dbc1157)

2022-01-18 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a change to branch branch-2.3
in repository https://gitbox.apache.org/repos/asf/carbondata.git.


 was dbc1157  [maven-release-plugin] prepare for next development iteration

This change permanently discards the following revisions:

 discard dbc1157  [maven-release-plugin] prepare for next development iteration


[carbondata] branch branch-2.3 created (now dbc1157)

2022-01-18 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a change to branch branch-2.3
in repository https://gitbox.apache.org/repos/asf/carbondata.git.


  at dbc1157  [maven-release-plugin] prepare for next development iteration

This branch includes the following new commits:

 new dbc1157  [maven-release-plugin] prepare for next development iteration

The 1 revisions listed above as "new" are entirely new to this
repository and will be described in separate emails.  The revisions
listed as "add" were already present in the repository and have only
been added to this reference.



[carbondata] 01/01: [maven-release-plugin] prepare for next development iteration

2022-01-18 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch branch-2.3
in repository https://gitbox.apache.org/repos/asf/carbondata.git

commit dbc11579397801b6ac31c1d7a41832e70f2cd905
Author: kunal642 
AuthorDate: Wed Jan 19 10:54:32 2022 +0530

[maven-release-plugin] prepare for next development iteration
---
 assembly/pom.xml| 2 +-
 common/pom.xml  | 2 +-
 core/pom.xml| 2 +-
 examples/flink/pom.xml  | 2 +-
 examples/spark/pom.xml  | 2 +-
 format/pom.xml  | 2 +-
 geo/pom.xml | 2 +-
 hadoop/pom.xml  | 2 +-
 index/bloom/pom.xml | 2 +-
 index/examples/pom.xml  | 2 +-
 index/lucene/pom.xml| 2 +-
 index/secondary-index/pom.xml   | 2 +-
 integration/flink-build/pom.xml | 2 +-
 integration/flink-proxy/pom.xml | 2 +-
 integration/flink/pom.xml   | 2 +-
 integration/hive/pom.xml| 2 +-
 integration/presto/pom.xml  | 2 +-
 integration/spark/pom.xml   | 2 +-
 mv/plan/pom.xml | 2 +-
 pom.xml | 4 ++--
 processing/pom.xml  | 2 +-
 sdk/sdk/pom.xml | 2 +-
 streaming/pom.xml   | 2 +-
 tools/cli/pom.xml   | 2 +-
 24 files changed, 25 insertions(+), 25 deletions(-)

diff --git a/assembly/pom.xml b/assembly/pom.xml
index 1107569..d8d03b7 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/common/pom.xml b/common/pom.xml
index 88dc1b8..6205781 100644
--- a/common/pom.xml
+++ b/common/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/core/pom.xml b/core/pom.xml
index 0e74e7a..a5f2e75 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/examples/flink/pom.xml b/examples/flink/pom.xml
index da34787..360fcff 100644
--- a/examples/flink/pom.xml
+++ b/examples/flink/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/examples/spark/pom.xml b/examples/spark/pom.xml
index ca666db..64f0d9f 100644
--- a/examples/spark/pom.xml
+++ b/examples/spark/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/format/pom.xml b/format/pom.xml
index a8c5b60..e992b2e 100644
--- a/format/pom.xml
+++ b/format/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/geo/pom.xml b/geo/pom.xml
index 83bbbd9..46f6695 100644
--- a/geo/pom.xml
+++ b/geo/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/hadoop/pom.xml b/hadoop/pom.xml
index 298337b..c38b0f9 100644
--- a/hadoop/pom.xml
+++ b/hadoop/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/index/bloom/pom.xml b/index/bloom/pom.xml
index 1096951..7ec45b4 100644
--- a/index/bloom/pom.xml
+++ b/index/bloom/pom.xml
@@ -4,7 +4,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/index/examples/pom.xml b/index/examples/pom.xml
index 3522097..bccbae1 100644
--- a/index/examples/pom.xml
+++ b/index/examples/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/index/lucene/pom.xml b/index/lucene/pom.xml
index 88d8afb..e23b94d 100644
--- a/index/lucene/pom.xml
+++ b/index/lucene/pom.xml
@@ -4,7 +4,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/index/secondary-index/pom.xml b/index/secondary-index/pom.xml
index f4ce62e..b1a2414 100644
--- a/index/secondary-index/pom.xml
+++ b/index/secondary-index/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/integration/flink-build/pom.xml b/integration/flink-build/pom.xml
index ba37c2e..a4b320e 100644
--- a/integration/flink-build/pom.xml
+++ b/integration/flink-build/pom.xml
@@ -4,7 +4,7 @@
 
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
 
 
diff --git a/integration/flink-proxy/pom.xml b/integration/flink-proxy/pom.xml
index 832d804..b7aee50 100644
--- a/integration/flink-proxy/pom.xml
+++ b/integration/flink-proxy

[carbondata] annotated tag apache-carbondata-2.3.0-rc2 created (now d53c1e7)

2022-01-18 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a change to annotated tag apache-carbondata-2.3.0-rc2
in repository https://gitbox.apache.org/repos/asf/carbondata.git.


  at d53c1e7  (tag)
 tagging 56a71caea37b51c6d045c5a0b386de842be799a3 (commit)
  by kunal642
  on Wed Jan 19 10:54:16 2022 +0530

- Log -
[maven-release-plugin] copy for tag apache-carbondata-2.3.0-rc2
---

This annotated tag includes the following new commits:

 new 56a71ca  [maven-release-plugin] prepare release 
apache-carbondata-2.3.0-rc2

The 1 revisions listed above as "new" are entirely new to this
repository and will be described in separate emails.  The revisions
listed as "add" were already present in the repository and have only
been added to this reference.



[carbondata] 01/01: [maven-release-plugin] prepare release apache-carbondata-2.3.0-rc2

2022-01-18 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to annotated tag apache-carbondata-2.3.0-rc2
in repository https://gitbox.apache.org/repos/asf/carbondata.git

commit 56a71caea37b51c6d045c5a0b386de842be799a3
Author: kunal642 
AuthorDate: Wed Jan 19 10:48:52 2022 +0530

[maven-release-plugin] prepare release apache-carbondata-2.3.0-rc2
---
 assembly/pom.xml| 2 +-
 common/pom.xml  | 2 +-
 core/pom.xml| 2 +-
 examples/flink/pom.xml  | 2 +-
 examples/spark/pom.xml  | 2 +-
 format/pom.xml  | 2 +-
 geo/pom.xml | 2 +-
 hadoop/pom.xml  | 2 +-
 index/bloom/pom.xml | 2 +-
 index/examples/pom.xml  | 2 +-
 index/lucene/pom.xml| 2 +-
 index/secondary-index/pom.xml   | 2 +-
 integration/flink-build/pom.xml | 2 +-
 integration/flink-proxy/pom.xml | 2 +-
 integration/flink/pom.xml   | 2 +-
 integration/hive/pom.xml| 2 +-
 integration/presto/pom.xml  | 2 +-
 integration/spark/pom.xml   | 2 +-
 mv/plan/pom.xml | 2 +-
 pom.xml | 4 ++--
 processing/pom.xml  | 2 +-
 sdk/sdk/pom.xml | 2 +-
 streaming/pom.xml   | 2 +-
 tools/cli/pom.xml   | 2 +-
 24 files changed, 25 insertions(+), 25 deletions(-)

diff --git a/assembly/pom.xml b/assembly/pom.xml
index d4c7aba..1107569 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/common/pom.xml b/common/pom.xml
index 68fea9c..88dc1b8 100644
--- a/common/pom.xml
+++ b/common/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/core/pom.xml b/core/pom.xml
index f4d2c16..0e74e7a 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/examples/flink/pom.xml b/examples/flink/pom.xml
index cd4b691..da34787 100644
--- a/examples/flink/pom.xml
+++ b/examples/flink/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/examples/spark/pom.xml b/examples/spark/pom.xml
index 0f804f4..ca666db 100644
--- a/examples/spark/pom.xml
+++ b/examples/spark/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/format/pom.xml b/format/pom.xml
index c547d37..a8c5b60 100644
--- a/format/pom.xml
+++ b/format/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/geo/pom.xml b/geo/pom.xml
index fcbd80c..83bbbd9 100644
--- a/geo/pom.xml
+++ b/geo/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/hadoop/pom.xml b/hadoop/pom.xml
index e8be41e..298337b 100644
--- a/hadoop/pom.xml
+++ b/hadoop/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/index/bloom/pom.xml b/index/bloom/pom.xml
index c4071ba..1096951 100644
--- a/index/bloom/pom.xml
+++ b/index/bloom/pom.xml
@@ -4,7 +4,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/index/examples/pom.xml b/index/examples/pom.xml
index 3873996..3522097 100644
--- a/index/examples/pom.xml
+++ b/index/examples/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/index/lucene/pom.xml b/index/lucene/pom.xml
index b10583d..88d8afb 100644
--- a/index/lucene/pom.xml
+++ b/index/lucene/pom.xml
@@ -4,7 +4,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/index/secondary-index/pom.xml b/index/secondary-index/pom.xml
index b998143..f4ce62e 100644
--- a/index/secondary-index/pom.xml
+++ b/index/secondary-index/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/integration/flink-build/pom.xml b/integration/flink-build/pom.xml
index c443497..ba37c2e 100644
--- a/integration/flink-build/pom.xml
+++ b/integration/flink-build/pom.xml
@@ -4,7 +4,7 @@
 
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
 
 
diff --git a/integration/flink-proxy/pom.xml b/integration/flink-proxy/pom.xml
index 23f5169..832d804 100644
--- a/integration/flink-proxy/pom.xml

[carbondata] branch branch-2.3 deleted (was 494ad6c)

2022-01-18 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a change to branch branch-2.3
in repository https://gitbox.apache.org/repos/asf/carbondata.git.


 was 494ad6c  [maven-release-plugin] prepare for next development iteration

This change permanently discards the following revisions:

 discard 494ad6c  [maven-release-plugin] prepare for next development iteration


[carbondata] branch branch-2.3 updated: [maven-release-plugin] prepare for next development iteration

2022-01-18 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch branch-2.3
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/branch-2.3 by this push:
 new 494ad6c  [maven-release-plugin] prepare for next development iteration
494ad6c is described below

commit 494ad6ce257354ba143c39c6c336bb02638d206a
Author: kunal642 
AuthorDate: Tue Jan 18 11:20:31 2022 +0530

[maven-release-plugin] prepare for next development iteration
---
 assembly/pom.xml| 2 +-
 common/pom.xml  | 2 +-
 core/pom.xml| 2 +-
 examples/flink/pom.xml  | 2 +-
 examples/spark/pom.xml  | 2 +-
 format/pom.xml  | 2 +-
 geo/pom.xml | 2 +-
 hadoop/pom.xml  | 2 +-
 index/bloom/pom.xml | 2 +-
 index/examples/pom.xml  | 2 +-
 index/lucene/pom.xml| 2 +-
 index/secondary-index/pom.xml   | 2 +-
 integration/flink-build/pom.xml | 2 +-
 integration/flink-proxy/pom.xml | 2 +-
 integration/flink/pom.xml   | 2 +-
 integration/hive/pom.xml| 2 +-
 integration/presto/pom.xml  | 2 +-
 integration/spark/pom.xml   | 2 +-
 mv/plan/pom.xml | 2 +-
 pom.xml | 4 ++--
 processing/pom.xml  | 2 +-
 sdk/sdk/pom.xml | 2 +-
 streaming/pom.xml   | 2 +-
 tools/cli/pom.xml   | 2 +-
 24 files changed, 25 insertions(+), 25 deletions(-)

diff --git a/assembly/pom.xml b/assembly/pom.xml
index 1107569..d8d03b7 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/common/pom.xml b/common/pom.xml
index 88dc1b8..6205781 100644
--- a/common/pom.xml
+++ b/common/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/core/pom.xml b/core/pom.xml
index 0e74e7a..a5f2e75 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/examples/flink/pom.xml b/examples/flink/pom.xml
index da34787..360fcff 100644
--- a/examples/flink/pom.xml
+++ b/examples/flink/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/examples/spark/pom.xml b/examples/spark/pom.xml
index ca666db..64f0d9f 100644
--- a/examples/spark/pom.xml
+++ b/examples/spark/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/format/pom.xml b/format/pom.xml
index a8c5b60..e992b2e 100644
--- a/format/pom.xml
+++ b/format/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/geo/pom.xml b/geo/pom.xml
index 83bbbd9..46f6695 100644
--- a/geo/pom.xml
+++ b/geo/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/hadoop/pom.xml b/hadoop/pom.xml
index 298337b..c38b0f9 100644
--- a/hadoop/pom.xml
+++ b/hadoop/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/index/bloom/pom.xml b/index/bloom/pom.xml
index 1096951..7ec45b4 100644
--- a/index/bloom/pom.xml
+++ b/index/bloom/pom.xml
@@ -4,7 +4,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/index/examples/pom.xml b/index/examples/pom.xml
index 3522097..bccbae1 100644
--- a/index/examples/pom.xml
+++ b/index/examples/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/index/lucene/pom.xml b/index/lucene/pom.xml
index 88d8afb..e23b94d 100644
--- a/index/lucene/pom.xml
+++ b/index/lucene/pom.xml
@@ -4,7 +4,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/index/secondary-index/pom.xml b/index/secondary-index/pom.xml
index f4ce62e..b1a2414 100644
--- a/index/secondary-index/pom.xml
+++ b/index/secondary-index/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/integration/flink-build/pom.xml b/integration/flink-build/pom.xml
index ba37c2e..a4b320e 100644
--- a/integration/flink-build/pom.xml
+++ b/integration/flink-build/pom.xml
@@ -4,7 +4,7 @@
 
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml

[carbondata] 01/01: [maven-release-plugin] prepare release apache-carbondata-1.1.0-rc3

2022-01-17 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch branch-2.3
in repository https://gitbox.apache.org/repos/asf/carbondata.git

commit 72178a479887e58d035e1d0869001f7704af3729
Author: kunal642 
AuthorDate: Tue Jan 18 11:13:31 2022 +0530

[maven-release-plugin] prepare release apache-carbondata-1.1.0-rc3
---
 assembly/pom.xml| 2 +-
 common/pom.xml  | 2 +-
 core/pom.xml| 2 +-
 examples/flink/pom.xml  | 2 +-
 examples/spark/pom.xml  | 2 +-
 format/pom.xml  | 2 +-
 geo/pom.xml | 2 +-
 hadoop/pom.xml  | 2 +-
 index/bloom/pom.xml | 2 +-
 index/examples/pom.xml  | 2 +-
 index/lucene/pom.xml| 2 +-
 index/secondary-index/pom.xml   | 2 +-
 integration/flink-build/pom.xml | 2 +-
 integration/flink-proxy/pom.xml | 2 +-
 integration/flink/pom.xml   | 2 +-
 integration/hive/pom.xml| 2 +-
 integration/presto/pom.xml  | 2 +-
 integration/spark/pom.xml   | 2 +-
 mv/plan/pom.xml | 2 +-
 pom.xml | 4 ++--
 processing/pom.xml  | 2 +-
 sdk/sdk/pom.xml | 2 +-
 streaming/pom.xml   | 2 +-
 tools/cli/pom.xml   | 2 +-
 24 files changed, 25 insertions(+), 25 deletions(-)

diff --git a/assembly/pom.xml b/assembly/pom.xml
index d4c7aba..1107569 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/common/pom.xml b/common/pom.xml
index 68fea9c..88dc1b8 100644
--- a/common/pom.xml
+++ b/common/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/core/pom.xml b/core/pom.xml
index f4d2c16..0e74e7a 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/examples/flink/pom.xml b/examples/flink/pom.xml
index cd4b691..da34787 100644
--- a/examples/flink/pom.xml
+++ b/examples/flink/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/examples/spark/pom.xml b/examples/spark/pom.xml
index 0f804f4..ca666db 100644
--- a/examples/spark/pom.xml
+++ b/examples/spark/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/format/pom.xml b/format/pom.xml
index c547d37..a8c5b60 100644
--- a/format/pom.xml
+++ b/format/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/geo/pom.xml b/geo/pom.xml
index fcbd80c..83bbbd9 100644
--- a/geo/pom.xml
+++ b/geo/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/hadoop/pom.xml b/hadoop/pom.xml
index e8be41e..298337b 100644
--- a/hadoop/pom.xml
+++ b/hadoop/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/index/bloom/pom.xml b/index/bloom/pom.xml
index c4071ba..1096951 100644
--- a/index/bloom/pom.xml
+++ b/index/bloom/pom.xml
@@ -4,7 +4,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/index/examples/pom.xml b/index/examples/pom.xml
index 3873996..3522097 100644
--- a/index/examples/pom.xml
+++ b/index/examples/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/index/lucene/pom.xml b/index/lucene/pom.xml
index b10583d..88d8afb 100644
--- a/index/lucene/pom.xml
+++ b/index/lucene/pom.xml
@@ -4,7 +4,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/index/secondary-index/pom.xml b/index/secondary-index/pom.xml
index b998143..f4ce62e 100644
--- a/index/secondary-index/pom.xml
+++ b/index/secondary-index/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/integration/flink-build/pom.xml b/integration/flink-build/pom.xml
index c443497..ba37c2e 100644
--- a/integration/flink-build/pom.xml
+++ b/integration/flink-build/pom.xml
@@ -4,7 +4,7 @@
 
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
 
 
diff --git a/integration/flink-proxy/pom.xml b/integration/flink-proxy/pom.xml
index 23f5169..832d804 100644
--- a/integration/flink-proxy/pom.xml
+++ b/integration/flink

[carbondata] branch master updated: [CARBONDATA-4318]Improve load overwrite performance for partition tables

2021-12-29 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new 308906e  [CARBONDATA-4318]Improve load overwrite performance for 
partition tables
308906e is described below

commit 308906e459607383eedf20c18236c7d17509959e
Author: akashrn5 
AuthorDate: Wed Dec 8 15:49:47 2021 +0530

[CARBONDATA-4318]Improve load overwrite performance for partition tables

Why is this PR needed?
With the increase in the number of overwrite loads for the partition table,
the time takes for each load keeps on increasing over time. This is because,

1. whenever a load overwrite for partition table is fired, it basically 
means
that we need to overwrite or drop the partitions if anything overlaps with
current partitions getting loaded. Since carbondata stores the partition
information in the segments file, to identify and drop partitions, it's
reading all the previous segment files to identify and drop the overwriting
partitions, which leads to a decrease in performance.

2. After partition load is completed, a cleanSegments method is called which
again reads segment file and table status file to identify MArked for delete
segments to clean. But Since the force clean is false and timeout also is
more than a day by default, it's not necessary to call this method.
Clean files should handle this part.

What changes were proposed in this PR?
1. we already have the information about current partitions, so with that 
first
identify if there are any partitions to overwrite, if yes then only we read 
segment
files to call dropParitition, else we don't read the segment files 
unnecessarily.
It also contains other refactoring to avoid reading table status file also.
2. no need to call clean segments after every load. Clean files will take 
care
to delete the expired ones.

This closes #4242
---
 .../carbondata/core/metadata/SegmentFileStore.java | 15 +++--
 .../hadoop/api/CarbonOutputCommitter.java  | 73 +-
 .../spark/rdd/CarbonDropPartitionRDD.scala |  6 +-
 .../command/management/CommonLoadUtils.scala   |  5 --
 .../CarbonAlterTableDropHivePartitionCommand.scala | 15 ++---
 .../allqueries/InsertIntoCarbonTableTestCase.scala | 24 ---
 6 files changed, 76 insertions(+), 62 deletions(-)

diff --git 
a/core/src/main/java/org/apache/carbondata/core/metadata/SegmentFileStore.java 
b/core/src/main/java/org/apache/carbondata/core/metadata/SegmentFileStore.java
index 67c061b..eda819b 100644
--- 
a/core/src/main/java/org/apache/carbondata/core/metadata/SegmentFileStore.java
+++ 
b/core/src/main/java/org/apache/carbondata/core/metadata/SegmentFileStore.java
@@ -1005,10 +1005,9 @@ public class SegmentFileStore {
* @param uniqueId
* @throws IOException
*/
-  public void dropPartitions(Segment segment, List 
partitionSpecs,
+  public void dropPartitions(String segmentNo, List partitionLocations,
   String uniqueId, List toBeDeletedSegments, List 
toBeUpdatedSegments)
   throws IOException {
-readSegment(tablePath, segment.getSegmentFileName());
 boolean updateSegment = false;
 for (Map.Entry entry : 
segmentFile.getLocationMap().entrySet()) {
   String location = entry.getKey();
@@ -1017,9 +1016,9 @@ public class SegmentFileStore {
   }
   Path path = new Path(location);
   // Update the status to delete if path equals
-  if (null != partitionSpecs) {
-for (PartitionSpec spec : partitionSpecs) {
-  if (path.equals(spec.getLocation())) {
+  if (null != partitionLocations) {
+for (String partitionLocation : partitionLocations) {
+  if (path.toString().equals(partitionLocation)) {
 
entry.getValue().setStatus(SegmentStatus.MARKED_FOR_DELETE.getMessage());
 updateSegment = true;
 break;
@@ -1031,7 +1030,7 @@ public class SegmentFileStore {
   String writePath = CarbonTablePath.getSegmentFilesLocation(tablePath);
   writePath =
   writePath + CarbonCommonConstants.FILE_SEPARATOR +
-  SegmentFileStore.genSegmentFileName(segment.getSegmentNo(),  
String.valueOf(uniqueId))
+  SegmentFileStore.genSegmentFileName(segmentNo,  
String.valueOf(uniqueId))
   + CarbonTablePath.SEGMENT_EXT;
   writeSegmentFile(segmentFile, writePath);
 }
@@ -1044,10 +1043,10 @@ public class SegmentFileStore {
   }
 }
 if (deleteSegment) {
-  toBeDeletedSegments.add(segment.getSegmentNo());
+  toBeDeletedSegments.add(segmentNo);
 }
 if (updateSegment) {
-  toBeUpdatedSegments.add(segment.getSegmentNo());
+  toBeUpdatedSegments.add(segmentNo);
 }
   }
 
diff --git 
a/hadoop/src/main/java/org/apache/carbondata/hadoop

[carbondata] branch master updated: [CARBONDATA-4319] Fixed clean files not deleteting stale delete delta files after horizontal compaction

2021-12-28 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new a072e7a  [CARBONDATA-4319] Fixed clean files not deleteting stale 
delete delta files after horizontal compaction
a072e7a is described below

commit a072e7a9e8af7491ee009c3f4005cd028801f02e
Author: Vikram Ahuja 
AuthorDate: Wed Dec 22 13:21:06 2021 +0530

[CARBONDATA-4319] Fixed clean files not deleteting stale delete delta files 
after horizontal compaction

Why is this PR needed?
After horizontal compaction was performed on partition and non partition 
tables, the clean files
operation was not deleting the stale delete delta files. the code was 
removed as the part of clean
files refactoring done previously.

What changes were proposed in this PR?
Clean files with force option now handles removal of these stale delta 
files as well as the stale
tableupdatestatus file for both partition and non partition table.

This closes #4245
---
 .../carbondata/core/mutate/CarbonUpdateUtil.java   | 122 +
 .../core/statusmanager/SegmentStatusManager.java   |  20 
 docs/clean-files.md|   8 +-
 .../apache/carbondata/trash/DataTrashManager.scala |  20 +++-
 .../cleanfiles/TestCleanFileCommand.scala  |  34 ++
 .../TestCleanFilesCommandPartitionTable.scala  |  68 
 6 files changed, 266 insertions(+), 6 deletions(-)

diff --git 
a/core/src/main/java/org/apache/carbondata/core/mutate/CarbonUpdateUtil.java 
b/core/src/main/java/org/apache/carbondata/core/mutate/CarbonUpdateUtil.java
index b814152..da8d329 100644
--- a/core/src/main/java/org/apache/carbondata/core/mutate/CarbonUpdateUtil.java
+++ b/core/src/main/java/org/apache/carbondata/core/mutate/CarbonUpdateUtil.java
@@ -25,6 +25,7 @@ import java.util.HashSet;
 import java.util.List;
 import java.util.Map;
 import java.util.Set;
+import java.util.stream.Collectors;
 
 import org.apache.carbondata.common.logging.LogServiceFactory;
 import org.apache.carbondata.core.constants.CarbonCommonConstants;
@@ -688,4 +689,125 @@ public class CarbonUpdateUtil {
 }
 return latestTimestamp;
   }
+
+
+  /**
+   * Handling of the clean up of old carbondata files, index files , delete 
delta,
+   * update status files.
+   *
+   * @param table   clean up will be handled on this table.
+   * @param isDryRun if clean files dryRun selected, then only size will be 
shown,
+   *files will not be deleted.
+   */
+  public static long cleanUpDeltaFiles(CarbonTable table, boolean isDryRun) 
throws IOException {
+
+SegmentStatusManager ssm = new 
SegmentStatusManager(table.getAbsoluteTableIdentifier());
+LoadMetadataDetails[] details =
+SegmentStatusManager.readLoadMetadata(table.getMetadataPath());
+long totalSizeDeleted = 0;
+ArrayList filesToBeDeleted = new ArrayList<>();
+SegmentUpdateStatusManager updateStatusManager = new 
SegmentUpdateStatusManager(table);
+SegmentUpdateDetails[] segmentUpdateDetails = 
updateStatusManager.getUpdateStatusDetails();
+// hold all the segments updated so that wen can check the delta files in 
them, ne need to
+// check the others.
+Set updatedSegments = new HashSet<>();
+for (SegmentUpdateDetails updateDetails : segmentUpdateDetails) {
+  updatedSegments.add(updateDetails.getSegmentName());
+}
+boolean isInvalidFile = false;
+// take the update status file name from 0th segment.
+String validUpdateStatusFile = ssm.getUpdateStatusFileName(details);
+// scan through each segment.
+
+if (table.isHivePartitionTable()) {
+  List partitionList = 
Arrays.stream(FileFactory.getCarbonFile(table
+  .getTablePath()).listFiles()).filter(partitionName -> 
partitionName.getName()
+  .contains("=")).collect(Collectors.toList());
+
+  List totalDeltaFiles = new ArrayList<>();
+
+  for (CarbonFile carbonFile : partitionList) {
+
totalDeltaFiles.addAll(carbonFile.listFiles(true).stream().filter(fileName -> 
fileName
+
.getName().endsWith(CarbonCommonConstants.DELETE_DELTA_FILE_EXT))
+.collect(Collectors.toList()));
+  }
+  SegmentUpdateDetails[] updateDetails = 
updateStatusManager.readLoadMetadata();
+  for (SegmentUpdateDetails block : updateDetails) {
+totalDeltaFiles.stream().filter(fileName -> 
fileName.getName().endsWith(block
+.getDeleteDeltaStartTimestamp() + 
CarbonCommonConstants.DELETE_DELTA_FILE_EXT))
+.collect(Collectors.toList()).forEach(fileName -> 
totalDeltaFiles.remove(fileName));
+  }
+  for (CarbonFile invalidFile: totalDeltaFiles) {
+totalSizeDeleted += invalidFile.getSize();
+

[carbondata] branch carbon3.0 created (now 0f1d2a4)

2021-12-23 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a change to branch carbon3.0
in repository https://gitbox.apache.org/repos/asf/carbondata.git.


  at 0f1d2a4  [CARBONDATA-4317] Fix TPCDS performance issues

No new revisions were added by this update.


[carbondata] branch master updated: [CARBONDATA-4317] Fix TPCDS performance issues

2021-12-22 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new 0f1d2a4  [CARBONDATA-4317] Fix TPCDS performance issues
0f1d2a4 is described below

commit 0f1d2a45e5f614fd123bd734ab37d7e453c21344
Author: Indhumathi27 
AuthorDate: Tue Dec 7 20:32:05 2021 +0530

[CARBONDATA-4317] Fix TPCDS performance issues

Why is this PR needed?
The following issues has degraded the TPCDS query performance
1. If dynamic filters is not present in partitionFilters Set, then that 
filter is skipped, to pushdown to spark.
2. In some cases, some nodes like Exchange / Shuffle is not reused, because 
the CarbonDataSourceSCan plan is not mached
3. While accessing the metadata on the canonicalized plan throws NPE

What changes were proposed in this PR?
1. Check if dynamic filters is present in PartitionFilters set. If not, 
pushdown the filter
2. Match the plans, by converting them to canonicalized and by normalising 
the expressions
3. Move variables used in metadata(), to avoid NPE while comparing plans

This closes #4241
---
 .../execution/strategy/CarbonDataSourceScan.scala  | 49 ++
 .../execution/strategy/CarbonSourceStrategy.scala  |  6 +--
 .../apache/spark/sql/CarbonToSparkAdapter.scala|  4 +-
 .../apache/spark/sql/CarbonToSparkAdapter.scala|  4 +-
 .../apache/spark/sql/CarbonToSparkAdapter.scala|  7 +++-
 5 files changed, 54 insertions(+), 16 deletions(-)

diff --git 
a/integration/spark/src/main/scala/org/apache/spark/sql/execution/strategy/CarbonDataSourceScan.scala
 
b/integration/spark/src/main/scala/org/apache/spark/sql/execution/strategy/CarbonDataSourceScan.scala
index 2e1bb96..31685b0 100644
--- 
a/integration/spark/src/main/scala/org/apache/spark/sql/execution/strategy/CarbonDataSourceScan.scala
+++ 
b/integration/spark/src/main/scala/org/apache/spark/sql/execution/strategy/CarbonDataSourceScan.scala
@@ -31,7 +31,6 @@ import 
org.apache.spark.sql.catalyst.plans.physical.{HashPartitioning, Partition
 import org.apache.spark.sql.execution.WholeStageCodegenExec
 
 import org.apache.carbondata.core.metadata.schema.BucketingInfo
-import org.apache.carbondata.core.readcommitter.ReadCommittedScope
 import org.apache.carbondata.core.scan.expression.Expression
 import org.apache.carbondata.hadoop.CarbonProjection
 
@@ -44,7 +43,6 @@ case class CarbonDataSourceScan(
 output: Seq[Attribute],
 partitionFiltersWithoutDpp: Seq[SparkExpression],
 dataFilters: Seq[SparkExpression],
-@transient readCommittedScope: ReadCommittedScope,
 @transient pushedDownProjection: CarbonProjection,
 @transient pushedDownFilters: Seq[Expression],
 directScanSupport: Boolean,
@@ -64,6 +62,10 @@ case class CarbonDataSourceScan(
 partitionFiltersWithDpp,
 segmentIds) {
 
+  val pushDownFiltersStr: String = 
seqToString(pushedDownFilters.map(_.getStatement))
+
+  val projectionColStr: String = 
seqToString(pushedDownProjection.getAllColumns)
+
   override lazy val (outputPartitioning, outputOrdering): (Partitioning, 
Seq[SortOrder]) = {
 val info: BucketingInfo = relation.carbonTable.getBucketingInfo
 if (info != null) {
@@ -91,15 +93,18 @@ case class CarbonDataSourceScan(
 }
   }
 
+  def seqToString(seq: Seq[Any]): String = seq.mkString("[", ", ", "]")
+
   override lazy val metadata: Map[String, String] = {
-def seqToString(seq: Seq[Any]) = seq.mkString("[", ", ", "]")
 val metadata =
   Map(
-"ReadSchema" -> seqToString(pushedDownProjection.getAllColumns),
+"ReadSchema" -> projectionColStr,
 "Batched" -> supportsBatchOrColumnar.toString,
 "DirectScan" -> (supportsBatchOrColumnar && 
directScanSupport).toString,
-"PushedFilters" -> seqToString(pushedDownFilters.map(_.getStatement)))
-if (relation.carbonTable.isHivePartitionTable) {
+"PushedFilters" -> pushDownFiltersStr)
+// if plan is canonicalized, then filter expressions will be normalized. 
In that case,
+// skip adding selected partitions to metadata
+if (!this.isCanonicalizedPlan && 
relation.carbonTable.isHivePartitionTable) {
   metadata + ("PartitionFilters" -> seqToString(partitionFiltersWithDpp)) +
 ("PartitionCount" -> selectedPartitions.size.toString)
 } else {
@@ -142,14 +147,40 @@ case class CarbonDataSourceScan(
   outputAttibutesAfterNormalizingExpressionIds,
   QueryPlan.normalizePredicates(partitionFiltersWithoutDpp, output),
   QueryPlan.normalizePredicates(dataFilters, output),
-  null,
-  null,
+  pushedDownProjection,
   Seq.empty,
   

[carbondata] branch master updated: [CARBONDATA-4316]Fix horizontal compaction failure for partition tables

2021-12-22 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new d629dc0  [CARBONDATA-4316]Fix horizontal compaction failure for 
partition tables
d629dc0 is described below

commit d629dc0b894a64bfbef762736775a182e40827fe
Author: akashrn5 
AuthorDate: Tue Dec 7 18:55:13 2021 +0530

[CARBONDATA-4316]Fix horizontal compaction failure for partition tables

Why is this PR needed?
Horizontal compaction fails for partition table leading to many delete
delta files for a single block, leading to slower query performance.
This is happening because during horizontal compaction the delta file
path prepared for the partition table is wrong which fails to identify
the path and fails the operation.

What changes were proposed in this PR?
If it is a partition table, read the segment file and identity the
partition where the block is present to prepare a proper partition path.

This closes #4240
---
 .../statusmanager/SegmentUpdateStatusManager.java  | 24 ++-
 .../mutation/CarbonProjectForDeleteCommand.scala   |  9 --
 .../command/mutation/HorizontalCompaction.scala|  4 +--
 .../testsuite/iud/DeleteCarbonTableTestCase.scala  | 34 ++
 .../processing/merger/CarbonDataMergerUtil.java|  9 +++---
 5 files changed, 69 insertions(+), 11 deletions(-)

diff --git 
a/core/src/main/java/org/apache/carbondata/core/statusmanager/SegmentUpdateStatusManager.java
 
b/core/src/main/java/org/apache/carbondata/core/statusmanager/SegmentUpdateStatusManager.java
index fe40494..8572ae5 100644
--- 
a/core/src/main/java/org/apache/carbondata/core/statusmanager/SegmentUpdateStatusManager.java
+++ 
b/core/src/main/java/org/apache/carbondata/core/statusmanager/SegmentUpdateStatusManager.java
@@ -19,6 +19,7 @@ package org.apache.carbondata.core.statusmanager;
 
 import java.io.*;
 import java.util.*;
+import java.util.stream.Collectors;
 
 import org.apache.carbondata.common.logging.LogServiceFactory;
 import org.apache.carbondata.core.constants.CarbonCommonConstants;
@@ -33,6 +34,7 @@ import org.apache.carbondata.core.locks.CarbonLockFactory;
 import org.apache.carbondata.core.locks.ICarbonLock;
 import org.apache.carbondata.core.locks.LockUsage;
 import org.apache.carbondata.core.metadata.AbsoluteTableIdentifier;
+import org.apache.carbondata.core.metadata.SegmentFileStore;
 import org.apache.carbondata.core.metadata.schema.table.CarbonTable;
 import org.apache.carbondata.core.mutate.CarbonUpdateUtil;
 import org.apache.carbondata.core.mutate.SegmentUpdateDetails;
@@ -364,7 +366,8 @@ public class SegmentUpdateStatusManager {
* @param blockName the specified block of the segment
* @return delete delta file list of the block
*/
-  public List getDeleteDeltaFilesList(final Segment segment, final 
String blockName) {
+  public List getDeleteDeltaFilesList(final Segment segment, final 
String blockName)
+  throws IOException {
 List deleteDeltaFileList = new ArrayList<>();
 String segmentPath = null;
 if (segment.isExternalSegment()) {
@@ -374,6 +377,25 @@ public class SegmentUpdateStatusManager {
   break;
 }
   }
+} else if (isPartitionTable) {
+  String segmentFileName = Arrays.stream(segmentDetails).filter(
+  loadMetaDataDetail -> loadMetaDataDetail.getLoadName()
+  
.equalsIgnoreCase(segment.getSegmentNo())).collect(Collectors.toList()).get(0)
+  .getSegmentFile();
+  SegmentFileStore segmentFileStore =
+  new SegmentFileStore(identifier.getTablePath(), segmentFileName);
+  segmentFileStore.readIndexFiles(SegmentStatus.SUCCESS, false, 
FileFactory.getConfiguration());
+  for (Map.Entry> entry : 
segmentFileStore.getIndexFilesMap().entrySet()) {
+List matchedBlocksInPartition = 
entry.getValue().stream().filter(blockFile -> {
+  String blockFileName = 
blockFile.substring(blockFile.lastIndexOf(File.separator) + 1);
+  return 
blockName.equalsIgnoreCase(CarbonUpdateUtil.getBlockName(blockFileName));
+}).collect(Collectors.toList());
+if (matchedBlocksInPartition.size() > 0) {
+  segmentPath = matchedBlocksInPartition.get(0)
+  .substring(0, 
matchedBlocksInPartition.get(0).lastIndexOf(File.separator));
+  break;
+}
+  }
 } else {
   segmentPath = CarbonTablePath.getSegmentPath(
   identifier.getTablePath(), segment.getSegmentNo());
diff --git 
a/integration/spark/src/main/scala/org/apache/spark/sql/execution/command/mutation/CarbonProjectForDeleteCommand.scala
 
b/integration/spark/src/main/scala/org/apache/spark/sql/execution/command/mutation/CarbonProjectForDeleteCommand.scala
index 87b1e41..7d54186 100644
--- 
a/integration/spark/

svn commit: r51639 - in /dev/carbondata/2.3.0-rc1: ./ apache-carbondata-2.3.0-source-release.zip apache-carbondata-2.3.0-source-release.zip.asc apache-carbondata-2.3.0-source-release.zip.md5 apache-ca

2021-12-20 Thread kunalkapoor
Author: kunalkapoor
Date: Mon Dec 20 13:36:25 2021
New Revision: 51639

Log:
Upload 2.3.0 rc1

Added:
dev/carbondata/2.3.0-rc1/
dev/carbondata/2.3.0-rc1/apache-carbondata-2.3.0-source-release.zip   (with 
props)
dev/carbondata/2.3.0-rc1/apache-carbondata-2.3.0-source-release.zip.asc
dev/carbondata/2.3.0-rc1/apache-carbondata-2.3.0-source-release.zip.md5
dev/carbondata/2.3.0-rc1/apache-carbondata-2.3.0-source-release.zip.sha512

Added: dev/carbondata/2.3.0-rc1/apache-carbondata-2.3.0-source-release.zip
==
Binary file - no diff available.

Propchange: dev/carbondata/2.3.0-rc1/apache-carbondata-2.3.0-source-release.zip
--
svn:mime-type = application/octet-stream

Added: dev/carbondata/2.3.0-rc1/apache-carbondata-2.3.0-source-release.zip.asc
==
--- dev/carbondata/2.3.0-rc1/apache-carbondata-2.3.0-source-release.zip.asc 
(added)
+++ dev/carbondata/2.3.0-rc1/apache-carbondata-2.3.0-source-release.zip.asc Mon 
Dec 20 13:36:25 2021
@@ -0,0 +1,14 @@
+-BEGIN PGP SIGNATURE-
+
+iQGzBAABCgAdFiEER4oGoEjiZ19CtKBtzP3MO+Fg6doFAmHAQJAACgkQzP3MO+Fg
+6drCRQv+Prlt6h4D9j7yDEnrbYtY1ZCEp+MeVvcPV4ovhsNLD8W79V9+W23e6gXD
+mwfwoDA8nJEyhtE4v3ku6OAMaQQK5FXn2e072qn2+XwPt8yq39vejIs/U4/72y9W
+2/z78076ThsUeO1CbnzCB5fy4A1RTFE6kKW6SWEXkjz6waG+Yrgze3Mti+z0Jre9
+TYk+xMQYZ4hGJ3ed62kv1ZDn/umOfN+fXxsAXlOpZi/Zb3K0QcskWFTjaC4as/gp
+Ddn3GNKoN1h84Scg5B/INrqobGZnpjZ9XNl/ZPx82mV7PPUYVwqL/FifNA7Zypse
+5XO/Jilja84VtYABDwV19F6xToOpX85CxRchUn2FF/XObjE7UKaJvNgXCBF22bem
+77aL3vCCycQMdRKWMp0wQU035WXtyWt12B8FzvFxKosI5yz1JjyiNhgCTc1pAIOU
++/T373fInbax15mIZbcNVOBRZvZFFu+ysPWcTnbNfFvD4ElM0/m8+v2XB3f56D56
+6hLbuHdo
+=Ohii
+-END PGP SIGNATURE-

Added: dev/carbondata/2.3.0-rc1/apache-carbondata-2.3.0-source-release.zip.md5
==
--- dev/carbondata/2.3.0-rc1/apache-carbondata-2.3.0-source-release.zip.md5 
(added)
+++ dev/carbondata/2.3.0-rc1/apache-carbondata-2.3.0-source-release.zip.md5 Mon 
Dec 20 13:36:25 2021
@@ -0,0 +1 @@
+8fd9a7e0246da32e73f8b936f7d88633  apache-carbondata-2.3.0-source-release.zip

Added: 
dev/carbondata/2.3.0-rc1/apache-carbondata-2.3.0-source-release.zip.sha512
==
--- dev/carbondata/2.3.0-rc1/apache-carbondata-2.3.0-source-release.zip.sha512 
(added)
+++ dev/carbondata/2.3.0-rc1/apache-carbondata-2.3.0-source-release.zip.sha512 
Mon Dec 20 13:36:25 2021
@@ -0,0 +1 @@
+7bd191c5a070c5380af89cef71052fbf9d4b161080fd92443b3457fe5890d053c99ad47dbe2d0002a39af5906f3f4aad3405fb9304817916c0ea40ce35e15003
  apache-carbondata-2.3.0-source-release.zip




[carbondata] branch branch-2.3 created (now 7006589)

2021-12-20 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a change to branch branch-2.3
in repository https://gitbox.apache.org/repos/asf/carbondata.git.


  at 7006589  [maven-release-plugin] prepare for next development iteration

This branch includes the following new commits:

 new 7006589  [maven-release-plugin] prepare for next development iteration

The 1 revisions listed above as "new" are entirely new to this
repository and will be described in separate emails.  The revisions
listed as "add" were already present in the repository and have only
been added to this reference.



[carbondata] 01/01: [maven-release-plugin] prepare for next development iteration

2021-12-20 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch branch-2.3
in repository https://gitbox.apache.org/repos/asf/carbondata.git

commit 70065894d02ce2e898b1ed3cd7b0b10f6305db44
Author: kunal642 
AuthorDate: Mon Dec 20 18:04:52 2021 +0530

[maven-release-plugin] prepare for next development iteration
---
 assembly/pom.xml| 2 +-
 common/pom.xml  | 2 +-
 core/pom.xml| 2 +-
 examples/flink/pom.xml  | 2 +-
 examples/spark/pom.xml  | 2 +-
 format/pom.xml  | 2 +-
 geo/pom.xml | 2 +-
 hadoop/pom.xml  | 2 +-
 index/bloom/pom.xml | 2 +-
 index/examples/pom.xml  | 2 +-
 index/lucene/pom.xml| 2 +-
 index/secondary-index/pom.xml   | 2 +-
 integration/flink-build/pom.xml | 2 +-
 integration/flink-proxy/pom.xml | 2 +-
 integration/flink/pom.xml   | 2 +-
 integration/hive/pom.xml| 2 +-
 integration/presto/pom.xml  | 2 +-
 integration/spark/pom.xml   | 2 +-
 mv/plan/pom.xml | 2 +-
 pom.xml | 4 ++--
 processing/pom.xml  | 2 +-
 sdk/sdk/pom.xml | 2 +-
 streaming/pom.xml   | 2 +-
 tools/cli/pom.xml   | 2 +-
 24 files changed, 25 insertions(+), 25 deletions(-)

diff --git a/assembly/pom.xml b/assembly/pom.xml
index 1107569..d8d03b7 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/common/pom.xml b/common/pom.xml
index 88dc1b8..6205781 100644
--- a/common/pom.xml
+++ b/common/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/core/pom.xml b/core/pom.xml
index 0e74e7a..a5f2e75 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/examples/flink/pom.xml b/examples/flink/pom.xml
index da34787..360fcff 100644
--- a/examples/flink/pom.xml
+++ b/examples/flink/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/examples/spark/pom.xml b/examples/spark/pom.xml
index ca666db..64f0d9f 100644
--- a/examples/spark/pom.xml
+++ b/examples/spark/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/format/pom.xml b/format/pom.xml
index a8c5b60..e992b2e 100644
--- a/format/pom.xml
+++ b/format/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/geo/pom.xml b/geo/pom.xml
index 83bbbd9..46f6695 100644
--- a/geo/pom.xml
+++ b/geo/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/hadoop/pom.xml b/hadoop/pom.xml
index 298337b..c38b0f9 100644
--- a/hadoop/pom.xml
+++ b/hadoop/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/index/bloom/pom.xml b/index/bloom/pom.xml
index 1096951..7ec45b4 100644
--- a/index/bloom/pom.xml
+++ b/index/bloom/pom.xml
@@ -4,7 +4,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/index/examples/pom.xml b/index/examples/pom.xml
index 3522097..bccbae1 100644
--- a/index/examples/pom.xml
+++ b/index/examples/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/index/lucene/pom.xml b/index/lucene/pom.xml
index 88d8afb..e23b94d 100644
--- a/index/lucene/pom.xml
+++ b/index/lucene/pom.xml
@@ -4,7 +4,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/index/secondary-index/pom.xml b/index/secondary-index/pom.xml
index f4ce62e..b1a2414 100644
--- a/index/secondary-index/pom.xml
+++ b/index/secondary-index/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/integration/flink-build/pom.xml b/integration/flink-build/pom.xml
index ba37c2e..a4b320e 100644
--- a/integration/flink-build/pom.xml
+++ b/integration/flink-build/pom.xml
@@ -4,7 +4,7 @@
 
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
 
 
diff --git a/integration/flink-proxy/pom.xml b/integration/flink-proxy/pom.xml
index 832d804..b7aee50 100644
--- a/integration/flink-proxy/pom.xml
+++ b/integration/flink-proxy

[carbondata] annotated tag apache-carbondata-2.3.0-rc1 created (now 6bb1227)

2021-12-20 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a change to annotated tag apache-carbondata-2.3.0-rc1
in repository https://gitbox.apache.org/repos/asf/carbondata.git.


  at 6bb1227  (tag)
 tagging 7fe1824733131785f78093567b34b6f94f1dc25a (commit)
  by kunal642
  on Mon Dec 20 18:04:23 2021 +0530

- Log -
[maven-release-plugin] copy for tag apache-carbondata-2.3.0-rc1
---

This annotated tag includes the following new commits:

 new 7fe1824  [maven-release-plugin] prepare release 
apache-carbondata-2.3.0-rc1

The 1 revisions listed above as "new" are entirely new to this
repository and will be described in separate emails.  The revisions
listed as "add" were already present in the repository and have only
been added to this reference.



[carbondata] 01/01: [maven-release-plugin] prepare release apache-carbondata-2.3.0-rc1

2021-12-20 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to annotated tag apache-carbondata-2.3.0-rc1
in repository https://gitbox.apache.org/repos/asf/carbondata.git

commit 7fe1824733131785f78093567b34b6f94f1dc25a
Author: kunal642 
AuthorDate: Mon Dec 20 17:59:41 2021 +0530

[maven-release-plugin] prepare release apache-carbondata-2.3.0-rc1
---
 assembly/pom.xml| 2 +-
 common/pom.xml  | 2 +-
 core/pom.xml| 2 +-
 examples/flink/pom.xml  | 2 +-
 examples/spark/pom.xml  | 2 +-
 format/pom.xml  | 2 +-
 geo/pom.xml | 2 +-
 hadoop/pom.xml  | 2 +-
 index/bloom/pom.xml | 2 +-
 index/examples/pom.xml  | 2 +-
 index/lucene/pom.xml| 2 +-
 index/secondary-index/pom.xml   | 2 +-
 integration/flink-build/pom.xml | 2 +-
 integration/flink-proxy/pom.xml | 2 +-
 integration/flink/pom.xml   | 2 +-
 integration/hive/pom.xml| 2 +-
 integration/presto/pom.xml  | 2 +-
 integration/spark/pom.xml   | 2 +-
 mv/plan/pom.xml | 2 +-
 pom.xml | 4 ++--
 processing/pom.xml  | 2 +-
 sdk/sdk/pom.xml | 2 +-
 streaming/pom.xml   | 2 +-
 tools/cli/pom.xml   | 2 +-
 24 files changed, 25 insertions(+), 25 deletions(-)

diff --git a/assembly/pom.xml b/assembly/pom.xml
index d4c7aba..1107569 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/common/pom.xml b/common/pom.xml
index 68fea9c..88dc1b8 100644
--- a/common/pom.xml
+++ b/common/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/core/pom.xml b/core/pom.xml
index f4d2c16..0e74e7a 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/examples/flink/pom.xml b/examples/flink/pom.xml
index cd4b691..da34787 100644
--- a/examples/flink/pom.xml
+++ b/examples/flink/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/examples/spark/pom.xml b/examples/spark/pom.xml
index 0f804f4..ca666db 100644
--- a/examples/spark/pom.xml
+++ b/examples/spark/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/format/pom.xml b/format/pom.xml
index c547d37..a8c5b60 100644
--- a/format/pom.xml
+++ b/format/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/geo/pom.xml b/geo/pom.xml
index fcbd80c..83bbbd9 100644
--- a/geo/pom.xml
+++ b/geo/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/hadoop/pom.xml b/hadoop/pom.xml
index e8be41e..298337b 100644
--- a/hadoop/pom.xml
+++ b/hadoop/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/index/bloom/pom.xml b/index/bloom/pom.xml
index c4071ba..1096951 100644
--- a/index/bloom/pom.xml
+++ b/index/bloom/pom.xml
@@ -4,7 +4,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/index/examples/pom.xml b/index/examples/pom.xml
index 3873996..3522097 100644
--- a/index/examples/pom.xml
+++ b/index/examples/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/index/lucene/pom.xml b/index/lucene/pom.xml
index b10583d..88d8afb 100644
--- a/index/lucene/pom.xml
+++ b/index/lucene/pom.xml
@@ -4,7 +4,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/index/secondary-index/pom.xml b/index/secondary-index/pom.xml
index b998143..f4ce62e 100644
--- a/index/secondary-index/pom.xml
+++ b/index/secondary-index/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/integration/flink-build/pom.xml b/integration/flink-build/pom.xml
index c443497..ba37c2e 100644
--- a/integration/flink-build/pom.xml
+++ b/integration/flink-build/pom.xml
@@ -4,7 +4,7 @@
 
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
 
 
diff --git a/integration/flink-proxy/pom.xml b/integration/flink-proxy/pom.xml
index 23f5169..832d804 100644
--- a/integration/flink-proxy/pom.xml

[carbondata] branch master updated: [maven-release-plugin] Reverted the pom changes to 2.2.0-SNAPSHOT

2021-12-20 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new 0ced3c8  [maven-release-plugin] Reverted the pom changes to 
2.2.0-SNAPSHOT
0ced3c8 is described below

commit 0ced3c864257d92780409088c77bba0a1652f1cb
Author: kunal642 
AuthorDate: Mon Dec 20 16:05:26 2021 +0530

[maven-release-plugin] Reverted the pom changes to 2.2.0-SNAPSHOT
---
 assembly/pom.xml| 2 +-
 common/pom.xml  | 2 +-
 core/pom.xml| 2 +-
 examples/flink/pom.xml  | 2 +-
 examples/spark/pom.xml  | 2 +-
 format/pom.xml  | 2 +-
 geo/pom.xml | 2 +-
 hadoop/pom.xml  | 2 +-
 index/bloom/pom.xml | 2 +-
 index/examples/pom.xml  | 2 +-
 index/lucene/pom.xml| 2 +-
 index/secondary-index/pom.xml   | 2 +-
 integration/flink-build/pom.xml | 2 +-
 integration/flink-proxy/pom.xml | 2 +-
 integration/flink/pom.xml   | 2 +-
 integration/hive/pom.xml| 2 +-
 integration/presto/pom.xml  | 2 +-
 integration/spark/pom.xml   | 2 +-
 mv/plan/pom.xml | 2 +-
 pom.xml | 2 +-
 processing/pom.xml  | 2 +-
 sdk/sdk/pom.xml | 2 +-
 streaming/pom.xml   | 2 +-
 tools/cli/pom.xml   | 2 +-
 24 files changed, 24 insertions(+), 24 deletions(-)

diff --git a/assembly/pom.xml b/assembly/pom.xml
index d8d03b7..d4c7aba 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.1-SNAPSHOT
+2.3.0-SNAPSHOT
 ../pom.xml
   
 
diff --git a/common/pom.xml b/common/pom.xml
index 6205781..68fea9c 100644
--- a/common/pom.xml
+++ b/common/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.1-SNAPSHOT
+2.3.0-SNAPSHOT
 ../pom.xml
   
 
diff --git a/core/pom.xml b/core/pom.xml
index a5f2e75..f4d2c16 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.1-SNAPSHOT
+2.3.0-SNAPSHOT
 ../pom.xml
   
 
diff --git a/examples/flink/pom.xml b/examples/flink/pom.xml
index 360fcff..cd4b691 100644
--- a/examples/flink/pom.xml
+++ b/examples/flink/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.1-SNAPSHOT
+2.3.0-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/examples/spark/pom.xml b/examples/spark/pom.xml
index 64f0d9f..0f804f4 100644
--- a/examples/spark/pom.xml
+++ b/examples/spark/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.1-SNAPSHOT
+2.3.0-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/format/pom.xml b/format/pom.xml
index e992b2e..c547d37 100644
--- a/format/pom.xml
+++ b/format/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.1-SNAPSHOT
+2.3.0-SNAPSHOT
 ../pom.xml
   
 
diff --git a/geo/pom.xml b/geo/pom.xml
index 46f6695..fcbd80c 100644
--- a/geo/pom.xml
+++ b/geo/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.1-SNAPSHOT
+2.3.0-SNAPSHOT
 ../pom.xml
   
 
diff --git a/hadoop/pom.xml b/hadoop/pom.xml
index c38b0f9..e8be41e 100644
--- a/hadoop/pom.xml
+++ b/hadoop/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.1-SNAPSHOT
+2.3.0-SNAPSHOT
 ../pom.xml
   
 
diff --git a/index/bloom/pom.xml b/index/bloom/pom.xml
index 7ec45b4..c4071ba 100644
--- a/index/bloom/pom.xml
+++ b/index/bloom/pom.xml
@@ -4,7 +4,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.1-SNAPSHOT
+2.3.0-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/index/examples/pom.xml b/index/examples/pom.xml
index bccbae1..3873996 100644
--- a/index/examples/pom.xml
+++ b/index/examples/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.1-SNAPSHOT
+2.3.0-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/index/lucene/pom.xml b/index/lucene/pom.xml
index e23b94d..b10583d 100644
--- a/index/lucene/pom.xml
+++ b/index/lucene/pom.xml
@@ -4,7 +4,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.1-SNAPSHOT
+2.3.0-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/index/secondary-index/pom.xml b/index/secondary-index/pom.xml
index b1a2414..b998143 100644
--- a/index/secondary-index/pom.xml
+++ b/index/secondary-index/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.1-SNAPSHOT
+2.3.0-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/integration/flink-build/pom.xml b/integration/flink-build/pom.xml
index a4b320e..c443497 100644
--- a/integration/flink-build/pom.xml
+++ b/integration/flink-build/pom.xml
@@ -4,7 +4,7

[carbondata] annotated tag apache-carbondata-2.3.0-rc1 deleted (was 189252b)

2021-12-20 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a change to annotated tag apache-carbondata-2.3.0-rc1
in repository https://gitbox.apache.org/repos/asf/carbondata.git.


*** WARNING: tag apache-carbondata-2.3.0-rc1 was deleted! ***

   tag was  189252b

The revisions that were on this annotated tag are still contained in
other references; therefore, this change does not discard any commits
from the repository.


[carbondata] branch master updated: [maven-release-plugin] prepare for next development iteration

2021-12-20 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new c0211fc  [maven-release-plugin] prepare for next development iteration
c0211fc is described below

commit c0211fcff16a79b8e9b4ce73196125526dfbc10b
Author: kunal642 
AuthorDate: Mon Dec 20 16:25:30 2021 +0530

[maven-release-plugin] prepare for next development iteration
---
 assembly/pom.xml| 2 +-
 common/pom.xml  | 2 +-
 core/pom.xml| 2 +-
 examples/flink/pom.xml  | 2 +-
 examples/spark/pom.xml  | 2 +-
 format/pom.xml  | 2 +-
 geo/pom.xml | 2 +-
 hadoop/pom.xml  | 2 +-
 index/bloom/pom.xml | 2 +-
 index/examples/pom.xml  | 2 +-
 index/lucene/pom.xml| 2 +-
 index/secondary-index/pom.xml   | 2 +-
 integration/flink-build/pom.xml | 2 +-
 integration/flink-proxy/pom.xml | 2 +-
 integration/flink/pom.xml   | 2 +-
 integration/hive/pom.xml| 2 +-
 integration/presto/pom.xml  | 2 +-
 integration/spark/pom.xml   | 2 +-
 mv/plan/pom.xml | 2 +-
 pom.xml | 4 ++--
 processing/pom.xml  | 2 +-
 sdk/sdk/pom.xml | 2 +-
 streaming/pom.xml   | 2 +-
 tools/cli/pom.xml   | 2 +-
 24 files changed, 25 insertions(+), 25 deletions(-)

diff --git a/assembly/pom.xml b/assembly/pom.xml
index 1107569..d8d03b7 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/common/pom.xml b/common/pom.xml
index 88dc1b8..6205781 100644
--- a/common/pom.xml
+++ b/common/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/core/pom.xml b/core/pom.xml
index 0e74e7a..a5f2e75 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/examples/flink/pom.xml b/examples/flink/pom.xml
index da34787..360fcff 100644
--- a/examples/flink/pom.xml
+++ b/examples/flink/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/examples/spark/pom.xml b/examples/spark/pom.xml
index ca666db..64f0d9f 100644
--- a/examples/spark/pom.xml
+++ b/examples/spark/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/format/pom.xml b/format/pom.xml
index a8c5b60..e992b2e 100644
--- a/format/pom.xml
+++ b/format/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/geo/pom.xml b/geo/pom.xml
index 83bbbd9..46f6695 100644
--- a/geo/pom.xml
+++ b/geo/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/hadoop/pom.xml b/hadoop/pom.xml
index 298337b..c38b0f9 100644
--- a/hadoop/pom.xml
+++ b/hadoop/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/index/bloom/pom.xml b/index/bloom/pom.xml
index 1096951..7ec45b4 100644
--- a/index/bloom/pom.xml
+++ b/index/bloom/pom.xml
@@ -4,7 +4,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/index/examples/pom.xml b/index/examples/pom.xml
index 3522097..bccbae1 100644
--- a/index/examples/pom.xml
+++ b/index/examples/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/index/lucene/pom.xml b/index/lucene/pom.xml
index 88d8afb..e23b94d 100644
--- a/index/lucene/pom.xml
+++ b/index/lucene/pom.xml
@@ -4,7 +4,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/index/secondary-index/pom.xml b/index/secondary-index/pom.xml
index f4ce62e..b1a2414 100644
--- a/index/secondary-index/pom.xml
+++ b/index/secondary-index/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/integration/flink-build/pom.xml b/integration/flink-build/pom.xml
index ba37c2e..a4b320e 100644
--- a/integration/flink-build/pom.xml
+++ b/integration/flink-build/pom.xml
@@ -4,7 +4,7 @@
 
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml

[carbondata] annotated tag apache-carbondata-2.3.0-rc1 created (now 189252b)

2021-12-20 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a change to annotated tag apache-carbondata-2.3.0-rc1
in repository https://gitbox.apache.org/repos/asf/carbondata.git.


  at 189252b  (tag)
 tagging 053d080cf54ec84be5a547eb62c948300b55a929 (commit)
  by kunal642
  on Mon Dec 20 16:25:16 2021 +0530

- Log -
[maven-release-plugin] copy for tag apache-carbondata-2.3.0-rc1
---

No new revisions were added by this update.


[carbondata] annotated tag apache-carbondata-2.3.0-rc1 deleted (was 9038617)

2021-12-20 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a change to annotated tag apache-carbondata-2.3.0-rc1
in repository https://gitbox.apache.org/repos/asf/carbondata.git.


*** WARNING: tag apache-carbondata-2.3.0-rc1 was deleted! ***

   tag was  9038617

The revisions that were on this annotated tag are still contained in
other references; therefore, this change does not discard any commits
from the repository.


[carbondata] branch master updated: [maven-release-plugin] prepare release apache-carbondata-2.3.0-rc1

2021-12-20 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new 053d080  [maven-release-plugin] prepare release 
apache-carbondata-2.3.0-rc1
053d080 is described below

commit 053d080cf54ec84be5a547eb62c948300b55a929
Author: kunal642 
AuthorDate: Mon Dec 20 16:20:15 2021 +0530

[maven-release-plugin] prepare release apache-carbondata-2.3.0-rc1
---
 assembly/pom.xml| 2 +-
 common/pom.xml  | 2 +-
 core/pom.xml| 2 +-
 examples/flink/pom.xml  | 2 +-
 examples/spark/pom.xml  | 2 +-
 format/pom.xml  | 2 +-
 geo/pom.xml | 2 +-
 hadoop/pom.xml  | 2 +-
 index/bloom/pom.xml | 2 +-
 index/examples/pom.xml  | 2 +-
 index/lucene/pom.xml| 2 +-
 index/secondary-index/pom.xml   | 2 +-
 integration/flink-build/pom.xml | 2 +-
 integration/flink-proxy/pom.xml | 2 +-
 integration/flink/pom.xml   | 2 +-
 integration/hive/pom.xml| 2 +-
 integration/presto/pom.xml  | 2 +-
 integration/spark/pom.xml   | 2 +-
 mv/plan/pom.xml | 2 +-
 pom.xml | 4 ++--
 processing/pom.xml  | 2 +-
 sdk/sdk/pom.xml | 2 +-
 streaming/pom.xml   | 2 +-
 tools/cli/pom.xml   | 2 +-
 24 files changed, 25 insertions(+), 25 deletions(-)

diff --git a/assembly/pom.xml b/assembly/pom.xml
index d4c7aba..1107569 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/common/pom.xml b/common/pom.xml
index 68fea9c..88dc1b8 100644
--- a/common/pom.xml
+++ b/common/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/core/pom.xml b/core/pom.xml
index f4d2c16..0e74e7a 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/examples/flink/pom.xml b/examples/flink/pom.xml
index cd4b691..da34787 100644
--- a/examples/flink/pom.xml
+++ b/examples/flink/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/examples/spark/pom.xml b/examples/spark/pom.xml
index 0f804f4..ca666db 100644
--- a/examples/spark/pom.xml
+++ b/examples/spark/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/format/pom.xml b/format/pom.xml
index c547d37..a8c5b60 100644
--- a/format/pom.xml
+++ b/format/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/geo/pom.xml b/geo/pom.xml
index fcbd80c..83bbbd9 100644
--- a/geo/pom.xml
+++ b/geo/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/hadoop/pom.xml b/hadoop/pom.xml
index e8be41e..298337b 100644
--- a/hadoop/pom.xml
+++ b/hadoop/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/index/bloom/pom.xml b/index/bloom/pom.xml
index c4071ba..1096951 100644
--- a/index/bloom/pom.xml
+++ b/index/bloom/pom.xml
@@ -4,7 +4,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/index/examples/pom.xml b/index/examples/pom.xml
index 3873996..3522097 100644
--- a/index/examples/pom.xml
+++ b/index/examples/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/index/lucene/pom.xml b/index/lucene/pom.xml
index b10583d..88d8afb 100644
--- a/index/lucene/pom.xml
+++ b/index/lucene/pom.xml
@@ -4,7 +4,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/index/secondary-index/pom.xml b/index/secondary-index/pom.xml
index b998143..f4ce62e 100644
--- a/index/secondary-index/pom.xml
+++ b/index/secondary-index/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/integration/flink-build/pom.xml b/integration/flink-build/pom.xml
index c443497..ba37c2e 100644
--- a/integration/flink-build/pom.xml
+++ b/integration/flink-build/pom.xml
@@ -4,7 +4,7 @@
 
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml

[carbondata] branch master updated: [maven-release-plugin] Reverted the pom changes to 2.2.0-SNAPSHOT

2021-12-20 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new 01f8e1a  [maven-release-plugin] Reverted the pom changes to 
2.2.0-SNAPSHOT
01f8e1a is described below

commit 01f8e1afb1bf5e25202b1fe7f264a9abaf0ceb75
Author: kunal642 
AuthorDate: Mon Dec 20 16:05:26 2021 +0530

[maven-release-plugin] Reverted the pom changes to 2.2.0-SNAPSHOT
---
 assembly/pom.xml| 2 +-
 common/pom.xml  | 2 +-
 core/pom.xml| 2 +-
 examples/flink/pom.xml  | 2 +-
 examples/spark/pom.xml  | 2 +-
 format/pom.xml  | 2 +-
 geo/pom.xml | 2 +-
 hadoop/pom.xml  | 2 +-
 index/bloom/pom.xml | 2 +-
 index/examples/pom.xml  | 2 +-
 index/lucene/pom.xml| 2 +-
 index/secondary-index/pom.xml   | 2 +-
 integration/flink-build/pom.xml | 2 +-
 integration/flink-proxy/pom.xml | 2 +-
 integration/flink/pom.xml   | 2 +-
 integration/hive/pom.xml| 2 +-
 integration/presto/pom.xml  | 2 +-
 integration/spark/pom.xml   | 2 +-
 mv/plan/pom.xml | 2 +-
 pom.xml | 2 +-
 processing/pom.xml  | 2 +-
 sdk/sdk/pom.xml | 2 +-
 streaming/pom.xml   | 2 +-
 tools/cli/pom.xml   | 2 +-
 24 files changed, 24 insertions(+), 24 deletions(-)

diff --git a/assembly/pom.xml b/assembly/pom.xml
index d8d03b7..d4c7aba 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.1-SNAPSHOT
+2.3.0-SNAPSHOT
 ../pom.xml
   
 
diff --git a/common/pom.xml b/common/pom.xml
index 6205781..68fea9c 100644
--- a/common/pom.xml
+++ b/common/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.1-SNAPSHOT
+2.3.0-SNAPSHOT
 ../pom.xml
   
 
diff --git a/core/pom.xml b/core/pom.xml
index a5f2e75..f4d2c16 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.1-SNAPSHOT
+2.3.0-SNAPSHOT
 ../pom.xml
   
 
diff --git a/examples/flink/pom.xml b/examples/flink/pom.xml
index 360fcff..cd4b691 100644
--- a/examples/flink/pom.xml
+++ b/examples/flink/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.1-SNAPSHOT
+2.3.0-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/examples/spark/pom.xml b/examples/spark/pom.xml
index 64f0d9f..0f804f4 100644
--- a/examples/spark/pom.xml
+++ b/examples/spark/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.1-SNAPSHOT
+2.3.0-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/format/pom.xml b/format/pom.xml
index e992b2e..c547d37 100644
--- a/format/pom.xml
+++ b/format/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.1-SNAPSHOT
+2.3.0-SNAPSHOT
 ../pom.xml
   
 
diff --git a/geo/pom.xml b/geo/pom.xml
index 46f6695..fcbd80c 100644
--- a/geo/pom.xml
+++ b/geo/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.1-SNAPSHOT
+2.3.0-SNAPSHOT
 ../pom.xml
   
 
diff --git a/hadoop/pom.xml b/hadoop/pom.xml
index c38b0f9..e8be41e 100644
--- a/hadoop/pom.xml
+++ b/hadoop/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.1-SNAPSHOT
+2.3.0-SNAPSHOT
 ../pom.xml
   
 
diff --git a/index/bloom/pom.xml b/index/bloom/pom.xml
index 7ec45b4..c4071ba 100644
--- a/index/bloom/pom.xml
+++ b/index/bloom/pom.xml
@@ -4,7 +4,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.1-SNAPSHOT
+2.3.0-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/index/examples/pom.xml b/index/examples/pom.xml
index bccbae1..3873996 100644
--- a/index/examples/pom.xml
+++ b/index/examples/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.1-SNAPSHOT
+2.3.0-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/index/lucene/pom.xml b/index/lucene/pom.xml
index e23b94d..b10583d 100644
--- a/index/lucene/pom.xml
+++ b/index/lucene/pom.xml
@@ -4,7 +4,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.1-SNAPSHOT
+2.3.0-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/index/secondary-index/pom.xml b/index/secondary-index/pom.xml
index b1a2414..b998143 100644
--- a/index/secondary-index/pom.xml
+++ b/index/secondary-index/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.1-SNAPSHOT
+2.3.0-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/integration/flink-build/pom.xml b/integration/flink-build/pom.xml
index a4b320e..c443497 100644
--- a/integration/flink-build/pom.xml
+++ b/integration/flink-build/pom.xml
@@ -4,7 +4,7

[carbondata] 01/01: [maven-release-plugin] prepare for next development iteration

2021-12-20 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git

commit 861fc679bb5c3176da9c8c2677cbb342b540e819
Author: kunal642 
AuthorDate: Mon Dec 20 15:15:19 2021 +0530

[maven-release-plugin] prepare for next development iteration
---
 assembly/pom.xml| 2 +-
 common/pom.xml  | 2 +-
 core/pom.xml| 2 +-
 examples/flink/pom.xml  | 2 +-
 examples/spark/pom.xml  | 2 +-
 format/pom.xml  | 2 +-
 geo/pom.xml | 2 +-
 hadoop/pom.xml  | 2 +-
 index/bloom/pom.xml | 2 +-
 index/examples/pom.xml  | 2 +-
 index/lucene/pom.xml| 2 +-
 index/secondary-index/pom.xml   | 2 +-
 integration/flink-build/pom.xml | 2 +-
 integration/flink-proxy/pom.xml | 2 +-
 integration/flink/pom.xml   | 2 +-
 integration/hive/pom.xml| 2 +-
 integration/presto/pom.xml  | 2 +-
 integration/spark/pom.xml   | 2 +-
 mv/plan/pom.xml | 2 +-
 pom.xml | 4 ++--
 processing/pom.xml  | 2 +-
 sdk/sdk/pom.xml | 2 +-
 streaming/pom.xml   | 2 +-
 tools/cli/pom.xml   | 2 +-
 24 files changed, 25 insertions(+), 25 deletions(-)

diff --git a/assembly/pom.xml b/assembly/pom.xml
index 1107569..d8d03b7 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/common/pom.xml b/common/pom.xml
index 88dc1b8..6205781 100644
--- a/common/pom.xml
+++ b/common/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/core/pom.xml b/core/pom.xml
index 0e74e7a..a5f2e75 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/examples/flink/pom.xml b/examples/flink/pom.xml
index da34787..360fcff 100644
--- a/examples/flink/pom.xml
+++ b/examples/flink/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/examples/spark/pom.xml b/examples/spark/pom.xml
index ca666db..64f0d9f 100644
--- a/examples/spark/pom.xml
+++ b/examples/spark/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/format/pom.xml b/format/pom.xml
index a8c5b60..e992b2e 100644
--- a/format/pom.xml
+++ b/format/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/geo/pom.xml b/geo/pom.xml
index 83bbbd9..46f6695 100644
--- a/geo/pom.xml
+++ b/geo/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/hadoop/pom.xml b/hadoop/pom.xml
index 298337b..c38b0f9 100644
--- a/hadoop/pom.xml
+++ b/hadoop/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/index/bloom/pom.xml b/index/bloom/pom.xml
index 1096951..7ec45b4 100644
--- a/index/bloom/pom.xml
+++ b/index/bloom/pom.xml
@@ -4,7 +4,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/index/examples/pom.xml b/index/examples/pom.xml
index 3522097..bccbae1 100644
--- a/index/examples/pom.xml
+++ b/index/examples/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/index/lucene/pom.xml b/index/lucene/pom.xml
index 88d8afb..e23b94d 100644
--- a/index/lucene/pom.xml
+++ b/index/lucene/pom.xml
@@ -4,7 +4,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/index/secondary-index/pom.xml b/index/secondary-index/pom.xml
index f4ce62e..b1a2414 100644
--- a/index/secondary-index/pom.xml
+++ b/index/secondary-index/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/integration/flink-build/pom.xml b/integration/flink-build/pom.xml
index ba37c2e..a4b320e 100644
--- a/integration/flink-build/pom.xml
+++ b/integration/flink-build/pom.xml
@@ -4,7 +4,7 @@
 
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
 
 
diff --git a/integration/flink-proxy/pom.xml b/integration/flink-proxy/pom.xml
index 832d804..b7aee50 100644
--- a/integration/flink-proxy/pom.xml
+++ b/integration/flink-proxy

[carbondata] branch master updated (18840af -> 861fc67)

2021-12-20 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git.


from 18840af  [CARBONDATA-4305] Support Carbondata Streamer tool for 
incremental fetch and merge from kafka and DFS Sources
 add fc3914f  [maven-release-plugin] prepare release 
apache-carbondata-2.3.0-rc1
 new 861fc67  [maven-release-plugin] prepare for next development iteration

The 1 revisions listed above as "new" are entirely new to this
repository and will be described in separate emails.  The revisions
listed as "add" were already present in the repository and have only
been added to this reference.


Summary of changes:
 assembly/pom.xml| 2 +-
 common/pom.xml  | 2 +-
 core/pom.xml| 2 +-
 examples/flink/pom.xml  | 2 +-
 examples/spark/pom.xml  | 2 +-
 format/pom.xml  | 2 +-
 geo/pom.xml | 2 +-
 hadoop/pom.xml  | 2 +-
 index/bloom/pom.xml | 2 +-
 index/examples/pom.xml  | 2 +-
 index/lucene/pom.xml| 2 +-
 index/secondary-index/pom.xml   | 2 +-
 integration/flink-build/pom.xml | 2 +-
 integration/flink-proxy/pom.xml | 2 +-
 integration/flink/pom.xml   | 2 +-
 integration/hive/pom.xml| 2 +-
 integration/presto/pom.xml  | 2 +-
 integration/spark/pom.xml   | 2 +-
 mv/plan/pom.xml | 2 +-
 pom.xml | 2 +-
 processing/pom.xml  | 2 +-
 sdk/sdk/pom.xml | 2 +-
 streaming/pom.xml   | 2 +-
 tools/cli/pom.xml   | 2 +-
 24 files changed, 24 insertions(+), 24 deletions(-)


[carbondata] annotated tag apache-carbondata-2.3.0-rc1 created (now 9038617)

2021-12-20 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a change to annotated tag apache-carbondata-2.3.0-rc1
in repository https://gitbox.apache.org/repos/asf/carbondata.git.


  at 9038617  (tag)
 tagging fc3914f3646177e1f0bccb6688be4688b5aaffd9 (commit)
  by kunal642
  on Mon Dec 20 15:15:03 2021 +0530

- Log -
[maven-release-plugin] copy for tag apache-carbondata-2.3.0-rc1
---

This annotated tag includes the following new commits:

 new fc3914f  [maven-release-plugin] prepare release 
apache-carbondata-2.3.0-rc1

The 1 revisions listed above as "new" are entirely new to this
repository and will be described in separate emails.  The revisions
listed as "add" were already present in the repository and have only
been added to this reference.



[carbondata] 01/01: [maven-release-plugin] prepare release apache-carbondata-2.3.0-rc1

2021-12-20 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to annotated tag apache-carbondata-2.3.0-rc1
in repository https://gitbox.apache.org/repos/asf/carbondata.git

commit fc3914f3646177e1f0bccb6688be4688b5aaffd9
Author: kunal642 
AuthorDate: Mon Dec 20 15:11:10 2021 +0530

[maven-release-plugin] prepare release apache-carbondata-2.3.0-rc1
---
 assembly/pom.xml| 2 +-
 common/pom.xml  | 2 +-
 core/pom.xml| 2 +-
 examples/flink/pom.xml  | 2 +-
 examples/spark/pom.xml  | 2 +-
 format/pom.xml  | 2 +-
 geo/pom.xml | 2 +-
 hadoop/pom.xml  | 2 +-
 index/bloom/pom.xml | 2 +-
 index/examples/pom.xml  | 2 +-
 index/lucene/pom.xml| 2 +-
 index/secondary-index/pom.xml   | 2 +-
 integration/flink-build/pom.xml | 2 +-
 integration/flink-proxy/pom.xml | 2 +-
 integration/flink/pom.xml   | 2 +-
 integration/hive/pom.xml| 2 +-
 integration/presto/pom.xml  | 2 +-
 integration/spark/pom.xml   | 2 +-
 mv/plan/pom.xml | 2 +-
 pom.xml | 4 ++--
 processing/pom.xml  | 2 +-
 sdk/sdk/pom.xml | 2 +-
 streaming/pom.xml   | 2 +-
 tools/cli/pom.xml   | 2 +-
 24 files changed, 25 insertions(+), 25 deletions(-)

diff --git a/assembly/pom.xml b/assembly/pom.xml
index d4c7aba..1107569 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/common/pom.xml b/common/pom.xml
index 68fea9c..88dc1b8 100644
--- a/common/pom.xml
+++ b/common/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/core/pom.xml b/core/pom.xml
index f4d2c16..0e74e7a 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/examples/flink/pom.xml b/examples/flink/pom.xml
index cd4b691..da34787 100644
--- a/examples/flink/pom.xml
+++ b/examples/flink/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/examples/spark/pom.xml b/examples/spark/pom.xml
index 0f804f4..ca666db 100644
--- a/examples/spark/pom.xml
+++ b/examples/spark/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/format/pom.xml b/format/pom.xml
index c547d37..a8c5b60 100644
--- a/format/pom.xml
+++ b/format/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/geo/pom.xml b/geo/pom.xml
index fcbd80c..83bbbd9 100644
--- a/geo/pom.xml
+++ b/geo/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/hadoop/pom.xml b/hadoop/pom.xml
index e8be41e..298337b 100644
--- a/hadoop/pom.xml
+++ b/hadoop/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/index/bloom/pom.xml b/index/bloom/pom.xml
index c4071ba..1096951 100644
--- a/index/bloom/pom.xml
+++ b/index/bloom/pom.xml
@@ -4,7 +4,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/index/examples/pom.xml b/index/examples/pom.xml
index 3873996..3522097 100644
--- a/index/examples/pom.xml
+++ b/index/examples/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/index/lucene/pom.xml b/index/lucene/pom.xml
index b10583d..88d8afb 100644
--- a/index/lucene/pom.xml
+++ b/index/lucene/pom.xml
@@ -4,7 +4,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/index/secondary-index/pom.xml b/index/secondary-index/pom.xml
index b998143..f4ce62e 100644
--- a/index/secondary-index/pom.xml
+++ b/index/secondary-index/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/integration/flink-build/pom.xml b/integration/flink-build/pom.xml
index c443497..ba37c2e 100644
--- a/integration/flink-build/pom.xml
+++ b/integration/flink-build/pom.xml
@@ -4,7 +4,7 @@
 
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
 
 
diff --git a/integration/flink-proxy/pom.xml b/integration/flink-proxy/pom.xml
index 23f5169..832d804 100644
--- a/integration/flink-proxy/pom.xml

[carbondata] annotated tag apache-carbondata-2.3.0-rc1 deleted (was a4a61a8)

2021-12-20 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a change to annotated tag apache-carbondata-2.3.0-rc1
in repository https://gitbox.apache.org/repos/asf/carbondata.git.


*** WARNING: tag apache-carbondata-2.3.0-rc1 was deleted! ***

   tag was  a4a61a8

This change permanently discards the following revisions:

 discard 232cbaa  [maven-release-plugin] prepare release 
apache-carbondata-2.3.0-rc1


[carbondata] branch branch-2.3 deleted (was b572b91)

2021-12-20 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a change to branch branch-2.3
in repository https://gitbox.apache.org/repos/asf/carbondata.git.


 was b572b91  [maven-release-plugin] prepare for next development iteration

This change permanently discards the following revisions:

 discard b572b91  [maven-release-plugin] prepare for next development iteration


[carbondata] 01/01: [maven-release-plugin] prepare for next development iteration

2021-12-20 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch branch-2.3
in repository https://gitbox.apache.org/repos/asf/carbondata.git

commit b572b919ef6eef30e5ca367deb714cdd66438721
Author: kunal642 
AuthorDate: Mon Dec 20 14:17:28 2021 +0530

[maven-release-plugin] prepare for next development iteration
---
 assembly/pom.xml| 2 +-
 common/pom.xml  | 2 +-
 core/pom.xml| 2 +-
 examples/flink/pom.xml  | 2 +-
 examples/spark/pom.xml  | 2 +-
 format/pom.xml  | 2 +-
 geo/pom.xml | 2 +-
 hadoop/pom.xml  | 2 +-
 index/bloom/pom.xml | 2 +-
 index/examples/pom.xml  | 2 +-
 index/lucene/pom.xml| 2 +-
 index/secondary-index/pom.xml   | 2 +-
 integration/flink-build/pom.xml | 2 +-
 integration/flink-proxy/pom.xml | 2 +-
 integration/flink/pom.xml   | 2 +-
 integration/hive/pom.xml| 2 +-
 integration/presto/pom.xml  | 2 +-
 integration/spark/pom.xml   | 2 +-
 mv/plan/pom.xml | 2 +-
 pom.xml | 4 ++--
 processing/pom.xml  | 2 +-
 sdk/sdk/pom.xml | 2 +-
 streaming/pom.xml   | 2 +-
 tools/cli/pom.xml   | 2 +-
 24 files changed, 25 insertions(+), 25 deletions(-)

diff --git a/assembly/pom.xml b/assembly/pom.xml
index 1107569..d8d03b7 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/common/pom.xml b/common/pom.xml
index 88dc1b8..6205781 100644
--- a/common/pom.xml
+++ b/common/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/core/pom.xml b/core/pom.xml
index 0e74e7a..a5f2e75 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/examples/flink/pom.xml b/examples/flink/pom.xml
index da34787..360fcff 100644
--- a/examples/flink/pom.xml
+++ b/examples/flink/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/examples/spark/pom.xml b/examples/spark/pom.xml
index ca666db..64f0d9f 100644
--- a/examples/spark/pom.xml
+++ b/examples/spark/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/format/pom.xml b/format/pom.xml
index a8c5b60..e992b2e 100644
--- a/format/pom.xml
+++ b/format/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/geo/pom.xml b/geo/pom.xml
index 83bbbd9..46f6695 100644
--- a/geo/pom.xml
+++ b/geo/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/hadoop/pom.xml b/hadoop/pom.xml
index 298337b..c38b0f9 100644
--- a/hadoop/pom.xml
+++ b/hadoop/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../pom.xml
   
 
diff --git a/index/bloom/pom.xml b/index/bloom/pom.xml
index 1096951..7ec45b4 100644
--- a/index/bloom/pom.xml
+++ b/index/bloom/pom.xml
@@ -4,7 +4,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/index/examples/pom.xml b/index/examples/pom.xml
index 3522097..bccbae1 100644
--- a/index/examples/pom.xml
+++ b/index/examples/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/index/lucene/pom.xml b/index/lucene/pom.xml
index 88d8afb..e23b94d 100644
--- a/index/lucene/pom.xml
+++ b/index/lucene/pom.xml
@@ -4,7 +4,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/index/secondary-index/pom.xml b/index/secondary-index/pom.xml
index f4ce62e..b1a2414 100644
--- a/index/secondary-index/pom.xml
+++ b/index/secondary-index/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/integration/flink-build/pom.xml b/integration/flink-build/pom.xml
index ba37c2e..a4b320e 100644
--- a/integration/flink-build/pom.xml
+++ b/integration/flink-build/pom.xml
@@ -4,7 +4,7 @@
 
 org.apache.carbondata
 carbondata-parent
-2.3.0
+2.3.1-SNAPSHOT
 ../../pom.xml
 
 
diff --git a/integration/flink-proxy/pom.xml b/integration/flink-proxy/pom.xml
index 832d804..b7aee50 100644
--- a/integration/flink-proxy/pom.xml
+++ b/integration/flink-proxy

[carbondata] branch branch-2.3 created (now b572b91)

2021-12-20 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a change to branch branch-2.3
in repository https://gitbox.apache.org/repos/asf/carbondata.git.


  at b572b91  [maven-release-plugin] prepare for next development iteration

This branch includes the following new commits:

 new b572b91  [maven-release-plugin] prepare for next development iteration

The 1 revisions listed above as "new" are entirely new to this
repository and will be described in separate emails.  The revisions
listed as "add" were already present in the repository and have only
been added to this reference.



[carbondata] 01/01: [maven-release-plugin] prepare release apache-carbondata-2.3.0-rc1

2021-12-20 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to annotated tag apache-carbondata-2.3.0-rc1
in repository https://gitbox.apache.org/repos/asf/carbondata.git

commit 232cbaae2b3d50450478e458b580195e9f830a9c
Author: kunal642 
AuthorDate: Mon Dec 20 14:15:36 2021 +0530

[maven-release-plugin] prepare release apache-carbondata-2.3.0-rc1
---
 assembly/pom.xml| 2 +-
 common/pom.xml  | 2 +-
 core/pom.xml| 2 +-
 examples/flink/pom.xml  | 2 +-
 examples/spark/pom.xml  | 2 +-
 format/pom.xml  | 2 +-
 geo/pom.xml | 2 +-
 hadoop/pom.xml  | 2 +-
 index/bloom/pom.xml | 2 +-
 index/examples/pom.xml  | 2 +-
 index/lucene/pom.xml| 2 +-
 index/secondary-index/pom.xml   | 2 +-
 integration/flink-build/pom.xml | 2 +-
 integration/flink-proxy/pom.xml | 2 +-
 integration/flink/pom.xml   | 2 +-
 integration/hive/pom.xml| 2 +-
 integration/presto/pom.xml  | 2 +-
 integration/spark/pom.xml   | 2 +-
 mv/plan/pom.xml | 2 +-
 pom.xml | 4 ++--
 processing/pom.xml  | 2 +-
 sdk/sdk/pom.xml | 2 +-
 streaming/pom.xml   | 2 +-
 tools/cli/pom.xml   | 2 +-
 24 files changed, 25 insertions(+), 25 deletions(-)

diff --git a/assembly/pom.xml b/assembly/pom.xml
index d4c7aba..1107569 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/common/pom.xml b/common/pom.xml
index 68fea9c..88dc1b8 100644
--- a/common/pom.xml
+++ b/common/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/core/pom.xml b/core/pom.xml
index f4d2c16..0e74e7a 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/examples/flink/pom.xml b/examples/flink/pom.xml
index cd4b691..da34787 100644
--- a/examples/flink/pom.xml
+++ b/examples/flink/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/examples/spark/pom.xml b/examples/spark/pom.xml
index 0f804f4..ca666db 100644
--- a/examples/spark/pom.xml
+++ b/examples/spark/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/format/pom.xml b/format/pom.xml
index c547d37..a8c5b60 100644
--- a/format/pom.xml
+++ b/format/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/geo/pom.xml b/geo/pom.xml
index fcbd80c..83bbbd9 100644
--- a/geo/pom.xml
+++ b/geo/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/hadoop/pom.xml b/hadoop/pom.xml
index e8be41e..298337b 100644
--- a/hadoop/pom.xml
+++ b/hadoop/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../pom.xml
   
 
diff --git a/index/bloom/pom.xml b/index/bloom/pom.xml
index c4071ba..1096951 100644
--- a/index/bloom/pom.xml
+++ b/index/bloom/pom.xml
@@ -4,7 +4,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/index/examples/pom.xml b/index/examples/pom.xml
index 3873996..3522097 100644
--- a/index/examples/pom.xml
+++ b/index/examples/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/index/lucene/pom.xml b/index/lucene/pom.xml
index b10583d..88d8afb 100644
--- a/index/lucene/pom.xml
+++ b/index/lucene/pom.xml
@@ -4,7 +4,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/index/secondary-index/pom.xml b/index/secondary-index/pom.xml
index b998143..f4ce62e 100644
--- a/index/secondary-index/pom.xml
+++ b/index/secondary-index/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
   
 
diff --git a/integration/flink-build/pom.xml b/integration/flink-build/pom.xml
index c443497..ba37c2e 100644
--- a/integration/flink-build/pom.xml
+++ b/integration/flink-build/pom.xml
@@ -4,7 +4,7 @@
 
 org.apache.carbondata
 carbondata-parent
-2.3.0-SNAPSHOT
+2.3.0
 ../../pom.xml
 
 
diff --git a/integration/flink-proxy/pom.xml b/integration/flink-proxy/pom.xml
index 23f5169..832d804 100644
--- a/integration/flink-proxy/pom.xml

[carbondata] annotated tag apache-carbondata-2.3.0-rc1 created (now a4a61a8)

2021-12-20 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a change to annotated tag apache-carbondata-2.3.0-rc1
in repository https://gitbox.apache.org/repos/asf/carbondata.git.


  at a4a61a8  (tag)
 tagging 232cbaae2b3d50450478e458b580195e9f830a9c (commit)
  by kunal642
  on Mon Dec 20 14:17:05 2021 +0530

- Log -
[maven-release-plugin] copy for tag apache-carbondata-2.3.0-rc1
---

This annotated tag includes the following new commits:

 new 232cbaa  [maven-release-plugin] prepare release 
apache-carbondata-2.3.0-rc1

The 1 revisions listed above as "new" are entirely new to this
repository and will be described in separate emails.  The revisions
listed as "add" were already present in the repository and have only
been added to this reference.



[carbondata] branch master updated: [CARBONDATA-4305] Support Carbondata Streamer tool for incremental fetch and merge from kafka and DFS Sources

2021-11-25 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new 18840af  [CARBONDATA-4305] Support Carbondata Streamer tool for 
incremental fetch and merge from kafka and DFS Sources
18840af is described below

commit 18840af9c1f7154b58e3c397dfc5a4440674bcee
Author: akashrn5 
AuthorDate: Wed Sep 1 19:24:24 2021 +0530

[CARBONDATA-4305] Support Carbondata Streamer tool for incremental fetch 
and merge from kafka and DFS Sources

Why is this PR needed?
In the current Carbondata CDC solution, if any user wants to integrate it 
with a streaming source then he
need to write a separate spark application to capture changes which is an 
overhead. We should be able to
incrementally capture the data changes from primary databases and should be 
able to incrementally ingest
the same in the data lake so that the overall latency decreases. The former 
is taken care of using
log-based CDC systems like Maxwell and Debezium. Here is a solution for the 
second aspect using Apache Carbondata.

What changes were proposed in this PR?
Carbondata streamer tool is a spark streaming application which enables 
users to incrementally ingest data
from various sources, like Kafka(standard pipeline would be like MYSQL => 
debezium => (kafka + Schema registry) => Carbondata Streamer tool)
and DFS into their data lakes. The tool comes with out-of-the-box support 
for almost all types of schema
evolution use cases. With the streamer tool only add column support is 
given with drop column and
other schema changes capability in line in the upcoming days. Please refer 
to design document for
more details about usage and working of the tool.

This closes #4235
---
 .../core/constants/CarbonCommonConstants.java  | 162 +++-
 .../carbondata/core/util/CarbonProperties.java |   4 +
 integration/spark/pom.xml  |  47 
 .../apache/carbondata/streamer/AvroDFSSource.scala |  58 +
 .../carbondata/streamer/AvroKafkaSource.scala  |  64 +
 .../apache/carbondata/streamer/CarbonDStream.scala |  56 
 .../carbondata/streamer/CarbonDataStreamer.scala   | 135 ++
 .../streamer/CarbonDataStreamerException.scala |  33 +++
 .../carbondata/streamer/CarbonStreamerConfig.scala | 282 +
 .../apache/carbondata/streamer/SchemaSource.scala  | 106 
 .../org/apache/carbondata/streamer/Source.scala| 235 +
 .../apache/carbondata/streamer/SourceFactory.scala |  54 
 .../mutation/merge/CarbonMergeDataSetCommand.scala |   5 +
 .../spark/testsuite/merge/MergeTestCase.scala  |  68 +
 pom.xml|   4 +
 15 files changed, 1312 insertions(+), 1 deletion(-)

diff --git 
a/core/src/main/java/org/apache/carbondata/core/constants/CarbonCommonConstants.java
 
b/core/src/main/java/org/apache/carbondata/core/constants/CarbonCommonConstants.java
index f24108a..919fc90 100644
--- 
a/core/src/main/java/org/apache/carbondata/core/constants/CarbonCommonConstants.java
+++ 
b/core/src/main/java/org/apache/carbondata/core/constants/CarbonCommonConstants.java
@@ -2684,6 +2684,150 @@ public final class CarbonCommonConstants {
   
//
   // CDC streamer configs start here
   
//
+  /**
+   * The database name where the target table is present to merge the incoming 
data. If not given by
+   * user, system will take the current database in the spark session.
+   */
+  @CarbonProperty
+  public static final String CARBON_STREAMER_DATABASE_NAME = 
"carbon.streamer.target.database";
+
+  /**
+   * The target carbondata table where the data has to be merged. If this is 
not configured by user,
+   * the operation will fail.
+   */
+  @CarbonProperty
+  public static final String CARBON_STREAMER_TABLE_NAME = 
"carbon.streamer.target.table";
+
+  /**
+   * Source type to ingest data from. It can be kafka or DFS
+   */
+  @CarbonProperty
+  public static final String CARBON_STREAMER_SOURCE_TYPE = 
"carbon.streamer.source.type";
+
+  public static final String CARBON_STREAMER_SOURCE_TYPE_DEFAULT = "kafka";
+
+  /**
+   * An absolute path on a given file system from where data needs to be read 
to ingest into the
+   * target carbondata table. Mandatory if the ingestion source type is DFS.
+   */
+  @CarbonProperty
+  public static final String CARBON_STREAMER_DFS_INPUT_PATH = 
"carbon.streamer.dfs.input.path";
+
+  /**
+   * Schema registry url in case schema registry is selected as schema 
provider.
+   */
+  @CarbonProperty
+  pub

[carbondata] branch master updated: [CARBONDATA-4296]: schema evolution, enforcement and deduplication utilities added

2021-11-15 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new 3be05d2  [CARBONDATA-4296]: schema evolution, enforcement and 
deduplication utilities added
3be05d2 is described below

commit 3be05d2a44d805cf763df05cbeacce2d90a44da0
Author: pratyakshsharma 
AuthorDate: Wed Oct 27 13:54:37 2021 +0530

[CARBONDATA-4296]: schema evolution, enforcement and deduplication 
utilities added

Why is this PR needed?
This PR adds schema enforcement, schema evolution and deduplication 
capabilities for
carbondata streamer tool specifically. For the existing IUD scenarios, some 
work
needs to be done to handle it completely, for example -
1. passing default values and storing them in table properties.

Changes proposed for the phase 2 -
1. Handling delete use cases with upsert operation/command itself. Right 
now we
consider update as delete + insert. With the new streamer tool, it is 
possible that
user sets upsert as the operation type and incoming stream has delete 
records as well.
What changes were proposed in this PR?

Configs and utility methods are added for the following use cases -
1. Schema enforcement
2. Schema evolution - add column, delete column, data type change scenario
3. Deduplicate the incoming dataset against incoming dataset itself. This 
is useful
in scenarios where incoming stream of data has multiple updates for the 
same record
and we want to pick the latest.
4. Deduplicate the incoming dataset against existing target dataset. This 
is useful
when operation type is set as INSERT and user does not want to insert 
duplicate records.

This closes #4227
---
 .../exceptions/sql/CarbonSchemaException.java  |  39 +++
 .../core/constants/CarbonCommonConstants.java  |  54 +++
 .../mutation/merge/CarbonMergeDataSetCommand.scala |  80 -
 .../mutation/merge/CarbonMergeDataSetUtil.scala| 390 -
 .../spark/sql/execution/strategy/DDLHelper.scala   |  97 +++--
 .../spark/testsuite/merge/MergeTestCase.scala  | 264 +++---
 6 files changed, 824 insertions(+), 100 deletions(-)

diff --git 
a/common/src/main/java/org/apache/carbondata/common/exceptions/sql/CarbonSchemaException.java
 
b/common/src/main/java/org/apache/carbondata/common/exceptions/sql/CarbonSchemaException.java
new file mode 100644
index 000..522f132
--- /dev/null
+++ 
b/common/src/main/java/org/apache/carbondata/common/exceptions/sql/CarbonSchemaException.java
@@ -0,0 +1,39 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements.  See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License.  You may obtain a copy of the License at
+ *
+ *http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.carbondata.common.exceptions.sql;
+
+import org.apache.carbondata.common.annotations.InterfaceAudience;
+import org.apache.carbondata.common.annotations.InterfaceStability;
+
+@InterfaceAudience.User
+@InterfaceStability.Stable
+public class CarbonSchemaException extends Exception {
+
+  private static final long serialVersionUID = 1L;
+
+  private final String message;
+
+  public CarbonSchemaException(String message) {
+super(message);
+this.message = message;
+  }
+
+  public String getMessage() {
+return this.message;
+  }
+}
diff --git 
a/core/src/main/java/org/apache/carbondata/core/constants/CarbonCommonConstants.java
 
b/core/src/main/java/org/apache/carbondata/core/constants/CarbonCommonConstants.java
index d72d6c1..f24108a 100644
--- 
a/core/src/main/java/org/apache/carbondata/core/constants/CarbonCommonConstants.java
+++ 
b/core/src/main/java/org/apache/carbondata/core/constants/CarbonCommonConstants.java
@@ -2681,4 +2681,58 @@ public final class CarbonCommonConstants {
 
   public static final String CARBON_CDC_MINMAX_PRUNING_ENABLED_DEFAULT = 
"false";
 
+  
//
+  // CDC streamer configs start here
+  
//
+
+  /**
+   * Name of the field from source schema whose value can be used f

[carbondata] branch master updated: [CARBONDATA-4306] Fix Query Performance issue for Spark 3.1

2021-10-23 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new 8953cde  [CARBONDATA-4306] Fix Query Performance issue for Spark 3.1
8953cde is described below

commit 8953cde6ad93ddd3622c60cd48117acfbbee56a2
Author: Indhumathi27 
AuthorDate: Thu Sep 30 14:41:35 2021 +0530

[CARBONDATA-4306] Fix Query Performance issue for Spark 3.1

Why is this PR needed?
Currently, with Spark 3.1, some rules are applied many times resulting in 
performance degrade.

What changes were proposed in this PR?
Changed Rules apply strategy from Fixed to Once and CarbonOptimizer can 
directly extend SparkOptimizer avoiding applying same rules many times

This Closes #4229
---
 .../execution/strategy/CarbonSourceStrategy.scala  |  2 +-
 .../apache/spark/sql/CarbonToSparkAdapter.scala| 23 ++
 2 files changed, 7 insertions(+), 18 deletions(-)

diff --git 
a/integration/spark/src/main/scala/org/apache/spark/sql/execution/strategy/CarbonSourceStrategy.scala
 
b/integration/spark/src/main/scala/org/apache/spark/sql/execution/strategy/CarbonSourceStrategy.scala
index 009d8c7..8a0779f 100644
--- 
a/integration/spark/src/main/scala/org/apache/spark/sql/execution/strategy/CarbonSourceStrategy.scala
+++ 
b/integration/spark/src/main/scala/org/apache/spark/sql/execution/strategy/CarbonSourceStrategy.scala
@@ -152,7 +152,7 @@ private[sql] object CarbonSourceStrategy extends 
SparkStrategy {
 var partitions : (Seq[CatalogTablePartition], Seq[PartitionSpec], 
Seq[Expression]) =
   (null, null, Seq.empty)
 var filterPredicates = allPredicates
-if(names.nonEmpty) {
+if(names.nonEmpty && partitionsFilter.nonEmpty) {
   partitions = CarbonFilters.getCatalogTablePartitions(
 partitionsFilter.filterNot(e => 
e.find(_.isInstanceOf[PlanExpression[_]]).isDefined),
 SparkSession.getActiveSession.get,
diff --git 
a/integration/spark/src/main/spark3.1/org/apache/spark/sql/CarbonToSparkAdapter.scala
 
b/integration/spark/src/main/spark3.1/org/apache/spark/sql/CarbonToSparkAdapter.scala
index 25a27ad..eaceb85 100644
--- 
a/integration/spark/src/main/spark3.1/org/apache/spark/sql/CarbonToSparkAdapter.scala
+++ 
b/integration/spark/src/main/spark3.1/org/apache/spark/sql/CarbonToSparkAdapter.scala
@@ -33,6 +33,7 @@ import 
org.apache.spark.sql.catalyst.expressions.codegen.Block._
 import org.apache.spark.sql.catalyst.optimizer.Optimizer
 import org.apache.spark.sql.catalyst.plans.logical.{LogicalPlan, SubqueryAlias}
 import org.apache.spark.sql.catalyst.rules.Rule
+import org.apache.spark.sql.execution.SparkOptimizer
 import org.apache.spark.sql.execution.datasources.{FilePartition, 
PartitionedFile}
 import org.apache.spark.sql.hive.HiveExternalCatalog
 import org.apache.spark.sql.optimizer.{CarbonIUDRule, CarbonUDFTransformRule, 
MVRewriteRule}
@@ -238,33 +239,21 @@ object CarbonToSparkAdapter extends SparkVersionAdapter {
 }
 
 class CarbonOptimizer(session: SparkSession, optimizer: Optimizer) extends
-  Optimizer(session.sessionState.catalogManager) {
+  SparkOptimizer(session.sessionState.catalogManager,
+session.sessionState.catalog,
+session.sessionState.experimentalMethods) {
 
   private lazy val mvRules = Seq(Batch("Materialized View Optimizers", Once,
 Seq(new MVRewriteRule(session)): _*))
 
-  private lazy val iudRule = Batch("IUD Optimizers", fixedPoint,
+  private lazy val iudRule = Batch("IUD Optimizers", Once,
 Seq(new CarbonIUDRule(), new CarbonUDFTransformRule(), new 
CarbonFileIndexReplaceRule()): _*)
 
   private lazy val secondaryIndexRule = Batch("SI Optimizers", Once,
 Seq(new CarbonSITransformationRule(session)): _*)
 
   override def defaultBatches: Seq[Batch] = {
-mvRules ++ convertedBatch() :+ iudRule :+ secondaryIndexRule
+mvRules ++ super.defaultBatches :+ iudRule :+ secondaryIndexRule
   }
 
-  def convertedBatch(): Seq[Batch] = {
-optimizer.batches.map { batch =>
-  Batch(
-batch.name,
-batch.strategy match {
-  case optimizer.Once =>
-Once
-  case _: optimizer.FixedPoint =>
-fixedPoint
-},
-batch.rules: _*
-  )
-}
-  }
 }


[carbondata] branch master updated: [CARBONDATA-4228] [CARBONDATA-4203] Fixed update/delete after alter add segment

2021-10-07 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new bca62cd  [CARBONDATA-4228] [CARBONDATA-4203] Fixed update/delete after 
alter add segment
bca62cd is described below

commit bca62cda806302fe10076c8d2f50e6fad95b147a
Author: nihal0107 
AuthorDate: Tue Sep 21 18:14:21 2021 +0530

[CARBONDATA-4228] [CARBONDATA-4203] Fixed update/delete after alter add 
segment

Why is this PR needed?
Deleted records are reappearing or updated records are showing old values 
in select
queries. It is because after horizontal compaction delete delta file for 
the external
segment is written to the default path which is Fact\part0\segment_x\ while 
if the
segment is an external segment then delete delta file should be written to 
the path
where the segment is present.

What changes were proposed in this PR?
After delete/update operation on the segment, horizontal compaction will be 
triggered.
Now after horizontal compaction for external segments, the delete delta 
file will be
written to the segment path at the place of the default path.

This closes #4220
---
 .../statusmanager/SegmentUpdateStatusManager.java  | 17 +--
 .../testsuite/addsegment/AddSegmentTestCase.scala  | 33 ++
 2 files changed, 47 insertions(+), 3 deletions(-)

diff --git 
a/core/src/main/java/org/apache/carbondata/core/statusmanager/SegmentUpdateStatusManager.java
 
b/core/src/main/java/org/apache/carbondata/core/statusmanager/SegmentUpdateStatusManager.java
index 31e253b..fe40494 100644
--- 
a/core/src/main/java/org/apache/carbondata/core/statusmanager/SegmentUpdateStatusManager.java
+++ 
b/core/src/main/java/org/apache/carbondata/core/statusmanager/SegmentUpdateStatusManager.java
@@ -366,8 +366,18 @@ public class SegmentUpdateStatusManager {
*/
   public List getDeleteDeltaFilesList(final Segment segment, final 
String blockName) {
 List deleteDeltaFileList = new ArrayList<>();
-String segmentPath = CarbonTablePath.getSegmentPath(
-identifier.getTablePath(), segment.getSegmentNo());
+String segmentPath = null;
+if (segment.isExternalSegment()) {
+  for (LoadMetadataDetails details : segmentDetails) {
+if (details.getLoadName().equals(segment.getSegmentNo())) {
+  segmentPath = details.getPath();
+  break;
+}
+  }
+} else {
+  segmentPath = CarbonTablePath.getSegmentPath(
+  identifier.getTablePath(), segment.getSegmentNo());
+}
 
 for (SegmentUpdateDetails block : updateDetails) {
   if ((block.getBlockName().equalsIgnoreCase(blockName)) &&
@@ -375,8 +385,9 @@ public class SegmentUpdateStatusManager {
   !CarbonUpdateUtil.isBlockInvalid(block.getSegmentStatus())) {
 Set deltaFileTimestamps = block.getDeltaFileStamps();
 if (deltaFileTimestamps != null && deltaFileTimestamps.size() > 0) {
+  String finalSegmentPath = segmentPath;
   deltaFileTimestamps.forEach(timestamp -> deleteDeltaFileList.add(
-  CarbonUpdateUtil.getDeleteDeltaFilePath(segmentPath, blockName, 
timestamp)));
+  CarbonUpdateUtil.getDeleteDeltaFilePath(finalSegmentPath, 
blockName, timestamp)));
 } else {
   // when the deltaFileTimestamps is null, then there is only one 
delta file
   // and the SegmentUpdateDetails will have same start and end 
timestamp,
diff --git 
a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/addsegment/AddSegmentTestCase.scala
 
b/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/addsegment/AddSegmentTestCase.scala
index 72b8134..5d2134f 100644
--- 
a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/addsegment/AddSegmentTestCase.scala
+++ 
b/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/addsegment/AddSegmentTestCase.scala
@@ -86,6 +86,39 @@ class AddSegmentTestCase extends QueryTest with 
BeforeAndAfterAll {
 FileFactory.deleteAllFilesOfDir(new File(newPath))
   }
 
+  test("test update/delete operation on added segment which required 
horizontal compaction") {
+sql("drop table if exists uniqdata")
+sql("""CREATE TABLE  uniqdata(empname String, designation String, doj 
Timestamp,
+  |  workgroupcategory int, workgroupcategoryname String, deptno int, 
deptname String,
+  |  projectcode int, projectjoindate Timestamp, projectenddate 
Date,attendance int,
+  |  utilization int,salary int, empno int)
+  | STORED AS carbondata""".stripMargin)
+
+sql(s"""LOAD DATA local inpath '$resourcesPath/data.csv' INTO TABLE 
uniqdata
+   | OPTIONS('DELIMITER'= ',', '

[carbondata] branch master updated: [CARBONDATA-4288][CARBONDATA-4289] Fix various issues with Index Server caching mechanism.

2021-09-20 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new ce860d0  [CARBONDATA-4288][CARBONDATA-4289] Fix various issues with 
Index Server caching mechanism.
ce860d0 is described below

commit ce860d0431f41e0594cee848149cdc344f728dd3
Author: Vikram Ahuja 
AuthorDate: Fri Sep 17 12:54:01 2021 +0530

[CARBONDATA-4288][CARBONDATA-4289] Fix various issues with Index Server 
caching mechanism.

Why is this PR needed?
There are 2 issues in the Index Server flow:
In case when there is a main table with a SI table with prepriming disabled 
and index serve
enabled, new load to main table and SI table put the cache for the main 
table in the index
server. Cache is also getting again when a select query is fired. This 
issue happens because
during load to SI table, getSplits is called on the main table segment 
which is in Insert In
Progress state. Index server considers this segment as a legacy segment 
because it's index
size = 0 and does not put it's entry in the tableToExecutor mapping. In the 
getsplits method
isRefreshneeded is false the first time getSplits is called. During the 
select query, in
getSplits method isRefreshNeeded is true and the previous loaded entry is 
removed from the
driver but since there is no entry for that table in tableToExecutor 
mapping, the previous
cache value becomes dead cache and always stays in the index server. The 
newly loaded cache
is loaded to a new executor and 2 copies of cache for the same segment is 
being mantained.
Concurrent select queries to the index server shows wrong cache values in 
the Index server.

What changes were proposed in this PR?
The following changes are proposed to the index server code:
Removing cache object from the index server in case the segment is INSERT 
IN PROGRESS and
in the case of legacy segment adding the value in tabeToExecutor mappping 
so that the cache
is also removed from the executor side.
Concurrent queries were able adding duplicate cache values to other 
executors. Changed logic
of assign executors method so that concurrent queries are not able to add 
cache for same segment
in other executors

This closes #4219
---
 .../indexserver/DistributedPruneRDD.scala  | 11 
 .../indexserver/DistributedRDDUtils.scala  | 76 +++---
 .../indexserver/DistributedRDDUtilsTest.scala  | 26 +++-
 3 files changed, 88 insertions(+), 25 deletions(-)

diff --git 
a/integration/spark/src/main/scala/org/apache/carbondata/indexserver/DistributedPruneRDD.scala
 
b/integration/spark/src/main/scala/org/apache/carbondata/indexserver/DistributedPruneRDD.scala
index dac47ff..bd0b5c9 100644
--- 
a/integration/spark/src/main/scala/org/apache/carbondata/indexserver/DistributedPruneRDD.scala
+++ 
b/integration/spark/src/main/scala/org/apache/carbondata/indexserver/DistributedPruneRDD.scala
@@ -115,6 +115,17 @@ private[indexserver] class DistributedPruneRDD(@transient 
private val ss: SparkS
   // scalastyle:on
   service.shutdownNow()
   val LOGGER = 
LogServiceFactory.getLogService(classOf[DistributedPruneRDD].getName)
+  // remove the cache of Inprogress segments, this case is required during 
the case of
+  // loading to SI. We do get splits of main table in case of Inprogress 
segment. No
+  // need to load it to the cache.
+  val inProgressSegments = 
indexInputFormat.getValidSegments.asScala.collect {
+case seg if DistributedRDDUtils
+  .isSegmentInProgress(indexInputFormat, seg.getSegmentNo) => 
seg.getSegmentNo
+  }
+  if (inProgressSegments.nonEmpty) {
+
IndexStoreManager.getInstance().clearInvalidSegments(indexInputFormat.getCarbonTable,
+  inProgressSegments.asJava)
+  }
   LOGGER.info(s"Time taken to collect ${ inputSplits.size } blocklets : " +
   (System.currentTimeMillis() - startTime))
   val cacheSize = if (CacheProvider.getInstance().getCarbonCache != null) {
diff --git 
a/integration/spark/src/main/scala/org/apache/carbondata/indexserver/DistributedRDDUtils.scala
 
b/integration/spark/src/main/scala/org/apache/carbondata/indexserver/DistributedRDDUtils.scala
index 1994e15..41db8f1 100644
--- 
a/integration/spark/src/main/scala/org/apache/carbondata/indexserver/DistributedRDDUtils.scala
+++ 
b/integration/spark/src/main/scala/org/apache/carbondata/indexserver/DistributedRDDUtils.scala
@@ -32,7 +32,7 @@ import 
org.apache.carbondata.core.indexstore.ExtendedBlockletWrapper
 import org.apache.carbondata.core.metadata.AbsoluteTableIdentifier
 import org.apache.carbondata.core.metadata.schema.table.CarbonTable
 import 
org.apache.carbondata.core.readcommitter.{LatestFilesReadC

[carbondata] branch master updated: [CARBONDATA-4285] Fix alter add complex columns with global sort compaction failure

2021-09-19 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new 22342f8  [CARBONDATA-4285] Fix alter add complex columns with global 
sort compaction failure
22342f8 is described below

commit 22342f847d7db515e5f8c17525522085f49bd2a5
Author: Mahesh Raju Somalaraju 
AuthorDate: Thu Sep 16 22:41:54 2021 +0530

[CARBONDATA-4285] Fix alter add complex columns with global sort compaction 
failure

Why is this PR needed?
Alter add complex columns with global sort compaction is failing due to

AOI exception : Currently creating default complex delimiter list in global 
sort compaction
with size of 3. For map case need extra complex delimiter for handling the 
key-value
bad record handling: When we add complex columns after insert the data, 
complex columns
has null data for previously loaded segments. this null value is going to 
treat as bad
record and compaction is failed.

What changes were proposed in this PR?
In Global sort compaction flow create default complex delimiter with 4, as 
already
doing in load flow.
Bad records handling pruned for compaction case. No need to check bad 
records for
compaction as they are already checked while loading. previously loaded 
segments data
we are inserting again in compaction case

This closes #4218
---
 .../hadoop/api/CarbonTableOutputFormat.java|  9 ++-
 .../spark/load/DataLoadProcessBuilderOnSpark.scala | 12 +++-
 .../spark/load/DataLoadProcessorStepOnSpark.scala  |  5 +-
 .../spark/rdd/CarbonTableCompactor.scala   |  3 +-
 .../alterTable/TestAlterTableAddColumns.scala  | 80 ++
 .../processing/loading/BadRecordsLogger.java   |  9 ++-
 .../loading/BadRecordsLoggerProvider.java  | 12 +++-
 .../loading/converter/impl/RowConverterImpl.java   | 14 ++--
 8 files changed, 128 insertions(+), 16 deletions(-)

diff --git 
a/hadoop/src/main/java/org/apache/carbondata/hadoop/api/CarbonTableOutputFormat.java
 
b/hadoop/src/main/java/org/apache/carbondata/hadoop/api/CarbonTableOutputFormat.java
index ed447a5..12f68d8 100644
--- 
a/hadoop/src/main/java/org/apache/carbondata/hadoop/api/CarbonTableOutputFormat.java
+++ 
b/hadoop/src/main/java/org/apache/carbondata/hadoop/api/CarbonTableOutputFormat.java
@@ -365,11 +365,16 @@ public class CarbonTableOutputFormat extends 
FileOutputFormat 2) {
+if (split.length > 3) {
+  model.setComplexDelimiter(split[1]);
+  model.setComplexDelimiter(split[2]);
+  model.setComplexDelimiter(split[3]);
+} else if (split.length > 2) {
   model.setComplexDelimiter(split[1]);
   model.setComplexDelimiter(split[2]);
 } else if (split.length > 1) {
diff --git 
a/integration/spark/src/main/scala/org/apache/carbondata/spark/load/DataLoadProcessBuilderOnSpark.scala
 
b/integration/spark/src/main/scala/org/apache/carbondata/spark/load/DataLoadProcessBuilderOnSpark.scala
index 83d1890..fe1de5c 100644
--- 
a/integration/spark/src/main/scala/org/apache/carbondata/spark/load/DataLoadProcessBuilderOnSpark.scala
+++ 
b/integration/spark/src/main/scala/org/apache/carbondata/spark/load/DataLoadProcessBuilderOnSpark.scala
@@ -73,7 +73,8 @@ object DataLoadProcessBuilderOnSpark {
   dataFrame: Option[DataFrame],
   model: CarbonLoadModel,
   hadoopConf: Configuration,
-  segmentMetaDataAccumulator: CollectionAccumulator[Map[String, 
SegmentMetaDataInfo]])
+  segmentMetaDataAccumulator: CollectionAccumulator[Map[String, 
SegmentMetaDataInfo]],
+  isCompactionFlow: Boolean = false)
   : Array[(String, (LoadMetadataDetails, ExecutionErrors))] = {
 var isLoadFromCSV = false
 val originRDD = if (dataFrame.isDefined) {
@@ -121,8 +122,13 @@ object DataLoadProcessBuilderOnSpark {
 // 2. Convert
 val convertRDD = inputRDD.mapPartitionsWithIndex { case (index, rows) =>
   ThreadLocalSessionInfo.setConfigurationToCurrentThread(conf.value.value)
-  DataLoadProcessorStepOnSpark.convertFunc(rows, index, modelBroadcast, 
partialSuccessAccum,
-convertStepRowCounter)
+  DataLoadProcessorStepOnSpark.convertFunc(rows,
+index,
+modelBroadcast,
+partialSuccessAccum,
+convertStepRowCounter,
+false,
+isCompactionFlow)
 }.filter(_ != null) // Filter the bad record
 
 // 3. Sort
diff --git 
a/integration/spark/src/main/scala/org/apache/carbondata/spark/load/DataLoadProcessorStepOnSpark.scala
 
b/integration/spark/src/main/scala/org/apache/carbondata/spark/load/DataLoadProcessorStepOnSpark.scala
index 1694579..8f59200 100644
--- 
a/integration/spark/src/main/scala/org/apache/carbondata/spark/load/DataLoadProcessorStepOnSpark.scala
+++ 
b/integration/spark/src/main/scala/org/apache/carbondat

[carbondata] branch master updated: [CARBONDATA-4284] Load/insert after alter add column on partition table with complex column fails

2021-09-16 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new 3b29bcb  [CARBONDATA-4284] Load/insert after alter add column on 
partition table with complex column fails
3b29bcb is described below

commit 3b29bcb1057166c85e05b67a061ec616e809831d
Author: ShreelekhyaG 
AuthorDate: Mon Sep 13 18:44:19 2021 +0530

[CARBONDATA-4284] Load/insert after alter add column on partition table 
with complex column fails

Why is this PR needed?
Insert after alter add column on partition table with complex column fails 
with bufferUnderFlowException
List of columns order in TableSchema is different after alter add column.
Ex: If partition is of dimension type, when table is created the schema 
columns order is as
dimension columns(partition column also) + complex column
After alter add, we are changing the order of columns in schema by moving 
the partition column to last.
complex column + partition column
Due to this change in order, while fillDimensionAndMeasureDetails, the 
indexing is wrong as it
expects complex column to be last always which causes 
bufferUnderFlowException while flattening complex row.

What changes were proposed in this PR?
After alter add, removed changes to add partition column at last.

This closes #4215
---
 .../spark/sql/execution/command/carbonTableSchemaCommon.scala |  5 -
 .../spark/testsuite/alterTable/TestAlterTableAddColumns.scala | 11 +++
 2 files changed, 11 insertions(+), 5 deletions(-)

diff --git 
a/integration/spark/src/main/scala/org/apache/spark/sql/execution/command/carbonTableSchemaCommon.scala
 
b/integration/spark/src/main/scala/org/apache/spark/sql/execution/command/carbonTableSchemaCommon.scala
index 2096e93..dff6ac1 100644
--- 
a/integration/spark/src/main/scala/org/apache/spark/sql/execution/command/carbonTableSchemaCommon.scala
+++ 
b/integration/spark/src/main/scala/org/apache/spark/sql/execution/command/carbonTableSchemaCommon.scala
@@ -387,11 +387,6 @@ class AlterTableColumnSchemaGenerator(
 
 allColumns = CarbonScalaUtil.reArrangeColumnSchema(allColumns)
 
-if (tableInfo.getFactTable.getPartitionInfo != null) {
-  val par = tableInfo.getFactTable.getPartitionInfo.getColumnSchemaList
-  allColumns = allColumns.filterNot(b => par.contains(b)) ++= par.asScala
-}
-
 def getLocalDictColumnList(tableProperties: mutable.Map[String, String],
 columns: mutable.ListBuffer[ColumnSchema]): 
(mutable.ListBuffer[ColumnSchema],
   mutable.ListBuffer[ColumnSchema]) = {
diff --git 
a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/alterTable/TestAlterTableAddColumns.scala
 
b/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/alterTable/TestAlterTableAddColumns.scala
index f5bfb32..bd4a112 100644
--- 
a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/alterTable/TestAlterTableAddColumns.scala
+++ 
b/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/alterTable/TestAlterTableAddColumns.scala
@@ -481,6 +481,17 @@ class TestAlterTableAddColumns extends QueryTest with 
BeforeAndAfterAll {
 sql("DROP TABLE IF EXISTS alter_com")
   }
 
+  test("test add column to partition table with complex column") {
+sql("drop table if exists alter_com")
+sql("create table alter_com(id int, map1 map) " +
+"partitioned by(name string) stored as carbondata")
+sql("insert into alter_com values( 1,map(1,2),'sh')")
+sql("ALTER TABLE alter_com ADD COLUMNS(intF int)")
+sql("insert into alter_com values(1,map(1,2),1,'df')")
+checkAnswer(sql("select * from alter_com"),
+  Seq(Row(1, Map(1 -> 2), null, "sh"), Row(1, Map(1 -> 2), 1, "df")))
+  }
+
   test("Validate default values of complex columns added by alter command") {
 sql("DROP TABLE IF EXISTS alter_com")
 sql("CREATE TABLE alter_com(doubleField double, arr1 array ) STORED 
AS carbondata")


[carbondata] branch master updated: [CARBONDATA-4271] Support DPP for carbon

2021-09-01 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new bdc9484  [CARBONDATA-4271] Support DPP for carbon
bdc9484 is described below

commit bdc9484ac8455e8f53e86367c0e5104364799068
Author: Indhumathi27 
AuthorDate: Tue Jul 13 17:07:39 2021 +0530

[CARBONDATA-4271] Support DPP for carbon

Why is this PR needed?
This PR enables Dynamic Partition Pruning for carbon.

What changes were proposed in this PR?
CarbonDatasourceHadoopRelation has to extend HadoopFsRelation,
because spark has added a check to use DPP only for relation matching 
HadoopFsRelation
Apply Dynamic filter and get runtimePartitions and set this to 
CarbonScanRDD for pruning

This closes #4199
---
 .../spark/sql/CarbonDataSourceScanHelper.scala |   3 +
 .../spark/rdd/CarbonDeltaRowScanRDD.scala  |   4 +-
 .../carbondata/spark/rdd/CarbonScanRDD.scala   |   2 +-
 .../spark/sql/CarbonDatasourceHadoopRelation.scala |  49 -
 .../scala/org/apache/spark/sql/CarbonSource.scala  |  24 ++-
 .../CarbonInsertIntoHadoopFsRelationCommand.scala  |   2 -
 .../table/CarbonCreateTableAsSelectCommand.scala   |   2 +-
 .../execution/strategy/CarbonDataSourceScan.scala  |  26 ++-
 .../sql/execution/strategy/CarbonPlanHelper.scala  |  44 ++--
 .../execution/strategy/CarbonSourceStrategy.scala  |  44 ++--
 .../spark/sql/execution/strategy/DMLStrategy.scala |   5 +-
 .../spark/sql/hive/CarbonFileMetastore.scala   |   2 +-
 .../apache/spark/sql/optimizer/CarbonFilters.scala |   6 +-
 .../spark/sql/CarbonDataSourceScanHelper.scala |  43 +++-
 .../apache/spark/sql/CarbonToSparkAdapter.scala|   8 +-
 .../dblocation/DBLocationCarbonTableTestCase.scala |   2 +-
 .../sql/DynamicPartitionPruningTestCase.scala  | 235 +
 17 files changed, 426 insertions(+), 75 deletions(-)

diff --git 
a/integration/spark/src/main/common2.3and2.4/org/apache/spark/sql/CarbonDataSourceScanHelper.scala
 
b/integration/spark/src/main/common2.3and2.4/org/apache/spark/sql/CarbonDataSourceScanHelper.scala
index 32a2e34..22183de 100644
--- 
a/integration/spark/src/main/common2.3and2.4/org/apache/spark/sql/CarbonDataSourceScanHelper.scala
+++ 
b/integration/spark/src/main/common2.3and2.4/org/apache/spark/sql/CarbonDataSourceScanHelper.scala
@@ -20,6 +20,7 @@ package org.apache.spark.sql
 import org.apache.spark.CarbonInputMetrics
 import org.apache.spark.rdd.RDD
 import org.apache.spark.sql.catalyst.InternalRow
+import org.apache.spark.sql.catalyst.catalog.CatalogTablePartition
 import org.apache.spark.sql.catalyst.expressions.{Attribute, Expression => 
SparkExpression}
 import org.apache.spark.sql.catalyst.plans.QueryPlan
 import org.apache.spark.sql.execution.{ColumnarBatchScan, DataSourceScanExec}
@@ -40,6 +41,8 @@ abstract class  CarbonDataSourceScanHelper(relation: 
CarbonDatasourceHadoopRelat
 pushedDownProjection: CarbonProjection,
 directScanSupport: Boolean,
 extraRDD: Option[(RDD[InternalRow], Boolean)],
+selectedCatalogPartitions: Seq[CatalogTablePartition],
+partitionFilterWithDpp: Seq[SparkExpression],
 segmentIds: Option[String])
   extends DataSourceScanExec with ColumnarBatchScan {
 
diff --git 
a/integration/spark/src/main/scala/org/apache/carbondata/spark/rdd/CarbonDeltaRowScanRDD.scala
 
b/integration/spark/src/main/scala/org/apache/carbondata/spark/rdd/CarbonDeltaRowScanRDD.scala
index de6aeb5..949216c 100644
--- 
a/integration/spark/src/main/scala/org/apache/carbondata/spark/rdd/CarbonDeltaRowScanRDD.scala
+++ 
b/integration/spark/src/main/scala/org/apache/carbondata/spark/rdd/CarbonDeltaRowScanRDD.scala
@@ -44,7 +44,7 @@ class CarbonDeltaRowScanRDD[T: ClassTag](
 @transient private val spark: SparkSession,
 @transient private val serializedTableInfo: Array[Byte],
 @transient private val tableInfo: TableInfo,
-@transient override val partitionNames: Seq[PartitionSpec],
+@transient private val newPartitionNames: Seq[PartitionSpec],
 override val columnProjection: CarbonProjection,
 var filter: IndexFilter,
 identifier: AbsoluteTableIdentifier,
@@ -62,7 +62,7 @@ class CarbonDeltaRowScanRDD[T: ClassTag](
 serializedTableInfo,
 tableInfo,
 inputMetricsStats,
-partitionNames,
+newPartitionNames,
 dataTypeConverterClz,
 readSupportClz) {
   override def internalGetPartitions: Array[Partition] = {
diff --git 
a/integration/spark/src/main/scala/org/apache/carbondata/spark/rdd/CarbonScanRDD.scala
 
b/integration/spark/src/main/scala/org/apache/carbondata/spark/rdd/CarbonScanRDD.scala
index fcb379c..90f1f8b 100644
--- 
a/integration/spark/src/main/scala/org/apache/carbondata/spark/rdd/CarbonScanRDD.scala
+++ 
b/integration/spark/src/main/scala/org/apache/carbondata/spark/rdd/CarbonScanRDD.scala
@@ -8

[carbondata] branch master updated: [CARBONDATA-4274] Fix create partition table error with spark 3.1

2021-08-31 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new ca659b5  [CARBONDATA-4274] Fix create partition table error with spark 
3.1
ca659b5 is described below

commit ca659b5cd3219de7c8a9603784559e20fdff1fda
Author: ShreelekhyaG 
AuthorDate: Thu Aug 19 20:26:00 2021 +0530

[CARBONDATA-4274] Fix create partition table error with spark 3.1

Why is this PR needed?
With spark 3.1, we can create a partition table by giving partition
columns from schema.
Like below example:
create table partitionTable(c1 int, c2 int, v1 string, v2 string)
stored as carbondata partitioned by (v2,c2)

When the table is created by SparkSession with CarbonExtension,
catalog table is created with the specified partitions.
But in cluster/ with carbon session, when we create partition
table with above syntax it is creating normal table with no partitions.

What changes were proposed in this PR?
partitionByStructFields is empty when we directly give partition
column names. So it was not creating a partition table. Made
changes to identify the partition column names and get the struct
field and datatype info from table columns.

This closes #4208
---
 .../org/apache/spark/sql/SparkVersionAdapter.scala | 16 -
 .../spark/sql/parser/CarbonSparkSqlParser.scala| 27 --
 .../StandardPartitionTableQueryTestCase.scala  | 21 +
 3 files changed, 50 insertions(+), 14 deletions(-)

diff --git 
a/integration/spark/src/main/spark3.1/org/apache/spark/sql/SparkVersionAdapter.scala
 
b/integration/spark/src/main/spark3.1/org/apache/spark/sql/SparkVersionAdapter.scala
index 5cf5c15..21d7586 100644
--- 
a/integration/spark/src/main/spark3.1/org/apache/spark/sql/SparkVersionAdapter.scala
+++ 
b/integration/spark/src/main/spark3.1/org/apache/spark/sql/SparkVersionAdapter.scala
@@ -233,17 +233,6 @@ trait SparkVersionAdapter {
 val partitionerFields = partitionByStructFields.map { structField =>
   PartitionerField(structField.name, Some(structField.dataType.toString), 
null)
 }
-// validate partition clause
-if (partitionerFields.nonEmpty) {
-  // partition columns should not be part of the schema
-  val badPartCols = 
partitionerFields.map(_.partitionColumn.toLowerCase).toSet
-.intersect(colNames.map(_.toLowerCase).toSet)
-  if (badPartCols.nonEmpty) {
-operationNotAllowed(s"Partition columns should not be specified in the 
schema: " +
-  badPartCols.map("\"" + _ + "\"").mkString("[", ",", "]")
-  , partitionColumns: PartitionFieldListContext)
-  }
-}
 partitionerFields
   }
 
@@ -280,7 +269,10 @@ trait SparkVersionAdapter {
 val options = new CarbonOption(properties)
 // validate streaming property
 validateStreamingProperty(options)
-var fields = parser.getFields(cols ++ partitionByStructFields)
+// with Spark 3.1, partitioned columns can be already present in schema.
+// Check and remove from fields and add partition columns at last
+val updatedCols = cols.filterNot(x => partitionByStructFields.contains(x))
+var fields = parser.getFields(updatedCols ++ partitionByStructFields)
 // validate for create table as select
 selectQuery match {
   case Some(q) =>
diff --git 
a/integration/spark/src/main/spark3.1/org/apache/spark/sql/parser/CarbonSparkSqlParser.scala
 
b/integration/spark/src/main/spark3.1/org/apache/spark/sql/parser/CarbonSparkSqlParser.scala
index ac0be49..49978d4 100644
--- 
a/integration/spark/src/main/spark3.1/org/apache/spark/sql/parser/CarbonSparkSqlParser.scala
+++ 
b/integration/spark/src/main/spark3.1/org/apache/spark/sql/parser/CarbonSparkSqlParser.scala
@@ -21,6 +21,7 @@ import scala.collection.mutable
 import org.antlr.v4.runtime.tree.TerminalNode
 import org.apache.spark.sql.{CarbonThreadUtil, CarbonToSparkAdapter, 
SparkSession}
 import org.apache.spark.sql.catalyst.parser.{AbstractSqlParser, SqlBaseParser}
+import org.apache.spark.sql.catalyst.parser.ParserUtils.operationNotAllowed
 import org.apache.spark.sql.catalyst.parser.SqlBaseParser._
 import org.apache.spark.sql.catalyst.plans.logical.LogicalPlan
 import org.apache.spark.sql.execution.{SparkSqlAstBuilder, SparkSqlParser}
@@ -130,8 +131,30 @@ class CarbonHelperSqlAstBuilder(conf: SQLConf,
 val tableProperties = convertPropertiesToLowercase(properties)
 
 // validate partition clause
-val partitionByStructFields = Option(partitionColumns).toSeq
-.flatMap(x => visitPartitionFieldList(x)._2)
+// There can be two scenarios for creating partition table with spark 3.1.
+// Scenario 1: create partition columns wit

[carbondata] branch master updated: [CARBONDATA-4204][CARBONDATA-4231] Fix add segment error message, index server failed testcases and dataload fail error on update

2021-07-27 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new c9a5231  [CARBONDATA-4204][CARBONDATA-4231] Fix add segment error 
message, index server failed testcases and dataload fail error on update
c9a5231 is described below

commit c9a5231ab73d64a0966a45a755ff7e9031c0201e
Author: ShreelekhyaG 
AuthorDate: Mon Jun 21 20:44:31 2021 +0530

[CARBONDATA-4204][CARBONDATA-4231] Fix add segment error message,
index server failed testcases and dataload fail error on update

Why is this PR needed?
1. When the path is empty in Carbon add segments then
StringIndexOutOfBoundsException is thrown.
2. Index server UT failures fix.
3. Update fails with dataload fail error if set bad
records action is specified to force with spark 3.1v.

What changes were proposed in this PR?
1. Added check to see if the path is empty and then throw
a valid error message.
2. Used checkAnswer instead of assert in test cases so
that the order of rows returned would be same with or
without index server. Excluded 2 test cases where explain
with query statistics is used, as we are not setting any
pruning info from index server.
3. On update command, dataframe.persist is called and with
latest 3.1 spark changes, spark returns a cloned
SparkSession from cacheManager with all specified
configurations disabled. As now it's using a different
sparkSession for 3.1 which is not initialized in CarbonEnv.
So CarbonEnv.init is called where new CarbonSessionInfo is
created with no sessionParams. So, the properties set were
not accessible. When a new carbonSessionInfo object is
getting created, made changes to set existing sessionparams
from currentThreadSessionInfo.

This closes #4157
---
 .../main/scala/org/apache/spark/sql/CarbonEnv.scala |  1 +
 .../command/management/CarbonAddLoadCommand.scala   |  3 +++
 .../testsuite/addsegment/AddSegmentTestCase.scala   |  9 +
 .../alterTable/TestAlterTableAddColumns.scala   | 21 ++---
 .../createTable/TestRenameTableWithIndex.scala  |  8 ++--
 .../AlterTableColumnRenameTestCase.scala| 19 ---
 6 files changed, 37 insertions(+), 24 deletions(-)

diff --git 
a/integration/spark/src/main/scala/org/apache/spark/sql/CarbonEnv.scala 
b/integration/spark/src/main/scala/org/apache/spark/sql/CarbonEnv.scala
index bb6e2ef..98c18ac 100644
--- a/integration/spark/src/main/scala/org/apache/spark/sql/CarbonEnv.scala
+++ b/integration/spark/src/main/scala/org/apache/spark/sql/CarbonEnv.scala
@@ -119,6 +119,7 @@ class CarbonEnv {
 val threadLevelCarbonSessionInfo = new CarbonSessionInfo()
 if (currentThreadSessionInfo != null) {
   
threadLevelCarbonSessionInfo.setThreadParams(currentThreadSessionInfo.getThreadParams)
+  
threadLevelCarbonSessionInfo.setSessionParams(currentThreadSessionInfo.getSessionParams)
 }
 ThreadLocalSessionInfo.setCarbonSessionInfo(threadLevelCarbonSessionInfo)
 ThreadLocalSessionInfo.setConfigurationToCurrentThread(
diff --git 
a/integration/spark/src/main/scala/org/apache/spark/sql/execution/command/management/CarbonAddLoadCommand.scala
 
b/integration/spark/src/main/scala/org/apache/spark/sql/execution/command/management/CarbonAddLoadCommand.scala
index e319185..ed40d9a 100644
--- 
a/integration/spark/src/main/scala/org/apache/spark/sql/execution/command/management/CarbonAddLoadCommand.scala
+++ 
b/integration/spark/src/main/scala/org/apache/spark/sql/execution/command/management/CarbonAddLoadCommand.scala
@@ -89,6 +89,9 @@ case class CarbonAddLoadCommand(
 
 var givenPath = options.getOrElse(
   "path", throw new UnsupportedOperationException("PATH is mandatory"))
+if (givenPath.length == 0) {
+  throw new UnsupportedOperationException("PATH cannot be empty")
+}
 // remove file separator if already present
 if (givenPath.charAt(givenPath.length - 1) == '/') {
   givenPath = givenPath.substring(0, givenPath.length - 1)
diff --git 
a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/addsegment/AddSegmentTestCase.scala
 
b/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/addsegment/AddSegmentTestCase.scala
index cc5a373..72b8134 100644
--- 
a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/addsegment/AddSegmentTestCase.scala
+++ 
b/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/addsegment/AddSegmentTestCase.scala
@@ -1120,6 +1120,15 @@ class AddSegmentTestCase extends QueryTest with 
BeforeAndAfterAll {
 assert(ex.getMessage.contains("can not add same segment path repeatedly"))
   }
 
+  test("Test add segment with empty 

[carbondata] branch master updated: [CARBONDATA-4210] Handle 3.1 parsing failures related to alter complex types

2021-07-13 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new 02e7723  [CARBONDATA-4210] Handle 3.1 parsing failures related to 
alter complex types
02e7723 is described below

commit 02e77234ebb9c1e18e194c1844aaf2d4a2473dcc
Author: akkio-97 
AuthorDate: Wed Jun 23 13:19:48 2021 +0530

[CARBONDATA-4210] Handle 3.1 parsing failures related to alter complex types

Why is this PR needed?
For 2.3 and 2.4 parsing of alter commands are done by spark. Which is not 
in the case of 3.1.

What changes were proposed in this PR?
So carbon is responsible for the parsing here.
Previously ignored test cases due to this issue are now enabled.

This closes #4162
---
 .../TestSIWithComplexArrayType.scala   | 254 +++---
 .../spark/sql/catalyst/CarbonParserUtil.scala  |  47 +++
 .../spark/sql/parser/CarbonSpark2SqlParser.scala   |  32 +-
 .../sql/parser/CarbonSparkSqlParserUtil.scala  |  29 +-
 .../alterTable/TestAlterTableAddColumns.scala  |   8 +-
 .../AlterTableColumnRenameTestCase.scala   | 387 ++---
 6 files changed, 390 insertions(+), 367 deletions(-)

diff --git 
a/index/secondary-index/src/test/scala/org/apache/carbondata/spark/testsuite/secondaryindex/TestSIWithComplexArrayType.scala
 
b/index/secondary-index/src/test/scala/org/apache/carbondata/spark/testsuite/secondaryindex/TestSIWithComplexArrayType.scala
index 46908b7..3fe1443 100644
--- 
a/index/secondary-index/src/test/scala/org/apache/carbondata/spark/testsuite/secondaryindex/TestSIWithComplexArrayType.scala
+++ 
b/index/secondary-index/src/test/scala/org/apache/carbondata/spark/testsuite/secondaryindex/TestSIWithComplexArrayType.scala
@@ -46,158 +46,142 @@ class TestSIWithComplexArrayType extends QueryTest with 
BeforeAndAfterEach {
   }
 
   test("Test restructured array and existing string column as index 
columns on SI with compaction") {
-// TODO: Support alter chnage column for complex type for SPARK 3.1.1
-// REFER: https://issues.apache.org/jira/browse/CARBONDATA-4210
-if (!sqlContext.sparkContext.version.startsWith("3.1")) {
-  sql("drop table if exists complextable")
-  sql("create table complextable (id string, country array, 
columnName string) stored as carbondata")
-  sql("insert into complextable select 1,array('china', 'us'), 'b'")
-  sql("insert into complextable select 2,array('pak'), 'v'")
+sql("drop table if exists complextable")
+sql("create table complextable (id string, country array, 
columnName string) stored as carbondata")
+sql("insert into complextable select 1,array('china', 'us'), 'b'")
+sql("insert into complextable select 2,array('pak'), 'v'")
 
-  sql("drop index if exists index_11 on complextable")
-  sql("ALTER TABLE complextable ADD COLUMNS(newArray array)")
-  sql("alter table complextable change newArray arr2 array")
-  sql("alter table complextable change columnName name string")
-  sql("insert into complextable select 3,array('china'), 
'f',array('hello','world')")
-  sql("insert into complextable select 
4,array('India'),'g',array('iron','man','jarvis')")
+sql("drop index if exists index_11 on complextable")
+sql("ALTER TABLE complextable ADD COLUMNS(newArray array)")
+sql("alter table complextable change newArray arr2 array")
+sql("alter table complextable change columnName name string")
+sql("insert into complextable select 3,array('china'), 
'f',array('hello','world')")
+sql("insert into complextable select 
4,array('India'),'g',array('iron','man','jarvis')")
 
-  checkAnswer(sql("select * from complextable where 
array_contains(arr2,'iron')"),
-Seq(Row("4", mutable.WrappedArray.make(Array("India")), "g",
-  mutable.WrappedArray.make(Array("iron", "man", "jarvis")
-  val result1 = sql("select * from complextable where 
array_contains(arr2,'iron') and name='g'")
-  val result2 = sql("select * from complextable where arr2[0]='iron' and 
name='f'")
-  sql("create index index_11 on table complextable(arr2, name) as 
'carbondata'")
-  sql("alter table complextable compact 'minor'")
-  val df1 = sql(" select * from complextable where 
array_contains(arr2,'iron') and name='g'")
-  val df2 = sql(" select * from complextable where arr2[0]='iron' and 
name='f'")
-  if (!isFilterPushedDownToSI(df1.queryExecution.sparkPlan)) {
-assert(fa

[carbondata] branch master updated: [HOTFIX]Revert wrong pom changes commit during prepare release process.

2021-07-05 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new c7a3d6d  [HOTFIX]Revert wrong pom changes commit during prepare 
release process.
c7a3d6d is described below

commit c7a3d6d13084c2a399f8a2a7356108c5161e51f3
Author: akashrn5 
AuthorDate: Mon Jul 5 22:00:52 2021 +0530

[HOTFIX]Revert wrong pom changes commit during prepare release process.

Why is this PR needed?
Due to wrong branch release, wrong pom changes are present.

What changes were proposed in this PR?
revert the pom changes.

This closes #4167
---
 assembly/pom.xml| 2 +-
 common/pom.xml  | 2 +-
 core/pom.xml| 2 +-
 examples/flink/pom.xml  | 2 +-
 examples/spark/pom.xml  | 2 +-
 format/pom.xml  | 2 +-
 geo/pom.xml | 2 +-
 hadoop/pom.xml  | 2 +-
 index/bloom/pom.xml | 2 +-
 index/examples/pom.xml  | 2 +-
 index/lucene/pom.xml| 2 +-
 index/secondary-index/pom.xml   | 2 +-
 integration/flink-build/pom.xml | 2 +-
 integration/flink-proxy/pom.xml | 2 +-
 integration/flink/pom.xml   | 2 +-
 integration/hive/pom.xml| 2 +-
 integration/presto/pom.xml  | 2 +-
 integration/spark/pom.xml   | 2 +-
 mv/plan/pom.xml | 2 +-
 pom.xml | 2 +-
 processing/pom.xml  | 2 +-
 sdk/sdk/pom.xml | 2 +-
 streaming/pom.xml   | 2 +-
 tools/cli/pom.xml   | 2 +-
 24 files changed, 24 insertions(+), 24 deletions(-)

diff --git a/assembly/pom.xml b/assembly/pom.xml
index b4054a2..5d96c9b 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.2.1-SNAPSHOT
+2.2.0-SNAPSHOT
 ../pom.xml
   
 
diff --git a/common/pom.xml b/common/pom.xml
index c533990..327708e 100644
--- a/common/pom.xml
+++ b/common/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.2.1-SNAPSHOT
+2.2.0-SNAPSHOT
 ../pom.xml
   
 
diff --git a/core/pom.xml b/core/pom.xml
index e559376..c144b11 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.2.1-SNAPSHOT
+2.2.0-SNAPSHOT
 ../pom.xml
   
 
diff --git a/examples/flink/pom.xml b/examples/flink/pom.xml
index 1a224eb..bce06c3 100644
--- a/examples/flink/pom.xml
+++ b/examples/flink/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.2.1-SNAPSHOT
+2.2.0-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/examples/spark/pom.xml b/examples/spark/pom.xml
index cbdb49e..3a95d84 100644
--- a/examples/spark/pom.xml
+++ b/examples/spark/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.2.1-SNAPSHOT
+2.2.0-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/format/pom.xml b/format/pom.xml
index 709f46f..ccfb03f 100644
--- a/format/pom.xml
+++ b/format/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.2.1-SNAPSHOT
+2.2.0-SNAPSHOT
 ../pom.xml
   
 
diff --git a/geo/pom.xml b/geo/pom.xml
index 5d623c1..f2c882a 100644
--- a/geo/pom.xml
+++ b/geo/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.2.1-SNAPSHOT
+2.2.0-SNAPSHOT
 ../pom.xml
   
 
diff --git a/hadoop/pom.xml b/hadoop/pom.xml
index 799472f..6b04990 100644
--- a/hadoop/pom.xml
+++ b/hadoop/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.2.1-SNAPSHOT
+2.2.0-SNAPSHOT
 ../pom.xml
   
 
diff --git a/index/bloom/pom.xml b/index/bloom/pom.xml
index 70673f7..57007b9 100644
--- a/index/bloom/pom.xml
+++ b/index/bloom/pom.xml
@@ -4,7 +4,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.2.1-SNAPSHOT
+2.2.0-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/index/examples/pom.xml b/index/examples/pom.xml
index 4cad0b7..aedaade 100644
--- a/index/examples/pom.xml
+++ b/index/examples/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.2.1-SNAPSHOT
+2.2.0-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/index/lucene/pom.xml b/index/lucene/pom.xml
index 48ebbb3..0e0baf5 100644
--- a/index/lucene/pom.xml
+++ b/index/lucene/pom.xml
@@ -4,7 +4,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.2.1-SNAPSHOT
+2.2.0-SNAPSHOT
 ../../pom.xml
   
 
diff --git a/index/secondary-index/pom.xml b/index/secondary-index/pom.xml
index 77f12fe..9754bd0 100644
--- a/index/secondary-index/pom.xml
+++ b/index/secondary-index/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.carbondata
 carbondata-parent
-2.2.1-SNAPSHOT
+2.2.0-SNAPSHOT
 ../../pom.xml
   
 
diff --git

[carbondata] branch branch-2.2 deleted (was 952cf82)

2021-07-05 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a change to branch branch-2.2
in repository https://gitbox.apache.org/repos/asf/carbondata.git.


 was 952cf82  [maven-release-plugin] prepare for next development iteration

This change permanently discards the following revisions:

 discard 952cf82  [maven-release-plugin] prepare for next development iteration


[carbondata] branch master updated: [CARBONDATA-4191] update table for primitive column not working when complex child column name and primitive column name match

2021-06-02 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new 4c04f7c  [CARBONDATA-4191] update table for primitive column not 
working when complex child column name and primitive column name match
4c04f7c is described below

commit 4c04f7c0e49efe5e3203ff2ce04bd7bf1b7fe6aa
Author: Mahesh Raju Somalaraju 
AuthorDate: Mon May 24 18:01:34 2021 +0530

[CARBONDATA-4191] update table for primitive column not working when 
complex child
column name and primitive column name match

Why is this PR needed?
Update primitive column not working when complex column child name and 
primitive
data type name same.
When an update for primitive is received, we are checking in complex child 
columns
if column name matches then returning UnsupportedOperationbException.

What changes were proposed in this PR?
Currently, we are ignoring the prefix of all columns and passing only 
columns/child
column info to the update command.
New Changes: Passing full column(alias name/table name.columnName) name 
which is given
by the user and added checks for handling the unsupported update operation 
of complex columns.

This closes #4139
---
 .../apache/spark/sql/optimizer/CarbonIUDRule.scala |  8 ---
 .../spark/sql/parser/CarbonSpark2SqlParser.scala   | 34 --
 .../complexType/TestComplexDataType.scala  | 72 +-
 3 files changed, 100 insertions(+), 14 deletions(-)

diff --git 
a/integration/spark/src/main/scala/org/apache/spark/sql/optimizer/CarbonIUDRule.scala
 
b/integration/spark/src/main/scala/org/apache/spark/sql/optimizer/CarbonIUDRule.scala
index da1ca55..77d85a8 100644
--- 
a/integration/spark/src/main/scala/org/apache/spark/sql/optimizer/CarbonIUDRule.scala
+++ 
b/integration/spark/src/main/scala/org/apache/spark/sql/optimizer/CarbonIUDRule.scala
@@ -41,14 +41,6 @@ class CarbonIUDRule extends Rule[LogicalPlan] with 
PredicateHelper {
   case Project(pList, child) if !isTransformed =>
 var (dest: Seq[NamedExpression], source: Seq[NamedExpression]) = 
pList
   .splitAt(pList.size - cols.size)
-// check complex column
-cols.foreach { col =>
-  val complexExists = "\"name\":\"" + col + "\""
-  if (dest.exists(m => m.dataType.json.contains(complexExists))) {
-throw new UnsupportedOperationException(
-  "Unsupported operation on Complex data type")
-  }
-}
 // check updated columns exists in table
 val diff = cols.diff(dest.map(_.name.toLowerCase))
 if (diff.nonEmpty) {
diff --git 
a/integration/spark/src/main/scala/org/apache/spark/sql/parser/CarbonSpark2SqlParser.scala
 
b/integration/spark/src/main/scala/org/apache/spark/sql/parser/CarbonSpark2SqlParser.scala
index b77f16f..945a420 100644
--- 
a/integration/spark/src/main/scala/org/apache/spark/sql/parser/CarbonSpark2SqlParser.scala
+++ 
b/integration/spark/src/main/scala/org/apache/spark/sql/parser/CarbonSpark2SqlParser.scala
@@ -17,8 +17,9 @@
 
 package org.apache.spark.sql.parser
 
-import scala.collection.JavaConverters.seqAsJavaListConverter
+import scala.collection.JavaConverters._
 import scala.collection.mutable
+import scala.collection.mutable.ListBuffer
 import scala.language.implicitConversions
 
 import org.apache.commons.lang3.StringUtils
@@ -257,6 +258,25 @@ class CarbonSpark2SqlParser extends CarbonDDLSqlParser {
 (SET ~> "(" ~> repsep(element, ",") <~ ")") ~
 ("=" ~> restInput) <~ opt(";") ^^ {
   case tab ~ columns ~ rest =>
+// If update is received for complex data types then throw exception
+var finalColumns = List.empty[String]
+var updateColumns = new ListBuffer[String]()
+columns.foreach { column =>
+  if (column.contains('.')) {
+val columnFullName = column.split('.')
+if (columnFullName.size >= 3) {
+  throw new UnsupportedOperationException("Unsupported operation 
on Complex data types")
+} else if ((tab._3.isDefined && 
tab._3.get.equals(columnFullName(0)))
+|| tab._4.table.equals(columnFullName(0))) {
+  updateColumns += columnFullName(1)
+} else {
+  throw new UnsupportedOperationException("Unsupported operation 
on Complex data types")
+}
+  } else {
+updateColumns += column
+  }
+}
+finalColumns = updateColumns.toList
 val (sel, where) = splitQuery(rest)
 val selectPattern = """^\s*select\s+"

[carbondata] branch master updated: [CARBONDATA-4186] Fixed insert failure when partition column present in local sort scope

2021-06-02 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new 01fd120  [CARBONDATA-4186] Fixed insert failure when partition column 
present in local sort scope
01fd120 is described below

commit 01fd12042e3300a5d18eef15a634df66b29580ee
Author: Nihal ojha 
AuthorDate: Wed May 12 12:26:42 2021 +0530

[CARBONDATA-4186] Fixed insert failure when partition column present in 
local sort scope

Why is this PR needed?
Currently when we create table with partition column and put the same 
column as part of
local sort scope then Insert query fails with ArrayIndexOutOfBounds 
exception.

What changes were proposed in this PR?
Handle ArrayIndexOutOfBound exception, earlier array size was not 
increasing because data
was inconsistence and in the wrong order for sortcolumn and 
isDimNoDictFlags.

This closes #4132
---
 .../carbondata/core/datastore/TableSpec.java   |  21 +++-
 .../StandardPartitionTableLoadingTestCase.scala|  18 
 .../loading/sort/unsafe/UnsafeCarbonRowPage.java   |  17 ++-
 .../processing/sort/sortdata/SortParameters.java   | 119 ++---
 .../processing/sort/sortdata/TableFieldStat.java   |  28 +++--
 .../processing/util/CarbonDataProcessorUtil.java   |   2 +-
 6 files changed, 174 insertions(+), 31 deletions(-)

diff --git 
a/core/src/main/java/org/apache/carbondata/core/datastore/TableSpec.java 
b/core/src/main/java/org/apache/carbondata/core/datastore/TableSpec.java
index 6cbf07c..68363aa 100644
--- a/core/src/main/java/org/apache/carbondata/core/datastore/TableSpec.java
+++ b/core/src/main/java/org/apache/carbondata/core/datastore/TableSpec.java
@@ -62,17 +62,28 @@ public class TableSpec {
 List dimensions = carbonTable.getVisibleDimensions();
 List measures = carbonTable.getVisibleMeasures();
 if (keepPartitionColumnsToEnd && carbonTable.getPartitionInfo() != null) {
-  // keep the partition columns in the end
+  // keep the partition columns in the end if that is not present in sort 
columns
   List reArrangedDimensions = new ArrayList<>();
   List reArrangedMeasures = new ArrayList<>();
   List partitionDimensions = new ArrayList<>();
   List partitionMeasures = new ArrayList<>();
   List columnSchemaList = 
carbonTable.getPartitionInfo().getColumnSchemaList();
+  String[] sortColumns = carbonTable.getTableInfo()
+  .getFactTable().getTableProperties().getOrDefault("sort_columns", 
"").split(",");
+  for (String col : sortColumns) {
+for (CarbonDimension dim : dimensions) {
+  if (dim.getColName().equalsIgnoreCase(col)) {
+reArrangedDimensions.add(dim);
+  }
+}
+  }
   for (CarbonDimension dim : dimensions) {
-if (columnSchemaList.contains(dim.getColumnSchema())) {
-  partitionDimensions.add(dim);
-} else {
-  reArrangedDimensions.add(dim);
+if (!dim.isSortColumn()) {
+  if (columnSchemaList.contains(dim.getColumnSchema())) {
+partitionDimensions.add(dim);
+  } else {
+reArrangedDimensions.add(dim);
+  }
 }
   }
   if (partitionDimensions.size() != 0) {
diff --git 
a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/standardpartition/StandardPartitionTableLoadingTestCase.scala
 
b/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/standardpartition/StandardPartitionTableLoadingTestCase.scala
index 3a143b6..bdeacc1 100644
--- 
a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/standardpartition/StandardPartitionTableLoadingTestCase.scala
+++ 
b/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/standardpartition/StandardPartitionTableLoadingTestCase.scala
@@ -517,6 +517,24 @@ class StandardPartitionTableLoadingTestCase extends 
QueryTest with BeforeAndAfte
 assert(result.get(0).get(6).equals(dataAndIndexSize._2))
   }
 
+  test("test partition column with different sort scope") {
+verifySortWithPartition("global_sort")
+verifySortWithPartition("no_sort")
+verifySortWithPartition("local_sort")
+  }
+
+  def verifySortWithPartition(scope: String): Unit = {
+sql("drop table if exists carbon_partition")
+sql(s"create table carbon_partition(id int, name string, salary double) " +
+"partitioned by(country string, id1 int)" +
+s"stored as carbondata 
tblproperties('sort_scope'='$scope','sort_columns'='country, id')")
+sql("insert into carbon_partition select 1, 'Ram',3500,'India', 20")
+checkAnswer(
+  sql("SELECT * FROM carbon_partit

[carbondata] branch master updated: [CARBONDATA-4185] Doc Changes for Heterogeneous format segments in carbondata

2021-05-20 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new 861ba2e  [CARBONDATA-4185] Doc Changes for Heterogeneous format 
segments in carbondata
861ba2e is described below

commit 861ba2e44f8087e15fbdba278e21e229355b7d99
Author: Mahesh Raju Somalaraju 
AuthorDate: Wed May 12 23:11:57 2021 +0530

[CARBONDATA-4185] Doc Changes for Heterogeneous format segments in 
carbondata

Why is this PR needed?
Heterogeneous format segments in carbondata documenation.

What changes were proposed in this PR?
Add segment feature background and impact on existed carbondata features

This closes #4134
---
 README.md|  1 +
 docs/addsegment-guide.md | 78 
 2 files changed, 79 insertions(+)

diff --git a/README.md b/README.md
index 4f8da5c..d1b47af 100644
--- a/README.md
+++ b/README.md
@@ -57,6 +57,7 @@ CarbonData is built using Apache Maven, to [build 
CarbonData](https://github.com
  * [CarbonData Lucene 
Index](https://github.com/apache/carbondata/blob/master/docs/index/lucene-index-guide.md)
  * [CarbonData 
MV](https://github.com/apache/carbondata/blob/master/docs/mv-guide.md)
 * [Carbondata Secondary 
Index](https://github.com/apache/carbondata/blob/master/docs/index/secondary-index-guide.md)
+* [Heterogeneous format segments in 
carbondata](https://github.com/apache/carbondata/blob/master/docs/addsegment-guide.md)
 * [SDK 
Guide](https://github.com/apache/carbondata/blob/master/docs/sdk-guide.md) 
 * [C++ SDK 
Guide](https://github.com/apache/carbondata/blob/master/docs/csdk-guide.md)
 * [Performance 
Tuning](https://github.com/apache/carbondata/blob/master/docs/performance-tuning.md)
 
diff --git a/docs/addsegment-guide.md b/docs/addsegment-guide.md
new file mode 100644
index 000..a45e6d6
--- /dev/null
+++ b/docs/addsegment-guide.md
@@ -0,0 +1,78 @@
+
+
+# Heterogeneous format segments in carbondata
+
+###Background
+In the industry, many users already adopted to data with different formats 
like ORC, Parquet, JSON, CSV etc.,  
+If users want to migrate to Carbondata for better performance or for better 
features then there is no direct way. 
+All the existing data needs to be converted to Carbondata to migrate.  
+This solution works out if the existing data is less, what if the existing 
data is more?   
+Heterogeneous format segments aims to solve this problem by avoiding data 
conversion.
+
+###Add segment with path and format
+Users can add the existing data as a segment to the carbon table provided the 
schema of the data
+ and the carbon table should be the same. 
+
+```
+Alter table table_name add segment options (‘path’= 
'hdfs://usr/oldtable,'format'=parquet)
+```
+In the above command user can add the existing data to the carbon table as a 
new segment and also
+ can provide the data format.
+
+During add segment, it will infer the schema from data and validates the 
schema against the carbon table. 
+If the schema doesn’t match it throws an exception.
+
+###Changes to tablestatus file
+Carbon adds the new segment by adding segment information to tablestatus file. 
In order to add the path and format information to tablestatus, we are going to 
add `segmentPath`  and ‘format’  to the tablestatus file. 
+And any extra `options` will be added to the segment file.
+
+
+###Changes to Spark Integration
+During select query carbon reads data through RDD which is created by
+  CarbonDatasourceHadoopRelation.buildScan, This RDD reads data from physical 
carbondata files and provides data to spark query plan.
+To support multiple formats per segment basis we can create multiple RDD using 
the existing Spark
+ file format scan class FileSourceScanExec . This class can generate scan RDD 
for all spark supported formats. We can union all these multi-format RDD and 
create a single RDD and provide it to spark query plan.
+
+Note: This integration will be clean as we use the sparks optimized reading, 
pruning and it
+ involves whole codegen and vector processing with unsafe support.
+
+###Changes to Presto Integration
+CarbondataSplitManager can create the splits for carbon and as well as for 
other formats and 
+ choose the page source as per the split.  
+
+### Impact on existed feature
+**Count(\*) query:**  In case if the segments are mixed with different formats 
then driver side
+ optimization for count(*) query will not work so it will be executed on 
executor side.
+
+**Index DataMaps:** Datamaps like block/blocklet datamap will only work for 
carbondata format
+ segments so there would not be any driver side pruning for other formats.
+
+**Update/Delete:** Update & Delete operations cannot be allowed on the table 
which has mixed formats
+But it can be allowed if the external segments are added with carbon

[carbondata] branch master updated: [CARBONDATA-4188] Fixed select query with small table page size after alter add column

2021-05-20 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new 41a756f  [CARBONDATA-4188] Fixed select query with small table page 
size after alter add column
41a756f is described below

commit 41a756fff7d8774fd919540561d1c39cfc24b9b4
Author: Nihal ojha 
AuthorDate: Mon May 17 15:30:24 2021 +0530

[CARBONDATA-4188] Fixed select query with small table page size after alter 
add column

Why is this PR needed?
Select query on table with long string data type and small page size throws
ArrayIndexOutOfBoudException after alter add columns.
Query fails because after changing the schema, the number of rows set in
bitsetGroup(RestructureIncludeFilterExecutorImpl.applyFilter()) for pages 
is not correct.

What changes were proposed in this PR?
Set the correct number of rows inside every page of bitsetGroup.

This closes #4137
---
 .../carbondata/core/scan/filter/FilterUtil.java| 29 ++-
 .../executer/RangeValueFilterExecutorImpl.java |  4 +-
 .../RestructureExcludeFilterExecutorImpl.java  |  6 +--
 .../RestructureIncludeFilterExecutorImpl.java  |  6 +--
 ...velRangeGreaterThanEqualFilterExecutorImpl.java |  4 +-
 ...RowLevelRangeGreaterThanFilterExecutorImpl.java |  4 +-
 ...wLevelRangeLessThanEqualFilterExecutorImpl.java |  4 +-
 .../RowLevelRangeLessThanFilterExecutorImpl.java   |  4 +-
 .../core/scan/filter/FilterUtilTest.java   | 43 --
 9 files changed, 52 insertions(+), 52 deletions(-)

diff --git 
a/core/src/main/java/org/apache/carbondata/core/scan/filter/FilterUtil.java 
b/core/src/main/java/org/apache/carbondata/core/scan/filter/FilterUtil.java
index 9dc6b5b..1acae54 100644
--- a/core/src/main/java/org/apache/carbondata/core/scan/filter/FilterUtil.java
+++ b/core/src/main/java/org/apache/carbondata/core/scan/filter/FilterUtil.java
@@ -30,7 +30,6 @@ import java.util.Set;
 
 import org.apache.carbondata.common.logging.LogServiceFactory;
 import org.apache.carbondata.core.constants.CarbonCommonConstants;
-import org.apache.carbondata.core.constants.CarbonV3DataFormatConstants;
 import org.apache.carbondata.core.datastore.block.SegmentProperties;
 import org.apache.carbondata.core.datastore.chunk.DimensionColumnPage;
 import org.apache.carbondata.core.datastore.chunk.impl.DimensionRawColumnChunk;
@@ -81,6 +80,7 @@ import 
org.apache.carbondata.core.scan.filter.resolver.RowLevelRangeFilterResolv
 import 
org.apache.carbondata.core.scan.filter.resolver.resolverinfo.ColumnResolvedFilterInfo;
 import 
org.apache.carbondata.core.scan.filter.resolver.resolverinfo.DimColumnResolvedFilterInfo;
 import 
org.apache.carbondata.core.scan.filter.resolver.resolverinfo.MeasureColumnResolvedFilterInfo;
+import org.apache.carbondata.core.scan.processor.RawBlockletColumnChunks;
 import org.apache.carbondata.core.scan.result.vector.CarbonDictionary;
 import org.apache.carbondata.core.util.BitSetGroup;
 import org.apache.carbondata.core.util.ByteUtil;
@@ -898,30 +898,23 @@ public final class FilterUtil {
   }
 
   /**
-   * This method will create default bitset group. Applicable for restructure 
scenarios.
+   * This method will create bit set group for particular raw blocklet column 
chunk.
+   * Applicable for restructure scenarios.
*
-   * @param pageCount
-   * @param totalRowCount
+   * @param rawBlockletColumnChunks
* @param defaultValue
* @return
*/
-  public static BitSetGroup createBitSetGroupWithDefaultValue(int pageCount, 
int totalRowCount,
-  boolean defaultValue) {
+  public static BitSetGroup 
createBitSetGroupWithColumnChunk(RawBlockletColumnChunks
+  rawBlockletColumnChunks, boolean defaultValue) {
+int pageCount = rawBlockletColumnChunks.getDataBlock().numberOfPages();
 BitSetGroup bitSetGroup = new BitSetGroup(pageCount);
-int numberOfRows = 
CarbonV3DataFormatConstants.NUMBER_OF_ROWS_PER_BLOCKLET_COLUMN_PAGE_DEFAULT;
-int pagesTobeFullFilled = totalRowCount / numberOfRows;
-int rowCountForLastPage = totalRowCount % numberOfRows;
-for (int i = 0; i < pagesTobeFullFilled; i++) {
-  BitSet bitSet = new BitSet(numberOfRows);
-  bitSet.set(0, numberOfRows, defaultValue);
+for (int i = 0; i < pageCount; i++) {
+  int pageRowCount = 
rawBlockletColumnChunks.getDataBlock().getPageRowCount(i);
+  BitSet bitSet = new BitSet(pageRowCount);
+  bitSet.set(0, pageRowCount, defaultValue);
   bitSetGroup.setBitSet(bitSet, i);
 }
-// create and fill bitset for the last page if any records are left
-if (rowCountForLastPage > 0) {
-  BitSet bitSet = new BitSet(rowCountForLastPage);
-  bitSet.set(0, rowCountForLastPage, defaultValue);
-  bitSetGroup.setBitSet(bitSet, pagesTobeFullFilled);
-}
 return bitSetGroup;
  

[carbondata] branch master updated: [CARBONDATA-4175] [CARBONDATA-4162] Leverage Secondary Index till segment level

2021-05-10 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new 8996369  [CARBONDATA-4175] [CARBONDATA-4162] Leverage Secondary Index 
till segment level
8996369 is described below

commit 899636910d7ebf914d7ec4e2443ce6a4e4ae2ef0
Author: Nihal ojha 
AuthorDate: Wed Mar 24 15:17:44 2021 +0530

[CARBONDATA-4175] [CARBONDATA-4162] Leverage Secondary Index till segment 
level

Why is this PR needed?
In the existing architecture, if the parent(main) table and SI table don’t 
have
the same valid segments then we disable the SI table. And then from the 
next query
onwards, we scan and prune only the parent table until we trigger the next 
load or
REINDEX command (as these commands will make the parent and SI table 
segments in sync).
Because of this, queries take more time to give the result when SI is 
disabled.

What changes were proposed in this PR?
Instead of disabling the SI table(when parent and child table segments are 
not in sync)
we will do pruning on SI tables for all the valid segments(segments with 
status success,
marked for update and load partial success) and the rest of the segments 
will be pruned by the parent table.
As of now, query on the SI table can be pruned in two ways:
a) With SI as data map.
b) WIth spark plan rewrite.
This PR contains changes to support both methods of SI to leverage till 
segment level.

This closes #4116
---
 .../apache/carbondata/core/index/IndexFilter.java  |  10 +++
 .../org/apache/carbondata/core/index/Segment.java  |  11 +++
 .../apache/carbondata/core/index/TableIndex.java   |  24 ++---
 .../carbondata/core/index/dev/IndexFactory.java|   1 +
 .../executer/RowLevelFilterExecutorImpl.java   |   4 +-
 .../carbondata/hadoop/api/CarbonInputFormat.java   |  15 
 .../TestCreateIndexWithLoadAndCompaction.scala |  32 ++-
 .../secondaryindex/TestIndexModelWithIUD.scala |  30 +++
 .../testsuite/secondaryindex/TestIndexRepair.scala |  20 ++---
 .../TestSIWithComplexArrayType.scala   |   9 +-
 .../secondaryindex/TestSIWithSecondaryIndex.scala  | 100 +++--
 .../carbondata/index/secondary/SecondaryIndex.java |  46 --
 .../index/secondary/SecondaryIndexFactory.java |   3 +
 .../command/index/ShowIndexesCommand.scala |  26 +-
 .../apache/spark/sql/index/CarbonIndexUtil.scala   |  28 --
 .../secondaryindex/command/SICreationCommand.scala |  26 +-
 .../joins/BroadCastSIFilterPushJoin.scala  |  18 +++-
 .../spark/sql/secondaryindex/load/Compactor.scala  |  14 ---
 .../optimizer/CarbonSecondaryIndexOptimizer.scala  |  30 ++-
 .../secondaryindex/rdd/SecondaryIndexCreator.scala |  18 
 .../testsuite/addsegment/AddSegmentTestCase.scala  |  36 
 21 files changed, 326 insertions(+), 175 deletions(-)

diff --git 
a/core/src/main/java/org/apache/carbondata/core/index/IndexFilter.java 
b/core/src/main/java/org/apache/carbondata/core/index/IndexFilter.java
index 1579a5e..af752f3 100644
--- a/core/src/main/java/org/apache/carbondata/core/index/IndexFilter.java
+++ b/core/src/main/java/org/apache/carbondata/core/index/IndexFilter.java
@@ -66,6 +66,8 @@ public class IndexFilter implements Serializable {
   // limit value used for row scanning, collected when carbon.mapOrderPushDown 
is enabled
   private int limit = -1;
 
+  private Set missingSISegments;
+
   public IndexFilter(CarbonTable table, Expression expression) {
 this(table, expression, false);
   }
@@ -283,4 +285,12 @@ public class IndexFilter implements Serializable {
   throw new RuntimeException("Error while resolving filter expression", e);
 }
   }
+
+  public void setMissingSISegments(Set missingSISegments) {
+this.missingSISegments = missingSISegments;
+  }
+
+  public Set getMissingSISegments() {
+return missingSISegments;
+  }
 }
diff --git a/core/src/main/java/org/apache/carbondata/core/index/Segment.java 
b/core/src/main/java/org/apache/carbondata/core/index/Segment.java
index 0a32984..4bbe543 100644
--- a/core/src/main/java/org/apache/carbondata/core/index/Segment.java
+++ b/core/src/main/java/org/apache/carbondata/core/index/Segment.java
@@ -28,6 +28,7 @@ import java.util.Map;
 import java.util.Objects;
 import java.util.Set;
 
+import org.apache.carbondata.core.indexstore.ExtendedBlocklet;
 import org.apache.carbondata.core.metadata.schema.table.Writable;
 import org.apache.carbondata.core.mutate.UpdateVO;
 import org.apache.carbondata.core.readcommitter.ReadCommittedScope;
@@ -96,6 +97,8 @@ public class Segment implements Serializable, Writable {
*/
   private SegmentMetaDataInfo segmentMetaDataInfo;
 
+  private List defaultIndexPrunedBlocklets;
+
   public Segment() {
 
   }
@@ -417,4 +420,12 @@ public cla

[carbondata] branch master updated: [CARBONDATA-4172] Select query having parent and child struct column in projection returns incorrect results

2021-04-26 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new 3b411bb  [CARBONDATA-4172] Select query having parent and child struct 
column in projection returns incorrect results
3b411bb is described below

commit 3b411bbea81ed9c70e88a4933dfff2742d686b6e
Author: Indhumathi27 
AuthorDate: Thu Apr 22 13:50:39 2021 +0530

[CARBONDATA-4172] Select query having parent and child struct column in 
projection returns incorrect results

Why is this PR needed?
After PR-3574, a scenario has been missed while code refactor.
Currently, if select query has both Parent and its child struct column in 
projection,
only child column is pushed down to carbon for filling result. For other 
columns in parent Struct, data output is null.

What changes were proposed in this PR?
If parent struct column is also present in projection, then push down only 
parent column to carbon.

This closes #4123
---
 .../sql/execution/strategy/PushDownHelper.scala| 26 ---
 .../complexType/TestComplexDataType.scala  | 51 +++---
 2 files changed, 26 insertions(+), 51 deletions(-)

diff --git 
a/integration/spark/src/main/scala/org/apache/spark/sql/execution/strategy/PushDownHelper.scala
 
b/integration/spark/src/main/scala/org/apache/spark/sql/execution/strategy/PushDownHelper.scala
index e532cd4..c8a8501 100644
--- 
a/integration/spark/src/main/scala/org/apache/spark/sql/execution/strategy/PushDownHelper.scala
+++ 
b/integration/spark/src/main/scala/org/apache/spark/sql/execution/strategy/PushDownHelper.scala
@@ -17,7 +17,7 @@
 
 package org.apache.spark.sql.execution.strategy
 
-import scala.collection.mutable.ArrayBuffer
+import scala.collection.mutable.{ArrayBuffer, ListBuffer}
 
 import org.apache.spark.sql.catalyst.expressions.{AttributeReference, 
GetArrayItem, GetMapValue, GetStructField, Literal, NamedExpression}
 import org.apache.spark.sql.types.{ArrayType, DataType, MapType, StructType}
@@ -35,15 +35,24 @@ object PushDownHelper {
 // parent/child field and push down the corresponding project column. In 
case of Array, Map,
 // ArrayOfStruct, StructOfArray, MapOfStruct or StructOfMap, push down 
parent column
 val output = ArrayBuffer[String]()
-projects.foreach(PushDownHelper.collectColumns(_, output))
+var parentColumns = ArrayBuffer[String]()
+projects.foreach(PushDownHelper.collectColumns(_, output, parentColumns))
+parentColumns = parentColumns.distinct
 if (output.isEmpty) {
   requiredColumns.foreach(projection.addColumn)
 } else {
   requiredColumns.map(_.toLowerCase).foreach { requiredColumn =>
 val childOption = output.filter(_.startsWith(requiredColumn + "."))
-childOption.isEmpty match {
-  case true => projection.addColumn(requiredColumn)
-  case false => childOption.foreach(projection.addColumn)
+if (childOption.isEmpty) {
+  projection.addColumn(requiredColumn)
+} else {
+  // If projection contains both parent and its child, then push down 
parent column
+  // itself instead of its child column
+  if (parentColumns.contains(requiredColumn)) {
+projection.addColumn(requiredColumn)
+  } else {
+childOption.foreach(projection.addColumn)
+  }
 }
   }
 }
@@ -51,7 +60,8 @@ object PushDownHelper {
 
   private def collectColumns(
   exp: NamedExpression,
-  pushDownColumns: ArrayBuffer[String]
+  pushDownColumns: ArrayBuffer[String],
+  parentColumns: ArrayBuffer[String]
   ): Unit = {
 exp transform {
   case struct: GetStructField =>
@@ -69,7 +79,9 @@ object PushDownHelper {
 pushDownColumns += getParentName(map)
 Literal.TrueLiteral
   case attr: AttributeReference =>
-pushDownColumns += attr.name.toLowerCase
+val attrName = attr.name.toLowerCase
+pushDownColumns += attrName
+parentColumns += attrName
 Literal.TrueLiteral
 }
   }
diff --git 
a/integration/spark/src/test/scala/org/apache/carbondata/integration/spark/testsuite/complexType/TestComplexDataType.scala
 
b/integration/spark/src/test/scala/org/apache/carbondata/integration/spark/testsuite/complexType/TestComplexDataType.scala
index 88f5613..6c457dd 100644
--- 
a/integration/spark/src/test/scala/org/apache/carbondata/integration/spark/testsuite/complexType/TestComplexDataType.scala
+++ 
b/integration/spark/src/test/scala/org/apache/carbondata/integration/spark/testsuite/complexType/TestComplexDataType.scala
@@ -611,51 +611,14 @@ class TestComplexDataType extends QueryTest with 
BeforeAndAfterAll {
 checkAnswer(sql("select a.d,a.f from table1"),
   Seq(Row(2, Row(3, "mno"

[carbondata] branch master updated: [CARBONDATA-4158]Add Secondary Index as a coarse-grain index and use secondary indexes for Presto queries

2021-04-22 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new 09ad509  [CARBONDATA-4158]Add Secondary Index as a coarse-grain index 
and use secondary indexes for Presto queries
09ad509 is described below

commit 09ad509d67fbe752892c7180fb8a10cd62fef465
Author: Venu Reddy 
AuthorDate: Wed Mar 10 00:23:07 2021 +0530

[CARBONDATA-4158]Add Secondary Index as a coarse-grain index and use 
secondary indexes for Presto queries

Why is this PR needed?
At present, secondary indexes are leveraged for query pruning via spark 
plan modification.
This approach is tightly coupled with spark because the plan modification 
is specific to
spark engine. In order to use secondary indexes for Presto or Hive queries, 
it is not
feasible to modify the query plans as we desire in the current approach. 
Thus need arises
for an engine agnostic approach to use secondary indexes in query pruning.

What changes were proposed in this PR?
1. Add Secondary Index as a coarse grain index.
2. Add a new insegment() UDF to support query within the particular segments
3. Control the use of Secondary Index as a coarse grain index pruning with
property('carbon.coarse.grain.secondary.index')
4. Use Index Server driver for Secondary Index pruning
5. Use Secondary Indexes with Presto Queries

This closes #4110
---
 README.md  |   1 +
 .../core/constants/CarbonCommonConstants.java  |  28 
 .../carbondata/core/index/AbstractIndexJob.java|  10 ++
 .../apache/carbondata/core/index/IndexChooser.java |  17 +-
 .../apache/carbondata/core/index/IndexFilter.java  |   2 +-
 .../carbondata/core/index/IndexInputFormat.java|  14 +-
 .../org/apache/carbondata/core/index/IndexJob.java |   2 +
 .../carbondata/core/index/IndexStoreManager.java   |  12 +-
 .../apache/carbondata/core/index/IndexUtil.java|  12 +-
 .../secondaryindex}/CarbonCostBasedOptimizer.java  |  50 +-
 .../carbondata/core/metadata/index/IndexType.java  |   4 +-
 .../metadata/schema/indextable/IndexTableInfo.java |  13 +-
 .../core/scan/expression/ExpressionResult.java |  24 +--
 .../core/scan/expression/LiteralExpression.java|  17 +-
 .../conditional/NotEqualsExpression.java   |   2 +-
 .../carbondata/core/util/CarbonProperties.java |  28 
 .../apache/carbondata/core/util/CarbonUtil.java|  34 
 .../apache/carbondata/core/util/SessionParams.java |   6 +-
 docs/index/secondary-index-guide.md|  29 
 .../carbondata/hadoop/api/CarbonInputFormat.java   |  46 +-
 .../org/apache/carbon/flink/TestCarbonWriter.scala |   2 +-
 .../presto/CarbondataPageSourceProvider.java   |   4 +
 .../carbondata/presto/CarbondataSplitManager.java  |   3 +
 .../carbondata/presto/impl/CarbonTableReader.java  |  56 +++
 .../presto/CarbondataPageSourceProvider.java   |   4 +
 .../carbondata/presto/CarbondataSplitManager.java  |   3 +
 .../carbondata/presto/impl/CarbonTableReader.java  |  56 +++
 .../carbondata/index/secondary/SecondaryIndex.java | 125 +++
 .../index/secondary/SecondaryIndexFactory.java | 172 +
 .../index/secondary/SecondaryIndexModel.java   | 103 
 .../indexserver/DistributedRDDUtils.scala  |  37 -
 .../indexserver/DistributedShowCacheRDD.scala  |   2 +-
 .../carbondata/indexserver/IndexServer.scala   |  23 ++-
 .../carbondata/spark/rdd/CarbonScanRDD.scala   |  13 +-
 .../scala/org/apache/spark/sql/CarbonEnv.scala |   3 +
 .../command/index/CarbonCreateIndexCommand.scala   |   4 +-
 .../command/index/CarbonRefreshIndexCommand.scala  |   4 +-
 .../execution/strategy/CarbonDataSourceScan.scala  |   6 +-
 .../execution/strategy/CarbonSourceStrategy.scala  |  13 +-
 .../execution/command/CarbonHiveCommands.scala |   3 +-
 .../apache/spark/sql/index/CarbonIndexUtil.scala   |  92 ++-
 .../secondaryindex/command/SICreationCommand.scala |   8 +
 .../jobs/StringProjectionQueryJob.scala|  34 ++--
 .../spark/sql/secondaryindex/load/Compactor.scala  |   6 +
 .../optimizer/CarbonSITransformationRule.scala |  13 +-
 .../optimizer/CarbonSecondaryIndexOptimizer.scala  |   1 +
 .../secondaryindex/rdd/SecondaryIndexCreator.scala |   8 +
 .../index/bloom/BloomCoarseGrainIndexSuite.scala   |   4 +-
 .../complexType/TestArrayContainsPushDown.scala|  12 +-
 .../processing/index/IndexWriterListener.java  |   6 +-
 50 files changed, 1067 insertions(+), 104 deletions(-)

diff --git a/README.md b/README.md
index f706e1c..4f47546 100644
--- a/README.md
+++ b/README.md
@@ -76,6 +76,7 @@ Some features are marked as experimental because the 
syntax/implementation might
 2. Accelerating performance using MV on parquet/orc.
 3

[carbondata] branch master updated: [CARBONDATA-4156] Fix Writing Segment Min max with all blocks of a segment

2021-03-25 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new 865ec9b  [CARBONDATA-4156] Fix Writing Segment Min max with all blocks 
of a segment
865ec9b is described below

commit 865ec9b908783a6169d380415aa6598c41b275d5
Author: Indhumathi27 
AuthorDate: Tue Mar 9 16:09:37 2021 +0530

[CARBONDATA-4156] Fix Writing Segment Min max with all blocks of a segment

Why is this PR needed?
PR-3999 has removed some code related to getting segment min max from all 
blocks.
Because of this, if segment has more than one block, currently, it is 
writing
min max considering one block only.

What changes were proposed in this PR?
Reverted specific code from above PR. Removed unwanted synchronization for 
some methods

This closes #4101
---
 .../core/segmentmeta/SegmentMetaDataInfoStats.java | 35 --
 .../carbondata/hadoop/testutil/StoreCreator.java   |  3 ++
 .../allqueries/TestPruneUsingSegmentMinMax.scala   |  1 +
 3 files changed, 30 insertions(+), 9 deletions(-)

diff --git 
a/core/src/main/java/org/apache/carbondata/core/segmentmeta/SegmentMetaDataInfoStats.java
 
b/core/src/main/java/org/apache/carbondata/core/segmentmeta/SegmentMetaDataInfoStats.java
index 4f11eb0..9c4ce63 100644
--- 
a/core/src/main/java/org/apache/carbondata/core/segmentmeta/SegmentMetaDataInfoStats.java
+++ 
b/core/src/main/java/org/apache/carbondata/core/segmentmeta/SegmentMetaDataInfoStats.java
@@ -22,6 +22,7 @@ import java.util.LinkedHashMap;
 import java.util.Map;
 
 import org.apache.carbondata.core.constants.CarbonCommonConstants;
+import org.apache.carbondata.core.indexstore.blockletindex.BlockIndex;
 import org.apache.carbondata.core.util.ByteUtil;
 
 /**
@@ -53,7 +54,7 @@ public class SegmentMetaDataInfoStats {
* @param segmentId get corresponding segment Id from map
* @return segmentMetaDataInfo for the corresponding segment
*/
-  public synchronized SegmentMetaDataInfo getTableSegmentMetaDataInfo(String 
tableName,
+  public SegmentMetaDataInfo getTableSegmentMetaDataInfo(String tableName,
   String segmentId) {
 Map segmentColumnMetaDataInfoMap = new 
LinkedHashMap<>();
 Map segmentMetaDataInfoMap =
@@ -85,14 +86,30 @@ public class SegmentMetaDataInfoStats {
   public synchronized void setBlockMetaDataInfo(String tableName, String 
segmentId,
   BlockColumnMetaDataInfo currentBlockColumnMetaInfo) {
 // check if tableName is present in tableSegmentMetaDataInfoMap
-Map segmentMinMaxMap = new HashMap<>();
-if (!this.tableSegmentMetaDataInfoMap.isEmpty()
-&& null != this.tableSegmentMetaDataInfoMap.get(tableName)
-&& !this.tableSegmentMetaDataInfoMap.get(tableName).isEmpty()) {
-  segmentMinMaxMap = this.tableSegmentMetaDataInfoMap.get(tableName);
+if (!this.tableSegmentMetaDataInfoMap.isEmpty() && null != 
this.tableSegmentMetaDataInfoMap
+.get(tableName) && 
!this.tableSegmentMetaDataInfoMap.get(tableName).isEmpty()
+&& null != 
this.tableSegmentMetaDataInfoMap.get(tableName).get(segmentId)) {
+  // get previous blockColumn metadata information
+  BlockColumnMetaDataInfo previousBlockColumnMetaInfo =
+  this.tableSegmentMetaDataInfoMap.get(tableName).get(segmentId);
+  // compare and get updated min and max values
+  byte[][] updatedMin = 
BlockIndex.compareAndUpdateMinMax(previousBlockColumnMetaInfo.getMin(),
+  currentBlockColumnMetaInfo.getMin(), true);
+  byte[][] updatedMax = 
BlockIndex.compareAndUpdateMinMax(previousBlockColumnMetaInfo.getMax(),
+  currentBlockColumnMetaInfo.getMax(), false);
+  // update the segment
+  this.tableSegmentMetaDataInfoMap.get(tableName).get(segmentId)
+  .setMinMax(updatedMin, updatedMax);
+} else {
+  Map segmentMinMaxMap = new HashMap<>();
+  if (!this.tableSegmentMetaDataInfoMap.isEmpty()
+  && null != this.tableSegmentMetaDataInfoMap.get(tableName)
+  && !this.tableSegmentMetaDataInfoMap.get(tableName).isEmpty()) {
+segmentMinMaxMap = this.tableSegmentMetaDataInfoMap.get(tableName);
+  }
+  segmentMinMaxMap.put(segmentId, currentBlockColumnMetaInfo);
+  this.tableSegmentMetaDataInfoMap.put(tableName, segmentMinMaxMap);
 }
-segmentMinMaxMap.put(segmentId, currentBlockColumnMetaInfo);
-this.tableSegmentMetaDataInfoMap.put(tableName, segmentMinMaxMap);
   }
 
   /**
@@ -112,7 +129,7 @@ public class SegmentMetaDataInfoStats {
   /**
* This method will do min/max comparison of values and update if required
*/
-  public synchronized byte[] compareAndUpdateMinMax(byte[] minMaxValueCompare1,
+  public byte[] compareAndUpdateMinMax(byte[] minMaxValu

[carbondata] branch master updated: [CARBONDATA-4155] Fix Create table like table with MV

2021-03-23 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new 0f53bdb  [CARBONDATA-4155] Fix Create table like table with MV
0f53bdb is described below

commit 0f53bdb0de7a501d6ee4bb07321599a05a0de1ee
Author: Indhumathi27 
AuthorDate: Mon Mar 22 20:42:35 2021 +0530

[CARBONDATA-4155] Fix Create table like table with MV

Why is this PR needed?
PR-4076 has added a new table property to fact table.
While executing create table like command, this property
is not excluded, which leads to parsing exception.

What changes were proposed in this PR?
Remove MV related info from destination table properties

This closes #4111
---
 .../command/table/CarbonCreateTableLikeCommand.scala |  4 
 .../carbondata/view/rewrite/TestAllOperationsOnMV.scala  | 16 
 2 files changed, 20 insertions(+)

diff --git 
a/integration/spark/src/main/scala/org/apache/spark/sql/execution/command/table/CarbonCreateTableLikeCommand.scala
 
b/integration/spark/src/main/scala/org/apache/spark/sql/execution/command/table/CarbonCreateTableLikeCommand.scala
index 926b951..8921601 100644
--- 
a/integration/spark/src/main/scala/org/apache/spark/sql/execution/command/table/CarbonCreateTableLikeCommand.scala
+++ 
b/integration/spark/src/main/scala/org/apache/spark/sql/execution/command/table/CarbonCreateTableLikeCommand.scala
@@ -26,6 +26,7 @@ import org.apache.spark.sql.execution.command.MetadataCommand
 
 import 
org.apache.carbondata.common.exceptions.sql.MalformedCarbonCommandException
 import org.apache.carbondata.common.logging.LogServiceFactory
+import org.apache.carbondata.core.constants.CarbonCommonConstants
 import org.apache.carbondata.core.metadata.schema.{SchemaEvolution, 
SchemaEvolutionEntry}
 import org.apache.carbondata.core.metadata.schema.table.{TableInfo, 
TableSchema}
 
@@ -59,6 +60,9 @@ case class CarbonCreateTableLikeCommand(
 dstTableSchema.setTableName(targetTable.table)
 dstTableSchema.setTableId(UUID.randomUUID().toString)
 
+// remove mv related info from source table tblProperties
+
dstTableSchema.getTableProperties.remove(CarbonCommonConstants.RELATED_MV_TABLES_MAP)
+
 val schemaEvol: SchemaEvolution = new SchemaEvolution
 val schEntryList: util.List[SchemaEvolutionEntry] = new 
util.ArrayList[SchemaEvolutionEntry]
 schemaEvol.setSchemaEvolutionEntryList(schEntryList)
diff --git 
a/integration/spark/src/test/scala/org/apache/carbondata/view/rewrite/TestAllOperationsOnMV.scala
 
b/integration/spark/src/test/scala/org/apache/carbondata/view/rewrite/TestAllOperationsOnMV.scala
index 167d628..251e460 100644
--- 
a/integration/spark/src/test/scala/org/apache/carbondata/view/rewrite/TestAllOperationsOnMV.scala
+++ 
b/integration/spark/src/test/scala/org/apache/carbondata/view/rewrite/TestAllOperationsOnMV.scala
@@ -708,6 +708,22 @@ class TestAllOperationsOnMV extends QueryTest with 
BeforeAndAfterEach {
 }.getMessage.contains("Materialized view default.does_not_exist does not 
exist")
   }
 
+  test("test create table like maintable having mv") {
+sql("drop table IF EXISTS maintable")
+sql("create table maintable(name string, c_code int, price int) STORED AS 
carbondata")
+sql("drop materialized view if exists mv_table ")
+sql("create materialized view mv_table  as select name, sum(price) from 
maintable group by name")
+sql("drop table if exists new_Table")
+sql("create table new_Table like maintable")
+sql("insert into table new_Table select 'abc',21,2000")
+checkAnswer(sql("select * from new_Table"), Seq(Row("abc", 21, 2000)))
+intercept[MalformedCarbonCommandException] {
+  sql("create table new_Table1 like mv_table")
+}.getMessage.contains("Unsupported operation on SI table or MV.")
+sql("drop table if exists new_Table")
+sql("drop table IF EXISTS maintable")
+  }
+
   test("drop meta cache on mv materialized view table") {
 defaultConfig()
 sql("drop table IF EXISTS maintable")


[carbondata] branch master updated: [CARBONDATA-4153] Fix DoNot Push down not equal to filter with Cast on SI

2021-03-23 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new fd0ff22  [CARBONDATA-4153] Fix DoNot Push down not equal to filter 
with Cast on SI
fd0ff22 is described below

commit fd0ff22a9d5834d34b39cfd7a781e7332802e4ad
Author: Indhumathi27 
AuthorDate: Tue Mar 16 15:04:25 2021 +0530

[CARBONDATA-4153] Fix DoNot Push down not equal to filter with Cast on SI

Why is this PR needed?
NOT EQUAL TO filter on SI index column, should not be pushed down to SI 
table.
Currently, where x!='2' is not pushing down to SI, but where x!=2 is pushed 
down to SI.

This is because "x != 2" will be wrapped in a CAST expression like NOT 
EQUAL TO(cast(x as int) = 2).

What changes were proposed in this PR?
Handle CAST case while checking DONOT PUSH DOWN to SI

This closes #4108
---
 .../secondaryindex/TestSIWithSecondaryIndex.scala| 16 
 .../optimizer/CarbonSecondaryIndexOptimizer.scala|  2 ++
 2 files changed, 18 insertions(+)

diff --git 
a/index/secondary-index/src/test/scala/org/apache/carbondata/spark/testsuite/secondaryindex/TestSIWithSecondaryIndex.scala
 
b/index/secondary-index/src/test/scala/org/apache/carbondata/spark/testsuite/secondaryindex/TestSIWithSecondaryIndex.scala
index e0dc1ce..a5c1d34 100644
--- 
a/index/secondary-index/src/test/scala/org/apache/carbondata/spark/testsuite/secondaryindex/TestSIWithSecondaryIndex.scala
+++ 
b/index/secondary-index/src/test/scala/org/apache/carbondata/spark/testsuite/secondaryindex/TestSIWithSecondaryIndex.scala
@@ -653,6 +653,22 @@ class TestSIWithSecondaryIndex extends QueryTest with 
BeforeAndAfterAll {
 sql("drop table if exists maintable")
   }
 
+  test("test SI with donot push down not equal to filter with Cast") {
+sql("drop table if exists maintable")
+sql("create table maintable (a string,b string,c string) STORED AS 
carbondata ")
+sql("insert into maintable values ('aa', '3', 'cc')")
+sql("create index indextable on table maintable(b) AS 'carbondata'")
+val df1 = sql("select * from maintable where b!=2")
+val df2 = sql("select * from maintable where b!='2'")
+if (isFilterPushedDownToSI(df1.queryExecution.sparkPlan) &&
+isFilterPushedDownToSI(df2.queryExecution.sparkPlan)) {
+  assert(false)
+} else {
+  assert(true)
+}
+sql("drop table if exists maintable")
+  }
+
   def createAndInsertDataIntoTable(): Unit = {
 sql("drop table if exists maintable2")
 sql("create table maintable2 (a string,b string,c int) STORED AS 
carbondata ")
diff --git 
a/integration/spark/src/main/scala/org/apache/spark/sql/secondaryindex/optimizer/CarbonSecondaryIndexOptimizer.scala
 
b/integration/spark/src/main/scala/org/apache/spark/sql/secondaryindex/optimizer/CarbonSecondaryIndexOptimizer.scala
index df56b24..23d17d2 100644
--- 
a/integration/spark/src/main/scala/org/apache/spark/sql/secondaryindex/optimizer/CarbonSecondaryIndexOptimizer.scala
+++ 
b/integration/spark/src/main/scala/org/apache/spark/sql/secondaryindex/optimizer/CarbonSecondaryIndexOptimizer.scala
@@ -513,6 +513,8 @@ class CarbonSecondaryIndexOptimizer(sparkSession: 
SparkSession) {
 val doNotPushToSI = condition match {
   case IsNotNull(child: AttributeReference) => !pushDownNotNullFilter
   case Not(EqualTo(left: AttributeReference, right: Literal)) => true
+  case Not(EqualTo(left: Cast, right: Literal))
+if left.child.isInstanceOf[AttributeReference] => true
   case Not(Like(left: AttributeReference, right: Literal)) => true
   case Not(In(left: AttributeReference, right: Seq[Expression])) => true
   case Not(Contains(left: AttributeReference, right: Literal)) => true


[carbondata] branch master updated: [CARBONDATA-4141] Index Server is not caching indexes for external tables with sdk segments

2021-03-09 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new 25c5687  [CARBONDATA-4141] Index Server is not caching indexes for 
external tables with sdk segments
25c5687 is described below

commit 25c5687e2c8514f54aae4617410d3909ad8840e8
Author: Karan980 
AuthorDate: Tue Mar 2 17:13:49 2021 +0530

[CARBONDATA-4141] Index Server is not caching indexes for external tables 
with sdk segments

Why is this PR needed?
Indexes cached in Executor cache are not dropped when drop table is called 
for external table
with SDK segments. Because, external tables with sdk segments will not have 
metadata like table
status file. So in drop table command we send zero segments to indexServer 
clearIndexes job,
which clears nothing from executor side. So when we drop this type of 
table, executor side
indexes are not dropped. Now when we again create external table with same 
location and do
select * or select count(*), it will not cache the indexes for this table, 
because indexes with
same loaction are already present. Now show metacache on this newly created 
table will use new tableId ,
but indexes present have the old tableId, whose table is already dropped. 
So show metacache will return
nothing, because of tableId mismatch.

What changes were proposed in this PR?
Prepared the validSegments from indexFiles present at external table 
location and send it to IndexServer clearIndexes job through IndexInputFormat.

This closes #4099
---
 .../apache/carbondata/core/index/IndexUtil.java| 36 +-
 .../blockletindex/BlockletIndexFactory.java|  8 +++--
 2 files changed, 34 insertions(+), 10 deletions(-)

diff --git a/core/src/main/java/org/apache/carbondata/core/index/IndexUtil.java 
b/core/src/main/java/org/apache/carbondata/core/index/IndexUtil.java
index c663b78..87d2a40 100644
--- a/core/src/main/java/org/apache/carbondata/core/index/IndexUtil.java
+++ b/core/src/main/java/org/apache/carbondata/core/index/IndexUtil.java
@@ -20,6 +20,7 @@ package org.apache.carbondata.core.index;
 import java.io.IOException;
 import java.lang.reflect.Constructor;
 import java.util.ArrayList;
+import java.util.Arrays;
 import java.util.HashMap;
 import java.util.HashSet;
 import java.util.List;
@@ -36,7 +37,10 @@ import org.apache.carbondata.core.indexstore.PartitionSpec;
 import org.apache.carbondata.core.metadata.AbsoluteTableIdentifier;
 import org.apache.carbondata.core.metadata.schema.table.CarbonTable;
 import org.apache.carbondata.core.metadata.schema.table.RelationIdentifier;
+import org.apache.carbondata.core.readcommitter.LatestFilesReadCommittedScope;
+import org.apache.carbondata.core.readcommitter.ReadCommittedScope;
 import org.apache.carbondata.core.scan.filter.resolver.FilterResolverIntf;
+import org.apache.carbondata.core.statusmanager.LoadMetadataDetails;
 import org.apache.carbondata.core.statusmanager.SegmentStatusManager;
 import org.apache.carbondata.core.util.BlockletIndexUtil;
 import org.apache.carbondata.core.util.CarbonProperties;
@@ -112,15 +116,31 @@ public class IndexUtil {
*/
   private static void executeClearIndexJob(IndexJob indexJob,
   CarbonTable carbonTable, String indexToClear) throws IOException {
-SegmentStatusManager.ValidAndInvalidSegmentsInfo 
validAndInvalidSegmentsInfo =
-getValidAndInvalidSegments(carbonTable, 
FileFactory.getConfiguration());
-List invalidSegment = new ArrayList<>();
-for (Segment segment : validAndInvalidSegmentsInfo.getInvalidSegments()) {
-  invalidSegment.add(segment.getSegmentNo());
+IndexInputFormat indexInputFormat;
+if (!carbonTable.isTransactionalTable()) {
+  ReadCommittedScope readCommittedScope =
+  new LatestFilesReadCommittedScope(carbonTable.getTablePath(),
+  FileFactory.getConfiguration());
+  LoadMetadataDetails[] loadMetadataDetails = 
readCommittedScope.getSegmentList();
+  List listOfValidSegments = new 
ArrayList<>(loadMetadataDetails.length);
+  Arrays.stream(loadMetadataDetails).forEach(segment -> {
+Segment seg = new Segment(segment.getLoadName(), 
segment.getSegmentFile());
+seg.setLoadMetadataDetails(segment);
+listOfValidSegments.add(seg);
+  });
+  indexInputFormat =
+  new IndexInputFormat(carbonTable, listOfValidSegments, new 
ArrayList<>(0), true,
+  indexToClear);
+} else {
+  SegmentStatusManager.ValidAndInvalidSegmentsInfo 
validAndInvalidSegmentsInfo =
+  getValidAndInvalidSegments(carbonTable, 
FileFactory.getConfiguration());
+  List invalidSegment = new ArrayList<>();
+  validAndInvalidSegmentsInfo.getInvalidSegments()
+  .forEach(segment -> invalidSe

[carbondata] branch master updated: [CARBONDATA-4121] Prepriming is not working in Index Server

2021-02-18 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new 59ad77a  [CARBONDATA-4121] Prepriming is not working in Index Server
59ad77a is described below

commit 59ad77a96bcbada2c62bc31cbda2acf24bfa7498
Author: Karan980 
AuthorDate: Fri Feb 5 16:56:23 2021 +0530

[CARBONDATA-4121] Prepriming is not working in Index Server

Why is this PR needed?
Prepriming is not working in Index Server. Server.getRemoteUser
returns null value in async call of prepriming which results in
NPE and crashes the indexServer application. Issue Induced after PR #3952

What changes were proposed in this PR?
Computed the Server.getRemoteUser value before making the async prepriming
call and then used the same value during async call. Code reset to code 
before PR #3952

This closes #4088
---
 .../main/scala/org/apache/carbondata/indexserver/IndexServer.scala | 7 +--
 1 file changed, 5 insertions(+), 2 deletions(-)

diff --git 
a/integration/spark/src/main/scala/org/apache/carbondata/indexserver/IndexServer.scala
 
b/integration/spark/src/main/scala/org/apache/carbondata/indexserver/IndexServer.scala
index 207fb85..089b857 100644
--- 
a/integration/spark/src/main/scala/org/apache/carbondata/indexserver/IndexServer.scala
+++ 
b/integration/spark/src/main/scala/org/apache/carbondata/indexserver/IndexServer.scala
@@ -128,6 +128,10 @@ object IndexServer extends ServerInterface {
   def getCount(request: IndexInputFormat): LongWritable = {
 doAs {
   val sparkSession = SparkSQLUtil.getSparkSession
+  var currentUser: String = null
+  if (!request.isFallbackJob) {
+currentUser = Server.getRemoteUser.getShortUserName
+  }
   lazy val getCountTask = {
 if (!request.isFallbackJob) {
   sparkSession.sparkContext.setLocalProperty("spark.jobGroup.id", 
request.getTaskGroupId)
@@ -139,8 +143,7 @@ object IndexServer extends ServerInterface {
   }
   sparkSession.sparkContext.setLocalProperty("spark.job.description", 
taskGroupDesc)
   // Fire Generic Event like ACLCheck..etc
-  val indexServerEvent = IndexServerEvent(sparkSession, 
request.getCarbonTable,
-Server.getRemoteUser.getShortUserName)
+  val indexServerEvent = IndexServerEvent(sparkSession, 
request.getCarbonTable, currentUser)
   OperationListenerBus.getInstance().fireEvent(indexServerEvent, 
operationContext)
 }
 val splits = new DistributedCountRDD(sparkSession, request).collect()



[carbondata] branch master updated: [CARBONDATA-4126] Concurrent compaction failed with load on table

2021-02-18 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new 5ec3536  [CARBONDATA-4126] Concurrent compaction failed with load on 
table
5ec3536 is described below

commit 5ec35367684a9b8e7024d52442eae138d166440e
Author: Karan980 
AuthorDate: Thu Feb 11 13:14:07 2021 +0530

[CARBONDATA-4126] Concurrent compaction failed with load on table

Why is this PR needed?
Concurrent compaction was failing when run in parallel with load.
During load we acquire SegmentLock for a particular segment, and
when this same lock we try to acquire during compaction, we were
not able to acquire this lock and compaction fails.

What changes were proposed in this PR?
Skipped compaction for segments for which we are not able to acquire
the SegmentLock instead of throwing the exception.

This closes #4093
---
 .../apache/carbondata/spark/rdd/CarbonTableCompactor.scala  | 13 -
 1 file changed, 8 insertions(+), 5 deletions(-)

diff --git 
a/integration/spark/src/main/scala/org/apache/carbondata/spark/rdd/CarbonTableCompactor.scala
 
b/integration/spark/src/main/scala/org/apache/carbondata/spark/rdd/CarbonTableCompactor.scala
index a381089..5db344b 100644
--- 
a/integration/spark/src/main/scala/org/apache/carbondata/spark/rdd/CarbonTableCompactor.scala
+++ 
b/integration/spark/src/main/scala/org/apache/carbondata/spark/rdd/CarbonTableCompactor.scala
@@ -92,19 +92,22 @@ class CarbonTableCompactor(
   val lastSegment = sortedSegments.get(sortedSegments.size() - 1)
   val compactedLoad = CarbonDataMergerUtil.getMergedLoadName(loadsToMerge)
   var segmentLocks: ListBuffer[ICarbonLock] = ListBuffer.empty
+  val validSegments = new java.util.ArrayList[LoadMetadataDetails]
   loadsToMerge.asScala.foreach { segmentId =>
 val segmentLock = CarbonLockFactory
   
.getCarbonLockObj(carbonLoadModel.getCarbonDataLoadSchema.getCarbonTable
 .getAbsoluteTableIdentifier,
 CarbonTablePath.addSegmentPrefix(segmentId.getLoadName) + 
LockUsage.LOCK)
-if (!segmentLock.lockWithRetries()) {
-  throw new Exception(s"Failed to acquire lock on segment 
${segmentId.getLoadName}," +
-s" during compaction of table 
${compactionModel.carbonTable.getQualifiedName}")
+if (segmentLock.lockWithRetries()) {
+  validSegments.add(segmentId)
+  segmentLocks += segmentLock
+} else {
+  LOGGER.warn(s"Failed to acquire lock on segment 
${segmentId.getLoadName}, " +
+  s"during compaction of table 
${compactionModel.carbonTable.getQualifiedName}")
 }
-segmentLocks += segmentLock
   }
   try {
-scanSegmentsAndSubmitJob(loadsToMerge, compactedSegments, 
compactedLoad)
+scanSegmentsAndSubmitJob(validSegments, compactedSegments, 
compactedLoad)
   } catch {
 case e: Exception =>
   LOGGER.error(s"Exception in compaction thread ${ e.getMessage }", e)



[carbondata] branch master updated: [CARBONDATA-4082] Fix alter table add segment query on adding a segment having delete delta files

2021-02-04 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new aa7efda  [CARBONDATA-4082] Fix alter table add segment query on adding 
a segment having delete delta files
aa7efda is described below

commit aa7efda9eec8485723be81455d96ba17851571e8
Author: Karan980 
AuthorDate: Mon Jan 4 12:10:05 2021 +0530

[CARBONDATA-4082] Fix alter table add segment query on adding a segment 
having delete delta files

Why is this PR needed?
When a segment is added to a carbon table by alter table add segment query
and that segment also have a deleteDelta file present in it, then on 
querying
the carbon table the deleted rows are coming in the result.

What changes were proposed in this PR?
Updating the tableStatus and tableUpdateStatus files in correct way for the
segments having delta delta files.

This closes #4070
---
 .../carbondata/core/mutate/CarbonUpdateUtil.java   |  21 +++-
 .../carbondata/core/util/path/CarbonTablePath.java |   2 +-
 .../hadoop/api/CarbonFileInputFormat.java  |   5 +-
 .../hadoop/api/CarbonOutputCommitter.java  |   6 +-
 .../hadoop/api/CarbonTableOutputFormat.java|   7 +-
 .../spark/rdd/CarbonDataRDDFactory.scala   |   2 +-
 .../command/management/CarbonAddLoadCommand.scala  | 126 -
 .../command/mutation/DeleteExecution.scala |   2 +-
 .../mutation/merge/CarbonMergeDataSetCommand.scala |   2 +-
 .../testsuite/addsegment/AddSegmentTestCase.scala  | 103 -
 .../processing/merger/CarbonDataMergerUtil.java|   2 +-
 .../processing/util/CarbonLoaderUtil.java  |  23 ++--
 12 files changed, 275 insertions(+), 26 deletions(-)

diff --git 
a/core/src/main/java/org/apache/carbondata/core/mutate/CarbonUpdateUtil.java 
b/core/src/main/java/org/apache/carbondata/core/mutate/CarbonUpdateUtil.java
index c5296dc..e78b630 100644
--- a/core/src/main/java/org/apache/carbondata/core/mutate/CarbonUpdateUtil.java
+++ b/core/src/main/java/org/apache/carbondata/core/mutate/CarbonUpdateUtil.java
@@ -47,6 +47,7 @@ import org.apache.carbondata.core.util.CarbonProperties;
 import org.apache.carbondata.core.util.CarbonUtil;
 import org.apache.carbondata.core.util.path.CarbonTablePath;
 
+import org.apache.commons.collections.CollectionUtils;
 import org.apache.hadoop.fs.Path;
 import org.apache.log4j.Logger;
 
@@ -145,9 +146,17 @@ public class CarbonUpdateUtil {
* @return
*/
   public static boolean updateSegmentStatus(List 
updateDetailsList,
-  CarbonTable table, String updateStatusFileIdentifier, boolean 
isCompaction) {
+  CarbonTable table, String updateStatusFileIdentifier, boolean 
isCompaction,
+  boolean isForceWrite) {
 boolean status = false;
 SegmentUpdateStatusManager segmentUpdateStatusManager = new 
SegmentUpdateStatusManager(table);
+if (isForceWrite && !CollectionUtils.isEmpty(updateDetailsList)) {
+  String segId = String.valueOf(SegmentStatusManager
+  
.createNewSegmentId(segmentUpdateStatusManager.getLoadMetadataDetails()));
+  for (SegmentUpdateDetails detail : updateDetailsList) {
+detail.setSegmentName(segId);
+  }
+}
 ICarbonLock updateLock = 
segmentUpdateStatusManager.getTableUpdateStatusLock();
 boolean lockStatus = false;
 
@@ -178,8 +187,14 @@ public class CarbonUpdateUtil {
 updateDetailsValidSeg.add(updateDetails);
   }
 }
-segmentUpdateStatusManager
-.writeLoadDetailsIntoFile(updateDetailsValidSeg, 
updateStatusFileIdentifier);
+// In case of ForceWrite, write the segmentUpdateDetails to the 
tableUpdateStatus file
+// without any validation of segments.
+if (isForceWrite) {
+  segmentUpdateStatusManager.writeLoadDetailsIntoFile(oldList, 
updateStatusFileIdentifier);
+} else {
+  segmentUpdateStatusManager
+  .writeLoadDetailsIntoFile(updateDetailsValidSeg, 
updateStatusFileIdentifier);
+}
 status = true;
   } else {
 LOGGER.error("Not able to acquire the segment update lock.");
diff --git 
a/core/src/main/java/org/apache/carbondata/core/util/path/CarbonTablePath.java 
b/core/src/main/java/org/apache/carbondata/core/util/path/CarbonTablePath.java
index 458ebdf..9967723 100644
--- 
a/core/src/main/java/org/apache/carbondata/core/util/path/CarbonTablePath.java
+++ 
b/core/src/main/java/org/apache/carbondata/core/util/path/CarbonTablePath.java
@@ -43,7 +43,7 @@ public class CarbonTablePath {
   private static final String FACT_DIR = "Fact";
   public static final String SEGMENT_PREFIX = "Segment_";
   private static final String PARTITION_PREFIX = "Part";
-  private static final String DATA_PART_PREFIX = "p

[carbondata] branch master updated: [CARBONDATA-4113] Partition prune and cache fix when carbon.read.partition.hive.direct is disabled

2021-02-02 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new 440ab03  [CARBONDATA-4113] Partition prune and cache fix when 
carbon.read.partition.hive.direct is disabled
440ab03 is described below

commit 440ab03822008509bc6547a6499049137eed5a72
Author: ShreelekhyaG 
AuthorDate: Thu Jan 28 15:40:27 2021 +0530

[CARBONDATA-4113] Partition prune and cache fix when 
carbon.read.partition.hive.direct is disabled

Why is this PR needed?
When carbon.read.partition.hive.direct is false then select queries on
partition table result is invalid . For a single partition, partition
values are appended to form the wrong path when loaded by the same segment.
Ex: For partition on column b, path: /tablepath/b=1/b=2

What changes were proposed in this PR?
In PartitionCacheManager, changes made to handle single and multiple 
partitions.
Encoded the URI path to handle space values in the string.

This closes #4084
---
 .../apache/spark/util/PartitionCacheManager.scala  | 27 --
 .../StandardPartitionTableLoadingTestCase.scala| 17 ++
 2 files changed, 32 insertions(+), 12 deletions(-)

diff --git 
a/integration/spark/src/main/scala/org/apache/spark/util/PartitionCacheManager.scala
 
b/integration/spark/src/main/scala/org/apache/spark/util/PartitionCacheManager.scala
index 39f33e5..2945dfa 100644
--- 
a/integration/spark/src/main/scala/org/apache/spark/util/PartitionCacheManager.scala
+++ 
b/integration/spark/src/main/scala/org/apache/spark/util/PartitionCacheManager.scala
@@ -23,6 +23,7 @@ import java.util
 import scala.collection.JavaConverters._
 import scala.collection.mutable
 
+import org.apache.commons.httpclient.util.URIUtil
 import org.apache.log4j.Logger
 import org.apache.spark.sql.catalyst.catalog.{CatalogStorageFormat, 
CatalogTablePartition}
 
@@ -96,7 +97,7 @@ object PartitionCacheManager extends Cache[PartitionCacheKey,
 } else if (invalidSegmentMap != null && invalidSegmentMap.nonEmpty) {
   CACHE.remove(identifier.tableId)
 }
-finalCache.values.flatMap(_._1).toList.asJava
+finalCache.values.flatMap(_._1).toSet.toList.asJava
   }
 
   override def getAll(keys: util.List[PartitionCacheKey]):
@@ -115,18 +116,20 @@ object PartitionCacheManager extends 
Cache[PartitionCacheKey,
 
   private def readPartition(identifier: PartitionCacheKey, segmentFilePath: 
String) = {
 val segmentFile = SegmentFileStore.readSegmentFile(segmentFilePath)
-val partitionPath = new mutable.StringBuilder()
 var partitionSpec: Map[String, String] = Map()
-segmentFile.getLocationMap.values().asScala
-  .flatMap(_.getPartitions.asScala).toSet.foreach { uniquePartition: 
String =>
-  
partitionPath.append(CarbonCommonConstants.FILE_SEPARATOR).append(uniquePartition)
-  val partitionSplit = uniquePartition.split("=")
-  partitionSpec = partitionSpec. +(partitionSplit(0) -> partitionSplit(1))
-}
-Seq(CatalogTablePartition(partitionSpec,
-  CatalogStorageFormat(
-Some(new URI(identifier.tablePath + partitionPath)),
-None, None, None, compressed = false, Map(
+segmentFile.getLocationMap.keySet().asScala
+  .map { uniquePartition: String =>
+val partitionSplit = uniquePartition.substring(1)
+  .split(CarbonCommonConstants.FILE_SEPARATOR)
+val storageFormat = CatalogStorageFormat(
+  Some(new URI(URIUtil.encodeQuery(identifier.tablePath + 
uniquePartition))),
+  None, None, None, compressed = false, Map())
+partitionSplit.foreach(partition => {
+  val partitionArray = partition.split("=")
+  partitionSpec = partitionSpec. + (partitionArray(0) -> 
partitionArray(1))
+})
+CatalogTablePartition(partitionSpec, storageFormat)
+  }.toSeq
   }
 
   override def put(key: PartitionCacheKey,
diff --git 
a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/standardpartition/StandardPartitionTableLoadingTestCase.scala
 
b/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/standardpartition/StandardPartitionTableLoadingTestCase.scala
index 6ab5e51..c8a0926 100644
--- 
a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/standardpartition/StandardPartitionTableLoadingTestCase.scala
+++ 
b/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/standardpartition/StandardPartitionTableLoadingTestCase.scala
@@ -633,6 +633,23 @@ class StandardPartitionTableLoadingTestCase extends 
QueryTest with BeforeAndAfte
 
CarbonProperties.getInstance().addProperty("carbon.read.partition.hive.direct", 
"true")
   }
 
+  test("test read hive partitions alternatively a

[carbondata] branch master updated: [CARBONDATA-4097] ColumnVectors should not be initialized as ColumnVectorWrapperDirect for alter tables

2021-01-26 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new f5e35cd  [CARBONDATA-4097] ColumnVectors should not be initialized as 
ColumnVectorWrapperDirect for alter tables
f5e35cd is described below

commit f5e35cd039109af3bbadd2375205e3c608052027
Author: Karan980 
AuthorDate: Tue Dec 22 18:51:23 2020 +0530

[CARBONDATA-4097] ColumnVectors should not be initialized as
ColumnVectorWrapperDirect for alter tables

Why is this PR needed?
Direct filling of column vectors is not allowed for alter tables,
But its column vectors were getting initialized as 
ColumnVectorWrapperDirect.

What changes were proposed in this PR?
Changed the initialization of column vectors to ColumnVectorWrapper
for alter tables.

This closes #4062
---
 .../scan/executor/impl/AbstractQueryExecutor.java  |  4 ++--
 .../core/scan/executor/util/RestructureUtil.java   |  7 +--
 .../scan/executor/util/RestructureUtilTest.java|  8 ++--
 .../vectorreader/AddColumnTestCases.scala  | 23 ++
 4 files changed, 36 insertions(+), 6 deletions(-)

diff --git 
a/core/src/main/java/org/apache/carbondata/core/scan/executor/impl/AbstractQueryExecutor.java
 
b/core/src/main/java/org/apache/carbondata/core/scan/executor/impl/AbstractQueryExecutor.java
index 3822f3e..72e3b31 100644
--- 
a/core/src/main/java/org/apache/carbondata/core/scan/executor/impl/AbstractQueryExecutor.java
+++ 
b/core/src/main/java/org/apache/carbondata/core/scan/executor/impl/AbstractQueryExecutor.java
@@ -445,7 +445,7 @@ public abstract class AbstractQueryExecutor implements 
QueryExecutor {
 blockExecutionInfo.getActualQueryDimensions(), 
segmentProperties.getDimensions(),
 segmentProperties.getComplexDimensions(),
 blockExecutionInfo.getActualQueryMeasures().length,
-queryModel.getTable().getTableInfo().isTransactionalTable());
+queryModel.getTable().getTableInfo().isTransactionalTable(), 
queryModel);
 boolean isStandardTable = 
CarbonUtil.isStandardCarbonTable(queryModel.getTable());
 String blockId = CarbonUtil
 .getBlockId(queryModel.getAbsoluteTableIdentifier(), filePath, 
segment.getSegmentNo(),
@@ -461,7 +461,7 @@ public abstract class AbstractQueryExecutor implements 
QueryExecutor {
 List projectionMeasures = RestructureUtil
 .createMeasureInfoAndGetCurrentBlockQueryMeasures(blockExecutionInfo,
 blockExecutionInfo.getActualQueryMeasures(), 
segmentProperties.getMeasures(),
-queryModel.getTable().getTableInfo().isTransactionalTable());
+queryModel.getTable().getTableInfo().isTransactionalTable(), 
queryModel);
 blockExecutionInfo.setProjectionMeasures(
 projectionMeasures.toArray(new 
ProjectionMeasure[projectionMeasures.size()]));
 blockExecutionInfo.setDataBlock(blockIndex);
diff --git 
a/core/src/main/java/org/apache/carbondata/core/scan/executor/util/RestructureUtil.java
 
b/core/src/main/java/org/apache/carbondata/core/scan/executor/util/RestructureUtil.java
index d08ff1b..2f2cdfe 100644
--- 
a/core/src/main/java/org/apache/carbondata/core/scan/executor/util/RestructureUtil.java
+++ 
b/core/src/main/java/org/apache/carbondata/core/scan/executor/util/RestructureUtil.java
@@ -67,7 +67,7 @@ public class RestructureUtil {
   public static List 
createDimensionInfoAndGetCurrentBlockQueryDimension(
   BlockExecutionInfo blockExecutionInfo, ProjectionDimension[] 
queryDimensions,
   List tableBlockDimensions, List 
tableComplexDimension,
-  int measureCount, boolean isTransactionalTable) {
+  int measureCount, boolean isTransactionalTable, QueryModel queryModel) {
 List presentDimension =
 new ArrayList<>(CarbonCommonConstants.DEFAULT_COLLECTION_SIZE);
 boolean[] isDimensionExists = new boolean[queryDimensions.length];
@@ -133,6 +133,7 @@ public class RestructureUtil {
 if (!isDimensionExists[dimIndex]) {
   defaultValues[dimIndex] = 
validateAndGetDefaultValue(queryDimension.getDimension());
   blockExecutionInfo.setRestructuredBlock(true);
+  queryModel.setDirectVectorFill(false);
   // set the flag to say whether a new dictionary column or no 
dictionary column
   // has been added. This will be useful after restructure for 
compaction scenarios where
   // newly added columns data need to be filled
@@ -401,7 +402,8 @@ public class RestructureUtil {
*/
   public static List 
createMeasureInfoAndGetCurrentBlockQueryMeasures(
   BlockExecutionInfo blockExecutionInfo, ProjectionMeasure[] queryMeasures,
-  List currentBlockMeasures, boolean isTransactionalTable) {
+  List currentBlockMeasures, boolean isTransactionalTable,
+  QueryModel quer

[carbondata] branch master updated: [CARBONDATA-4096] SDK read fails from cluster and sdk read filter query on sort column giving wrong result with IndexServer

2021-01-22 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new 7585656  [CARBONDATA-4096] SDK read fails from cluster and sdk read 
filter query on sort column giving wrong result with IndexServer
7585656 is described below

commit 7585656b565d7eb76802fafcc84a28daba89b25e
Author: ShreelekhyaG 
AuthorDate: Tue Dec 22 18:44:43 2020 +0530

[CARBONDATA-4096] SDK read fails from cluster and sdk read filter query on
sort column giving wrong result with IndexServer

Why is this PR needed?
1. Create a table and read from sdk written files fails in cluster with
java.nio.file.NoSuchFileException: 
hdfs:/hacluster/user/hive/warehouse/carbon.store/default/sdk.
2. After fixing the above path issue, filter query on sort column gives
the wrong result with IndexServer.

What changes were proposed in this PR?
1. In getAllDeleteDeltaFiles , used CarbonFiles.listFiles instead of 
Files.walk
to handle custom file types.
2. In PruneWithFilter , isResolvedOnSegment is used in filterResolver step.
Have set table and expression on executor side, so indexserver can use this
in filterResolver step.

This closes #4064
---
 .../carbondata/core/index/IndexInputFormat.java|  8 ++--
 .../hadoop/api/CarbonFileInputFormat.java  | 23 --
 2 files changed, 19 insertions(+), 12 deletions(-)

diff --git 
a/core/src/main/java/org/apache/carbondata/core/index/IndexInputFormat.java 
b/core/src/main/java/org/apache/carbondata/core/index/IndexInputFormat.java
index dbb5b4f..072dbbc 100644
--- a/core/src/main/java/org/apache/carbondata/core/index/IndexInputFormat.java
+++ b/core/src/main/java/org/apache/carbondata/core/index/IndexInputFormat.java
@@ -154,8 +154,12 @@ public class IndexInputFormat extends 
FileInputFormat
 if (indexLevel == null) {
   TableIndex defaultIndex = IndexStoreManager.getInstance()
   .getIndex(table, 
distributable.getDistributable().getIndexSchema());
-  blocklets = defaultIndex
-  .prune(segmentsToLoad, new IndexFilter(filterResolverIntf), 
partitions);
+  IndexFilter filter = new IndexFilter(filterResolverIntf);
+  filter.setTable(table);
+  if (filterResolverIntf != null) {
+filter.setExpression(filterResolverIntf.getFilterExpression());
+  }
+  blocklets = defaultIndex.prune(segmentsToLoad, filter, partitions);
   blocklets = IndexUtil
   .pruneIndexes(table, filterResolverIntf, segmentsToLoad, 
partitions, blocklets,
   indexChooser);
diff --git 
a/hadoop/src/main/java/org/apache/carbondata/hadoop/api/CarbonFileInputFormat.java
 
b/hadoop/src/main/java/org/apache/carbondata/hadoop/api/CarbonFileInputFormat.java
index 2a655e8..91116b4 100644
--- 
a/hadoop/src/main/java/org/apache/carbondata/hadoop/api/CarbonFileInputFormat.java
+++ 
b/hadoop/src/main/java/org/apache/carbondata/hadoop/api/CarbonFileInputFormat.java
@@ -20,21 +20,17 @@ package org.apache.carbondata.hadoop.api;
 import java.io.File;
 import java.io.IOException;
 import java.io.Serializable;
-import java.nio.file.Files;
-import java.nio.file.Path;
-import java.nio.file.Paths;
 import java.util.ArrayList;
 import java.util.Comparator;
 import java.util.LinkedList;
 import java.util.List;
 import java.util.regex.Pattern;
-import java.util.stream.Collectors;
-import java.util.stream.Stream;
 
 import org.apache.carbondata.common.annotations.InterfaceAudience;
 import org.apache.carbondata.common.annotations.InterfaceStability;
 import org.apache.carbondata.core.constants.CarbonCommonConstants;
 import org.apache.carbondata.core.datastore.filesystem.CarbonFile;
+import org.apache.carbondata.core.datastore.filesystem.CarbonFileFilter;
 import org.apache.carbondata.core.datastore.impl.FileFactory;
 import org.apache.carbondata.core.index.IndexFilter;
 import org.apache.carbondata.core.index.Segment;
@@ -265,11 +261,18 @@ public class CarbonFileInputFormat extends 
CarbonInputFormat implements Se
   }
 
   private List getAllDeleteDeltaFiles(String path) {
-List deltaFiles = null;
-try (Stream walk = Files.walk(Paths.get(path))) {
-  deltaFiles = walk.map(x -> x.toString())
-  .filter(f -> f.endsWith(CarbonCommonConstants.DELETE_DELTA_FILE_EXT))
-  .collect(Collectors.toList());
+List deltaFiles = new ArrayList<>();
+try {
+  FileFactory.getCarbonFile(path).listFiles(true, new CarbonFileFilter() {
+@Override
+public boolean accept(CarbonFile file) {
+  if 
(file.getName().endsWith(CarbonCommonConstants.DELETE_DELTA_FILE_EXT)) {
+deltaFiles.add(file.getAbsolutePath());
+return true;
+  }
+  

[carbondata] branch master updated: [CARBONDATA-4055]Fix creation of empty segment directory and meta entry when there is no update/insert data

2021-01-20 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new aa2121e  [CARBONDATA-4055]Fix creation of empty segment directory and 
meta entry when there is no update/insert data
aa2121e is described below

commit aa2121e761f620e60470f5f856bc29141777f97a
Author: akashrn5 
AuthorDate: Mon Nov 23 19:18:06 2020 +0530

[CARBONDATA-4055]Fix creation of empty segment directory and meta
entry when there is no update/insert data

Why is this PR needed?
1. After #3999 when an update happens on the table, a new segment
is created for updated data. But when there is no data to update,
still the segments are created and the table status has in progress
entries for those empty segments. This leads to unnecessary segment
dirs and an increase in table status entries.
2. after this, clean files don't clean these empty segments.
3. when the source table do not have data, CTAS will result in same
problem mentioned.

What changes were proposed in this PR?
when the data is not present during update, make the segment as marked
for delete so that the clean files take care to delete the segment,
for cats already handled, added test cases.

This closes #4018
---
 .../org/apache/carbondata/spark/rdd/CarbonDataRDDFactory.scala |  3 ++-
 .../spark/testsuite/createTable/TestCreateTableAsSelect.scala  | 10 ++
 .../spark/testsuite/iud/UpdateCarbonTableTestCase.scala|  8 +++-
 3 files changed, 19 insertions(+), 2 deletions(-)

diff --git 
a/integration/spark/src/main/scala/org/apache/carbondata/spark/rdd/CarbonDataRDDFactory.scala
 
b/integration/spark/src/main/scala/org/apache/carbondata/spark/rdd/CarbonDataRDDFactory.scala
index 09059b1..f62aa64 100644
--- 
a/integration/spark/src/main/scala/org/apache/carbondata/spark/rdd/CarbonDataRDDFactory.scala
+++ 
b/integration/spark/src/main/scala/org/apache/carbondata/spark/rdd/CarbonDataRDDFactory.scala
@@ -315,7 +315,8 @@ object CarbonDataRDDFactory {
 try {
   if (!carbonLoadModel.isCarbonTransactionalTable || 
segmentLock.lockWithRetries()) {
 if (updateModel.isDefined && dataFrame.get.rdd.isEmpty()) {
-  // if the rowToBeUpdated is empty, do nothing
+  // if the rowToBeUpdated is empty, mark created segment as marked 
for delete and return
+  CarbonLoaderUtil.updateTableStatusForFailure(carbonLoadModel, "")
 } else {
   status = if (scanResultRdd.isDefined) {
 val colSchema = carbonLoadModel
diff --git 
a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/createTable/TestCreateTableAsSelect.scala
 
b/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/createTable/TestCreateTableAsSelect.scala
index 8384939..0bd8f73 100644
--- 
a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/createTable/TestCreateTableAsSelect.scala
+++ 
b/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/createTable/TestCreateTableAsSelect.scala
@@ -27,7 +27,9 @@ import org.scalatest.BeforeAndAfterAll
 import org.apache.carbondata.core.constants.CarbonCommonConstants
 import org.apache.carbondata.core.datastore.filesystem.CarbonFile
 import org.apache.carbondata.core.datastore.impl.FileFactory
+import org.apache.carbondata.core.statusmanager.SegmentStatusManager
 import org.apache.carbondata.core.util.CarbonProperties
+import org.apache.carbondata.core.util.path.CarbonTablePath
 
 /**
  * test functionality for create table as select command
@@ -209,6 +211,7 @@ class TestCreateTableAsSelect extends QueryTest with 
BeforeAndAfterAll {
 
   test("test create table as select with " +
"where clause in select from hive/orc table that does not return data") 
{
+
CarbonProperties.getInstance().addProperty("carbon.clean.file.force.allowed", 
"true")
 sql("DROP TABLE IF EXISTS ctas_select_where_orc")
 sql(
   """
@@ -216,8 +219,15 @@ class TestCreateTableAsSelect extends QueryTest with 
BeforeAndAfterAll {
 | STORED AS carbondata
 | AS SELECT * FROM orc_ctas_test
 | where key=300""".stripMargin)
+val carbonTable = CarbonEnv.getCarbonTable(Some("default"), 
"ctas_select_where_orc")(sqlContext
+  .sparkSession)
+val segmentPath = 
FileFactory.getCarbonFile(CarbonTablePath.getSegmentPath(carbonTable
+  .getTablePath, "0"))
+sql("clean files for table ctas_select_where_orc options('force' = 
'true')")
+assert(!segmentPath.exists())
 checkAnswer(sql("SELECT * FROM ctas_select_where_orc"),
   sql("SELECT * FROM orc_ctas_test where key=300"))
+
CarbonProperti

[carbondata] branch master updated: [CARBONDATA-4080] Wrong results for select count on invalid segments

2020-12-17 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new 35c73e6  [CARBONDATA-4080] Wrong results for select count on invalid 
segments
35c73e6 is described below

commit 35c73e61c55f39f1bf38105e6bf32ff3673396fb
Author: akkio-97 
AuthorDate: Wed Dec 9 15:15:51 2020 +0530

[CARBONDATA-4080] Wrong results for select count on invalid segments

Why is this PR needed?
Wrong results for
1. select count on marked for delete segment
2. select count on compacted segment
Issue comes only when the user explicitly sets deleted/compacted
segments using the property carbon.input.segments.
As select * on such segments gives 0 rows as output, in order to
maintain consistency, select count should also give 0 rows.

What changes were proposed in this PR?
So this PR focuses on considering only valid segments to be accessed
during select count query.

This closes #4050
---
 .../hadoop/api/CarbonTableInputFormat.java |  15 ++-
 .../segmentreading/TestSegmentReading.scala| 104 +
 2 files changed, 117 insertions(+), 2 deletions(-)

diff --git 
a/hadoop/src/main/java/org/apache/carbondata/hadoop/api/CarbonTableInputFormat.java
 
b/hadoop/src/main/java/org/apache/carbondata/hadoop/api/CarbonTableInputFormat.java
index a2c162e..f24a2f1 100644
--- 
a/hadoop/src/main/java/org/apache/carbondata/hadoop/api/CarbonTableInputFormat.java
+++ 
b/hadoop/src/main/java/org/apache/carbondata/hadoop/api/CarbonTableInputFormat.java
@@ -55,6 +55,7 @@ import 
org.apache.carbondata.core.readcommitter.TableStatusReadCommittedScope;
 import org.apache.carbondata.core.scan.filter.resolver.FilterResolverIntf;
 import org.apache.carbondata.core.statusmanager.FileFormat;
 import org.apache.carbondata.core.statusmanager.LoadMetadataDetails;
+import org.apache.carbondata.core.statusmanager.SegmentStatus;
 import org.apache.carbondata.core.statusmanager.SegmentStatusManager;
 import org.apache.carbondata.core.statusmanager.SegmentUpdateStatusManager;
 import org.apache.carbondata.core.statusmanager.StageInputCollector;
@@ -220,7 +221,7 @@ public class CarbonTableInputFormat extends 
CarbonInputFormat {
* `INPUT_SEGMENT_NUMBERS` in job configuration
*/
   private List getFilteredSegment(JobContext job, List 
validSegments,
-  boolean validationRequired, ReadCommittedScope readCommittedScope) {
+  boolean validationRequired, ReadCommittedScope readCommittedScope) 
throws IOException {
 Segment[] segmentsToAccess = getSegmentsToAccess(job, readCommittedScope);
 if (segmentsToAccess.length == 0 || 
segmentsToAccess[0].getSegmentNo().equalsIgnoreCase("*")) {
   return validSegments;
@@ -241,7 +242,17 @@ public class CarbonTableInputFormat extends 
CarbonInputFormat {
 }
 if (!validationRequired && filteredSegmentToAccess.size() != 
segmentToAccessMap.size()) {
   for (Segment segment : segmentToAccessMap.values()) {
-if (!filteredSegmentToAccess.containsKey(segment.getSegmentNo())) {
+boolean isSegmentValid = true;
+LoadMetadataDetails[] segmentList = 
readCommittedScope.getSegmentList();
+for (LoadMetadataDetails validSegment : segmentList) {
+  if (validSegment.getLoadName().equals(segment.getSegmentNo()) && (
+  
validSegment.getSegmentStatus().equals(SegmentStatus.MARKED_FOR_DELETE)
+  || 
validSegment.getSegmentStatus().equals(SegmentStatus.COMPACTED))) {
+isSegmentValid = false;
+break;
+  }
+}
+if (isSegmentValid && 
!filteredSegmentToAccess.containsKey(segment.getSegmentNo())) {
   filteredSegmentToAccess.put(segment.getSegmentNo(), segment);
 }
   }
diff --git 
a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/segmentreading/TestSegmentReading.scala
 
b/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/segmentreading/TestSegmentReading.scala
index 8ce8cc5..5c0336c 100644
--- 
a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/segmentreading/TestSegmentReading.scala
+++ 
b/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/segmentreading/TestSegmentReading.scala
@@ -22,6 +22,8 @@ import org.apache.spark.sql.test.util.QueryTest
 import org.scalatest.BeforeAndAfterAll
 
 import 
org.apache.carbondata.common.exceptions.sql.MalformedCarbonCommandException
+import org.apache.carbondata.core.util.CarbonProperties
+import org.apache.carbondata.view.rewrite.TestUtil
 
 /**
  * Created by rahul on 19/9/17.
@@ -420,4 +422,106 @@ class TestSegmentReading extends QueryTest with 
BeforeAndAfterAll {
 
 sql("set spark.sql.adaptive.enabled=false")
   }
+
+  test(&

[carbondata] branch master updated: [CARBONDATA-4077] Refactor and Fix Insert into partition issue with FileMergeSortComparator

2020-12-17 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new 29ecd5f  [CARBONDATA-4077] Refactor and Fix Insert into partition 
issue with FileMergeSortComparator
29ecd5f is described below

commit 29ecd5fee7af97633b756bfd944251a18dc76c18
Author: Indhumathi27 
AuthorDate: Sat Dec 5 19:35:58 2020 +0530

[CARBONDATA-4077] Refactor and Fix Insert into partition issue with 
FileMergeSortComparator

Why is this PR needed?
From PR-3995 changes, insert into partition flow scenario is missed. Using 
Map for getting
Dict/No-Dict sort column info during final sort task will affect load 
performance,
if number of sort columns is more.

What changes were proposed in this PR?
Handled the insert into partition flow
Refactored the code, to use list of only Dict/No-Dict sort column indexes 
instead of
Map to fix performance issue.

This closes #4039
---
 .../query/SecondaryIndexQueryResultProcessor.java  |   2 -
 .../unsafe/holder/UnsafeFinalMergePageHolder.java  |   6 +-
 .../sort/unsafe/holder/UnsafeInmemoryHolder.java   |   6 +-
 .../holder/UnsafeSortTempFileChunkHolder.java  |   6 +-
 .../merger/CompactionResultSortProcessor.java  |  11 +-
 .../sort/sortdata/FileMergeSortComparator.java | 100 
 .../processing/sort/sortdata/SortParameters.java   |  69 +++
 .../sort/sortdata/SortTempFileChunkHolder.java |  12 +-
 .../processing/sort/sortdata/TableFieldStat.java   |  27 +++--
 .../processing/util/CarbonDataProcessorUtil.java   | 127 ++---
 .../sort/sortdata/FileMergeSortComparatorTest.java |  27 ++---
 11 files changed, 239 insertions(+), 154 deletions(-)

diff --git 
a/integration/spark/src/main/java/org/apache/spark/sql/secondaryindex/query/SecondaryIndexQueryResultProcessor.java
 
b/integration/spark/src/main/java/org/apache/spark/sql/secondaryindex/query/SecondaryIndexQueryResultProcessor.java
index 4d045f0..41a5c43 100644
--- 
a/integration/spark/src/main/java/org/apache/spark/sql/secondaryindex/query/SecondaryIndexQueryResultProcessor.java
+++ 
b/integration/spark/src/main/java/org/apache/spark/sql/secondaryindex/query/SecondaryIndexQueryResultProcessor.java
@@ -528,8 +528,6 @@ public class SecondaryIndexQueryResultProcessor {
 CarbonCommonConstants.FILE_SEPARATOR, 
CarbonCommonConstants.SORT_TEMP_FILE_LOCATION);
 sortParameters.setNoDictionarySortColumn(
 CarbonDataProcessorUtil.getNoDictSortColMapping(indexTable));
-sortParameters.setSortColumnSchemaOrderMap(
-CarbonDataProcessorUtil.getSortColSchemaOrderMapping(indexTable));
 finalMerger = new SingleThreadFinalSortFilesMerger(sortTempFileLocation,
 indexTable.getTableName(), sortParameters);
   }
diff --git 
a/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/holder/UnsafeFinalMergePageHolder.java
 
b/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/holder/UnsafeFinalMergePageHolder.java
index 717bb91..7abd3a9 100644
--- 
a/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/holder/UnsafeFinalMergePageHolder.java
+++ 
b/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/holder/UnsafeFinalMergePageHolder.java
@@ -61,10 +61,10 @@ public class UnsafeFinalMergePageHolder implements 
SortTempChunkHolder {
 }
 this.noDictDataType = rowPages[0].getTableFieldStat().getNoDictDataType();
 LOGGER.info("Processing unsafe inmemory rows page with size : " + 
actualSize);
-this.comparator = new 
FileMergeSortComparator(tableFieldStat.getIsSortColNoDictFlags(),
-tableFieldStat.getNoDictSchemaDataType(),
+this.comparator = new 
FileMergeSortComparator(tableFieldStat.getNoDictSchemaDataType(),
 tableFieldStat.getNoDictSortColumnSchemaOrderMapping(),
-tableFieldStat.getSortColSchemaOrderMap());
+tableFieldStat.getNoDictSortColIdxSchemaOrderMapping(),
+tableFieldStat.getDictSortColIdxSchemaOrderMapping());
   }
 
   public boolean hasNext() {
diff --git 
a/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/holder/UnsafeInmemoryHolder.java
 
b/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/holder/UnsafeInmemoryHolder.java
index a46811f..a21e802 100644
--- 
a/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/holder/UnsafeInmemoryHolder.java
+++ 
b/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/holder/UnsafeInmemoryHolder.java
@@ -48,10 +48,10 @@ public class UnsafeInmemoryHolder implements 
SortTempChunkHolder {
 this.rowPage = rowPage;
 LOGGER.info("Processing unsafe inmemory rows page with size : " + 
actualSize);

[carbondata] branch master updated: [CARBONDATA-4050]Avoid redundant RPC calls to get file status when CarbonFile is instantiated with fileStatus constructor

2020-12-03 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new c5f464f  [CARBONDATA-4050]Avoid redundant RPC calls to get file status 
when CarbonFile is instantiated with fileStatus constructor
c5f464f is described below

commit c5f464fab7117000aa76150c30a616ca02b5d443
Author: Venu Reddy 
AuthorDate: Thu Nov 19 00:33:20 2020 +0530

[CARBONDATA-4050]Avoid redundant RPC calls to get file status when 
CarbonFile is
instantiated with fileStatus constructor

Why is this PR needed?
In createCarbonDataFileBlockMetaInfoMapping method, we get list of 
carbondata
files in the segment, loop through all the carbon files and make a map of
fileNameToMetaInfoMapping

In that carbon files loop, if the file is of AbstractDFSCarbonFile type, we 
get
the org.apache.hadoop.fs.FileStatus thrice for each file. And the method to 
get
file status is an RPC call(fileSystem.getFileStatus(path)). It takes ~2ms 
in the
cluster for each call. Thus, incur an overhead of ~6ms per file. So overall 
driver
side query processing time has increased significantly when there are more 
carbon
files. Hence caused TPC-DS queries performance degradation.

What changes were proposed in this PR?
Avoided redundant RPC calls to get file status in getAbsolutePath(), 
getSize() and
getLocations() methods when CarbonFile is instantiated with FileStatus 
constructor

This closes #4010
---
 .../core/datastore/filesystem/AbstractDFSCarbonFile.java  | 11 ++-
 1 file changed, 10 insertions(+), 1 deletion(-)

diff --git 
a/core/src/main/java/org/apache/carbondata/core/datastore/filesystem/AbstractDFSCarbonFile.java
 
b/core/src/main/java/org/apache/carbondata/core/datastore/filesystem/AbstractDFSCarbonFile.java
index 338076e..25337c7 100644
--- 
a/core/src/main/java/org/apache/carbondata/core/datastore/filesystem/AbstractDFSCarbonFile.java
+++ 
b/core/src/main/java/org/apache/carbondata/core/datastore/filesystem/AbstractDFSCarbonFile.java
@@ -107,6 +107,9 @@ public abstract class AbstractDFSCarbonFile implements 
CarbonFile {
   @Override
   public String getAbsolutePath() {
 try {
+  if (fileStatus != null) {
+return fileStatus.getPath().toString();
+  }
   return fileSystem.getFileStatus(path).getPath().toString();
 } catch (IOException e) {
   throw new CarbonFileException("Unable to get file status: ", e);
@@ -155,6 +158,9 @@ public abstract class AbstractDFSCarbonFile implements 
CarbonFile {
   @Override
   public long getSize() {
 try {
+  if (fileStatus != null) {
+return fileStatus.getLen();
+  }
   return fileSystem.getFileStatus(path).getLen();
 } catch (IOException e) {
   throw new CarbonFileException("Unable to get file status for " + 
path.toString(), e);
@@ -541,7 +547,10 @@ public abstract class AbstractDFSCarbonFile implements 
CarbonFile {
   @Override
   public String[] getLocations() throws IOException {
 BlockLocation[] blkLocations;
-FileStatus fileStatus = fileSystem.getFileStatus(path);
+FileStatus fileStatus = this.fileStatus;
+if (fileStatus == null) {
+  fileStatus = fileSystem.getFileStatus(path);
+}
 if (fileStatus instanceof LocatedFileStatus) {
   blkLocations = ((LocatedFileStatus) fileStatus).getBlockLocations();
 } else {



[carbondata] branch master updated: [CARBONDATA-4046] Handled multiple partition columns for partition cache

2020-12-03 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new b379ebe  [CARBONDATA-4046] Handled multiple partition columns for 
partition cache
b379ebe is described below

commit b379ebec3cc4f251f784f704cefff345ef17c032
Author: Nihal ojha 
AuthorDate: Tue Nov 3 10:47:02 2020 +0530

[CARBONDATA-4046] Handled multiple partition columns for partition cache

Why is this PR needed?
1. Currently when property carbon.read.partition.hive.direct is false then
select count * fails on table which contains multiple partition columns.
2. Subtraction of the different data types.
3. If the final cache is empty and the invalid segment list is non-empty
then clear the cache.

What changes were proposed in this PR?
1. Handled multiple partition columns.
2. Handled subtraction of the different data types.
3. If the final cache is empty and the invalid segment list is non-empty 
then clear the cache.

This closes #4002
---
 .../apache/spark/util/PartitionCacheManager.scala  | 23 ++
 .../StandardPartitionTableLoadingTestCase.scala| 15 ++
 2 files changed, 30 insertions(+), 8 deletions(-)

diff --git 
a/integration/spark/src/main/scala/org/apache/spark/util/PartitionCacheManager.scala
 
b/integration/spark/src/main/scala/org/apache/spark/util/PartitionCacheManager.scala
index 411cbe2..39f33e5 100644
--- 
a/integration/spark/src/main/scala/org/apache/spark/util/PartitionCacheManager.scala
+++ 
b/integration/spark/src/main/scala/org/apache/spark/util/PartitionCacheManager.scala
@@ -79,9 +79,10 @@ object PartitionCacheManager extends Cache[PartitionCacheKey,
   segmentFilePath.getAbsolutePath), segmentFileModifiedTime))
   }
 }.toMap
+val invalidSegmentMap = validInvalidSegments.getInvalidSegments.asScala
+  .map(seg => (seg.getSegmentNo, seg)).toMap
 // remove all invalid segment entries from cache
-val finalCache = cacheablePartitionSpecs --
- 
validInvalidSegments.getInvalidSegments.asScala.map(_.getSegmentNo)
+val finalCache = cacheablePartitionSpecs -- invalidSegmentMap.keySet
 val cacheObject = CacheablePartitionSpec(finalCache)
 if (finalCache.nonEmpty) {
   // remove the existing cache as new cache values may be added.
@@ -92,6 +93,8 @@ object PartitionCacheManager extends Cache[PartitionCacheKey,
 cacheObject,
 cacheObject.getMemorySize,
 identifier.expirationTime)
+} else if (invalidSegmentMap != null && invalidSegmentMap.nonEmpty) {
+  CACHE.remove(identifier.tableId)
 }
 finalCache.values.flatMap(_._1).toList.asJava
   }
@@ -112,14 +115,18 @@ object PartitionCacheManager extends 
Cache[PartitionCacheKey,
 
   private def readPartition(identifier: PartitionCacheKey, segmentFilePath: 
String) = {
 val segmentFile = SegmentFileStore.readSegmentFile(segmentFilePath)
+val partitionPath = new mutable.StringBuilder()
+var partitionSpec: Map[String, String] = Map()
 segmentFile.getLocationMap.values().asScala
-  .flatMap(_.getPartitions.asScala).toSet.map { uniquePartition: String =>
+  .flatMap(_.getPartitions.asScala).toSet.foreach { uniquePartition: 
String =>
+  
partitionPath.append(CarbonCommonConstants.FILE_SEPARATOR).append(uniquePartition)
   val partitionSplit = uniquePartition.split("=")
-  val storageFormat = CatalogStorageFormat(
-Some(new URI(identifier.tablePath + "/" + uniquePartition)),
-None, None, None, compressed = false, Map())
-  CatalogTablePartition(Map(partitionSplit(0) -> partitionSplit(1)), 
storageFormat)
-}.toSeq
+  partitionSpec = partitionSpec. +(partitionSplit(0) -> partitionSplit(1))
+}
+Seq(CatalogTablePartition(partitionSpec,
+  CatalogStorageFormat(
+Some(new URI(identifier.tablePath + partitionPath)),
+None, None, None, compressed = false, Map(
   }
 
   override def put(key: PartitionCacheKey,
diff --git 
a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/standardpartition/StandardPartitionTableLoadingTestCase.scala
 
b/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/standardpartition/StandardPartitionTableLoadingTestCase.scala
index 128274e..6ab5e51 100644
--- 
a/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/standardpartition/StandardPartitionTableLoadingTestCase.scala
+++ 
b/integration/spark/src/test/scala/org/apache/carbondata/spark/testsuite/standardpartition/StandardPartitionTableLoadingTestCase.scala
@@ -618,6 +618,21 @@ class StandardPartitionTableLoadingTestCase extends 
QueryTest with BeforeAndAfte
 
CarbonProperties.getInstance().addProperty("carbon.read.parti

[carbondata] branch master updated: [CARBONDATA-4022] Fix invalid path issue for segment added through alter add segment query

2020-12-03 Thread kunalkapoor
This is an automated email from the ASF dual-hosted git repository.

kunalkapoor pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
 new e91a645  [CARBONDATA-4022] Fix invalid path issue for segment added 
through alter add segment query
e91a645 is described below

commit e91a645f53793f54a8f8b6e1a7256dc7c29c7a09
Author: Karan980 
AuthorDate: Fri Nov 20 17:11:16 2020 +0530

[CARBONDATA-4022] Fix invalid path issue for segment added through alter 
add segment query

Why is this PR needed?
Segment Added through alter table add segment query doesn't require 
tablepath in its
filepath, because the segment is present at some external location which is 
not inside
the table.

What changes were proposed in this PR?
Filepath for normal segments (which are not added through alter table add 
segment query)
is tablePath + filePath. But for segments added through alter table add 
segment query
doesn't require tablePath, so we directly use the filePath of the segment.

This closes #4017
---
 .../org/apache/carbondata/core/indexstore/ExtendedBlocklet.java   | 8 +++-
 1 file changed, 7 insertions(+), 1 deletion(-)

diff --git 
a/core/src/main/java/org/apache/carbondata/core/indexstore/ExtendedBlocklet.java
 
b/core/src/main/java/org/apache/carbondata/core/indexstore/ExtendedBlocklet.java
index edb921b..e0b21d5 100644
--- 
a/core/src/main/java/org/apache/carbondata/core/indexstore/ExtendedBlocklet.java
+++ 
b/core/src/main/java/org/apache/carbondata/core/indexstore/ExtendedBlocklet.java
@@ -20,6 +20,7 @@ package org.apache.carbondata.core.indexstore;
 import java.io.DataInput;
 import java.io.DataOutput;
 import java.io.DataOutputStream;
+import java.io.File;
 import java.io.IOException;
 import java.util.List;
 import java.util.Map;
@@ -219,7 +220,12 @@ public class ExtendedBlocklet extends Blocklet {
 if (in.readBoolean()) {
   indexUniqueId = in.readUTF();
 }
-setFilePath(tablePath + getPath());
+String filePath = getPath();
+if (filePath.startsWith(File.separator)) {
+  setFilePath(tablePath + filePath);
+} else {
+  setFilePath(filePath);
+}
 boolean isSplitPresent = in.readBoolean();
 if (isSplitPresent) {
   // getting the length of the data



svn commit: r44383 - in /release/carbondata: ./ 2.1.0/

2020-11-08 Thread kunalkapoor
Author: kunalkapoor
Date: Sun Nov  8 16:35:05 2020
New Revision: 44383

Log:
Upload 2.1.0 release

Added:
release/carbondata/2.1.0/

release/carbondata/2.1.0/apache-carbondata-2.1.0-bin-spark2.3.4-hadoop2.7.2.jar 
  (with props)

release/carbondata/2.1.0/apache-carbondata-2.1.0-bin-spark2.3.4-hadoop2.7.2.jar.asc
   (with props)

release/carbondata/2.1.0/apache-carbondata-2.1.0-bin-spark2.3.4-hadoop2.7.2.jar.sha512

release/carbondata/2.1.0/apache-carbondata-2.1.0-bin-spark2.4.5-hadoop2.7.2.jar 
  (with props)

release/carbondata/2.1.0/apache-carbondata-2.1.0-bin-spark2.4.5-hadoop2.7.2.jar.asc
   (with props)

release/carbondata/2.1.0/apache-carbondata-2.1.0-bin-spark2.4.5-hadoop2.7.2.jar.sha512
release/carbondata/2.1.0/apache-carbondata-2.1.0-source-release.zip   (with 
props)
release/carbondata/2.1.0/apache-carbondata-2.1.0-source-release.zip.asc   
(with props)
release/carbondata/2.1.0/apache-carbondata-2.1.0-source-release.zip.md5
release/carbondata/2.1.0/apache-carbondata-2.1.0-source-release.zip.sha512
Modified:
release/carbondata/KEYS

Added: 
release/carbondata/2.1.0/apache-carbondata-2.1.0-bin-spark2.3.4-hadoop2.7.2.jar
==
Binary file - no diff available.

Propchange: 
release/carbondata/2.1.0/apache-carbondata-2.1.0-bin-spark2.3.4-hadoop2.7.2.jar
--
svn:mime-type = application/java-archive

Added: 
release/carbondata/2.1.0/apache-carbondata-2.1.0-bin-spark2.3.4-hadoop2.7.2.jar.asc
==
Binary file - no diff available.

Propchange: 
release/carbondata/2.1.0/apache-carbondata-2.1.0-bin-spark2.3.4-hadoop2.7.2.jar.asc
--
svn:mime-type = application/pgp-signature

Added: 
release/carbondata/2.1.0/apache-carbondata-2.1.0-bin-spark2.3.4-hadoop2.7.2.jar.sha512
==
--- 
release/carbondata/2.1.0/apache-carbondata-2.1.0-bin-spark2.3.4-hadoop2.7.2.jar.sha512
 (added)
+++ 
release/carbondata/2.1.0/apache-carbondata-2.1.0-bin-spark2.3.4-hadoop2.7.2.jar.sha512
 Sun Nov  8 16:35:05 2020
@@ -0,0 +1 @@
+1dee1960178f947156cae8cb66c6e66b2ae594906b16668a6532ca8913af6fc104898d344359b7e3f2d6cad445b2d49cb244499c9b1810c4be7f6a6043a1022c
  apache-carbondata-2.1.0-bin-spark2.3.4-hadoop2.7.2.jar

Added: 
release/carbondata/2.1.0/apache-carbondata-2.1.0-bin-spark2.4.5-hadoop2.7.2.jar
==
Binary file - no diff available.

Propchange: 
release/carbondata/2.1.0/apache-carbondata-2.1.0-bin-spark2.4.5-hadoop2.7.2.jar
--
svn:mime-type = application/java-archive

Added: 
release/carbondata/2.1.0/apache-carbondata-2.1.0-bin-spark2.4.5-hadoop2.7.2.jar.asc
==
Binary file - no diff available.

Propchange: 
release/carbondata/2.1.0/apache-carbondata-2.1.0-bin-spark2.4.5-hadoop2.7.2.jar.asc
--
svn:mime-type = application/pgp-signature

Added: 
release/carbondata/2.1.0/apache-carbondata-2.1.0-bin-spark2.4.5-hadoop2.7.2.jar.sha512
==
--- 
release/carbondata/2.1.0/apache-carbondata-2.1.0-bin-spark2.4.5-hadoop2.7.2.jar.sha512
 (added)
+++ 
release/carbondata/2.1.0/apache-carbondata-2.1.0-bin-spark2.4.5-hadoop2.7.2.jar.sha512
 Sun Nov  8 16:35:05 2020
@@ -0,0 +1 @@
+7ca85fb6bfd4572f65cd3cba0cb99ed9f80dc11ec04ba6d9c2a345adc8c69d1521b77a207b265df2a443d0ccb94d411ca9b2aefb54d651e8fad5aa5f80f5f2f2
  apache-carbondata-2.1.0-bin-spark2.4.5-hadoop2.7.2.jar

Added: release/carbondata/2.1.0/apache-carbondata-2.1.0-source-release.zip
==
Binary file - no diff available.

Propchange: release/carbondata/2.1.0/apache-carbondata-2.1.0-source-release.zip
--
svn:mime-type = application/zip

Added: release/carbondata/2.1.0/apache-carbondata-2.1.0-source-release.zip.asc
==
Binary file - no diff available.

Propchange: 
release/carbondata/2.1.0/apache-carbondata-2.1.0-source-release.zip.asc
--
svn:mime-type = application/pgp-signature

Added: release/carbondata/2.1.0/apache-carbondata-2.1.0-source-release.zip.md5
==
--- release/carbondata/2.1.0/apache-carbondata-2.1.0-source

svn commit: r42277 - in /dev/carbondata/2.1.0-rc2: ./ apache-carbondata-2.1.0-source-release.zip apache-carbondata-2.1.0-source-release.zip.asc apache-carbondata-2.1.0-source-release.zip.md5 apache-ca

2020-11-03 Thread kunalkapoor
Author: kunalkapoor
Date: Tue Nov  3 21:37:14 2020
New Revision: 42277

Log:
Upload 2.1.0-rc2

Added:
dev/carbondata/2.1.0-rc2/
dev/carbondata/2.1.0-rc2/apache-carbondata-2.1.0-source-release.zip   (with 
props)
dev/carbondata/2.1.0-rc2/apache-carbondata-2.1.0-source-release.zip.asc   
(with props)
dev/carbondata/2.1.0-rc2/apache-carbondata-2.1.0-source-release.zip.md5
dev/carbondata/2.1.0-rc2/apache-carbondata-2.1.0-source-release.zip.sha512

Added: dev/carbondata/2.1.0-rc2/apache-carbondata-2.1.0-source-release.zip
==
Binary file - no diff available.

Propchange: dev/carbondata/2.1.0-rc2/apache-carbondata-2.1.0-source-release.zip
--
svn:mime-type = application/zip

Added: dev/carbondata/2.1.0-rc2/apache-carbondata-2.1.0-source-release.zip.asc
==
Binary file - no diff available.

Propchange: 
dev/carbondata/2.1.0-rc2/apache-carbondata-2.1.0-source-release.zip.asc
--
svn:mime-type = application/pgp-signature

Added: dev/carbondata/2.1.0-rc2/apache-carbondata-2.1.0-source-release.zip.md5
==
--- dev/carbondata/2.1.0-rc2/apache-carbondata-2.1.0-source-release.zip.md5 
(added)
+++ dev/carbondata/2.1.0-rc2/apache-carbondata-2.1.0-source-release.zip.md5 Tue 
Nov  3 21:37:14 2020
@@ -0,0 +1 @@
+e3f64d72b204360ab753c2e961624f35  apache-carbondata-2.1.0-source-release.zip

Added: 
dev/carbondata/2.1.0-rc2/apache-carbondata-2.1.0-source-release.zip.sha512
==
--- dev/carbondata/2.1.0-rc2/apache-carbondata-2.1.0-source-release.zip.sha512 
(added)
+++ dev/carbondata/2.1.0-rc2/apache-carbondata-2.1.0-source-release.zip.sha512 
Tue Nov  3 21:37:14 2020
@@ -0,0 +1 @@
+a1f23b309059c3dc75c8b562a4d27011d65ebda9094a9f0f5a91eb4153feacb6ad36536ff6ab9840f93a4af1d5d8d92936edf9a377b52f34c5fe130110c6da10
  apache-carbondata-2.1.0-source-release.zip




  1   2   3   4   5   >