[GitHub] carbondata issue #2785: [CARBONDATA-2986] Table Properties are lost when mul...

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2785
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder2.1/694/



---


[GitHub] carbondata issue #2785: [CARBONDATA-2986] Table Properties are lost when mul...

2018-10-03 Thread Indhumathi27
Github user Indhumathi27 commented on the issue:

https://github.com/apache/carbondata/pull/2785
  
retest this please


---


[GitHub] carbondata issue #2772: [CARBONDATA-2977] Write uncompress_size to ChunkComp...

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2772
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder2.1/693/



---


[GitHub] carbondata issue #2793: [CARBONDATA-2990] Fixed JVM crash when rebuilding bl...

2018-10-03 Thread xuchuanyin
Github user xuchuanyin commented on the issue:

https://github.com/apache/carbondata/pull/2793
  
@ravipesala Why does it only occur in cluster mode? Just in curious


---


[GitHub] carbondata issue #2772: [CARBONDATA-2977] Write uncompress_size to ChunkComp...

2018-10-03 Thread jackylk
Github user jackylk commented on the issue:

https://github.com/apache/carbondata/pull/2772
  
retest this please


---


[GitHub] carbondata pull request #2776: [HOTFIX] Avoid adding empty table property

2018-10-03 Thread jackylk
Github user jackylk closed the pull request at:

https://github.com/apache/carbondata/pull/2776


---


[GitHub] carbondata issue #2768: [CARBONDATA-2594] Do not add InvertedIndex in Encodi...

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2768
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder2.1/692/



---


[GitHub] carbondata pull request #2791: [HOTFIX]correct the exception handling in loo...

2018-10-03 Thread akashrn5
Github user akashrn5 commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2791#discussion_r222537606
  
--- Diff: 
integration/spark2/src/main/scala/org/apache/spark/sql/hive/CarbonFileMetastore.scala
 ---
@@ -23,6 +23,7 @@ import java.net.URI
 import scala.collection.mutable.ArrayBuffer
 
 import org.apache.hadoop.fs.permission.{FsAction, FsPermission}
+import org.apache.hadoop.hive.ql.metadata.HiveException
--- End diff --

scalastyle fails if i give the space, i have checked it



---


[GitHub] carbondata pull request #2791: [HOTFIX]correct the exception handling in loo...

2018-10-03 Thread akashrn5
Github user akashrn5 commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2791#discussion_r222537366
  
--- Diff: 
integration/spark2/src/main/scala/org/apache/spark/sql/hive/CarbonFileMetastore.scala
 ---
@@ -208,7 +209,10 @@ class CarbonFileMetastore extends CarbonMetaStore {
 try {
   lookupRelation(tableIdentifier)(sparkSession)
 } catch {
-  case _: Exception =>
+  case ex: Exception =>
+if (ex.getCause.isInstanceOf[HiveException]) {
+  throw ex
+}
--- End diff --

here we arereturning false directly after catching exception, telling the 
table does not exists, this is wrong, i have check the flow , we might get 
hiveException(regarding permission) also if the user is not allowed to access 
the table, that time also we return false and we will not get the proper error.


---


[GitHub] carbondata issue #2791: [HOTFIX]correct the exception handling in lookup rel...

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2791
  
Build Failed with Spark 2.2.1, Please check CI 
http://95.216.28.178:8080/job/ApacheCarbonPRBuilder1/888/



---


[GitHub] carbondata issue #2791: [HOTFIX]correct the exception handling in lookup rel...

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2791
  
Build Success with Spark 2.3.1, Please check CI 
http://136.243.101.176:8080/job/carbondataprbuilder2.3/8956/



---


[GitHub] carbondata issue #2796: [CARBONDATA-2991]NegativeArraySizeException during q...

2018-10-03 Thread sujith71955
Github user sujith71955 commented on the issue:

https://github.com/apache/carbondata/pull/2796
  
@BJangir @jackylk @chenliang613 @ravipesala 
I think UUID.randomUUID().toString() has some performance bottle necks even 
though its negligible, as it internally uses SecureRandom API which is a 
synchronized method. 
if possible explore more on this as our system is more performance driven.
What about using JUG?
There is also a time bound uuid generation which seems to work lot faster 
compare to normal UUID generation supported by java.
Following link will give you a fair idea regarding implementation and 
performance comparison.
http://www.dcalabresi.com/blog/java/generate-java-uuid-performance/


---


[GitHub] carbondata issue #2796: [CARBONDATA-2991]NegativeArraySizeException during q...

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2796
  
Build Success with Spark 2.2.1, Please check CI 
http://95.216.28.178:8080/job/ApacheCarbonPRBuilder1/887/



---


[GitHub] carbondata issue #2796: [CARBONDATA-2991]NegativeArraySizeException during q...

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2796
  
Build Success with Spark 2.3.1, Please check CI 
http://136.243.101.176:8080/job/carbondataprbuilder2.3/8955/



---


[GitHub] carbondata pull request #2791: [HOTFIX]correct the exception handling in loo...

2018-10-03 Thread sujith71955
Github user sujith71955 commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2791#discussion_r222416107
  
--- Diff: 
integration/spark2/src/main/scala/org/apache/spark/sql/hive/CarbonFileMetastore.scala
 ---
@@ -23,6 +23,7 @@ import java.net.URI
 import scala.collection.mutable.ArrayBuffer
 
 import org.apache.hadoop.fs.permission.{FsAction, FsPermission}
+import org.apache.hadoop.hive.ql.metadata.HiveException
--- End diff --

nit: space is required, please check the rule


---


[GitHub] carbondata pull request #2791: [HOTFIX]correct the exception handling in loo...

2018-10-03 Thread sujith71955
Github user sujith71955 commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2791#discussion_r222415572
  
--- Diff: 
integration/spark2/src/main/scala/org/apache/spark/sql/hive/CarbonFileMetastore.scala
 ---
@@ -208,7 +209,10 @@ class CarbonFileMetastore extends CarbonMetaStore {
 try {
   lookupRelation(tableIdentifier)(sparkSession)
 } catch {
-  case _: Exception =>
+  case ex: Exception =>
+if (ex.getCause.isInstanceOf[HiveException]) {
+  throw ex
+}
--- End diff --

Didn't got your context of explanation is  for which review comment?


---


[GitHub] carbondata issue #2791: [HOTFIX]correct the exception handling in lookup rel...

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2791
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder2.1/691/



---


[GitHub] carbondata pull request #2791: [HOTFIX]correct the exception handling in loo...

2018-10-03 Thread akashrn5
Github user akashrn5 commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2791#discussion_r222409727
  
--- Diff: 
integration/spark2/src/main/scala/org/apache/spark/sql/hive/CarbonFileMetastore.scala
 ---
@@ -208,7 +209,10 @@ class CarbonFileMetastore extends CarbonMetaStore {
 try {
   lookupRelation(tableIdentifier)(sparkSession)
 } catch {
-  case _: Exception =>
+  case ex: Exception =>
+if (ex.getCause.isInstanceOf[HiveException]) {
+  throw ex
+}
--- End diff --

i checked the flow and permission errors will be thrown by Hive as 
HiveException and wrapped as analysis exception, and all the other exceptions 
will be same as i have described in description, during that time we cannot 
just catch exception and return false as table not exists. so i m checking as 
HiveException


---


[GitHub] carbondata issue #2796: [CARBONDATA-2991]NegativeArraySizeException during q...

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2796
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder2.1/690/



---


[GitHub] carbondata pull request #2796: [CARBONDATA-2991]NegativeArraySizeException d...

2018-10-03 Thread BJangir
GitHub user BJangir opened a pull request:

https://github.com/apache/carbondata/pull/2796

[CARBONDATA-2991]NegativeArraySizeException during query execution

Issue :- During Query Execution sometime NegativeArraySizeException  
Exception in Some Tasks . And sometime Executor is lost (JVM crash)

Root Cause :- It is because existing memoryblock is removed while it was 
in-use. This happened  because duplicate taskid generated. Sometime freed same 
memory addresses are assigned to another task which will initialize memory 
block to0 and this cause NegativeSizeArrayException whereas sometime freed 
memory will not be used any task of executor process but running task will try 
to access it and as that address is not part of process so JVM crash will 
happen.

Solution :-  Change taskID generation to UUID based instead of 
System.nanoTime()

Be sure to do all of the following checklist to help us incorporate 
your contribution quickly and easily:

 - [ ] Any interfaces changed?
 NA
 - [ ] Any backward compatibility impacted?
 NA
 - [ ] Document update required?
NA
 - [ ] Testing done
Please provide details on 
- Whether new unit test cases have been added or why no new tests 
are required?
- How it is tested? Please attach test report.
- Is it a performance related change? Please attach the performance 
test report.
- Any additional information to help reviewers in testing this 
change.
 
Manual Testing Done  
 - [ ] For large changes, please consider breaking it into sub-tasks under 
an umbrella JIRA. 
NA


You can merge this pull request into a Git repository by running:

$ git pull https://github.com/BJangir/incubator-carbondata CARBONDATA-2991

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/carbondata/pull/2796.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #2796


commit b41b35f9fdd86e7080e2d03d3219e260bcad7d31
Author: BJangir 
Date:   2018-10-03T17:35:42Z

[CARBONDATA-2991]NegativeArraySizeException during query execution




---


[GitHub] carbondata issue #2729: [WIP]Store optimization

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2729
  
Build Failed with Spark 2.2.1, Please check CI 
http://95.216.28.178:8080/job/ApacheCarbonPRBuilder1/886/



---


[GitHub] carbondata issue #2729: [WIP]Store optimization

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2729
  
Build Failed  with Spark 2.3.1, Please check CI 
http://136.243.101.176:8080/job/carbondataprbuilder2.3/8954/



---


[GitHub] carbondata issue #2729: [WIP]Store optimization

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2729
  
Build Failed  with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder2.1/689/



---


[GitHub] carbondata issue #2729: [WIP]Store optimization

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2729
  
Build Failed  with Spark 2.3.1, Please check CI 
http://136.243.101.176:8080/job/carbondataprbuilder2.3/8953/



---


[GitHub] carbondata issue #2729: [WIP]Store optimization

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2729
  
Build Failed with Spark 2.2.1, Please check CI 
http://95.216.28.178:8080/job/ApacheCarbonPRBuilder1/885/



---


[GitHub] carbondata issue #2729: [WIP]Store optimization

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2729
  
Build Failed  with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder2.1/688/



---


[GitHub] carbondata issue #2729: [WIP]Store optimization

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2729
  
Build Failed  with Spark 2.3.1, Please check CI 
http://136.243.101.176:8080/job/carbondataprbuilder2.3/8952/



---


[GitHub] carbondata issue #2729: [WIP]Store optimization

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2729
  
Build Failed with Spark 2.2.1, Please check CI 
http://95.216.28.178:8080/job/ApacheCarbonPRBuilder1/884/



---


[GitHub] carbondata issue #2730: [WIP] Performance improvements for Fileformat and Pr...

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2730
  
Build Failed with Spark 2.2.1, Please check CI 
http://95.216.28.178:8080/job/ApacheCarbonPRBuilder1/883/



---


[GitHub] carbondata issue #2730: [WIP] Performance improvements for Fileformat and Pr...

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2730
  
Build Failed  with Spark 2.3.1, Please check CI 
http://136.243.101.176:8080/job/carbondataprbuilder2.3/8951/



---


[GitHub] carbondata issue #2787: [HOTFIX] Fixed Between Query Data Mismatch issue for...

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2787
  
Build Success with Spark 2.2.1, Please check CI 
http://95.216.28.178:8080/job/ApacheCarbonPRBuilder1/881/



---


[GitHub] carbondata issue #2730: [WIP] Performance improvements for Fileformat and Pr...

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2730
  
Build Failed  with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder2.1/687/



---


[GitHub] carbondata issue #2787: [HOTFIX] Fixed Between Query Data Mismatch issue for...

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2787
  
Build Success with Spark 2.3.1, Please check CI 
http://136.243.101.176:8080/job/carbondataprbuilder2.3/8949/



---


[GitHub] carbondata pull request #2795: [HOTFIX] Changes to align printing of informa...

2018-10-03 Thread asfgit
Github user asfgit closed the pull request at:

https://github.com/apache/carbondata/pull/2795


---


[GitHub] carbondata issue #2795: [HOTFIX] Changes to align printing of information in...

2018-10-03 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2795
  
LGTM


---


[jira] [Resolved] (CARBONDATA-2976) Support dumping column chunk meta in CarbonCli

2018-10-03 Thread Ravindra Pesala (JIRA)


 [ 
https://issues.apache.org/jira/browse/CARBONDATA-2976?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ravindra Pesala resolved CARBONDATA-2976.
-
   Resolution: Fixed
Fix Version/s: 1.5.0

> Support dumping column chunk meta in CarbonCli
> --
>
> Key: CARBONDATA-2976
> URL: https://issues.apache.org/jira/browse/CARBONDATA-2976
> Project: CarbonData
>  Issue Type: New Feature
>Reporter: Jacky Li
>Assignee: Jacky Li
>Priority: Major
> Fix For: 1.5.0
>
>  Time Spent: 50m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] carbondata pull request #2771: [CARBONDATA-2976] Support dumping column chun...

2018-10-03 Thread asfgit
Github user asfgit closed the pull request at:

https://github.com/apache/carbondata/pull/2771


---


[jira] [Resolved] (CARBONDATA-2984) streaming throw NPE when there is no data in the task of a batch

2018-10-03 Thread Ravindra Pesala (JIRA)


 [ 
https://issues.apache.org/jira/browse/CARBONDATA-2984?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ravindra Pesala resolved CARBONDATA-2984.
-
   Resolution: Fixed
Fix Version/s: 1.5.0

> streaming throw NPE when there is no data in the task of a batch
> 
>
> Key: CARBONDATA-2984
> URL: https://issues.apache.org/jira/browse/CARBONDATA-2984
> Project: CarbonData
>  Issue Type: Bug
>Reporter: QiangCai
>Priority: Major
> Fix For: 1.5.0
>
>  Time Spent: 2h 10m
>  Remaining Estimate: 0h
>
> !746438440.jpg!



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] carbondata pull request #2782: [CARBONDATA-2984][Streaming] Fix NPE when the...

2018-10-03 Thread asfgit
Github user asfgit closed the pull request at:

https://github.com/apache/carbondata/pull/2782


---


[GitHub] carbondata issue #2788: [Documentation] Readme updated with latest topics an...

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2788
  
Build Failed  with Spark 2.3.1, Please check CI 
http://136.243.101.176:8080/job/carbondataprbuilder2.3/8948/



---


[GitHub] carbondata issue #2782: [CARBONDATA-2984][Streaming] Fix NPE when there is n...

2018-10-03 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2782
  
LGTM


---


[jira] [Resolved] (CARBONDATA-2987) Data mismatch after compaction with measure sort columns

2018-10-03 Thread Ravindra Pesala (JIRA)


 [ 
https://issues.apache.org/jira/browse/CARBONDATA-2987?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ravindra Pesala resolved CARBONDATA-2987.
-
Resolution: Fixed

> Data mismatch after compaction with measure sort columns
> 
>
> Key: CARBONDATA-2987
> URL: https://issues.apache.org/jira/browse/CARBONDATA-2987
> Project: CarbonData
>  Issue Type: Improvement
>Reporter: Ajantha Bhat
>Assignee: Ajantha Bhat
>Priority: Critical
>  Time Spent: 2h 40m
>  Remaining Estimate: 0h
>
> problem: Data mismatch after compaction with measure sort columns
> root cause : In compaction flow (DictionaryBasedResultCollector), in 
> ColumnPageWrapper inverted index mapping is not handled. Because of this, row 
> of no dictionary dimension columns gets data form other rows.
> Hence the data mismatch
>  
> solution: Handle inverted index mapping for  DictionaryBasedResultCollector 
> flow in ColumnPageWrapper
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] carbondata pull request #2784: [CARBONDATA-2987] Data mismatch after compact...

2018-10-03 Thread asfgit
Github user asfgit closed the pull request at:

https://github.com/apache/carbondata/pull/2784


---


[GitHub] carbondata pull request #2790: [HOTFIX] carbon reader support open another r...

2018-10-03 Thread asfgit
Github user asfgit closed the pull request at:

https://github.com/apache/carbondata/pull/2790


---


[GitHub] carbondata issue #2790: [HOTFIX] carbon reader support open another reader w...

2018-10-03 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2790
  
LGTM


---


[jira] [Resolved] (CARBONDATA-2967) Select is failing on pre-aggregate datamap when thrift server is restarted.

2018-10-03 Thread Ravindra Pesala (JIRA)


 [ 
https://issues.apache.org/jira/browse/CARBONDATA-2967?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ravindra Pesala resolved CARBONDATA-2967.
-
   Resolution: Fixed
Fix Version/s: 1.5.0

> Select is failing on pre-aggregate datamap when thrift server is restarted.
> ---
>
> Key: CARBONDATA-2967
> URL: https://issues.apache.org/jira/browse/CARBONDATA-2967
> Project: CarbonData
>  Issue Type: Bug
>Reporter: Kunal Kapoor
>Assignee: Kunal Kapoor
>Priority: Major
> Fix For: 1.5.0
>
>  Time Spent: 4h 40m
>  Remaining Estimate: 0h
>
> *Problem:* 
> NullPointerException is thrown when select query is fired on a datamap. This 
> is because to access dictionary files of the parent table the child table 
> tries to get tablePath from CarbonTable object of the parent. Because the 
> metadata is not populated therefore NullpointerException is thrown.
>  
> 1, 10.2.3.19, executor 1): java.lang.RuntimeException: Error while resolving 
> filter expression 
>         at 
> org.apache.carbondata.core.metadata.schema.table.CarbonTable.resolveFilter(CarbonTable.java:1043)
>  
>         at 
> org.apache.carbondata.core.scan.model.QueryModelBuilder.build(QueryModelBuilder.java:322)
>  
>         at 
> org.apache.carbondata.hadoop.api.CarbonInputFormat.createQueryModel(CarbonInputFormat.java:632)
>  
>         at 
> org.apache.carbondata.spark.rdd.CarbonScanRDD.internalCompute(CarbonScanRDD.scala:419)
>  
>         at 
> org.apache.carbondata.spark.rdd.CarbonRDD.compute(CarbonRDD.scala:78) 
>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323) 
>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:287) 
>         at 
> org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) 
>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323) 
>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:287) 
>         at 
> org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) 
>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323) 
>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:287) 
>         at 
> org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96) 
>         at 
> org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53) 
>         at org.apache.spark.scheduler.Task.run(Task.scala:109) 
>         at 
> org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:338) 
>         at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>  
>         at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>  
>         at java.lang.Thread.run(Thread.java:748) 
> Caused by: java.lang.NullPointerException 
>         at 
> org.apache.carbondata.core.scan.executor.util.QueryUtil.getTableIdentifierForColumn(QueryUtil.java:401)
>  
>         at 
> org.apache.carbondata.core.scan.filter.FilterUtil.getForwardDictionaryCache(FilterUtil.java:1416)
>  
>         at 
> org.apache.carbondata.core.scan.filter.FilterUtil.getFilterValues(FilterUtil.java:712)
>  
>         at 
> org.apache.carbondata.core.scan.filter.resolver.resolverinfo.visitor.DictionaryColumnVisitor.populateFilterResolvedInfo(DictionaryColumnVisitor.java:60)
>  
>         at 
> org.apache.carbondata.core.scan.filter.resolver.resolverinfo.DimColumnResolvedFilterInfo.populateFilterInfoBasedOnColumnType(DimColumnResolvedFilterInfo.java:119)
>  
>         at 
> org.apache.carbondata.core.scan.filter.resolver.ConditionalFilterResolverImpl.resolve(ConditionalFilterResolverImpl.java:107)
>  
>         at 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] carbondata pull request #2786: [CARBONDATA-2967] Fixed NPE for preaggregate ...

2018-10-03 Thread asfgit
Github user asfgit closed the pull request at:

https://github.com/apache/carbondata/pull/2786


---


[GitHub] carbondata issue #2788: [Documentation] Readme updated with latest topics an...

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2788
  
Build Success with Spark 2.2.1, Please check CI 
http://95.216.28.178:8080/job/ApacheCarbonPRBuilder1/880/



---


[GitHub] carbondata issue #2786: [CARBONDATA-2967] Fixed NPE for preaggregate queries

2018-10-03 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2786
  
LGTM


---


[jira] [Resolved] (CARBONDATA-2979) select count fails when carbondata file is written through SDK and read through sparkfileformat for complex datatype map(struct->array->map)

2018-10-03 Thread Ravindra Pesala (JIRA)


 [ 
https://issues.apache.org/jira/browse/CARBONDATA-2979?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ravindra Pesala resolved CARBONDATA-2979.
-
   Resolution: Fixed
Fix Version/s: 1.5.0

> select count fails when carbondata file is written through SDK and read 
> through sparkfileformat for complex datatype map(struct->array->map)
> 
>
> Key: CARBONDATA-2979
> URL: https://issues.apache.org/jira/browse/CARBONDATA-2979
> Project: CarbonData
>  Issue Type: Bug
>  Components: file-format
>Affects Versions: 1.5.0
>Reporter: Rahul Singha
>Assignee: Manish Gupta
>Priority: Minor
> Fix For: 1.5.0
>
> Attachments: MapSchema_15_int.avsc
>
>  Time Spent: 1h 50m
>  Remaining Estimate: 0h
>
> *Steps:*
> create carabondata and carbonindex file using SDK
> place the files in a hdfs location
> Read files using spark file format
> create table schema15_int using carbon location 
> 'hdfs://hacluster/user/rahul/map/mapschema15_int';
> Select count(*) from  schema15_int;
> *Actual Result:*
> Error: org.apache.spark.SparkException: Job aborted due to stage failure: 
> Task 0 in stage 24.0 failed 4 times, most recent failure: Lost task 0.3 in 
> stage 24.0 (TID 34, BLR114238, executor 3): java.io.IOException: All the 
> files doesn't have same schema. Unsupported operation on nonTransactional 
> table. Check logs.
>  at 
> org.apache.carbondata.core.scan.executor.impl.AbstractQueryExecutor.updateColumns(AbstractQueryExecutor.java:276)
>  at 
> org.apache.carbondata.core.scan.executor.impl.AbstractQueryExecutor.getDataBlocks(AbstractQueryExecutor.java:234)
>  at 
> org.apache.carbondata.core.scan.executor.impl.AbstractQueryExecutor.initQuery(AbstractQueryExecutor.java:141)
>  at 
> org.apache.carbondata.core.scan.executor.impl.AbstractQueryExecutor.getBlockExecutionInfos(AbstractQueryExecutor.java:401)
>  at 
> org.apache.carbondata.core.scan.executor.impl.VectorDetailQueryExecutor.execute(VectorDetailQueryExecutor.java:44)
>  at 
> org.apache.carbondata.spark.vectorreader.VectorizedCarbonRecordReader.initialize(VectorizedCarbonRecordReader.java:143)
>  at 
> org.apache.spark.sql.carbondata.execution.datasources.SparkCarbonFileFormat$$anonfun$buildReaderWithPartitionValues$2.apply(SparkCarbonFileFormat.scala:395)
>  at 
> org.apache.spark.sql.carbondata.execution.datasources.SparkCarbonFileFormat$$anonfun$buildReaderWithPartitionValues$2.apply(SparkCarbonFileFormat.scala:361)
>  at 
> org.apache.spark.sql.execution.datasources.FileScanRDD$$anon$1.org$apache$spark$sql$execution$datasources$FileScanRDD$$anon$$readCurrentFile(FileScanRDD.scala:124)
>  at 
> org.apache.spark.sql.execution.datasources.FileScanRDD$$anon$1.nextIterator(FileScanRDD.scala:174)
>  at 
> org.apache.spark.sql.execution.datasources.FileScanRDD$$anon$1.hasNext(FileScanRDD.scala:105)
>  at 
> org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.scan_nextBatch$(Unknown
>  Source)
>  at 
> org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.agg_doAggregateWithoutKey$(Unknown
>  Source)
>  at 
> org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.processNext(Unknown
>  Source)
>  at 
> org.apache.spark.sql.execution.BufferedRowIterator.hasNext(BufferedRowIterator.java:43)
>  at 
> org.apache.spark.sql.execution.WholeStageCodegenExec$$anonfun$8$$anon$1.hasNext(WholeStageCodegenExec.scala:395)
>  at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
>  at 
> org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:125)
>  at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)
>  at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
>  at org.apache.spark.scheduler.Task.run(Task.scala:108)
>  at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:338)
>  at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>  at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>  at java.lang.Thread.run(Thread.java:745)



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] carbondata pull request #2774: [CARBONDATA-2979] select count fails when car...

2018-10-03 Thread asfgit
Github user asfgit closed the pull request at:

https://github.com/apache/carbondata/pull/2774


---


[GitHub] carbondata issue #2730: [WIP] Performance improvements for Fileformat and Pr...

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2730
  
Build Failed  with Spark 2.3.1, Please check CI 
http://136.243.101.176:8080/job/carbondataprbuilder2.3/8950/



---


[GitHub] carbondata issue #2730: [WIP] Performance improvements for Fileformat and Pr...

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2730
  
Build Failed with Spark 2.2.1, Please check CI 
http://95.216.28.178:8080/job/ApacheCarbonPRBuilder1/882/



---


[GitHub] carbondata issue #2730: [WIP] Performance improvements for Fileformat and Pr...

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2730
  
Build Failed  with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder2.1/686/



---


[GitHub] carbondata pull request #2794: [CARBONDATA-2985]Fix issues in Table level co...

2018-10-03 Thread KanakaKumar
Github user KanakaKumar commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2794#discussion_r222330949
  
--- Diff: 
processing/src/main/java/org/apache/carbondata/processing/merger/CarbonDataMergerUtil.java
 ---
@@ -744,6 +744,7 @@ private static long getSizeOfSegment(String tablePath, 
String segId) {
 if (size >= 2) {
   level1Size = noOfSegmentLevelsCount[0];
   level2Size = noOfSegmentLevelsCount[1];
+  level2Size = level2Size == 1 ? 0 : level2Size;
--- End diff --

Please add a comment why to handle this


---


[GitHub] carbondata issue #2774: [CARBONDATA-2979] select count fails when carbondata...

2018-10-03 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2774
  
LGTM


---


[GitHub] carbondata pull request #2793: [CARBONDATA-2990] Fixed JVM crash when rebuil...

2018-10-03 Thread manishgupta88
Github user manishgupta88 commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2793#discussion_r222318367
  
--- Diff: 
core/src/main/java/org/apache/carbondata/core/memory/UnsafeMemoryManager.java 
---
@@ -217,6 +212,17 @@ public static MemoryBlock 
allocateMemoryWithRetry(MemoryAllocator memoryAllocato
 return baseBlock;
   }
 
+  private MemoryAllocator getMemoryAllocator(MemoryType memoryType) {
+switch (memoryType) {
+  case ONHEAP:
+return MemoryAllocator.HEAP;
+  case OFFHEAP:
+return MemoryAllocator.UNSAFE;
+  default:
+return MemoryAllocator.UNSAFE;
--- End diff --

combine OFFHEAP and default case as one
`case OFFHEAP:
default:
return MemoryAllocator.UNSAFE;`


---


[jira] [Updated] (CARBONDATA-2991) NegativeArraySizeException during query execution

2018-10-03 Thread Babulal (JIRA)


 [ 
https://issues.apache.org/jira/browse/CARBONDATA-2991?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Babulal updated CARBONDATA-2991:

Description: 
During Query Execution sometime NegativeArraySizeException  Exception in Some 
Tasks . And sometime Executor is lost (JVM crash)

 

ava.lang.NegativeArraySizeException at 
org.apache.carbondata.core.datastore.chunk.store.impl.unsafe.UnsafeVariableLengthDimesionDataChunkStore.getRow(UnsafeVariableLengthDimesionDataChunkStore.java:157)
 at 
org.apache.carbondata.core.datastore.chunk.impl.AbstractDimensionDataChunk.getChunkData(AbstractDimensionDataChunk.java:46)
 at 
org.apache.carbondata.core.scan.result.AbstractScannedResult.getNoDictionaryKeyArray(AbstractScannedResult.java:470)
 at 
org.apache.carbondata.core.scan.result.impl.NonFilterQueryScannedResult.getNoDictionaryKeyArray(NonFilterQueryScannedResult.java:102)
 at 
org.apache.carbondata.core.scan.collector.impl.DictionaryBasedResultCollector.collectData(DictionaryBasedResultCollector.java:101)
 at 
org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:51)
 at 
org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:32)
 at 
org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.getBatchResult(DetailQueryResultIterator.java:49)

 

 

 

Issue Analysis :- 

Possible Root Cause :- It is because existing memoryblock is removed while it 
was in-use. This happened  because duplicate taskid generated. Sometime freed 
same memory addresses are assigned to another task which will initialize memory 
block to0 and this cause NegativeSizeArrayException whereas sometime freed 
memory will not be used any task of executor process but running task will try 
to access it and as that address is not part of process so JVM crash will 
happen.

 

*Steps to find cause* 

*Add the code to create tasklist ids and add taskid to list on 
setCarbonTaskInfo()  , and if it is duplicate then Log a Warn message.*

Please check attachment. 

 

 Run the Query multiple time and found warn message in executor logs

2018-09-29 14:48:41,840 | INFO | [[Executor task launch worker for task 
435242][partitionID:1;queryID:29971946625611231]] | [Executor task launch 
worker for task 435242][partitionID:1;queryID:29971946625611231] Total memory 
used after task 29971946381679677 is 0 Current tasks running now are : [] | 
org.apache.carbondata.common.logging.impl.StandardLogService.logInfoMessage(StandardLogService.java:150)
2018-09-29 14:48:41,840 | INFO | [[Executor task launch worker for task 
435242][partitionID:1;queryID:29971946625611231]] | Finished task 17091.0 in 
stage 22.0 (TID 435242). 1412 bytes result sent to driver | 
org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)
{color:#FF}*2018-09-29 14:48:41,842 | WARN | [Executor task launch worker 
for task 435393] | Executor task launch worker for task 435393 Already This 
Task is is Present29971946637094373 | 
org.apache.carbondata.common.logging.impl.StandardLogService.logWarnMessage(StandardLogService.java:168)*{color}
2018-09-29 14:48:41,842 | INFO | [dispatcher-event-loop-13] | Got assigned task 
435395 | org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)

  was:
During Query Execution sometime NegativeArraySizeException  Exception in Some 
Tasks . And sometime Executor is lost (JVM crash)

 

ava.lang.NegativeArraySizeException at 
org.apache.carbondata.core.datastore.chunk.store.impl.unsafe.UnsafeVariableLengthDimesionDataChunkStore.getRow(UnsafeVariableLengthDimesionDataChunkStore.java:157)
 at 
org.apache.carbondata.core.datastore.chunk.impl.AbstractDimensionDataChunk.getChunkData(AbstractDimensionDataChunk.java:46)
 at 
org.apache.carbondata.core.scan.result.AbstractScannedResult.getNoDictionaryKeyArray(AbstractScannedResult.java:470)
 at 
org.apache.carbondata.core.scan.result.impl.NonFilterQueryScannedResult.getNoDictionaryKeyArray(NonFilterQueryScannedResult.java:102)
 at 
org.apache.carbondata.core.scan.collector.impl.DictionaryBasedResultCollector.collectData(DictionaryBasedResultCollector.java:101)
 at 
org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:51)
 at 
org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:32)
 at 
org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.getBatchResult(DetailQueryResultIterator.java:49)

 

 

 

Issue Analysis :- 

Possible Root Cause :- It is because existing memoryblock is removed while it 
was in-use. This happened  because duplicate taskid generated. Sometime freed 
same memory addresses are assigned to another task which will initialize memory 
block to0 and this cause NegativeSizeArrayException whereas sometime freed 
memory will not be used any task of executor process but running task will try 

[GitHub] carbondata issue #2795: [HOTFIX] Changes to align printing of information in...

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2795
  
Build Failed with Spark 2.2.1, Please check CI 
http://95.216.28.178:8080/job/ApacheCarbonPRBuilder1/878/



---


[GitHub] carbondata issue #2795: [HOTFIX] Changes to align printing of information in...

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2795
  
Build Success with Spark 2.3.1, Please check CI 
http://136.243.101.176:8080/job/carbondataprbuilder2.3/8946/



---


[jira] [Updated] (CARBONDATA-2991) NegativeArraySizeException during query execution

2018-10-03 Thread Babulal (JIRA)


 [ 
https://issues.apache.org/jira/browse/CARBONDATA-2991?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Babulal updated CARBONDATA-2991:

Attachment: Root_Cause_Find_Step.JPG

> NegativeArraySizeException during query execution 
> --
>
> Key: CARBONDATA-2991
> URL: https://issues.apache.org/jira/browse/CARBONDATA-2991
> Project: CarbonData
>  Issue Type: Bug
>Affects Versions: 1.4.0, 1.3.1
>Reporter: Babulal
>Assignee: Babulal
>Priority: Major
> Attachments: Root_Cause_Find_Step.JPG
>
>
> During Query Execution sometime NegativeArraySizeException  Exception in Some 
> Tasks . And sometime Executor is lost (JVM crash)
>  
> ava.lang.NegativeArraySizeException at 
> org.apache.carbondata.core.datastore.chunk.store.impl.unsafe.UnsafeVariableLengthDimesionDataChunkStore.getRow(UnsafeVariableLengthDimesionDataChunkStore.java:157)
>  at 
> org.apache.carbondata.core.datastore.chunk.impl.AbstractDimensionDataChunk.getChunkData(AbstractDimensionDataChunk.java:46)
>  at 
> org.apache.carbondata.core.scan.result.AbstractScannedResult.getNoDictionaryKeyArray(AbstractScannedResult.java:470)
>  at 
> org.apache.carbondata.core.scan.result.impl.NonFilterQueryScannedResult.getNoDictionaryKeyArray(NonFilterQueryScannedResult.java:102)
>  at 
> org.apache.carbondata.core.scan.collector.impl.DictionaryBasedResultCollector.collectData(DictionaryBasedResultCollector.java:101)
>  at 
> org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:51)
>  at 
> org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:32)
>  at 
> org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.getBatchResult(DetailQueryResultIterator.java:49)
>  
>  
>  
> Issue Analysis :- 
> Possible Root Cause :- It is because existing memoryblock is removed while it 
> was in-use. This happened  because duplicate taskid generated. Sometime freed 
> same memory addresses are assigned to another task which will initialize 
> memory block to0 and this cause NegativeSizeArrayException whereas sometime 
> freed memory will not be used any task of executor process but running task 
> will try to access it and as that address is not part of process so JVM crash 
> will happen.
>  
>  
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (CARBONDATA-2991) NegativeArraySizeException during query execution

2018-10-03 Thread Babulal (JIRA)


 [ 
https://issues.apache.org/jira/browse/CARBONDATA-2991?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Babulal updated CARBONDATA-2991:

Description: 
During Query Execution sometime NegativeArraySizeException  Exception in Some 
Tasks . And sometime Executor is lost (JVM crash)

 

ava.lang.NegativeArraySizeException at 
org.apache.carbondata.core.datastore.chunk.store.impl.unsafe.UnsafeVariableLengthDimesionDataChunkStore.getRow(UnsafeVariableLengthDimesionDataChunkStore.java:157)
 at 
org.apache.carbondata.core.datastore.chunk.impl.AbstractDimensionDataChunk.getChunkData(AbstractDimensionDataChunk.java:46)
 at 
org.apache.carbondata.core.scan.result.AbstractScannedResult.getNoDictionaryKeyArray(AbstractScannedResult.java:470)
 at 
org.apache.carbondata.core.scan.result.impl.NonFilterQueryScannedResult.getNoDictionaryKeyArray(NonFilterQueryScannedResult.java:102)
 at 
org.apache.carbondata.core.scan.collector.impl.DictionaryBasedResultCollector.collectData(DictionaryBasedResultCollector.java:101)
 at 
org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:51)
 at 
org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:32)
 at 
org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.getBatchResult(DetailQueryResultIterator.java:49)

 

 

 

Issue Analysis :- 

Possible Root Cause :- It is because existing memoryblock is removed while it 
was in-use. This happened  because duplicate taskid generated. Sometime freed 
same memory addresses are assigned to another task which will initialize memory 
block to0 and this cause NegativeSizeArrayException whereas sometime freed 
memory will not be used any task of executor process but running task will try 
to access it and as that address is not part of process so JVM crash will 
happen.

 

 

 

  was:
During Query Execution sometime NegativeArraySizeException  Exception in Some 
Tasks . And sometime Executor is lost (JVM crash)

 

ava.lang.NegativeArraySizeException at 
org.apache.carbondata.core.datastore.chunk.store.impl.unsafe.UnsafeVariableLengthDimesionDataChunkStore.getRow(UnsafeVariableLengthDimesionDataChunkStore.java:157)
 at 
org.apache.carbondata.core.datastore.chunk.impl.AbstractDimensionDataChunk.getChunkData(AbstractDimensionDataChunk.java:46)
 at 
org.apache.carbondata.core.scan.result.AbstractScannedResult.getNoDictionaryKeyArray(AbstractScannedResult.java:470)
 at 
org.apache.carbondata.core.scan.result.impl.NonFilterQueryScannedResult.getNoDictionaryKeyArray(NonFilterQueryScannedResult.java:102)
 at 
org.apache.carbondata.core.scan.collector.impl.DictionaryBasedResultCollector.collectData(DictionaryBasedResultCollector.java:101)
 at 
org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:51)
 at 
org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:32)
 at 
org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.getBatchResult(DetailQueryResultIterator.java:49)

 

 

 

Issue Analysis :- 

Possible Root Cause :- It is because existing memoryblock is removed while it 
was in-use. This happened  because duplicate taskid generated. Sometime freed 
same memory addresses are assigned to another task which will initialize memory 
block to0 and this cause NegativeSizeArrayException whereas sometime freed 
memory will not be used any task of executor process but running task will try 
to access it and as that address is not part of process so JVM crash will 
happen.

 

Method to find Cause :- 

 

!Root_Cause_Find_Step.JPG!

 

!Log_Message.JPG!

 


> NegativeArraySizeException during query execution 
> --
>
> Key: CARBONDATA-2991
> URL: https://issues.apache.org/jira/browse/CARBONDATA-2991
> Project: CarbonData
>  Issue Type: Bug
>Affects Versions: 1.4.0, 1.3.1
>Reporter: Babulal
>Priority: Major
>
> During Query Execution sometime NegativeArraySizeException  Exception in Some 
> Tasks . And sometime Executor is lost (JVM crash)
>  
> ava.lang.NegativeArraySizeException at 
> org.apache.carbondata.core.datastore.chunk.store.impl.unsafe.UnsafeVariableLengthDimesionDataChunkStore.getRow(UnsafeVariableLengthDimesionDataChunkStore.java:157)
>  at 
> org.apache.carbondata.core.datastore.chunk.impl.AbstractDimensionDataChunk.getChunkData(AbstractDimensionDataChunk.java:46)
>  at 
> org.apache.carbondata.core.scan.result.AbstractScannedResult.getNoDictionaryKeyArray(AbstractScannedResult.java:470)
>  at 
> org.apache.carbondata.core.scan.result.impl.NonFilterQueryScannedResult.getNoDictionaryKeyArray(NonFilterQueryScannedResult.java:102)
>  at 
> 

[jira] [Assigned] (CARBONDATA-2991) NegativeArraySizeException during query execution

2018-10-03 Thread Babulal (JIRA)


 [ 
https://issues.apache.org/jira/browse/CARBONDATA-2991?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Babulal reassigned CARBONDATA-2991:
---

Assignee: Babulal

> NegativeArraySizeException during query execution 
> --
>
> Key: CARBONDATA-2991
> URL: https://issues.apache.org/jira/browse/CARBONDATA-2991
> Project: CarbonData
>  Issue Type: Bug
>Affects Versions: 1.4.0, 1.3.1
>Reporter: Babulal
>Assignee: Babulal
>Priority: Major
>
> During Query Execution sometime NegativeArraySizeException  Exception in Some 
> Tasks . And sometime Executor is lost (JVM crash)
>  
> ava.lang.NegativeArraySizeException at 
> org.apache.carbondata.core.datastore.chunk.store.impl.unsafe.UnsafeVariableLengthDimesionDataChunkStore.getRow(UnsafeVariableLengthDimesionDataChunkStore.java:157)
>  at 
> org.apache.carbondata.core.datastore.chunk.impl.AbstractDimensionDataChunk.getChunkData(AbstractDimensionDataChunk.java:46)
>  at 
> org.apache.carbondata.core.scan.result.AbstractScannedResult.getNoDictionaryKeyArray(AbstractScannedResult.java:470)
>  at 
> org.apache.carbondata.core.scan.result.impl.NonFilterQueryScannedResult.getNoDictionaryKeyArray(NonFilterQueryScannedResult.java:102)
>  at 
> org.apache.carbondata.core.scan.collector.impl.DictionaryBasedResultCollector.collectData(DictionaryBasedResultCollector.java:101)
>  at 
> org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:51)
>  at 
> org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:32)
>  at 
> org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.getBatchResult(DetailQueryResultIterator.java:49)
>  
>  
>  
> Issue Analysis :- 
> Possible Root Cause :- It is because existing memoryblock is removed while it 
> was in-use. This happened  because duplicate taskid generated. Sometime freed 
> same memory addresses are assigned to another task which will initialize 
> memory block to0 and this cause NegativeSizeArrayException whereas sometime 
> freed memory will not be used any task of executor process but running task 
> will try to access it and as that address is not part of process so JVM crash 
> will happen.
>  
>  
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] carbondata issue #2793: [CARBONDATA-2990] Fixed JVM crash when rebuilding bl...

2018-10-03 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2793
  
@xuchuanyin It can happen only in a cluster, that's why it is not caught in 
UT. 


---


[GitHub] carbondata issue #2787: [HOTFIX] Fixed Between Query Data Mismatch issue for...

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2787
  
Build Failed  with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder2.1/685/



---


[jira] [Created] (CARBONDATA-2991) NegativeArraySizeException during query execution

2018-10-03 Thread Babulal (JIRA)
Babulal created CARBONDATA-2991:
---

 Summary: NegativeArraySizeException during query execution 
 Key: CARBONDATA-2991
 URL: https://issues.apache.org/jira/browse/CARBONDATA-2991
 Project: CarbonData
  Issue Type: Bug
Affects Versions: 1.3.1, 1.4.0
Reporter: Babulal


During Query Execution sometime NegativeArraySizeException  Exception in Some 
Tasks . And sometime Executor is lost (JVM crash)

 

ava.lang.NegativeArraySizeException at 
org.apache.carbondata.core.datastore.chunk.store.impl.unsafe.UnsafeVariableLengthDimesionDataChunkStore.getRow(UnsafeVariableLengthDimesionDataChunkStore.java:157)
 at 
org.apache.carbondata.core.datastore.chunk.impl.AbstractDimensionDataChunk.getChunkData(AbstractDimensionDataChunk.java:46)
 at 
org.apache.carbondata.core.scan.result.AbstractScannedResult.getNoDictionaryKeyArray(AbstractScannedResult.java:470)
 at 
org.apache.carbondata.core.scan.result.impl.NonFilterQueryScannedResult.getNoDictionaryKeyArray(NonFilterQueryScannedResult.java:102)
 at 
org.apache.carbondata.core.scan.collector.impl.DictionaryBasedResultCollector.collectData(DictionaryBasedResultCollector.java:101)
 at 
org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:51)
 at 
org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:32)
 at 
org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.getBatchResult(DetailQueryResultIterator.java:49)

 

 

 

Issue Analysis :- 

Possible Root Cause :- It is because existing memoryblock is removed while it 
was in-use. This happened  because duplicate taskid generated. Sometime freed 
same memory addresses are assigned to another task which will initialize memory 
block to0 and this cause NegativeSizeArrayException whereas sometime freed 
memory will not be used any task of executor process but running task will try 
to access it and as that address is not part of process so JVM crash will 
happen.

 

Method to find Cause :- 

 

!Root_Cause_Find_Step.JPG!

 

!Log_Message.JPG!

 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] carbondata issue #2788: [Documentation] Readme updated with latest topics an...

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2788
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder2.1/684/



---


[GitHub] carbondata issue #2788: [Documentation] Readme updated with latest topics an...

2018-10-03 Thread sgururajshetty
Github user sgururajshetty commented on the issue:

https://github.com/apache/carbondata/pull/2788
  
@sraghunandan & @kunal642 kindly review and merge the doc


---


[GitHub] carbondata issue #2795: [HOTFIX] Changes to align printing of information in...

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2795
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder2.1/683/



---


[GitHub] carbondata pull request #2795: [HOTFIX] Changes to align printing of informa...

2018-10-03 Thread manishgupta88
GitHub user manishgupta88 opened a pull request:

https://github.com/apache/carbondata/pull/2795

[HOTFIX] Changes to align printing of information in explain command based 
on enable.query.statistics flag

Now user need to set the flag enable.query.statistics = true to print the 
pruning and dataMap related information in the explain command.

 - [ ] Any interfaces changed?
 
 - [ ] Any backward compatibility impacted?
 
 - [ ] Document update required?

 - [ ] Testing done
   
 - [ ] For large changes, please consider breaking it into sub-tasks under 
an umbrella JIRA. 



You can merge this pull request into a Git repository by running:

$ git pull https://github.com/manishgupta88/carbondata 
explain_profiler_disable_default

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/carbondata/pull/2795.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #2795


commit 2b03fe91efd2ba144e6ef988750fc1e8c63d40f5
Author: manishgupta88 
Date:   2018-10-03T12:43:45Z

[HOTFIX] Changes to align printing of information in explain command based 
on enable.query.statistics flag




---


[GitHub] carbondata issue #2792: [CARBONDATA-2981] Support read primitive data type i...

2018-10-03 Thread kunal642
Github user kunal642 commented on the issue:

https://github.com/apache/carbondata/pull/2792
  
@xubo245 Please add link for CSDK-guide in README file.


---


[jira] [Resolved] (CARBONDATA-2952) Provide CarbonReader C++ interface for SDK

2018-10-03 Thread Kunal Kapoor (JIRA)


 [ 
https://issues.apache.org/jira/browse/CARBONDATA-2952?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kunal Kapoor resolved CARBONDATA-2952.
--
Resolution: Fixed

> Provide CarbonReader C++ interface for SDK
> --
>
> Key: CARBONDATA-2952
> URL: https://issues.apache.org/jira/browse/CARBONDATA-2952
> Project: CarbonData
>  Issue Type: Sub-task
>  Components: other
>Affects Versions: 1.5.0
>Reporter: xubo245
>Assignee: xubo245
>Priority: Major
> Fix For: 1.5.0
>
>  Time Spent: 15h 40m
>  Remaining Estimate: 0h
>
> Provide CarbonReader C++ interface for SDK
> 1.init carbonreader
> 2.config data path and tablename
> 3.configure projection
> 4.build carbon reader
> 5.hasNext
> 6.readNextRow
> 7.close



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] carbondata pull request #2738: [CARBONDATA-2952] Provide c++ interface for S...

2018-10-03 Thread asfgit
Github user asfgit closed the pull request at:

https://github.com/apache/carbondata/pull/2738


---


[GitHub] carbondata issue #2738: [CARBONDATA-2952] Provide c++ interface for SDK Carb...

2018-10-03 Thread kunal642
Github user kunal642 commented on the issue:

https://github.com/apache/carbondata/pull/2738
  
LGTM. 
One minor comment -> Reference for CSDK-guide.md is missing from README due 
to which there is no way to navigate to the guide. Please handle this as part 
of #2792 


---


[GitHub] carbondata issue #2738: [CARBONDATA-2952] Provide c++ interface for SDK Carb...

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2738
  
Build Success with Spark 2.2.1, Please check CI 
http://95.216.28.178:8080/job/ApacheCarbonPRBuilder1/877/



---


[GitHub] carbondata issue #2738: [CARBONDATA-2952] Provide c++ interface for SDK Carb...

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2738
  
Build Success with Spark 2.3.1, Please check CI 
http://136.243.101.176:8080/job/carbondataprbuilder2.3/8945/



---


[jira] [Resolved] (CARBONDATA-2982) CarbonSchemaReader don't support Array

2018-10-03 Thread Manish Gupta (JIRA)


 [ 
https://issues.apache.org/jira/browse/CARBONDATA-2982?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Manish Gupta resolved CARBONDATA-2982.
--
Resolution: Fixed

> CarbonSchemaReader don't support Array
> --
>
> Key: CARBONDATA-2982
> URL: https://issues.apache.org/jira/browse/CARBONDATA-2982
> Project: CarbonData
>  Issue Type: Bug
>  Components: other
>Affects Versions: 1.5.0
>Reporter: xubo245
>Assignee: xubo245
>Priority: Major
> Fix For: 1.5.0
>
>  Time Spent: 4h 20m
>  Remaining Estimate: 0h
>
> CarbonSchemaReader don't support Array
> When we read schema from index file and the data include array data 
> type
> run org.apache.carbondata.examples.sdk.CarbonReaderExample :
> {code:java}
> Schema schema = CarbonSchemaReader
> .readSchemaInIndexFile(dataFiles[0].getAbsolutePath())
> .asOriginOrder();
> // Transform the schema
> String[] strings = new String[schema.getFields().length];
> for (int i = 0; i < schema.getFields().length; i++) {
> strings[i] = (schema.getFields())[i].getFieldName();
> System.out.println(strings[i] + "\t" + 
> schema.getFields()[i].getSchemaOrdinal());
> }
> {code}
> and throw some exception:
> {code:java}
> log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for more 
> info.
> arrayfield.val0   -1
> stringfield   0
> shortfield1
> intfield  2
> longfield 3
> doublefield   4
> boolfield 5
> datefield 6
> timefield 7
> decimalfield  8
> varcharfield  9
> arrayfield10
> Complex child columns projection NOT supported through CarbonReader
> java.lang.UnsupportedOperationException: Complex child columns projection NOT 
> supported through CarbonReader
>   at 
> org.apache.carbondata.sdk.file.CarbonReaderBuilder.build(CarbonReaderBuilder.java:155)
>   at 
> org.apache.carbondata.examples.sdk.CarbonReaderExample.main(CarbonReaderExample.java:110)
> {code}
> It print arrayfield.val0  -1, child schema



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] carbondata pull request #2780: [CARBONDATA-2982] CarbonSchemaReader support ...

2018-10-03 Thread asfgit
Github user asfgit closed the pull request at:

https://github.com/apache/carbondata/pull/2780


---


[GitHub] carbondata issue #2738: [CARBONDATA-2952] Provide c++ interface for SDK Carb...

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2738
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder2.1/682/



---


[GitHub] carbondata pull request #2789: [HOTFIX] Fixed S3 metrics issue.

2018-10-03 Thread asfgit
Github user asfgit closed the pull request at:

https://github.com/apache/carbondata/pull/2789


---


[GitHub] carbondata issue #2738: [CARBONDATA-2952] Provide c++ interface for SDK Carb...

2018-10-03 Thread kunal642
Github user kunal642 commented on the issue:

https://github.com/apache/carbondata/pull/2738
  
retest this please


---


[GitHub] carbondata issue #2787: [HOTFIX] Fixed Between Query Data Mismatch issue for...

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2787
  
Build Failed  with Spark 2.3.1, Please check CI 
http://136.243.101.176:8080/job/carbondataprbuilder2.3/8944/



---


[GitHub] carbondata issue #2789: [HOTFIX] Fixed S3 metrics issue.

2018-10-03 Thread kumarvishal09
Github user kumarvishal09 commented on the issue:

https://github.com/apache/carbondata/pull/2789
  
LGTM


---


[GitHub] carbondata issue #2789: [HOTFIX] Fixed S3 metrics issue.

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2789
  
Build Success with Spark 2.3.1, Please check CI 
http://136.243.101.176:8080/job/carbondataprbuilder2.3/8943/



---


[GitHub] carbondata issue #2787: [HOTFIX] Fixed Between Query Data Mismatch issue for...

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2787
  
Build Failed with Spark 2.2.1, Please check CI 
http://95.216.28.178:8080/job/ApacheCarbonPRBuilder1/876/



---


[GitHub] carbondata issue #2789: [HOTFIX] Fixed S3 metrics issue.

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2789
  
Build Success with Spark 2.2.1, Please check CI 
http://95.216.28.178:8080/job/ApacheCarbonPRBuilder1/875/



---


[GitHub] carbondata issue #2780: [CARBONDATA-2982] CarbonSchemaReader support array

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2780
  
Build Failed  with Spark 2.3.1, Please check CI 
http://136.243.101.176:8080/job/carbondataprbuilder2.3/8942/



---


[GitHub] carbondata issue #2786: [CARBONDATA-2967] Fixed NPE for preaggregate queries

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2786
  
Build Success with Spark 2.2.1, Please check CI 
http://95.216.28.178:8080/job/ApacheCarbonPRBuilder1/872/



---


[GitHub] carbondata issue #2789: [HOTFIX] Fixed S3 metrics issue.

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2789
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder2.1/681/



---


[GitHub] carbondata issue #2780: [CARBONDATA-2982] CarbonSchemaReader support array

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2780
  
Build Success with Spark 2.2.1, Please check CI 
http://95.216.28.178:8080/job/ApacheCarbonPRBuilder1/874/



---


[GitHub] carbondata pull request #2791: [HOTFIX]correct the exception handling in loo...

2018-10-03 Thread KanakaKumar
Github user KanakaKumar commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2791#discussion_r52487
  
--- Diff: 
integration/spark2/src/main/spark2.1/org/apache/spark/sql/hive/CarbonSessionState.scala
 ---
@@ -141,7 +141,13 @@ class CarbonHiveSessionCatalog(
*/
   override def lookupRelation(name: TableIdentifier,
   alias: Option[String]): LogicalPlan = {
-val rtnRelation = super.lookupRelation(name, alias)
+val rtnRelation =
+  try {
+super.lookupRelation(name, alias)
+  } catch {
+case ex: Exception =>
--- End diff --

This change is not required. Catch and throwing same exception. 


---


[GitHub] carbondata issue #2786: [CARBONDATA-2967] Fixed NPE for preaggregate queries

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2786
  
Build Success with Spark 2.3.1, Please check CI 
http://136.243.101.176:8080/job/carbondataprbuilder2.3/8940/



---


[GitHub] carbondata issue #2787: [HOTFIX] Fixed Between Query Data Mismatch issue for...

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2787
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder2.1/680/



---


[GitHub] carbondata issue #2789: [HOTFIX] Fixed S3 metrics issue.

2018-10-03 Thread kumarvishal09
Github user kumarvishal09 commented on the issue:

https://github.com/apache/carbondata/pull/2789
  
retest this please


---


[GitHub] carbondata issue #2729: [WIP]Store optimization

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2729
  
Build Failed  with Spark 2.3.1, Please check CI 
http://136.243.101.176:8080/job/carbondataprbuilder2.3/8941/



---


[GitHub] carbondata issue #2787: [HOTFIX] Fixed Between Query Data Mismatch issue for...

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2787
  
Build Failed  with Spark 2.3.1, Please check CI 
http://136.243.101.176:8080/job/carbondataprbuilder2.3/8939/



---


[GitHub] carbondata issue #2780: [CARBONDATA-2982] CarbonSchemaReader support array

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2780
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder2.1/679/



---


[GitHub] carbondata issue #2729: [WIP]Store optimization

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2729
  
Build Failed with Spark 2.2.1, Please check CI 
http://95.216.28.178:8080/job/ApacheCarbonPRBuilder1/873/



---


[GitHub] carbondata issue #2729: [WIP]Store optimization

2018-10-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2729
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder2.1/678/



---


[GitHub] carbondata issue #2780: [CARBONDATA-2982] CarbonSchemaReader support array

2018-10-03 Thread manishgupta88
Github user manishgupta88 commented on the issue:

https://github.com/apache/carbondata/pull/2780
  
retest this please


---


  1   2   >