[GitHub] incubator-carbondata issue #783: [CARBONDATA-903] data load is not failing e...

2017-04-11 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/incubator-carbondata/pull/783
  
Build Failed  with Spark 1.6.2, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/1580/



---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Created] (CARBONDATA-908) bitmap encode

2017-04-11 Thread Jarck (JIRA)
Jarck created CARBONDATA-908:


 Summary: bitmap encode
 Key: CARBONDATA-908
 URL: https://issues.apache.org/jira/browse/CARBONDATA-908
 Project: CarbonData
  Issue Type: New Feature
  Components: core, data-load, data-query
Reporter: Jarck
Assignee: Jarck


for frequent filter queries on low cardinality columns, use bitmap encode can 
speed up query



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[GitHub] incubator-carbondata issue #751: [CARBONDATA-816] Added Example for Hive Int...

2017-04-11 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/incubator-carbondata/pull/751
  
Build Success with Spark 1.6.2, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/1579/



---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] incubator-carbondata issue #751: [CARBONDATA-816] Added Example for Hive Int...

2017-04-11 Thread anubhav100
Github user anubhav100 commented on the issue:

https://github.com/apache/incubator-carbondata/pull/751
  
@chenliang613 i added the code for stoping application automatically 
,tested it on my system and rebased all commits into one commit can it be 
merged?


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] incubator-carbondata issue #781: [CARBONDATA-900] Is null query on a newly a...

2017-04-11 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/incubator-carbondata/pull/781
  
Build Success with Spark 1.6.2, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/1578/



---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Commented] (CARBONDATA-907) The grammar for DELETE SEGMENT FOR DATE in website is not correct

2017-04-11 Thread chenerlu (JIRA)

[ 
https://issues.apache.org/jira/browse/CARBONDATA-907?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15965298#comment-15965298
 ] 

chenerlu commented on CARBONDATA-907:
-

I have raised a PR below:
https://github.com/apache/incubator-carbondata/pull/785

> The grammar for DELETE SEGMENT FOR DATE in website is not correct 
> --
>
> Key: CARBONDATA-907
> URL: https://issues.apache.org/jira/browse/CARBONDATA-907
> Project: CarbonData
>  Issue Type: Bug
>Reporter: chenerlu
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> The grammar for DELETE SEGMENT FOR DATE in website is not correct 



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[GitHub] incubator-carbondata issue #785: [CARBONDATA-907] The grammar for DELETE SEG...

2017-04-11 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/incubator-carbondata/pull/785
  
Can one of the admins verify this patch?


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] incubator-carbondata pull request #785: [CARBONDATA-907] The grammar for DEL...

2017-04-11 Thread chenerlu
GitHub user chenerlu opened a pull request:

https://github.com/apache/incubator-carbondata/pull/785

[CARBONDATA-907] The grammar for DELETE SEGMENT FOR DATE in website is not 
correct

Problem
As I check open source CarbonData code, the grammar for DELETE SEGMENT FOR 
DATE in website is not correct, while the corresponding example is correct.
Solution
Update the doc for for DELETE SEGMENT FOR DATE in website.

You can merge this pull request into a Git repository by running:

$ git pull https://github.com/chenerlu/incubator-carbondata patch-2

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/incubator-carbondata/pull/785.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #785


commit 94ce539fd3966451778ea8db5dca61ea22352177
Author: chenerlu 
Date:   2017-04-12T02:03:29Z

Update dml-operation-on-carbondata.md

commit 9940f22cdb080ed2c970b6c617698666f23673da
Author: chenerlu 
Date:   2017-04-12T02:48:00Z

Update dml-operation-on-carbondata.md




---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Created] (CARBONDATA-907) The grammar for DELETE SEGMENT FOR DATE in website is not correct

2017-04-11 Thread chenerlu (JIRA)
chenerlu created CARBONDATA-907:
---

 Summary: The grammar for DELETE SEGMENT FOR DATE in website is not 
correct 
 Key: CARBONDATA-907
 URL: https://issues.apache.org/jira/browse/CARBONDATA-907
 Project: CarbonData
  Issue Type: Bug
Reporter: chenerlu


The grammar for DELETE SEGMENT FOR DATE in website is not correct 



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[jira] [Commented] (CARBONDATA-906) Always OOM error when import large dataset (100milion rows)

2017-04-11 Thread Crabo Yang (JIRA)

[ 
https://issues.apache.org/jira/browse/CARBONDATA-906?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15965243#comment-15965243
 ] 

Crabo Yang commented on CARBONDATA-906:
---

We keep turning Carbondata for a few months, sucessed on loading 5million row 
(about 100 columns per row)
web have a 20 nodes cluster and 32GB ram each.

CarbonData have slow loading data performance and we should split the job into 
really small batch to load.
Always a hard time to play with .

can you guys help me out of this OOM hell?

> Always OOM error when import large dataset (100milion rows)
> ---
>
> Key: CARBONDATA-906
> URL: https://issues.apache.org/jira/browse/CARBONDATA-906
> Project: CarbonData
>  Issue Type: Bug
>  Components: data-load
>Affects Versions: 1.0.0-incubating
>Reporter: Crabo Yang
>
> java.lang.OutOfMemoryError: GC overhead limit exceeded
>   at 
> java.util.concurrent.ConcurrentHashMap$Segment.put(ConcurrentHashMap.java:457)
>   at 
> java.util.concurrent.ConcurrentHashMap.put(ConcurrentHashMap.java:1130)
>   at 
> org.apache.carbondata.core.cache.dictionary.ColumnReverseDictionaryInfo.addDataToDictionaryMap(ColumnReverseDictionaryInfo.java:101)
>   at 
> org.apache.carbondata.core.cache.dictionary.ColumnReverseDictionaryInfo.addDictionaryChunk(ColumnReverseDictionaryInfo.java:88)
>   at 
> org.apache.carbondata.core.cache.dictionary.DictionaryCacheLoaderImpl.fillDictionaryValuesAndAddToDictionaryChunks(DictionaryCacheLoaderImpl.java:113)
>   at 
> org.apache.carbondata.core.cache.dictionary.DictionaryCacheLoaderImpl.load(DictionaryCacheLoaderImpl.java:81)
>   at 
> org.apache.carbondata.core.cache.dictionary.AbstractDictionaryCache.loadDictionaryData(AbstractDictionaryCache.java:236)
>   at 
> org.apache.carbondata.core.cache.dictionary.AbstractDictionaryCache.checkAndLoadDictionaryData(AbstractDictionaryCache.java:186)
>   at 
> org.apache.carbondata.core.cache.dictionary.ReverseDictionaryCache.getDictionary(ReverseDictionaryCache.java:174)
>   at 
> org.apache.carbondata.core.cache.dictionary.ReverseDictionaryCache.get(ReverseDictionaryCache.java:67)
>   at 
> org.apache.carbondata.core.cache.dictionary.ReverseDictionaryCache.get(ReverseDictionaryCache.java:38)
>   at 
> org.apache.carbondata.processing.newflow.converter.impl.DictionaryFieldConverterImpl.(DictionaryFieldConverterImpl.java:92)
>   at 
> org.apache.carbondata.processing.newflow.converter.impl.FieldEncoderFactory.createFieldEncoder(FieldEncoderFactory.java:77)
>   at 
> org.apache.carbondata.processing.newflow.converter.impl.RowConverterImpl.initialize(RowConverterImpl.java:102)
>   at 
> org.apache.carbondata.processing.newflow.steps.DataConverterProcessorStepImpl.initialize(DataConverterProcessorStepImpl.java:69)
>   at 
> org.apache.carbondata.processing.newflow.steps.SortProcessorStepImpl.initialize(SortProcessorStepImpl.java:57)
>   at 
> org.apache.carbondata.processing.newflow.steps.DataWriterProcessorStepImpl.initialize(DataWriterProcessorStepImpl.java:79)
>   at 
> org.apache.carbondata.processing.newflow.DataLoadExecutor.execute(DataLoadExecutor.java:45)
>   at 
> org.apache.carbondata.spark.rdd.NewDataFrameLoaderRDD$$anon$2.(NewCarbonDataLoadRDD.scala:425)
>   at 
> org.apache.carbondata.spark.rdd.NewDataFrameLoaderRDD.compute(NewCarbonDataLoadRDD.scala:383)
>   at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
>   at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
>   at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
>   at org.apache.spark.scheduler.Task.run(Task.scala:89)
>   at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>   at java.lang.Thread.run(Thread.java:745)



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[jira] [Created] (CARBONDATA-906) Always OOM error when import large dataset (100milion rows)

2017-04-11 Thread Crabo Yang (JIRA)
Crabo Yang created CARBONDATA-906:
-

 Summary: Always OOM error when import large dataset (100milion 
rows)
 Key: CARBONDATA-906
 URL: https://issues.apache.org/jira/browse/CARBONDATA-906
 Project: CarbonData
  Issue Type: Bug
  Components: data-load
Affects Versions: 1.0.0-incubating
Reporter: Crabo Yang


java.lang.OutOfMemoryError: GC overhead limit exceeded
at 
java.util.concurrent.ConcurrentHashMap$Segment.put(ConcurrentHashMap.java:457)
at 
java.util.concurrent.ConcurrentHashMap.put(ConcurrentHashMap.java:1130)
at 
org.apache.carbondata.core.cache.dictionary.ColumnReverseDictionaryInfo.addDataToDictionaryMap(ColumnReverseDictionaryInfo.java:101)
at 
org.apache.carbondata.core.cache.dictionary.ColumnReverseDictionaryInfo.addDictionaryChunk(ColumnReverseDictionaryInfo.java:88)
at 
org.apache.carbondata.core.cache.dictionary.DictionaryCacheLoaderImpl.fillDictionaryValuesAndAddToDictionaryChunks(DictionaryCacheLoaderImpl.java:113)
at 
org.apache.carbondata.core.cache.dictionary.DictionaryCacheLoaderImpl.load(DictionaryCacheLoaderImpl.java:81)
at 
org.apache.carbondata.core.cache.dictionary.AbstractDictionaryCache.loadDictionaryData(AbstractDictionaryCache.java:236)
at 
org.apache.carbondata.core.cache.dictionary.AbstractDictionaryCache.checkAndLoadDictionaryData(AbstractDictionaryCache.java:186)
at 
org.apache.carbondata.core.cache.dictionary.ReverseDictionaryCache.getDictionary(ReverseDictionaryCache.java:174)
at 
org.apache.carbondata.core.cache.dictionary.ReverseDictionaryCache.get(ReverseDictionaryCache.java:67)
at 
org.apache.carbondata.core.cache.dictionary.ReverseDictionaryCache.get(ReverseDictionaryCache.java:38)
at 
org.apache.carbondata.processing.newflow.converter.impl.DictionaryFieldConverterImpl.(DictionaryFieldConverterImpl.java:92)
at 
org.apache.carbondata.processing.newflow.converter.impl.FieldEncoderFactory.createFieldEncoder(FieldEncoderFactory.java:77)
at 
org.apache.carbondata.processing.newflow.converter.impl.RowConverterImpl.initialize(RowConverterImpl.java:102)
at 
org.apache.carbondata.processing.newflow.steps.DataConverterProcessorStepImpl.initialize(DataConverterProcessorStepImpl.java:69)
at 
org.apache.carbondata.processing.newflow.steps.SortProcessorStepImpl.initialize(SortProcessorStepImpl.java:57)
at 
org.apache.carbondata.processing.newflow.steps.DataWriterProcessorStepImpl.initialize(DataWriterProcessorStepImpl.java:79)
at 
org.apache.carbondata.processing.newflow.DataLoadExecutor.execute(DataLoadExecutor.java:45)
at 
org.apache.carbondata.spark.rdd.NewDataFrameLoaderRDD$$anon$2.(NewCarbonDataLoadRDD.scala:425)
at 
org.apache.carbondata.spark.rdd.NewDataFrameLoaderRDD.compute(NewCarbonDataLoadRDD.scala:383)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
at org.apache.spark.scheduler.Task.run(Task.scala:89)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)




--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[GitHub] incubator-carbondata issue #783: [CARBONDATA-903] data load is not failing e...

2017-04-11 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/incubator-carbondata/pull/783
  
@mohammadshahidkhan Please create a Abstract class and move the common 
methods to there.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] incubator-carbondata issue #755: [CARBONDATA-881] Load status is successful ...

2017-04-11 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/incubator-carbondata/pull/755
  
@mohammadshahidkhan Please check all caller methods of 
`writeLoadDetailsIntoFile` and do exception handling there also, otherwise same 
issue happens if they catch the error and just log it.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] incubator-carbondata issue #755: [CARBONDATA-881] Load status is successful ...

2017-04-11 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/incubator-carbondata/pull/755
  
LGTM


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] incubator-carbondata issue #782: [CARBONDATA-901] Fix some spelling mistakes

2017-04-11 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/incubator-carbondata/pull/782
  
Build Success with Spark 1.6.2, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/1577/



---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] incubator-carbondata pull request #782: [CARBONDATA-901] Fix some spelling m...

2017-04-11 Thread asfgit
Github user asfgit closed the pull request at:

https://github.com/apache/incubator-carbondata/pull/782


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] incubator-carbondata issue #782: [CARBONDATA-901] Fix some spelling mistakes

2017-04-11 Thread chenliang613
Github user chenliang613 commented on the issue:

https://github.com/apache/incubator-carbondata/pull/782
  
LGTM


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Commented] (CARBONDATA-901) Fix some spelling mistakes

2017-04-11 Thread Liang Chen (JIRA)

[ 
https://issues.apache.org/jira/browse/CARBONDATA-901?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15964830#comment-15964830
 ] 

Liang Chen commented on CARBONDATA-901:
---

Hi Cao Gaofei

Already gave the contributor right to you , so now you can assign issue to 
yourself.

> Fix some spelling mistakes
> --
>
> Key: CARBONDATA-901
> URL: https://issues.apache.org/jira/browse/CARBONDATA-901
> Project: CarbonData
>  Issue Type: Improvement
>  Components: core
>Reporter: Cao Gaofei
>Assignee: Cao Gaofei
>Priority: Trivial
>  Time Spent: 0.5h
>  Remaining Estimate: 0h
>
> https://github.com/apache/incubator-carbondata/pull/782



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[GitHub] incubator-carbondata pull request #783: [CARBONDATA-903] data load is not fa...

2017-04-11 Thread mohammadshahidkhan
Github user mohammadshahidkhan closed the pull request at:

https://github.com/apache/incubator-carbondata/pull/783


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] incubator-carbondata pull request #783: [CARBONDATA-903] data load is not fa...

2017-04-11 Thread mohammadshahidkhan
GitHub user mohammadshahidkhan reopened a pull request:

https://github.com/apache/incubator-carbondata/pull/783

[CARBONDATA-903] data load is not failing even though bad records exists in 
the data in case of unsafe sort or batch sort



You can merge this pull request into a Git repository by running:

$ git pull https://github.com/mohammadshahidkhan/incubator-carbondata 
dataload_termination_on_fail

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/incubator-carbondata/pull/783.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #783


commit db9cf1b714f28fdf5fd75911f54ed7a2ad6c3dec
Author: mohammadshahidkhan 
Date:   2017-04-11T12:56:51Z

[CARBONDATA-903] data load is not failing even though bad records exists in 
the data in case of unsafe sort or batch sort




---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] incubator-carbondata pull request #783: [CARBONDATA-903] data load is not fa...

2017-04-11 Thread mohammadshahidkhan
Github user mohammadshahidkhan closed the pull request at:

https://github.com/apache/incubator-carbondata/pull/783


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] incubator-carbondata pull request #783: [CARBONDATA-903] data load is not fa...

2017-04-11 Thread mohammadshahidkhan
Github user mohammadshahidkhan closed the pull request at:

https://github.com/apache/incubator-carbondata/pull/783


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] incubator-carbondata pull request #783: [CARBONDATA-903] data load is not fa...

2017-04-11 Thread mohammadshahidkhan
GitHub user mohammadshahidkhan reopened a pull request:

https://github.com/apache/incubator-carbondata/pull/783

[CARBONDATA-903] data load is not failing even though bad records exists in 
the data in case of unsafe sort or batch sort



You can merge this pull request into a Git repository by running:

$ git pull https://github.com/mohammadshahidkhan/incubator-carbondata 
dataload_termination_on_fail

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/incubator-carbondata/pull/783.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #783


commit db9cf1b714f28fdf5fd75911f54ed7a2ad6c3dec
Author: mohammadshahidkhan 
Date:   2017-04-11T12:56:51Z

[CARBONDATA-903] data load is not failing even though bad records exists in 
the data in case of unsafe sort or batch sort




---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] incubator-carbondata issue #784: [WIP]Pr730

2017-04-11 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/incubator-carbondata/pull/784
  
Build Success with Spark 1.6.2, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/1575/



---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] incubator-carbondata issue #784: [WIP]Pr730

2017-04-11 Thread QiangCai
Github user QiangCai commented on the issue:

https://github.com/apache/incubator-carbondata/pull/784
  
retest this please


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] incubator-carbondata issue #769: [CARBONDATA-888] Added include and exclude ...

2017-04-11 Thread jackylk
Github user jackylk commented on the issue:

https://github.com/apache/incubator-carbondata/pull/769
  
I will merge it later as we are preparing version 1.1 release, code is 
freezed for master branch. Or you can create another PR to merge into 12-dev 
branch. Thanks.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] incubator-carbondata issue #783: [CARBONDATA-903] data load is not failing e...

2017-04-11 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/incubator-carbondata/pull/783
  
Build Success with Spark 1.6.2, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/1573/



---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] incubator-carbondata issue #783: [CARBONDATA-903] data load is not failing e...

2017-04-11 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/incubator-carbondata/pull/783
  
Build Failed  with Spark 1.6.2, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/1572/



---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] incubator-carbondata issue #720: [CARBONDATA-642] Delete Subquery is not wor...

2017-04-11 Thread jackylk
Github user jackylk commented on the issue:

https://github.com/apache/incubator-carbondata/pull/720
  
LGTM


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Updated] (CARBONDATA-904) ArrayIndexOutOfBoundsException

2017-04-11 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-904?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO updated CARBONDATA-904:
-
  Attachment: Test_Data1.csv
  Test_Data1_h1.csv
 Environment: Spark1.6
Request participants:   (was: )
   Fix Version/s: 1.1.0-incubating

> ArrayIndexOutOfBoundsException 
> ---
>
> Key: CARBONDATA-904
> URL: https://issues.apache.org/jira/browse/CARBONDATA-904
> Project: CarbonData
>  Issue Type: Bug
> Environment: Spark1.6
>Reporter: SWATI RAO
> Fix For: 1.1.0-incubating
>
> Attachments: Test_Data1_h1.csv, Test_Data1_h1.csv
>
>
> Or operator is not working properly.
> When we execute these query in hive it is working fine but when we execute 
> the same in carbondata it throws an exception:
> java.lang.ArrayIndexOutOfBoundsException
> HIVE:
> 0: jdbc:hive2://hadoop-master:1> create table Test_Boundary_h1 (c1_int 
> int,c2_Bigint Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string 
> string,c6_Timestamp Timestamp,c7_Datatype_Desc string) ROW FORMAT DELIMITED 
> FIELDS TERMINATED BY ',' ;
> +-+--+
> | result  |
> +-+--+
> +-+--+
> No rows selected (1.177 seconds)
> 0: jdbc:hive2://hadoop-master:1> load data local inpath 
> '/opt/Carbon/CarbonData/TestData/Data/Test_Data1_h1.csv' OVERWRITE INTO TABLE 
> Test_Boundary_h1 ;
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> No rows selected (0.437 seconds)
> 0: jdbc:hive2://hadoop-master:1> select c6_Timestamp,max(c6_Timestamp) 
> from Test_Boundary_h1 where c6_Timestamp ='2017-07-01 12:07:28' or 
> c6_Timestamp ='2019-07-05 13:07:30' or c6_Timestamp = '1999-01-06 10:05:29' 
> group by c6_Timestamp ;
> +++--+
> |  c6_Timestamp  |  _c1   |
> +++--+
> | 2017-07-01 12:07:28.0  | 2017-07-01 12:07:28.0  |
> +++--+
> 1 row selected (1.637 seconds)
> CARBONDATA:
> 0: jdbc:hive2://hadoop-master:1> create table Test_Boundary (c1_int 
> int,c2_Bigint Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string 
> string,c6_Timestamp Timestamp,c7_Datatype_Desc string) STORED BY 
> 'org.apache.carbondata.format' ;
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> No rows selected (4.48 seconds)
> 0: jdbc:hive2://hadoop-master:1> LOAD DATA INPATH 
> 'hdfs://192.168.2.145:54310/BabuStore/Data/Test_Data1.csv' INTO table 
> Test_Boundary 
> OPTIONS('DELIMITER'=',','QUOTECHAR'='','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='')
>  ;
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> No rows selected (4.445 seconds)
> 0: jdbc:hive2://hadoop-master:1> select c6_Timestamp,max(c6_Timestamp) 
> from Test_Boundary where c6_Timestamp ='2017-07-01 12:07:28' or c6_Timestamp 
> =' 2019-07-05 13:07:30' or c6_Timestamp = '1999-01-06 10:05:29' group by 
> c6_Timestamp ;
> Error: org.apache.spark.SparkException: Job aborted due to stage failure: 
> Task 0 in stage 5.0 failed 4 times, most recent failure: Lost task 0.3 in 
> stage 5.0 (TID 8, hadoop-master): java.lang.RuntimeException: 
> java.util.concurrent.ExecutionException: 
> java.lang.ArrayIndexOutOfBoundsException: 0
>   at 
> org.apache.carbondata.core.scan.processor.AbstractDataBlockIterator.updateScanner(AbstractDataBlockIterator.java:136)
>   at 
> org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:50)
>   at 
> org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:32)
>   at 
> org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.getBatchResult(DetailQueryResultIterator.java:50)
>   at 
> org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.next(DetailQueryResultIterator.java:41)
>   at 
> org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.next(DetailQueryResultIterator.java:31)
>   at 
> org.apache.carbondata.core.scan.result.iterator.ChunkRowIterator.(ChunkRowIterator.java:41)
>   at 
> org.apache.carbondata.hadoop.CarbonRecordReader.initialize(CarbonRecordReader.java:79)
>   at 
> org.apache.carbondata.spark.rdd.CarbonScanRDD.compute(CarbonScanRDD.scala:204)
>   at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
>   at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
>   at 
> org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
>   at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
>   at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
>   at 
> 

[jira] [Updated] (CARBONDATA-904) ArrayIndexOutOfBoundsException

2017-04-11 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-904?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO updated CARBONDATA-904:
-
Attachment: (was: Test_Data1.csv)

> ArrayIndexOutOfBoundsException 
> ---
>
> Key: CARBONDATA-904
> URL: https://issues.apache.org/jira/browse/CARBONDATA-904
> Project: CarbonData
>  Issue Type: Bug
> Environment: Spark1.6
>Reporter: SWATI RAO
> Fix For: 1.1.0-incubating
>
> Attachments: Test_Data1_h1.csv, Test_Data1_h1.csv
>
>
> Or operator is not working properly.
> When we execute these query in hive it is working fine but when we execute 
> the same in carbondata it throws an exception:
> java.lang.ArrayIndexOutOfBoundsException
> HIVE:
> 0: jdbc:hive2://hadoop-master:1> create table Test_Boundary_h1 (c1_int 
> int,c2_Bigint Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string 
> string,c6_Timestamp Timestamp,c7_Datatype_Desc string) ROW FORMAT DELIMITED 
> FIELDS TERMINATED BY ',' ;
> +-+--+
> | result  |
> +-+--+
> +-+--+
> No rows selected (1.177 seconds)
> 0: jdbc:hive2://hadoop-master:1> load data local inpath 
> '/opt/Carbon/CarbonData/TestData/Data/Test_Data1_h1.csv' OVERWRITE INTO TABLE 
> Test_Boundary_h1 ;
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> No rows selected (0.437 seconds)
> 0: jdbc:hive2://hadoop-master:1> select c6_Timestamp,max(c6_Timestamp) 
> from Test_Boundary_h1 where c6_Timestamp ='2017-07-01 12:07:28' or 
> c6_Timestamp ='2019-07-05 13:07:30' or c6_Timestamp = '1999-01-06 10:05:29' 
> group by c6_Timestamp ;
> +++--+
> |  c6_Timestamp  |  _c1   |
> +++--+
> | 2017-07-01 12:07:28.0  | 2017-07-01 12:07:28.0  |
> +++--+
> 1 row selected (1.637 seconds)
> CARBONDATA:
> 0: jdbc:hive2://hadoop-master:1> create table Test_Boundary (c1_int 
> int,c2_Bigint Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string 
> string,c6_Timestamp Timestamp,c7_Datatype_Desc string) STORED BY 
> 'org.apache.carbondata.format' ;
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> No rows selected (4.48 seconds)
> 0: jdbc:hive2://hadoop-master:1> LOAD DATA INPATH 
> 'hdfs://192.168.2.145:54310/BabuStore/Data/Test_Data1.csv' INTO table 
> Test_Boundary 
> OPTIONS('DELIMITER'=',','QUOTECHAR'='','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='')
>  ;
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> No rows selected (4.445 seconds)
> 0: jdbc:hive2://hadoop-master:1> select c6_Timestamp,max(c6_Timestamp) 
> from Test_Boundary where c6_Timestamp ='2017-07-01 12:07:28' or c6_Timestamp 
> =' 2019-07-05 13:07:30' or c6_Timestamp = '1999-01-06 10:05:29' group by 
> c6_Timestamp ;
> Error: org.apache.spark.SparkException: Job aborted due to stage failure: 
> Task 0 in stage 5.0 failed 4 times, most recent failure: Lost task 0.3 in 
> stage 5.0 (TID 8, hadoop-master): java.lang.RuntimeException: 
> java.util.concurrent.ExecutionException: 
> java.lang.ArrayIndexOutOfBoundsException: 0
>   at 
> org.apache.carbondata.core.scan.processor.AbstractDataBlockIterator.updateScanner(AbstractDataBlockIterator.java:136)
>   at 
> org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:50)
>   at 
> org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:32)
>   at 
> org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.getBatchResult(DetailQueryResultIterator.java:50)
>   at 
> org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.next(DetailQueryResultIterator.java:41)
>   at 
> org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.next(DetailQueryResultIterator.java:31)
>   at 
> org.apache.carbondata.core.scan.result.iterator.ChunkRowIterator.(ChunkRowIterator.java:41)
>   at 
> org.apache.carbondata.hadoop.CarbonRecordReader.initialize(CarbonRecordReader.java:79)
>   at 
> org.apache.carbondata.spark.rdd.CarbonScanRDD.compute(CarbonScanRDD.scala:204)
>   at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
>   at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
>   at 
> org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
>   at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
>   at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
>   at 
> org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
>   at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
>   at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
>   at 
> 

[jira] [Updated] (CARBONDATA-904) ArrayIndexOutOfBoundsException

2017-04-11 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-904?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO updated CARBONDATA-904:
-
Attachment: (was: Test_Data1.csv)

> ArrayIndexOutOfBoundsException 
> ---
>
> Key: CARBONDATA-904
> URL: https://issues.apache.org/jira/browse/CARBONDATA-904
> Project: CarbonData
>  Issue Type: Bug
> Environment: Spark1.6
>Reporter: SWATI RAO
> Fix For: 1.1.0-incubating
>
> Attachments: Test_Data1_h1.csv, Test_Data1_h1.csv
>
>
> Or operator is not working properly.
> When we execute these query in hive it is working fine but when we execute 
> the same in carbondata it throws an exception:
> java.lang.ArrayIndexOutOfBoundsException
> HIVE:
> 0: jdbc:hive2://hadoop-master:1> create table Test_Boundary_h1 (c1_int 
> int,c2_Bigint Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string 
> string,c6_Timestamp Timestamp,c7_Datatype_Desc string) ROW FORMAT DELIMITED 
> FIELDS TERMINATED BY ',' ;
> +-+--+
> | result  |
> +-+--+
> +-+--+
> No rows selected (1.177 seconds)
> 0: jdbc:hive2://hadoop-master:1> load data local inpath 
> '/opt/Carbon/CarbonData/TestData/Data/Test_Data1_h1.csv' OVERWRITE INTO TABLE 
> Test_Boundary_h1 ;
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> No rows selected (0.437 seconds)
> 0: jdbc:hive2://hadoop-master:1> select c6_Timestamp,max(c6_Timestamp) 
> from Test_Boundary_h1 where c6_Timestamp ='2017-07-01 12:07:28' or 
> c6_Timestamp ='2019-07-05 13:07:30' or c6_Timestamp = '1999-01-06 10:05:29' 
> group by c6_Timestamp ;
> +++--+
> |  c6_Timestamp  |  _c1   |
> +++--+
> | 2017-07-01 12:07:28.0  | 2017-07-01 12:07:28.0  |
> +++--+
> 1 row selected (1.637 seconds)
> CARBONDATA:
> 0: jdbc:hive2://hadoop-master:1> create table Test_Boundary (c1_int 
> int,c2_Bigint Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string 
> string,c6_Timestamp Timestamp,c7_Datatype_Desc string) STORED BY 
> 'org.apache.carbondata.format' ;
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> No rows selected (4.48 seconds)
> 0: jdbc:hive2://hadoop-master:1> LOAD DATA INPATH 
> 'hdfs://192.168.2.145:54310/BabuStore/Data/Test_Data1.csv' INTO table 
> Test_Boundary 
> OPTIONS('DELIMITER'=',','QUOTECHAR'='','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='')
>  ;
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> No rows selected (4.445 seconds)
> 0: jdbc:hive2://hadoop-master:1> select c6_Timestamp,max(c6_Timestamp) 
> from Test_Boundary where c6_Timestamp ='2017-07-01 12:07:28' or c6_Timestamp 
> =' 2019-07-05 13:07:30' or c6_Timestamp = '1999-01-06 10:05:29' group by 
> c6_Timestamp ;
> Error: org.apache.spark.SparkException: Job aborted due to stage failure: 
> Task 0 in stage 5.0 failed 4 times, most recent failure: Lost task 0.3 in 
> stage 5.0 (TID 8, hadoop-master): java.lang.RuntimeException: 
> java.util.concurrent.ExecutionException: 
> java.lang.ArrayIndexOutOfBoundsException: 0
>   at 
> org.apache.carbondata.core.scan.processor.AbstractDataBlockIterator.updateScanner(AbstractDataBlockIterator.java:136)
>   at 
> org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:50)
>   at 
> org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:32)
>   at 
> org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.getBatchResult(DetailQueryResultIterator.java:50)
>   at 
> org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.next(DetailQueryResultIterator.java:41)
>   at 
> org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.next(DetailQueryResultIterator.java:31)
>   at 
> org.apache.carbondata.core.scan.result.iterator.ChunkRowIterator.(ChunkRowIterator.java:41)
>   at 
> org.apache.carbondata.hadoop.CarbonRecordReader.initialize(CarbonRecordReader.java:79)
>   at 
> org.apache.carbondata.spark.rdd.CarbonScanRDD.compute(CarbonScanRDD.scala:204)
>   at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
>   at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
>   at 
> org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
>   at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
>   at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
>   at 
> org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
>   at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
>   at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
>   at 
> 

[GitHub] incubator-carbondata pull request #775: [CARBONDATA-660] Bad Records Logs an...

2017-04-11 Thread asfgit
Github user asfgit closed the pull request at:

https://github.com/apache/incubator-carbondata/pull/775


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] incubator-carbondata issue #775: [CARBONDATA-660] Bad Records Logs and Raw C...

2017-04-11 Thread jackylk
Github user jackylk commented on the issue:

https://github.com/apache/incubator-carbondata/pull/775
  
LGTM


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] incubator-carbondata issue #737: [CARBONDATA-882] Add SORT_COLUMNS support i...

2017-04-11 Thread jackylk
Github user jackylk commented on the issue:

https://github.com/apache/incubator-carbondata/pull/737
  
please retest this


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Created] (CARBONDATA-904) ArrayIndexOutOfBoundsException

2017-04-11 Thread SWATI RAO (JIRA)
SWATI RAO created CARBONDATA-904:


 Summary: ArrayIndexOutOfBoundsException 
 Key: CARBONDATA-904
 URL: https://issues.apache.org/jira/browse/CARBONDATA-904
 Project: CarbonData
  Issue Type: Bug
Reporter: SWATI RAO
 Attachments: Test_Data1.csv, Test_Data1_h1.csv

Or operator is not working properly.

When we execute these query in hive it is working fine but when we execute the 
same in carbondata it throws an exception:
java.lang.ArrayIndexOutOfBoundsException

HIVE:
0: jdbc:hive2://hadoop-master:1> create table Test_Boundary_h1 (c1_int 
int,c2_Bigint Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string 
string,c6_Timestamp Timestamp,c7_Datatype_Desc string) ROW FORMAT DELIMITED 
FIELDS TERMINATED BY ',' ;
+-+--+
| result  |
+-+--+
+-+--+
No rows selected (1.177 seconds)
0: jdbc:hive2://hadoop-master:1> load data local inpath 
'/opt/Carbon/CarbonData/TestData/Data/Test_Data1_h1.csv' OVERWRITE INTO TABLE 
Test_Boundary_h1 ;
+-+--+
| Result  |
+-+--+
+-+--+
No rows selected (0.437 seconds)
0: jdbc:hive2://hadoop-master:1> select c6_Timestamp,max(c6_Timestamp) from 
Test_Boundary_h1 where c6_Timestamp ='2017-07-01 12:07:28' or c6_Timestamp 
='2019-07-05 13:07:30' or c6_Timestamp = '1999-01-06 10:05:29' group by 
c6_Timestamp ;
+++--+
|  c6_Timestamp  |  _c1   |
+++--+
| 2017-07-01 12:07:28.0  | 2017-07-01 12:07:28.0  |
+++--+
1 row selected (1.637 seconds)

CARBONDATA:
0: jdbc:hive2://hadoop-master:1> create table Test_Boundary (c1_int 
int,c2_Bigint Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string 
string,c6_Timestamp Timestamp,c7_Datatype_Desc string) STORED BY 
'org.apache.carbondata.format' ;
+-+--+
| Result  |
+-+--+
+-+--+
No rows selected (4.48 seconds)

0: jdbc:hive2://hadoop-master:1> LOAD DATA INPATH 
'hdfs://192.168.2.145:54310/BabuStore/Data/Test_Data1.csv' INTO table 
Test_Boundary 
OPTIONS('DELIMITER'=',','QUOTECHAR'='','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='')
 ;
+-+--+
| Result  |
+-+--+
+-+--+
No rows selected (4.445 seconds)
0: jdbc:hive2://hadoop-master:1> select c6_Timestamp,max(c6_Timestamp) from 
Test_Boundary where c6_Timestamp ='2017-07-01 12:07:28' or c6_Timestamp =' 
2019-07-05 13:07:30' or c6_Timestamp = '1999-01-06 10:05:29' group by 
c6_Timestamp ;
Error: org.apache.spark.SparkException: Job aborted due to stage failure: Task 
0 in stage 5.0 failed 4 times, most recent failure: Lost task 0.3 in stage 5.0 
(TID 8, hadoop-master): java.lang.RuntimeException: 
java.util.concurrent.ExecutionException: 
java.lang.ArrayIndexOutOfBoundsException: 0
at 
org.apache.carbondata.core.scan.processor.AbstractDataBlockIterator.updateScanner(AbstractDataBlockIterator.java:136)
at 
org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:50)
at 
org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:32)
at 
org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.getBatchResult(DetailQueryResultIterator.java:50)
at 
org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.next(DetailQueryResultIterator.java:41)
at 
org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.next(DetailQueryResultIterator.java:31)
at 
org.apache.carbondata.core.scan.result.iterator.ChunkRowIterator.(ChunkRowIterator.java:41)
at 
org.apache.carbondata.hadoop.CarbonRecordReader.initialize(CarbonRecordReader.java:79)
at 
org.apache.carbondata.spark.rdd.CarbonScanRDD.compute(CarbonScanRDD.scala:204)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)
at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
at 

[GitHub] incubator-carbondata issue #783: [CARBONDATA-903] data load is not failing e...

2017-04-11 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/incubator-carbondata/pull/783
  
Build Success with Spark 1.6.2, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/1570/



---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] incubator-carbondata pull request #709: [CARBONDATA-861] Improvements in que...

2017-04-11 Thread asfgit
Github user asfgit closed the pull request at:

https://github.com/apache/incubator-carbondata/pull/709


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Assigned] (CARBONDATA-902) NoClassDefFoundError for Decimal datatype during select queries

2017-04-11 Thread anubhav tarar (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-902?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

anubhav tarar reassigned CARBONDATA-902:


Assignee: anubhav tarar

> NoClassDefFoundError for Decimal datatype during select queries
> ---
>
> Key: CARBONDATA-902
> URL: https://issues.apache.org/jira/browse/CARBONDATA-902
> Project: CarbonData
>  Issue Type: Bug
>  Components: data-query
> Environment: Spark 2.1, Hive 1.2.1
>Reporter: Neha Bhardwaj
>Assignee: anubhav tarar
>Priority: Minor
> Attachments: testHive1.csv
>
>
> Decimal data type raises exception while selecting the data from the table in 
> hive.
> Steps to reproduce:
> 1) In Spark Shell :
>  a) Create Table -
> import org.apache.spark.sql.SparkSession
> import org.apache.spark.sql.CarbonSession._
> val carbon = 
> SparkSession.builder().config(sc.getConf).getOrCreateCarbonSession("hdfs://localhost:54310/opt/data")
>  
>  scala> carbon.sql(""" create table testHive1(id int,name string,dob 
> timestamp,experience decimal,salary double,incentive bigint) stored 
> by'carbondata' """).show 
>  b) Load Data - 
> scala> carbon.sql(""" load data inpath 
> 'hdfs://localhost:54310/Files/testHive1.csv' into table testHive1 """ ).show
> 2) In Hive : 
>  a) Add Jars - 
> add jar 
> /home/neha/incubator-carbondata/integration/hive/carbondata-hive-1.1.0-incubating-SNAPSHOT.jar;
> add jar 
> /home/neha/incubator-carbondata/assembly/target/scala-2.11/carbondata_2.11-1.1.0-incubating-SNAPSHOT-shade-hadoop2.7.2.jar;
>  
>  b) Set Properties - 
> set hive.mapred.supports.subdirectories=true;
> set mapreduce.input.fileinputformat.input.dir.recursive=true;
> c) Alter location - 
> hive> alter table testHive1 set LOCATION 
> 'hdfs://localhost:54310/opt/data/default/testhive1' ;
> d) Alter FileFormat -
> alter table testHive1 set FILEFORMAT
> INPUTFORMAT "org.apache.carbondata.hive.MapredCarbonInputFormat"
> OUTPUTFORMAT "org.apache.carbondata.hive.MapredCarbonOutputFormat"
> SERDE "org.apache.carbondata.hive.CarbonHiveSerDe";
>  e) Create Table -
> create table testHive1(id int,name string,dob timestamp,experience 
> decimal,salary double,incentive bigint);
>  f) Execute Queries - 
> select * from testHive1;
> 3) Query :
> hive> select * from testHive1;
> Expected Output : 
> ResultSet should display all the data present in the table.
> Result:
> Exception in thread "[main][partitionID:testhive1;queryID:8945394553892]" 
> java.lang.NoClassDefFoundError: org/apache/spark/sql/types/Decimal
>   at 
> org.apache.carbondata.core.scan.collector.impl.AbstractScannedResultCollector.getMeasureData(AbstractScannedResultCollector.java:109)
>   at 
> org.apache.carbondata.core.scan.collector.impl.AbstractScannedResultCollector.fillMeasureData(AbstractScannedResultCollector.java:78)
>   at 
> org.apache.carbondata.core.scan.collector.impl.DictionaryBasedResultCollector.fillMeasureData(DictionaryBasedResultCollector.java:158)
>   at 
> org.apache.carbondata.core.scan.collector.impl.DictionaryBasedResultCollector.collectData(DictionaryBasedResultCollector.java:115)
>   at 
> org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:51)
>   at 
> org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:32)
>   at 
> org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.getBatchResult(DetailQueryResultIterator.java:50)
>   at 
> org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.next(DetailQueryResultIterator.java:41)
>   at 
> org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.next(DetailQueryResultIterator.java:31)
>   at 
> org.apache.carbondata.core.scan.result.iterator.ChunkRowIterator.(ChunkRowIterator.java:41)
>   at 
> org.apache.carbondata.hive.CarbonHiveRecordReader.initialize(CarbonHiveRecordReader.java:84)
>   at 
> org.apache.carbondata.hive.CarbonHiveRecordReader.(CarbonHiveRecordReader.java:66)
>   at 
> org.apache.carbondata.hive.MapredCarbonInputFormat.getRecordReader(MapredCarbonInputFormat.java:68)
>   at 
> org.apache.hadoop.hive.ql.exec.FetchOperator$FetchInputFormatSplit.getRecordReader(FetchOperator.java:673)
>   at 
> org.apache.hadoop.hive.ql.exec.FetchOperator.getRecordReader(FetchOperator.java:323)
>   at 
> org.apache.hadoop.hive.ql.exec.FetchOperator.getNextRow(FetchOperator.java:445)
>   at 
> org.apache.hadoop.hive.ql.exec.FetchOperator.pushRow(FetchOperator.java:414)
>   at org.apache.hadoop.hive.ql.exec.FetchTask.fetch(FetchTask.java:140)
>   at org.apache.hadoop.hive.ql.Driver.getResults(Driver.java:1670)
>   at 
> 

[GitHub] incubator-carbondata issue #709: [CARBONDATA-861] Improvements in query

2017-04-11 Thread jackylk
Github user jackylk commented on the issue:

https://github.com/apache/incubator-carbondata/pull/709
  
LGTM


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] incubator-carbondata pull request #783: [CARBONDATA-903] data load is not fa...

2017-04-11 Thread mohammadshahidkhan
GitHub user mohammadshahidkhan opened a pull request:

https://github.com/apache/incubator-carbondata/pull/783

[CARBONDATA-903] data load is not failing even though bad records exists in 
the data in case of unsafe sort or batch sort



You can merge this pull request into a Git repository by running:

$ git pull https://github.com/mohammadshahidkhan/incubator-carbondata 
dataload_termination_on_fail

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/incubator-carbondata/pull/783.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #783


commit 39d329a3ca6eda1d9e372a0dbabc804b5b7ad493
Author: mohammadshahidkhan 
Date:   2017-04-11T12:56:51Z

[CARBONDATA-903] data load is not failing even though bad records exists in 
the data in case of unsafe sort or batch sort




---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Updated] (CARBONDATA-903) data load is not failing even though bad records exists in the data in case of unsafe sort or batch sort

2017-04-11 Thread Mohammad Shahid Khan (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-903?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mohammad Shahid Khan updated CARBONDATA-903:

Summary: data load is not failing even though bad records exists in the 
data in case of unsafe sort or batch sort  (was: data load is not failing even 
though bad records exists in the data)

> data load is not failing even though bad records exists in the data in case 
> of unsafe sort or batch sort
> 
>
> Key: CARBONDATA-903
> URL: https://issues.apache.org/jira/browse/CARBONDATA-903
> Project: CarbonData
>  Issue Type: Bug
>Reporter: Mohammad Shahid Khan
>Assignee: Mohammad Shahid Khan
>Priority: Critical
>




--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[jira] [Commented] (CARBONDATA-903) data load is not failing even though bad records exists in the data

2017-04-11 Thread Mohammad Shahid Khan (JIRA)

[ 
https://issues.apache.org/jira/browse/CARBONDATA-903?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15964327#comment-15964327
 ] 

Mohammad Shahid Khan commented on CARBONDATA-903:
-

When carbon.load.use.batch.sort or enable.unsafe.sort or table bucket 
configured 
the data loading is not failing even in though the bad records exists in the 
data being loaded.

> data load is not failing even though bad records exists in the data
> ---
>
> Key: CARBONDATA-903
> URL: https://issues.apache.org/jira/browse/CARBONDATA-903
> Project: CarbonData
>  Issue Type: Bug
>Reporter: Mohammad Shahid Khan
>Assignee: Mohammad Shahid Khan
>Priority: Critical
>




--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[jira] [Assigned] (CARBONDATA-903) data load is not failing even though bad records exists in the data

2017-04-11 Thread Mohammad Shahid Khan (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-903?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mohammad Shahid Khan reassigned CARBONDATA-903:
---

Assignee: Mohammad Shahid Khan

> data load is not failing even though bad records exists in the data
> ---
>
> Key: CARBONDATA-903
> URL: https://issues.apache.org/jira/browse/CARBONDATA-903
> Project: CarbonData
>  Issue Type: Bug
>Reporter: Mohammad Shahid Khan
>Assignee: Mohammad Shahid Khan
>Priority: Critical
>




--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[jira] [Created] (CARBONDATA-903) data load is not failing even though bad records exists in the data

2017-04-11 Thread Mohammad Shahid Khan (JIRA)
Mohammad Shahid Khan created CARBONDATA-903:
---

 Summary: data load is not failing even though bad records exists 
in the data
 Key: CARBONDATA-903
 URL: https://issues.apache.org/jira/browse/CARBONDATA-903
 Project: CarbonData
  Issue Type: Bug
Reporter: Mohammad Shahid Khan
Priority: Critical






--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[jira] [Created] (CARBONDATA-902) NoClassDefFoundError for Decimal datatype during select queries

2017-04-11 Thread Neha Bhardwaj (JIRA)
Neha Bhardwaj created CARBONDATA-902:


 Summary: NoClassDefFoundError for Decimal datatype during select 
queries
 Key: CARBONDATA-902
 URL: https://issues.apache.org/jira/browse/CARBONDATA-902
 Project: CarbonData
  Issue Type: Bug
  Components: data-query
 Environment: Spark 2.1, Hive 1.2.1
Reporter: Neha Bhardwaj
Priority: Minor
 Attachments: testHive1.csv

Decimal data type raises exception while selecting the data from the table in 
hive.

Steps to reproduce:
1) In Spark Shell :

 a) Create Table -
import org.apache.spark.sql.SparkSession
import org.apache.spark.sql.CarbonSession._

val carbon = 
SparkSession.builder().config(sc.getConf).getOrCreateCarbonSession("hdfs://localhost:54310/opt/data")
 
 scala> carbon.sql(""" create table testHive1(id int,name string,dob 
timestamp,experience decimal,salary double,incentive bigint) stored 
by'carbondata' """).show 

 b) Load Data - 
scala> carbon.sql(""" load data inpath 
'hdfs://localhost:54310/Files/testHive1.csv' into table testHive1 """ ).show


2) In Hive : 

 a) Add Jars - 
add jar 
/home/neha/incubator-carbondata/integration/hive/carbondata-hive-1.1.0-incubating-SNAPSHOT.jar;
add jar 
/home/neha/incubator-carbondata/assembly/target/scala-2.11/carbondata_2.11-1.1.0-incubating-SNAPSHOT-shade-hadoop2.7.2.jar;
 
 b) Set Properties - 
set hive.mapred.supports.subdirectories=true;
set mapreduce.input.fileinputformat.input.dir.recursive=true;

c) Alter location - 
hive> alter table testHive1 set LOCATION 
'hdfs://localhost:54310/opt/data/default/testhive1' ;

d) Alter FileFormat -
alter table testHive1 set FILEFORMAT
INPUTFORMAT "org.apache.carbondata.hive.MapredCarbonInputFormat"
OUTPUTFORMAT "org.apache.carbondata.hive.MapredCarbonOutputFormat"
SERDE "org.apache.carbondata.hive.CarbonHiveSerDe";

 e) Create Table -
create table testHive1(id int,name string,dob timestamp,experience 
decimal,salary double,incentive bigint);

 f) Execute Queries - 
select * from testHive1;

3) Query :
hive> select * from testHive1;

Expected Output : 
ResultSet should display all the data present in the table.

Result:
Exception in thread "[main][partitionID:testhive1;queryID:8945394553892]" 
java.lang.NoClassDefFoundError: org/apache/spark/sql/types/Decimal
at 
org.apache.carbondata.core.scan.collector.impl.AbstractScannedResultCollector.getMeasureData(AbstractScannedResultCollector.java:109)
at 
org.apache.carbondata.core.scan.collector.impl.AbstractScannedResultCollector.fillMeasureData(AbstractScannedResultCollector.java:78)
at 
org.apache.carbondata.core.scan.collector.impl.DictionaryBasedResultCollector.fillMeasureData(DictionaryBasedResultCollector.java:158)
at 
org.apache.carbondata.core.scan.collector.impl.DictionaryBasedResultCollector.collectData(DictionaryBasedResultCollector.java:115)
at 
org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:51)
at 
org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:32)
at 
org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.getBatchResult(DetailQueryResultIterator.java:50)
at 
org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.next(DetailQueryResultIterator.java:41)
at 
org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.next(DetailQueryResultIterator.java:31)
at 
org.apache.carbondata.core.scan.result.iterator.ChunkRowIterator.(ChunkRowIterator.java:41)
at 
org.apache.carbondata.hive.CarbonHiveRecordReader.initialize(CarbonHiveRecordReader.java:84)
at 
org.apache.carbondata.hive.CarbonHiveRecordReader.(CarbonHiveRecordReader.java:66)
at 
org.apache.carbondata.hive.MapredCarbonInputFormat.getRecordReader(MapredCarbonInputFormat.java:68)
at 
org.apache.hadoop.hive.ql.exec.FetchOperator$FetchInputFormatSplit.getRecordReader(FetchOperator.java:673)
at 
org.apache.hadoop.hive.ql.exec.FetchOperator.getRecordReader(FetchOperator.java:323)
at 
org.apache.hadoop.hive.ql.exec.FetchOperator.getNextRow(FetchOperator.java:445)
at 
org.apache.hadoop.hive.ql.exec.FetchOperator.pushRow(FetchOperator.java:414)
at org.apache.hadoop.hive.ql.exec.FetchTask.fetch(FetchTask.java:140)
at org.apache.hadoop.hive.ql.Driver.getResults(Driver.java:1670)
at 
org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:233)
at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:165)
at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:376)
at 
org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:736)
at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:681)
at 

[jira] [Commented] (CARBONDATA-688) Abnormal behaviour of double datatype when used in DICTIONARY_INCLUDE and filtering null values

2017-04-11 Thread Rahul Kumar (JIRA)

[ 
https://issues.apache.org/jira/browse/CARBONDATA-688?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15964302#comment-15964302
 ] 

Rahul Kumar commented on CARBONDATA-688:


[~geetikagupta],
This is working fine now . You can close it.

> Abnormal behaviour of double datatype when used in DICTIONARY_INCLUDE and 
> filtering null values
> ---
>
> Key: CARBONDATA-688
> URL: https://issues.apache.org/jira/browse/CARBONDATA-688
> Project: CarbonData
>  Issue Type: Bug
>  Components: data-query
>Affects Versions: 1.1.0-incubating
> Environment: Spark 2.1
>Reporter: Geetika Gupta
>Priority: Minor
> Attachments: 100_olap_C20.csv
>
>
> I tried to create a table having double as a column and load null values into 
> that table. When I performed the select query on the table, it is displaying 
> wrong data.
> Below are the commands used:
> Create table :
> create table  Comp_VMALL_DICTIONARY_INCLUDE (imei string,deviceInformationId 
> int,MAC string,deviceColor string,device_backColor string,modelId 
> string,marketName string,AMSize string,ROMSize string,CUPAudit 
> string,CPIClocked string,series string,productionDate timestamp,bomCode 
> string,internalModels string, deliveryTime string, channelsId string, 
> channelsName string , deliveryAreaId string, deliveryCountry string, 
> deliveryProvince string, deliveryCity string,deliveryDistrict string, 
> deliveryStreet string, oxSingleNumber string, ActiveCheckTime string, 
> ActiveAreaId string, ActiveCountry string, ActiveProvince string, Activecity 
> string, ActiveDistrict string, ActiveStreet string, ActiveOperatorId string, 
> Active_releaseId string, Active_EMUIVersion string, Active_operaSysVersion 
> string, Active_BacVerNumber string, Active_BacFlashVer string, 
> Active_webUIVersion string, Active_webUITypeCarrVer 
> string,Active_webTypeDataVerNumber string, Active_operatorsVersion string, 
> Active_phonePADPartitionedVersions string, Latest_YEAR int, Latest_MONTH int, 
> Latest_DAY Decimal(30,10), Latest_HOUR string, Latest_areaId string, 
> Latest_country string, Latest_province string, Latest_city string, 
> Latest_district string, Latest_street string, Latest_releaseId string, 
> Latest_EMUIVersion string, Latest_operaSysVersion string, Latest_BacVerNumber 
> string, Latest_BacFlashVer string, Latest_webUIVersion string, 
> Latest_webUITypeCarrVer string, Latest_webTypeDataVerNumber string, 
> Latest_operatorsVersion string, Latest_phonePADPartitionedVersions string, 
> Latest_operatorId string, gamePointDescription string,gamePointId 
> double,contractNumber BigInt)  STORED BY 'org.apache.carbondata.format' 
> TBLPROPERTIES('DICTIONARY_INCLUDE'='imei,deviceInformationId,productionDate,gamePointId,Latest_DAY,contractNumber');
> Load command:
> LOAD DATA INPATH  'hdfs://localhost:54311/BabuStore/DATA/100_olap_C20.csv' 
> INTO table Comp_VMALL_DICTIONARY_INCLUDE options ('DELIMITER'=',', 
> 'QUOTECHAR'='"', 
> 'BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='imei,deviceInformationId,MAC,deviceColor,device_backColor,modelId,marketName,AMSize,ROMSize,CUPAudit,CPIClocked,series,productionDate,bomCode,internalModels,deliveryTime,channelsId,channelsName,deliveryAreaId,deliveryCountry,deliveryProvince,deliveryCity,deliveryDistrict,deliveryStreet,oxSingleNumber,contractNumber,ActiveCheckTime,ActiveAreaId,ActiveCountry,ActiveProvince,Activecity,ActiveDistrict,ActiveStreet,ActiveOperatorId,Active_releaseId,Active_EMUIVersion,Active_operaSysVersion,Active_BacVerNumber,Active_BacFlashVer,Active_webUIVersion,Active_webUITypeCarrVer,Active_webTypeDataVerNumber,Active_operatorsVersion,Active_phonePADPartitionedVersions,Latest_YEAR,Latest_MONTH,Latest_DAY,Latest_HOUR,Latest_areaId,Latest_country,Latest_province,Latest_city,Latest_district,Latest_street,Latest_releaseId,Latest_EMUIVersion,Latest_operaSysVersion,Latest_BacVerNumber,Latest_BacFlashVer,Latest_webUIVersion,Latest_webUITypeCarrVer,Latest_webTypeDataVerNumber,Latest_operatorsVersion,Latest_phonePADPartitionedVersions,Latest_operatorId,gamePointId,gamePointDescription');
> Select query:
> select gamePointId  from Comp_VMALL_DICTIONARY_INCLUDE where gamePointId IS 
> NOT NULL order by gamePointId;
> select gamePointId from Comp_VMALL_DICTIONARY_INCLUDE where gamePointId is 
> NULL;
> The first select command displays null values as well and the second command 
> displays no values.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[GitHub] incubator-carbondata pull request #781: [CARBONDATA-900] Is null query on a ...

2017-04-11 Thread manishgupta88
GitHub user manishgupta88 reopened a pull request:

https://github.com/apache/incubator-carbondata/pull/781

[CARBONDATA-900] Is null query on a newly added measure column is not 
returning proper results



Analysis: When is null query is executed on newly added measure column, 
control goes to RowLevelFilterExecuterImpl class, where measure existence is 
checked. In case the measure is not found, bitset group is not getting 
populated with default values due to which that block is not returning any 
result.

Solution: When query is on a restructured block where newly added column 
does not exist, create the default bitset group so that based on default value 
existence default bitset group is created and results are returned based on 
that.

You can merge this pull request into a Git repository by running:

$ git pull https://github.com/manishgupta88/incubator-carbondata 
alter_measure_query_fix

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/incubator-carbondata/pull/781.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #781


commit 1c68a23448e9a3a3a5fca9331125f7bd7aac11a4
Author: manishgupta88 
Date:   2017-04-11T11:24:03Z

Problem: Is null query on a newly added measure column is not returning 
proper results.

Analysis: When is null query is executed on newly added measure column, 
control goes to RowLevelFilterExecuterImpl class, where measure existence is 
checked. In case the measure is not found, bitset group is not getting 
populated with default values due to which that block is not returning any 
result.

Solution: When query is on a restructured block where newly added column 
does not exist, create the default bitset group so that based on default value 
existence default bitset group is created and results are returned based on 
that.




---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] incubator-carbondata pull request #781: [CARBONDATA-900] Is null query on a ...

2017-04-11 Thread manishgupta88
Github user manishgupta88 closed the pull request at:

https://github.com/apache/incubator-carbondata/pull/781


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Updated] (CARBONDATA-901) Fix some spelling mistakes

2017-04-11 Thread Cao Gaofei (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-901?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Cao Gaofei updated CARBONDATA-901:
--
Request participants:   (was: )
 Summary: Fix some spelling mistakes  (was: fix some spelling 
mistakes)

> Fix some spelling mistakes
> --
>
> Key: CARBONDATA-901
> URL: https://issues.apache.org/jira/browse/CARBONDATA-901
> Project: CarbonData
>  Issue Type: Improvement
>  Components: core
>Reporter: Cao Gaofei
>Priority: Trivial
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> https://github.com/apache/incubator-carbondata/pull/782



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[GitHub] incubator-carbondata issue #780: [CARBONDATA-892] Fixed NullPointerException...

2017-04-11 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/incubator-carbondata/pull/780
  
Build Success with Spark 1.6.2, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/1569/



---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] incubator-carbondata issue #781: [CARBONDATA-900] Is null query on a newly a...

2017-04-11 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/incubator-carbondata/pull/781
  
Build Success with Spark 1.6.2, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/1568/



---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Created] (CARBONDATA-900) Is null query on a newly added measure column is not returning proper results

2017-04-11 Thread Manish Gupta (JIRA)
Manish Gupta created CARBONDATA-900:
---

 Summary:  Is null query on a newly added measure column is not 
returning proper results
 Key: CARBONDATA-900
 URL: https://issues.apache.org/jira/browse/CARBONDATA-900
 Project: CarbonData
  Issue Type: Bug
Reporter: Manish Gupta
Assignee: Manish Gupta
Priority: Minor
 Fix For: 1.1.0-incubating


When is null query is executed on newly added measure column, control goes to 
RowLevelFilterExecuterImpl class, where measure existence is checked. In case 
the measure is not found, bitset group is not getting populated with default 
values due to which that block is not returning any result.
Below queries can be executed to reproduce the issue:

CREATE TABLE uniqdata110 (CUST_ID int,CUST_NAME String) STORED BY 'carbondata'
LOAD DATA INPATH '' into table uniqdata110 
OPTIONS('BAD_RECORDS_LOGGER_ENABLE'='TRUE', 
'BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME')
ALTER TABLE uniqdata110  ADD COLUMNS (a6 int)
LOAD DATA INPATH '' into table uniqdata110 
OPTIONS('BAD_RECORDS_LOGGER_ENABLE'='TRUE', 
'BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,a6')
select * from uniqdata110
select * from uniqdata110 where a6 is null

Data:
7,hello1
8,welcome1
bye,11




--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[GitHub] incubator-carbondata pull request #779: [CARBONDATA-899] Added Support for D...

2017-04-11 Thread bhavya411
GitHub user bhavya411 opened a pull request:

https://github.com/apache/incubator-carbondata/pull/779

[CARBONDATA-899] Added Support for DecimalType and Timestamp for spark-2.1

Added Support for DecimalType and Timestamp with Spark-2.1
Local Build is running fine.
Manually Tested the functionality on Presto CLI

You can merge this pull request into a Git repository by running:

$ git pull https://github.com/bhavya411/incubator-carbondata CARBONDATA-899

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/incubator-carbondata/pull/779.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #779


commit 2a1df96f100445b26287cb37549740cb45761990
Author: Bhavya 
Date:   2017-04-04T08:49:49Z

Fixed Carbondata-848

commit ef9269b962b1f2332c25bc95a3af3d0aadc6b19e
Author: Bhavya 
Date:   2017-04-04T10:41:20Z

remove Presto module '

commit 30017786f948b0f88a385e54029dee00a85f5648
Author: Bhavya 
Date:   2017-04-11T09:46:39Z

Added support for DecimalType and Fixed Date and time issues




---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Created] (CARBONDATA-899) Added Support for Decimal data type and Fixed the timestamp and date issues for Spark-2.1

2017-04-11 Thread Bhavya Aggarwal (JIRA)
Bhavya Aggarwal created CARBONDATA-899:
--

 Summary: Added Support for Decimal data type and Fixed the 
timestamp and date issues for Spark-2.1
 Key: CARBONDATA-899
 URL: https://issues.apache.org/jira/browse/CARBONDATA-899
 Project: CarbonData
  Issue Type: Improvement
  Components: presto-integration
Reporter: Bhavya Aggarwal
Assignee: Bhavya Aggarwal
Priority: Minor


The Decimal Type correct support is added as well as issues related to 
timestamp and date are resolved.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[GitHub] incubator-carbondata issue #777: [CARBONDATA-863] Moved AlterTableAddColumnR...

2017-04-11 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/incubator-carbondata/pull/777
  
Build Success with Spark 1.6.2, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/1564/



---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] incubator-carbondata issue #778: [WIP] Added getAll dictionary to codegen of...

2017-04-11 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/incubator-carbondata/pull/778
  
Build Success with Spark 1.6.2, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/1563/



---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] incubator-carbondata issue #777: [CARBONDATA-863] Moved AlterTableAddColumnR...

2017-04-11 Thread kunal642
Github user kunal642 commented on the issue:

https://github.com/apache/incubator-carbondata/pull/777
  
retest this please


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] incubator-carbondata pull request #778: [WIP] Added getAll dictionary to cod...

2017-04-11 Thread ravipesala
GitHub user ravipesala opened a pull request:

https://github.com/apache/incubator-carbondata/pull/778

[WIP] Added getAll dictionary to codegen of dictionary decoder.



You can merge this pull request into a Git repository by running:

$ git pull https://github.com/ravipesala/incubator-carbondata 
getAlldictionary-codegen

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/incubator-carbondata/pull/778.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #778






---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] incubator-carbondata pull request #751: [CARBONDATA-816] Added Example for H...

2017-04-11 Thread anubhav100
Github user anubhav100 commented on a diff in the pull request:


https://github.com/apache/incubator-carbondata/pull/751#discussion_r110853402
  
--- Diff: integration/hive/pom.xml ---
@@ -64,6 +64,79 @@
 compile
 
 
+org.apache.spark
+spark-network-common_2.10
+2.1.0
+
+
+org.apache.hadoop
+hadoop-hdfs
+2.7.3
--- End diff --

resolved


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] incubator-carbondata pull request #751: [CARBONDATA-816] Added Example for H...

2017-04-11 Thread anubhav100
Github user anubhav100 commented on a diff in the pull request:


https://github.com/apache/incubator-carbondata/pull/751#discussion_r110853360
  
--- Diff: integration/hive/pom.xml ---
@@ -64,6 +64,79 @@
 compile
 
 
+org.apache.spark
+spark-network-common_2.10
+2.1.0
--- End diff --

resolved


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] incubator-carbondata pull request #751: [CARBONDATA-816] Added Example for H...

2017-04-11 Thread anubhav100
Github user anubhav100 commented on a diff in the pull request:


https://github.com/apache/incubator-carbondata/pull/751#discussion_r110853283
  
--- Diff: integration/hive/pom.xml ---
@@ -64,6 +64,79 @@
 compile
 
 
+org.apache.spark
+spark-network-common_2.10
+2.1.0
+
+
+org.apache.hadoop
+hadoop-hdfs
+2.7.3
+
+
+ch.qos.logback
+logback-classic
+
+
+javax.servlet
+servlet-api
+
+
+
+
+org.apache.zookeeper
+zookeeper
+3.4.7
+
+
+jline
+jline
+
+
+
+
+org.apache.carbondata
+carbondata-spark2
+${project.version}
+
+
+org.apache.spark
+spark-sql_${scala.binary.version}
+
+
+org.apache.spark
+
spark-hive-thriftserver_${scala.binary.version}
+
+
+org.apache.spark
+spark-repl_${scala.binary.version}
+
+
+org.apache.hadoop
+hadoop-common
+2.7.3
+
+
+org.apache.httpcomponents
+httpclient
+4.3.4
+
+
+org.apache.httpcomponents
+httpcore
+4.3-alpha1
+
+
+org.apache.hadoop
--- End diff --

done


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] incubator-carbondata issue #777: [CARBONDATA-863] Moved AlterTableAddColumnR...

2017-04-11 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/incubator-carbondata/pull/777
  
Build Failed  with Spark 1.6.2, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/1562/



---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] incubator-carbondata issue #777: [CARBONDATA-863] Moved AlterTableAddColumnR...

2017-04-11 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/incubator-carbondata/pull/777
  
Build Failed  with Spark 1.6.2, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/1561/



---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Updated] (CARBONDATA-896) throws nullpointerException while insert data

2017-04-11 Thread sehriff (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-896?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

sehriff updated CARBONDATA-896:
---
Affects Version/s: 1.0.0-incubating

> throws nullpointerException while insert data
> -
>
> Key: CARBONDATA-896
> URL: https://issues.apache.org/jira/browse/CARBONDATA-896
> Project: CarbonData
>  Issue Type: Bug
>Affects Versions: 1.0.0-incubating
>Reporter: sehriff
>
> insert data into carbon table from hive using sql like:
> cc.sql("insert into carbon.table_carbon select * from hivetable").show
> and got the following error:
> Job aborted due to stage failure: Task 0 in stage 26.3 failed 4 times, most 
> recent failure: Lost task 0.3 in stage 26.3 (TID 5628, HDD013): 
> java.lang.NullPointerException
>   at 
> org.apache.carbondata.core.datastore.filesystem.AbstractDFSCarbonFile.getLastModifiedTime(AbstractDFSCarbonFile.java:135)
>   at 
> org.apache.carbondata.core.datastore.filesystem.AbstractDFSCarbonFile.isFileModified(AbstractDFSCarbonFile.java:210)
>   at 
> org.apache.carbondata.core.cache.dictionary.AbstractDictionaryCache.isDictionaryMetaFileModified(AbstractDictionaryCache.java:119)
>   at 
> org.apache.carbondata.core.cache.dictionary.AbstractDictionaryCache.checkAndLoadDictionaryData(AbstractDictionaryCache.java:158)
>   at 
> org.apache.carbondata.core.cache.dictionary.ReverseDictionaryCache.getDictionary(ReverseDictionaryCache.java:174)
>   at 
> org.apache.carbondata.core.cache.dictionary.ReverseDictionaryCache.get(ReverseDictionaryCache.java:67)
>   at 
> org.apache.carbondata.core.cache.dictionary.ReverseDictionaryCache.get(ReverseDictionaryCache.java:38)
>   at 
> org.apache.carbondata.spark.load.CarbonLoaderUtil.getDictionary(CarbonLoaderUtil.java:463)
>   at 
> org.apache.carbondata.spark.load.CarbonLoaderUtil.getDictionary(CarbonLoaderUtil.java:469)
>   at 
> org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD$$anon$1.(CarbonGlobalDictionaryRDD.scala:413)
>   at 
> org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD.compute(CarbonGlobalDictionaryRDD.scala:342)
>   at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
>   at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
>   at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
>   at org.apache.spark.scheduler.Task.run(Task.scala:89)
>   at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>   at java.lang.Thread.run(Thread.java:745)



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[GitHub] incubator-carbondata pull request #777: [CARBONDATA-863] Moved AlterTableAdd...

2017-04-11 Thread kunal642
GitHub user kunal642 opened a pull request:

https://github.com/apache/incubator-carbondata/pull/777

[CARBONDATA-863] Moved AlterTableAddColumnRDD call to AlterTableCommands



You can merge this pull request into a Git repository by running:

$ git pull https://github.com/kunal642/incubator-carbondata CARBONDATA-863

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/incubator-carbondata/pull/777.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #777


commit d83cec14d02bf3ef828fc978c5a255f2c34ef40a
Author: kunal642 
Date:   2017-04-11T08:45:10Z

added AlterTableAddColumnRDD to AlterTableCommands




---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] incubator-carbondata issue #776: [CARBONDATA-897] Removed redundant fields f...

2017-04-11 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/incubator-carbondata/pull/776
  
Can one of the admins verify this patch?


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Updated] (CARBONDATA-887) lazy rdd iterator for InsertInto

2017-04-11 Thread Venkata Ramana G (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-887?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Venkata Ramana G updated CARBONDATA-887:

Priority: Minor  (was: Major)

> lazy rdd iterator for InsertInto
> 
>
> Key: CARBONDATA-887
> URL: https://issues.apache.org/jira/browse/CARBONDATA-887
> Project: CarbonData
>  Issue Type: Improvement
>Reporter: QiangCai
>Assignee: QiangCai
>Priority: Minor
> Fix For: 1.1.0-incubating
>
>  Time Spent: 2.5h
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[jira] [Resolved] (CARBONDATA-887) lazy rdd iterator for InsertInto

2017-04-11 Thread Venkata Ramana G (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-887?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Venkata Ramana G resolved CARBONDATA-887.
-
   Resolution: Fixed
Fix Version/s: 1.1.0-incubating

> lazy rdd iterator for InsertInto
> 
>
> Key: CARBONDATA-887
> URL: https://issues.apache.org/jira/browse/CARBONDATA-887
> Project: CarbonData
>  Issue Type: Improvement
>Reporter: QiangCai
>Assignee: QiangCai
>Priority: Minor
> Fix For: 1.1.0-incubating
>
>  Time Spent: 2.5h
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[GitHub] incubator-carbondata pull request #776: [CARBONDATA-897] Removed redundant f...

2017-04-11 Thread jatin9896
GitHub user jatin9896 opened a pull request:

https://github.com/apache/incubator-carbondata/pull/776

[CARBONDATA-897] Removed redundant fields from table in 
configuration-parameters.md



You can merge this pull request into a Git repository by running:

$ git pull https://github.com/jatin9896/incubator-carbondata 
docs/issuesInMdFileTable

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/incubator-carbondata/pull/776.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #776


commit 8e8135711c45e088c7b0cf1d47a6467ca6b62529
Author: jatin 
Date:   2017-04-11T08:08:10Z

Removed redundant fields from table in configuration-parameters.md




---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Created] (CARBONDATA-898) When select query and alter table rename table is triggered concurrently, NullPointerException is getting thrown

2017-04-11 Thread Naresh P R (JIRA)
Naresh P R created CARBONDATA-898:
-

 Summary: When select query and alter table rename table is 
triggered concurrently, NullPointerException is getting thrown
 Key: CARBONDATA-898
 URL: https://issues.apache.org/jira/browse/CARBONDATA-898
 Project: CarbonData
  Issue Type: Bug
 Environment: Spark 2.1
Reporter: Naresh P R
Assignee: Naresh P R
Priority: Minor


When user triggers select query and alter table rename table command 
concurrently, Select query is throwning NullPointerException if the files does 
not exist in hdfs.

When dictionary file or schema file does not exist, File not found exception 
should be thrown



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[GitHub] incubator-carbondata pull request #765: [CARBONDATA-887]lazy rdd iterator fo...

2017-04-11 Thread asfgit
Github user asfgit closed the pull request at:

https://github.com/apache/incubator-carbondata/pull/765


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Created] (CARBONDATA-897) Redundant Fields Inside * **Global Dictionary Configurations** in Configuration-parameters.md

2017-04-11 Thread Pallavi Singh (JIRA)
Pallavi Singh created CARBONDATA-897:


 Summary: Redundant Fields Inside  * **Global Dictionary 
Configurations** in Configuration-parameters.md
 Key: CARBONDATA-897
 URL: https://issues.apache.org/jira/browse/CARBONDATA-897
 Project: CarbonData
  Issue Type: Bug
  Components: docs
Reporter: Pallavi Singh
Assignee: Pallavi Singh
Priority: Minor
 Attachments: Configurations.png

In the Configuration-parameters.md file under the table Global Dictionary 
Configurations the row for field  high.cardinality.threshold has extra columns 
with redundant values in the md file.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[jira] [Created] (CARBONDATA-896) throws nullpointerException while insert data

2017-04-11 Thread sehriff (JIRA)
sehriff created CARBONDATA-896:
--

 Summary: throws nullpointerException while insert data
 Key: CARBONDATA-896
 URL: https://issues.apache.org/jira/browse/CARBONDATA-896
 Project: CarbonData
  Issue Type: Bug
Reporter: sehriff


insert data into carbon table from hive using sql like:
cc.sql("insert into carbon.table_carbon select * from hivetable").show
and got the following error:
Job aborted due to stage failure: Task 0 in stage 26.3 failed 4 times, most 
recent failure: Lost task 0.3 in stage 26.3 (TID 5628, HDD013): 
java.lang.NullPointerException
at 
org.apache.carbondata.core.datastore.filesystem.AbstractDFSCarbonFile.getLastModifiedTime(AbstractDFSCarbonFile.java:135)
at 
org.apache.carbondata.core.datastore.filesystem.AbstractDFSCarbonFile.isFileModified(AbstractDFSCarbonFile.java:210)
at 
org.apache.carbondata.core.cache.dictionary.AbstractDictionaryCache.isDictionaryMetaFileModified(AbstractDictionaryCache.java:119)
at 
org.apache.carbondata.core.cache.dictionary.AbstractDictionaryCache.checkAndLoadDictionaryData(AbstractDictionaryCache.java:158)
at 
org.apache.carbondata.core.cache.dictionary.ReverseDictionaryCache.getDictionary(ReverseDictionaryCache.java:174)
at 
org.apache.carbondata.core.cache.dictionary.ReverseDictionaryCache.get(ReverseDictionaryCache.java:67)
at 
org.apache.carbondata.core.cache.dictionary.ReverseDictionaryCache.get(ReverseDictionaryCache.java:38)
at 
org.apache.carbondata.spark.load.CarbonLoaderUtil.getDictionary(CarbonLoaderUtil.java:463)
at 
org.apache.carbondata.spark.load.CarbonLoaderUtil.getDictionary(CarbonLoaderUtil.java:469)
at 
org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD$$anon$1.(CarbonGlobalDictionaryRDD.scala:413)
at 
org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD.compute(CarbonGlobalDictionaryRDD.scala:342)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
at org.apache.spark.scheduler.Task.run(Task.scala:89)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[GitHub] incubator-carbondata issue #765: [CARBONDATA-887]lazy rdd iterator for Inser...

2017-04-11 Thread gvramana
Github user gvramana commented on the issue:

https://github.com/apache/incubator-carbondata/pull/765
  
LGTM


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---