[GitHub] incubator-carbondata issue #783: [CARBONDATA-903] data load is not failing e...
Github user CarbonDataQA commented on the issue: https://github.com/apache/incubator-carbondata/pull/783 Build Failed with Spark 1.6.2, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/1580/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[jira] [Created] (CARBONDATA-908) bitmap encode
Jarck created CARBONDATA-908: Summary: bitmap encode Key: CARBONDATA-908 URL: https://issues.apache.org/jira/browse/CARBONDATA-908 Project: CarbonData Issue Type: New Feature Components: core, data-load, data-query Reporter: Jarck Assignee: Jarck for frequent filter queries on low cardinality columns, use bitmap encode can speed up query -- This message was sent by Atlassian JIRA (v6.3.15#6346)
[GitHub] incubator-carbondata issue #751: [CARBONDATA-816] Added Example for Hive Int...
Github user CarbonDataQA commented on the issue: https://github.com/apache/incubator-carbondata/pull/751 Build Success with Spark 1.6.2, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/1579/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] incubator-carbondata issue #751: [CARBONDATA-816] Added Example for Hive Int...
Github user anubhav100 commented on the issue: https://github.com/apache/incubator-carbondata/pull/751 @chenliang613 i added the code for stoping application automatically ,tested it on my system and rebased all commits into one commit can it be merged? --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] incubator-carbondata issue #781: [CARBONDATA-900] Is null query on a newly a...
Github user CarbonDataQA commented on the issue: https://github.com/apache/incubator-carbondata/pull/781 Build Success with Spark 1.6.2, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/1578/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[jira] [Commented] (CARBONDATA-907) The grammar for DELETE SEGMENT FOR DATE in website is not correct
[ https://issues.apache.org/jira/browse/CARBONDATA-907?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15965298#comment-15965298 ] chenerlu commented on CARBONDATA-907: - I have raised a PR below: https://github.com/apache/incubator-carbondata/pull/785 > The grammar for DELETE SEGMENT FOR DATE in website is not correct > -- > > Key: CARBONDATA-907 > URL: https://issues.apache.org/jira/browse/CARBONDATA-907 > Project: CarbonData > Issue Type: Bug >Reporter: chenerlu > Time Spent: 20m > Remaining Estimate: 0h > > The grammar for DELETE SEGMENT FOR DATE in website is not correct -- This message was sent by Atlassian JIRA (v6.3.15#6346)
[GitHub] incubator-carbondata issue #785: [CARBONDATA-907] The grammar for DELETE SEG...
Github user CarbonDataQA commented on the issue: https://github.com/apache/incubator-carbondata/pull/785 Can one of the admins verify this patch? --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] incubator-carbondata pull request #785: [CARBONDATA-907] The grammar for DEL...
GitHub user chenerlu opened a pull request: https://github.com/apache/incubator-carbondata/pull/785 [CARBONDATA-907] The grammar for DELETE SEGMENT FOR DATE in website is not correct Problem As I check open source CarbonData code, the grammar for DELETE SEGMENT FOR DATE in website is not correct, while the corresponding example is correct. Solution Update the doc for for DELETE SEGMENT FOR DATE in website. You can merge this pull request into a Git repository by running: $ git pull https://github.com/chenerlu/incubator-carbondata patch-2 Alternatively you can review and apply these changes as the patch at: https://github.com/apache/incubator-carbondata/pull/785.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #785 commit 94ce539fd3966451778ea8db5dca61ea22352177 Author: chenerluDate: 2017-04-12T02:03:29Z Update dml-operation-on-carbondata.md commit 9940f22cdb080ed2c970b6c617698666f23673da Author: chenerlu Date: 2017-04-12T02:48:00Z Update dml-operation-on-carbondata.md --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[jira] [Created] (CARBONDATA-907) The grammar for DELETE SEGMENT FOR DATE in website is not correct
chenerlu created CARBONDATA-907: --- Summary: The grammar for DELETE SEGMENT FOR DATE in website is not correct Key: CARBONDATA-907 URL: https://issues.apache.org/jira/browse/CARBONDATA-907 Project: CarbonData Issue Type: Bug Reporter: chenerlu The grammar for DELETE SEGMENT FOR DATE in website is not correct -- This message was sent by Atlassian JIRA (v6.3.15#6346)
[jira] [Commented] (CARBONDATA-906) Always OOM error when import large dataset (100milion rows)
[ https://issues.apache.org/jira/browse/CARBONDATA-906?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15965243#comment-15965243 ] Crabo Yang commented on CARBONDATA-906: --- We keep turning Carbondata for a few months, sucessed on loading 5million row (about 100 columns per row) web have a 20 nodes cluster and 32GB ram each. CarbonData have slow loading data performance and we should split the job into really small batch to load. Always a hard time to play with . can you guys help me out of this OOM hell? > Always OOM error when import large dataset (100milion rows) > --- > > Key: CARBONDATA-906 > URL: https://issues.apache.org/jira/browse/CARBONDATA-906 > Project: CarbonData > Issue Type: Bug > Components: data-load >Affects Versions: 1.0.0-incubating >Reporter: Crabo Yang > > java.lang.OutOfMemoryError: GC overhead limit exceeded > at > java.util.concurrent.ConcurrentHashMap$Segment.put(ConcurrentHashMap.java:457) > at > java.util.concurrent.ConcurrentHashMap.put(ConcurrentHashMap.java:1130) > at > org.apache.carbondata.core.cache.dictionary.ColumnReverseDictionaryInfo.addDataToDictionaryMap(ColumnReverseDictionaryInfo.java:101) > at > org.apache.carbondata.core.cache.dictionary.ColumnReverseDictionaryInfo.addDictionaryChunk(ColumnReverseDictionaryInfo.java:88) > at > org.apache.carbondata.core.cache.dictionary.DictionaryCacheLoaderImpl.fillDictionaryValuesAndAddToDictionaryChunks(DictionaryCacheLoaderImpl.java:113) > at > org.apache.carbondata.core.cache.dictionary.DictionaryCacheLoaderImpl.load(DictionaryCacheLoaderImpl.java:81) > at > org.apache.carbondata.core.cache.dictionary.AbstractDictionaryCache.loadDictionaryData(AbstractDictionaryCache.java:236) > at > org.apache.carbondata.core.cache.dictionary.AbstractDictionaryCache.checkAndLoadDictionaryData(AbstractDictionaryCache.java:186) > at > org.apache.carbondata.core.cache.dictionary.ReverseDictionaryCache.getDictionary(ReverseDictionaryCache.java:174) > at > org.apache.carbondata.core.cache.dictionary.ReverseDictionaryCache.get(ReverseDictionaryCache.java:67) > at > org.apache.carbondata.core.cache.dictionary.ReverseDictionaryCache.get(ReverseDictionaryCache.java:38) > at > org.apache.carbondata.processing.newflow.converter.impl.DictionaryFieldConverterImpl.(DictionaryFieldConverterImpl.java:92) > at > org.apache.carbondata.processing.newflow.converter.impl.FieldEncoderFactory.createFieldEncoder(FieldEncoderFactory.java:77) > at > org.apache.carbondata.processing.newflow.converter.impl.RowConverterImpl.initialize(RowConverterImpl.java:102) > at > org.apache.carbondata.processing.newflow.steps.DataConverterProcessorStepImpl.initialize(DataConverterProcessorStepImpl.java:69) > at > org.apache.carbondata.processing.newflow.steps.SortProcessorStepImpl.initialize(SortProcessorStepImpl.java:57) > at > org.apache.carbondata.processing.newflow.steps.DataWriterProcessorStepImpl.initialize(DataWriterProcessorStepImpl.java:79) > at > org.apache.carbondata.processing.newflow.DataLoadExecutor.execute(DataLoadExecutor.java:45) > at > org.apache.carbondata.spark.rdd.NewDataFrameLoaderRDD$$anon$2.(NewCarbonDataLoadRDD.scala:425) > at > org.apache.carbondata.spark.rdd.NewDataFrameLoaderRDD.compute(NewCarbonDataLoadRDD.scala:383) > at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) > at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) > at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66) > at org.apache.spark.scheduler.Task.run(Task.scala:89) > at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) > at java.lang.Thread.run(Thread.java:745) -- This message was sent by Atlassian JIRA (v6.3.15#6346)
[jira] [Created] (CARBONDATA-906) Always OOM error when import large dataset (100milion rows)
Crabo Yang created CARBONDATA-906: - Summary: Always OOM error when import large dataset (100milion rows) Key: CARBONDATA-906 URL: https://issues.apache.org/jira/browse/CARBONDATA-906 Project: CarbonData Issue Type: Bug Components: data-load Affects Versions: 1.0.0-incubating Reporter: Crabo Yang java.lang.OutOfMemoryError: GC overhead limit exceeded at java.util.concurrent.ConcurrentHashMap$Segment.put(ConcurrentHashMap.java:457) at java.util.concurrent.ConcurrentHashMap.put(ConcurrentHashMap.java:1130) at org.apache.carbondata.core.cache.dictionary.ColumnReverseDictionaryInfo.addDataToDictionaryMap(ColumnReverseDictionaryInfo.java:101) at org.apache.carbondata.core.cache.dictionary.ColumnReverseDictionaryInfo.addDictionaryChunk(ColumnReverseDictionaryInfo.java:88) at org.apache.carbondata.core.cache.dictionary.DictionaryCacheLoaderImpl.fillDictionaryValuesAndAddToDictionaryChunks(DictionaryCacheLoaderImpl.java:113) at org.apache.carbondata.core.cache.dictionary.DictionaryCacheLoaderImpl.load(DictionaryCacheLoaderImpl.java:81) at org.apache.carbondata.core.cache.dictionary.AbstractDictionaryCache.loadDictionaryData(AbstractDictionaryCache.java:236) at org.apache.carbondata.core.cache.dictionary.AbstractDictionaryCache.checkAndLoadDictionaryData(AbstractDictionaryCache.java:186) at org.apache.carbondata.core.cache.dictionary.ReverseDictionaryCache.getDictionary(ReverseDictionaryCache.java:174) at org.apache.carbondata.core.cache.dictionary.ReverseDictionaryCache.get(ReverseDictionaryCache.java:67) at org.apache.carbondata.core.cache.dictionary.ReverseDictionaryCache.get(ReverseDictionaryCache.java:38) at org.apache.carbondata.processing.newflow.converter.impl.DictionaryFieldConverterImpl.(DictionaryFieldConverterImpl.java:92) at org.apache.carbondata.processing.newflow.converter.impl.FieldEncoderFactory.createFieldEncoder(FieldEncoderFactory.java:77) at org.apache.carbondata.processing.newflow.converter.impl.RowConverterImpl.initialize(RowConverterImpl.java:102) at org.apache.carbondata.processing.newflow.steps.DataConverterProcessorStepImpl.initialize(DataConverterProcessorStepImpl.java:69) at org.apache.carbondata.processing.newflow.steps.SortProcessorStepImpl.initialize(SortProcessorStepImpl.java:57) at org.apache.carbondata.processing.newflow.steps.DataWriterProcessorStepImpl.initialize(DataWriterProcessorStepImpl.java:79) at org.apache.carbondata.processing.newflow.DataLoadExecutor.execute(DataLoadExecutor.java:45) at org.apache.carbondata.spark.rdd.NewDataFrameLoaderRDD$$anon$2.(NewCarbonDataLoadRDD.scala:425) at org.apache.carbondata.spark.rdd.NewDataFrameLoaderRDD.compute(NewCarbonDataLoadRDD.scala:383) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66) at org.apache.spark.scheduler.Task.run(Task.scala:89) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) at java.lang.Thread.run(Thread.java:745) -- This message was sent by Atlassian JIRA (v6.3.15#6346)
[GitHub] incubator-carbondata issue #783: [CARBONDATA-903] data load is not failing e...
Github user ravipesala commented on the issue: https://github.com/apache/incubator-carbondata/pull/783 @mohammadshahidkhan Please create a Abstract class and move the common methods to there. --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] incubator-carbondata issue #755: [CARBONDATA-881] Load status is successful ...
Github user ravipesala commented on the issue: https://github.com/apache/incubator-carbondata/pull/755 @mohammadshahidkhan Please check all caller methods of `writeLoadDetailsIntoFile` and do exception handling there also, otherwise same issue happens if they catch the error and just log it. --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] incubator-carbondata issue #755: [CARBONDATA-881] Load status is successful ...
Github user ravipesala commented on the issue: https://github.com/apache/incubator-carbondata/pull/755 LGTM --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] incubator-carbondata issue #782: [CARBONDATA-901] Fix some spelling mistakes
Github user CarbonDataQA commented on the issue: https://github.com/apache/incubator-carbondata/pull/782 Build Success with Spark 1.6.2, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/1577/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] incubator-carbondata pull request #782: [CARBONDATA-901] Fix some spelling m...
Github user asfgit closed the pull request at: https://github.com/apache/incubator-carbondata/pull/782 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] incubator-carbondata issue #782: [CARBONDATA-901] Fix some spelling mistakes
Github user chenliang613 commented on the issue: https://github.com/apache/incubator-carbondata/pull/782 LGTM --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[jira] [Commented] (CARBONDATA-901) Fix some spelling mistakes
[ https://issues.apache.org/jira/browse/CARBONDATA-901?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15964830#comment-15964830 ] Liang Chen commented on CARBONDATA-901: --- Hi Cao Gaofei Already gave the contributor right to you , so now you can assign issue to yourself. > Fix some spelling mistakes > -- > > Key: CARBONDATA-901 > URL: https://issues.apache.org/jira/browse/CARBONDATA-901 > Project: CarbonData > Issue Type: Improvement > Components: core >Reporter: Cao Gaofei >Assignee: Cao Gaofei >Priority: Trivial > Time Spent: 0.5h > Remaining Estimate: 0h > > https://github.com/apache/incubator-carbondata/pull/782 -- This message was sent by Atlassian JIRA (v6.3.15#6346)
[GitHub] incubator-carbondata pull request #783: [CARBONDATA-903] data load is not fa...
Github user mohammadshahidkhan closed the pull request at: https://github.com/apache/incubator-carbondata/pull/783 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] incubator-carbondata pull request #783: [CARBONDATA-903] data load is not fa...
GitHub user mohammadshahidkhan reopened a pull request: https://github.com/apache/incubator-carbondata/pull/783 [CARBONDATA-903] data load is not failing even though bad records exists in the data in case of unsafe sort or batch sort You can merge this pull request into a Git repository by running: $ git pull https://github.com/mohammadshahidkhan/incubator-carbondata dataload_termination_on_fail Alternatively you can review and apply these changes as the patch at: https://github.com/apache/incubator-carbondata/pull/783.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #783 commit db9cf1b714f28fdf5fd75911f54ed7a2ad6c3dec Author: mohammadshahidkhanDate: 2017-04-11T12:56:51Z [CARBONDATA-903] data load is not failing even though bad records exists in the data in case of unsafe sort or batch sort --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] incubator-carbondata pull request #783: [CARBONDATA-903] data load is not fa...
Github user mohammadshahidkhan closed the pull request at: https://github.com/apache/incubator-carbondata/pull/783 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] incubator-carbondata pull request #783: [CARBONDATA-903] data load is not fa...
Github user mohammadshahidkhan closed the pull request at: https://github.com/apache/incubator-carbondata/pull/783 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] incubator-carbondata pull request #783: [CARBONDATA-903] data load is not fa...
GitHub user mohammadshahidkhan reopened a pull request: https://github.com/apache/incubator-carbondata/pull/783 [CARBONDATA-903] data load is not failing even though bad records exists in the data in case of unsafe sort or batch sort You can merge this pull request into a Git repository by running: $ git pull https://github.com/mohammadshahidkhan/incubator-carbondata dataload_termination_on_fail Alternatively you can review and apply these changes as the patch at: https://github.com/apache/incubator-carbondata/pull/783.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #783 commit db9cf1b714f28fdf5fd75911f54ed7a2ad6c3dec Author: mohammadshahidkhanDate: 2017-04-11T12:56:51Z [CARBONDATA-903] data load is not failing even though bad records exists in the data in case of unsafe sort or batch sort --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] incubator-carbondata issue #784: [WIP]Pr730
Github user CarbonDataQA commented on the issue: https://github.com/apache/incubator-carbondata/pull/784 Build Success with Spark 1.6.2, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/1575/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] incubator-carbondata issue #784: [WIP]Pr730
Github user QiangCai commented on the issue: https://github.com/apache/incubator-carbondata/pull/784 retest this please --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] incubator-carbondata issue #769: [CARBONDATA-888] Added include and exclude ...
Github user jackylk commented on the issue: https://github.com/apache/incubator-carbondata/pull/769 I will merge it later as we are preparing version 1.1 release, code is freezed for master branch. Or you can create another PR to merge into 12-dev branch. Thanks. --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] incubator-carbondata issue #783: [CARBONDATA-903] data load is not failing e...
Github user CarbonDataQA commented on the issue: https://github.com/apache/incubator-carbondata/pull/783 Build Success with Spark 1.6.2, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/1573/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] incubator-carbondata issue #783: [CARBONDATA-903] data load is not failing e...
Github user CarbonDataQA commented on the issue: https://github.com/apache/incubator-carbondata/pull/783 Build Failed with Spark 1.6.2, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/1572/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] incubator-carbondata issue #720: [CARBONDATA-642] Delete Subquery is not wor...
Github user jackylk commented on the issue: https://github.com/apache/incubator-carbondata/pull/720 LGTM --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[jira] [Updated] (CARBONDATA-904) ArrayIndexOutOfBoundsException
[ https://issues.apache.org/jira/browse/CARBONDATA-904?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] SWATI RAO updated CARBONDATA-904: - Attachment: Test_Data1.csv Test_Data1_h1.csv Environment: Spark1.6 Request participants: (was: ) Fix Version/s: 1.1.0-incubating > ArrayIndexOutOfBoundsException > --- > > Key: CARBONDATA-904 > URL: https://issues.apache.org/jira/browse/CARBONDATA-904 > Project: CarbonData > Issue Type: Bug > Environment: Spark1.6 >Reporter: SWATI RAO > Fix For: 1.1.0-incubating > > Attachments: Test_Data1_h1.csv, Test_Data1_h1.csv > > > Or operator is not working properly. > When we execute these query in hive it is working fine but when we execute > the same in carbondata it throws an exception: > java.lang.ArrayIndexOutOfBoundsException > HIVE: > 0: jdbc:hive2://hadoop-master:1> create table Test_Boundary_h1 (c1_int > int,c2_Bigint Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string > string,c6_Timestamp Timestamp,c7_Datatype_Desc string) ROW FORMAT DELIMITED > FIELDS TERMINATED BY ',' ; > +-+--+ > | result | > +-+--+ > +-+--+ > No rows selected (1.177 seconds) > 0: jdbc:hive2://hadoop-master:1> load data local inpath > '/opt/Carbon/CarbonData/TestData/Data/Test_Data1_h1.csv' OVERWRITE INTO TABLE > Test_Boundary_h1 ; > +-+--+ > | Result | > +-+--+ > +-+--+ > No rows selected (0.437 seconds) > 0: jdbc:hive2://hadoop-master:1> select c6_Timestamp,max(c6_Timestamp) > from Test_Boundary_h1 where c6_Timestamp ='2017-07-01 12:07:28' or > c6_Timestamp ='2019-07-05 13:07:30' or c6_Timestamp = '1999-01-06 10:05:29' > group by c6_Timestamp ; > +++--+ > | c6_Timestamp | _c1 | > +++--+ > | 2017-07-01 12:07:28.0 | 2017-07-01 12:07:28.0 | > +++--+ > 1 row selected (1.637 seconds) > CARBONDATA: > 0: jdbc:hive2://hadoop-master:1> create table Test_Boundary (c1_int > int,c2_Bigint Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string > string,c6_Timestamp Timestamp,c7_Datatype_Desc string) STORED BY > 'org.apache.carbondata.format' ; > +-+--+ > | Result | > +-+--+ > +-+--+ > No rows selected (4.48 seconds) > 0: jdbc:hive2://hadoop-master:1> LOAD DATA INPATH > 'hdfs://192.168.2.145:54310/BabuStore/Data/Test_Data1.csv' INTO table > Test_Boundary > OPTIONS('DELIMITER'=',','QUOTECHAR'='','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='') > ; > +-+--+ > | Result | > +-+--+ > +-+--+ > No rows selected (4.445 seconds) > 0: jdbc:hive2://hadoop-master:1> select c6_Timestamp,max(c6_Timestamp) > from Test_Boundary where c6_Timestamp ='2017-07-01 12:07:28' or c6_Timestamp > =' 2019-07-05 13:07:30' or c6_Timestamp = '1999-01-06 10:05:29' group by > c6_Timestamp ; > Error: org.apache.spark.SparkException: Job aborted due to stage failure: > Task 0 in stage 5.0 failed 4 times, most recent failure: Lost task 0.3 in > stage 5.0 (TID 8, hadoop-master): java.lang.RuntimeException: > java.util.concurrent.ExecutionException: > java.lang.ArrayIndexOutOfBoundsException: 0 > at > org.apache.carbondata.core.scan.processor.AbstractDataBlockIterator.updateScanner(AbstractDataBlockIterator.java:136) > at > org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:50) > at > org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:32) > at > org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.getBatchResult(DetailQueryResultIterator.java:50) > at > org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.next(DetailQueryResultIterator.java:41) > at > org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.next(DetailQueryResultIterator.java:31) > at > org.apache.carbondata.core.scan.result.iterator.ChunkRowIterator.(ChunkRowIterator.java:41) > at > org.apache.carbondata.hadoop.CarbonRecordReader.initialize(CarbonRecordReader.java:79) > at > org.apache.carbondata.spark.rdd.CarbonScanRDD.compute(CarbonScanRDD.scala:204) > at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) > at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) > at > org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) > at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) > at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) > at >
[jira] [Updated] (CARBONDATA-904) ArrayIndexOutOfBoundsException
[ https://issues.apache.org/jira/browse/CARBONDATA-904?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] SWATI RAO updated CARBONDATA-904: - Attachment: (was: Test_Data1.csv) > ArrayIndexOutOfBoundsException > --- > > Key: CARBONDATA-904 > URL: https://issues.apache.org/jira/browse/CARBONDATA-904 > Project: CarbonData > Issue Type: Bug > Environment: Spark1.6 >Reporter: SWATI RAO > Fix For: 1.1.0-incubating > > Attachments: Test_Data1_h1.csv, Test_Data1_h1.csv > > > Or operator is not working properly. > When we execute these query in hive it is working fine but when we execute > the same in carbondata it throws an exception: > java.lang.ArrayIndexOutOfBoundsException > HIVE: > 0: jdbc:hive2://hadoop-master:1> create table Test_Boundary_h1 (c1_int > int,c2_Bigint Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string > string,c6_Timestamp Timestamp,c7_Datatype_Desc string) ROW FORMAT DELIMITED > FIELDS TERMINATED BY ',' ; > +-+--+ > | result | > +-+--+ > +-+--+ > No rows selected (1.177 seconds) > 0: jdbc:hive2://hadoop-master:1> load data local inpath > '/opt/Carbon/CarbonData/TestData/Data/Test_Data1_h1.csv' OVERWRITE INTO TABLE > Test_Boundary_h1 ; > +-+--+ > | Result | > +-+--+ > +-+--+ > No rows selected (0.437 seconds) > 0: jdbc:hive2://hadoop-master:1> select c6_Timestamp,max(c6_Timestamp) > from Test_Boundary_h1 where c6_Timestamp ='2017-07-01 12:07:28' or > c6_Timestamp ='2019-07-05 13:07:30' or c6_Timestamp = '1999-01-06 10:05:29' > group by c6_Timestamp ; > +++--+ > | c6_Timestamp | _c1 | > +++--+ > | 2017-07-01 12:07:28.0 | 2017-07-01 12:07:28.0 | > +++--+ > 1 row selected (1.637 seconds) > CARBONDATA: > 0: jdbc:hive2://hadoop-master:1> create table Test_Boundary (c1_int > int,c2_Bigint Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string > string,c6_Timestamp Timestamp,c7_Datatype_Desc string) STORED BY > 'org.apache.carbondata.format' ; > +-+--+ > | Result | > +-+--+ > +-+--+ > No rows selected (4.48 seconds) > 0: jdbc:hive2://hadoop-master:1> LOAD DATA INPATH > 'hdfs://192.168.2.145:54310/BabuStore/Data/Test_Data1.csv' INTO table > Test_Boundary > OPTIONS('DELIMITER'=',','QUOTECHAR'='','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='') > ; > +-+--+ > | Result | > +-+--+ > +-+--+ > No rows selected (4.445 seconds) > 0: jdbc:hive2://hadoop-master:1> select c6_Timestamp,max(c6_Timestamp) > from Test_Boundary where c6_Timestamp ='2017-07-01 12:07:28' or c6_Timestamp > =' 2019-07-05 13:07:30' or c6_Timestamp = '1999-01-06 10:05:29' group by > c6_Timestamp ; > Error: org.apache.spark.SparkException: Job aborted due to stage failure: > Task 0 in stage 5.0 failed 4 times, most recent failure: Lost task 0.3 in > stage 5.0 (TID 8, hadoop-master): java.lang.RuntimeException: > java.util.concurrent.ExecutionException: > java.lang.ArrayIndexOutOfBoundsException: 0 > at > org.apache.carbondata.core.scan.processor.AbstractDataBlockIterator.updateScanner(AbstractDataBlockIterator.java:136) > at > org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:50) > at > org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:32) > at > org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.getBatchResult(DetailQueryResultIterator.java:50) > at > org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.next(DetailQueryResultIterator.java:41) > at > org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.next(DetailQueryResultIterator.java:31) > at > org.apache.carbondata.core.scan.result.iterator.ChunkRowIterator.(ChunkRowIterator.java:41) > at > org.apache.carbondata.hadoop.CarbonRecordReader.initialize(CarbonRecordReader.java:79) > at > org.apache.carbondata.spark.rdd.CarbonScanRDD.compute(CarbonScanRDD.scala:204) > at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) > at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) > at > org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) > at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) > at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) > at > org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) > at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) > at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) > at >
[jira] [Updated] (CARBONDATA-904) ArrayIndexOutOfBoundsException
[ https://issues.apache.org/jira/browse/CARBONDATA-904?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] SWATI RAO updated CARBONDATA-904: - Attachment: (was: Test_Data1.csv) > ArrayIndexOutOfBoundsException > --- > > Key: CARBONDATA-904 > URL: https://issues.apache.org/jira/browse/CARBONDATA-904 > Project: CarbonData > Issue Type: Bug > Environment: Spark1.6 >Reporter: SWATI RAO > Fix For: 1.1.0-incubating > > Attachments: Test_Data1_h1.csv, Test_Data1_h1.csv > > > Or operator is not working properly. > When we execute these query in hive it is working fine but when we execute > the same in carbondata it throws an exception: > java.lang.ArrayIndexOutOfBoundsException > HIVE: > 0: jdbc:hive2://hadoop-master:1> create table Test_Boundary_h1 (c1_int > int,c2_Bigint Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string > string,c6_Timestamp Timestamp,c7_Datatype_Desc string) ROW FORMAT DELIMITED > FIELDS TERMINATED BY ',' ; > +-+--+ > | result | > +-+--+ > +-+--+ > No rows selected (1.177 seconds) > 0: jdbc:hive2://hadoop-master:1> load data local inpath > '/opt/Carbon/CarbonData/TestData/Data/Test_Data1_h1.csv' OVERWRITE INTO TABLE > Test_Boundary_h1 ; > +-+--+ > | Result | > +-+--+ > +-+--+ > No rows selected (0.437 seconds) > 0: jdbc:hive2://hadoop-master:1> select c6_Timestamp,max(c6_Timestamp) > from Test_Boundary_h1 where c6_Timestamp ='2017-07-01 12:07:28' or > c6_Timestamp ='2019-07-05 13:07:30' or c6_Timestamp = '1999-01-06 10:05:29' > group by c6_Timestamp ; > +++--+ > | c6_Timestamp | _c1 | > +++--+ > | 2017-07-01 12:07:28.0 | 2017-07-01 12:07:28.0 | > +++--+ > 1 row selected (1.637 seconds) > CARBONDATA: > 0: jdbc:hive2://hadoop-master:1> create table Test_Boundary (c1_int > int,c2_Bigint Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string > string,c6_Timestamp Timestamp,c7_Datatype_Desc string) STORED BY > 'org.apache.carbondata.format' ; > +-+--+ > | Result | > +-+--+ > +-+--+ > No rows selected (4.48 seconds) > 0: jdbc:hive2://hadoop-master:1> LOAD DATA INPATH > 'hdfs://192.168.2.145:54310/BabuStore/Data/Test_Data1.csv' INTO table > Test_Boundary > OPTIONS('DELIMITER'=',','QUOTECHAR'='','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='') > ; > +-+--+ > | Result | > +-+--+ > +-+--+ > No rows selected (4.445 seconds) > 0: jdbc:hive2://hadoop-master:1> select c6_Timestamp,max(c6_Timestamp) > from Test_Boundary where c6_Timestamp ='2017-07-01 12:07:28' or c6_Timestamp > =' 2019-07-05 13:07:30' or c6_Timestamp = '1999-01-06 10:05:29' group by > c6_Timestamp ; > Error: org.apache.spark.SparkException: Job aborted due to stage failure: > Task 0 in stage 5.0 failed 4 times, most recent failure: Lost task 0.3 in > stage 5.0 (TID 8, hadoop-master): java.lang.RuntimeException: > java.util.concurrent.ExecutionException: > java.lang.ArrayIndexOutOfBoundsException: 0 > at > org.apache.carbondata.core.scan.processor.AbstractDataBlockIterator.updateScanner(AbstractDataBlockIterator.java:136) > at > org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:50) > at > org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:32) > at > org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.getBatchResult(DetailQueryResultIterator.java:50) > at > org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.next(DetailQueryResultIterator.java:41) > at > org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.next(DetailQueryResultIterator.java:31) > at > org.apache.carbondata.core.scan.result.iterator.ChunkRowIterator.(ChunkRowIterator.java:41) > at > org.apache.carbondata.hadoop.CarbonRecordReader.initialize(CarbonRecordReader.java:79) > at > org.apache.carbondata.spark.rdd.CarbonScanRDD.compute(CarbonScanRDD.scala:204) > at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) > at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) > at > org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) > at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) > at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) > at > org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) > at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) > at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) > at >
[GitHub] incubator-carbondata pull request #775: [CARBONDATA-660] Bad Records Logs an...
Github user asfgit closed the pull request at: https://github.com/apache/incubator-carbondata/pull/775 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] incubator-carbondata issue #775: [CARBONDATA-660] Bad Records Logs and Raw C...
Github user jackylk commented on the issue: https://github.com/apache/incubator-carbondata/pull/775 LGTM --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] incubator-carbondata issue #737: [CARBONDATA-882] Add SORT_COLUMNS support i...
Github user jackylk commented on the issue: https://github.com/apache/incubator-carbondata/pull/737 please retest this --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[jira] [Created] (CARBONDATA-904) ArrayIndexOutOfBoundsException
SWATI RAO created CARBONDATA-904: Summary: ArrayIndexOutOfBoundsException Key: CARBONDATA-904 URL: https://issues.apache.org/jira/browse/CARBONDATA-904 Project: CarbonData Issue Type: Bug Reporter: SWATI RAO Attachments: Test_Data1.csv, Test_Data1_h1.csv Or operator is not working properly. When we execute these query in hive it is working fine but when we execute the same in carbondata it throws an exception: java.lang.ArrayIndexOutOfBoundsException HIVE: 0: jdbc:hive2://hadoop-master:1> create table Test_Boundary_h1 (c1_int int,c2_Bigint Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string string,c6_Timestamp Timestamp,c7_Datatype_Desc string) ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' ; +-+--+ | result | +-+--+ +-+--+ No rows selected (1.177 seconds) 0: jdbc:hive2://hadoop-master:1> load data local inpath '/opt/Carbon/CarbonData/TestData/Data/Test_Data1_h1.csv' OVERWRITE INTO TABLE Test_Boundary_h1 ; +-+--+ | Result | +-+--+ +-+--+ No rows selected (0.437 seconds) 0: jdbc:hive2://hadoop-master:1> select c6_Timestamp,max(c6_Timestamp) from Test_Boundary_h1 where c6_Timestamp ='2017-07-01 12:07:28' or c6_Timestamp ='2019-07-05 13:07:30' or c6_Timestamp = '1999-01-06 10:05:29' group by c6_Timestamp ; +++--+ | c6_Timestamp | _c1 | +++--+ | 2017-07-01 12:07:28.0 | 2017-07-01 12:07:28.0 | +++--+ 1 row selected (1.637 seconds) CARBONDATA: 0: jdbc:hive2://hadoop-master:1> create table Test_Boundary (c1_int int,c2_Bigint Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string string,c6_Timestamp Timestamp,c7_Datatype_Desc string) STORED BY 'org.apache.carbondata.format' ; +-+--+ | Result | +-+--+ +-+--+ No rows selected (4.48 seconds) 0: jdbc:hive2://hadoop-master:1> LOAD DATA INPATH 'hdfs://192.168.2.145:54310/BabuStore/Data/Test_Data1.csv' INTO table Test_Boundary OPTIONS('DELIMITER'=',','QUOTECHAR'='','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='') ; +-+--+ | Result | +-+--+ +-+--+ No rows selected (4.445 seconds) 0: jdbc:hive2://hadoop-master:1> select c6_Timestamp,max(c6_Timestamp) from Test_Boundary where c6_Timestamp ='2017-07-01 12:07:28' or c6_Timestamp =' 2019-07-05 13:07:30' or c6_Timestamp = '1999-01-06 10:05:29' group by c6_Timestamp ; Error: org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 5.0 failed 4 times, most recent failure: Lost task 0.3 in stage 5.0 (TID 8, hadoop-master): java.lang.RuntimeException: java.util.concurrent.ExecutionException: java.lang.ArrayIndexOutOfBoundsException: 0 at org.apache.carbondata.core.scan.processor.AbstractDataBlockIterator.updateScanner(AbstractDataBlockIterator.java:136) at org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:50) at org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:32) at org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.getBatchResult(DetailQueryResultIterator.java:50) at org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.next(DetailQueryResultIterator.java:41) at org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.next(DetailQueryResultIterator.java:31) at org.apache.carbondata.core.scan.result.iterator.ChunkRowIterator.(ChunkRowIterator.java:41) at org.apache.carbondata.hadoop.CarbonRecordReader.initialize(CarbonRecordReader.java:79) at org.apache.carbondata.spark.rdd.CarbonScanRDD.compute(CarbonScanRDD.scala:204) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41) at
[GitHub] incubator-carbondata issue #783: [CARBONDATA-903] data load is not failing e...
Github user CarbonDataQA commented on the issue: https://github.com/apache/incubator-carbondata/pull/783 Build Success with Spark 1.6.2, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/1570/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] incubator-carbondata pull request #709: [CARBONDATA-861] Improvements in que...
Github user asfgit closed the pull request at: https://github.com/apache/incubator-carbondata/pull/709 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[jira] [Assigned] (CARBONDATA-902) NoClassDefFoundError for Decimal datatype during select queries
[ https://issues.apache.org/jira/browse/CARBONDATA-902?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] anubhav tarar reassigned CARBONDATA-902: Assignee: anubhav tarar > NoClassDefFoundError for Decimal datatype during select queries > --- > > Key: CARBONDATA-902 > URL: https://issues.apache.org/jira/browse/CARBONDATA-902 > Project: CarbonData > Issue Type: Bug > Components: data-query > Environment: Spark 2.1, Hive 1.2.1 >Reporter: Neha Bhardwaj >Assignee: anubhav tarar >Priority: Minor > Attachments: testHive1.csv > > > Decimal data type raises exception while selecting the data from the table in > hive. > Steps to reproduce: > 1) In Spark Shell : > a) Create Table - > import org.apache.spark.sql.SparkSession > import org.apache.spark.sql.CarbonSession._ > val carbon = > SparkSession.builder().config(sc.getConf).getOrCreateCarbonSession("hdfs://localhost:54310/opt/data") > > scala> carbon.sql(""" create table testHive1(id int,name string,dob > timestamp,experience decimal,salary double,incentive bigint) stored > by'carbondata' """).show > b) Load Data - > scala> carbon.sql(""" load data inpath > 'hdfs://localhost:54310/Files/testHive1.csv' into table testHive1 """ ).show > 2) In Hive : > a) Add Jars - > add jar > /home/neha/incubator-carbondata/integration/hive/carbondata-hive-1.1.0-incubating-SNAPSHOT.jar; > add jar > /home/neha/incubator-carbondata/assembly/target/scala-2.11/carbondata_2.11-1.1.0-incubating-SNAPSHOT-shade-hadoop2.7.2.jar; > > b) Set Properties - > set hive.mapred.supports.subdirectories=true; > set mapreduce.input.fileinputformat.input.dir.recursive=true; > c) Alter location - > hive> alter table testHive1 set LOCATION > 'hdfs://localhost:54310/opt/data/default/testhive1' ; > d) Alter FileFormat - > alter table testHive1 set FILEFORMAT > INPUTFORMAT "org.apache.carbondata.hive.MapredCarbonInputFormat" > OUTPUTFORMAT "org.apache.carbondata.hive.MapredCarbonOutputFormat" > SERDE "org.apache.carbondata.hive.CarbonHiveSerDe"; > e) Create Table - > create table testHive1(id int,name string,dob timestamp,experience > decimal,salary double,incentive bigint); > f) Execute Queries - > select * from testHive1; > 3) Query : > hive> select * from testHive1; > Expected Output : > ResultSet should display all the data present in the table. > Result: > Exception in thread "[main][partitionID:testhive1;queryID:8945394553892]" > java.lang.NoClassDefFoundError: org/apache/spark/sql/types/Decimal > at > org.apache.carbondata.core.scan.collector.impl.AbstractScannedResultCollector.getMeasureData(AbstractScannedResultCollector.java:109) > at > org.apache.carbondata.core.scan.collector.impl.AbstractScannedResultCollector.fillMeasureData(AbstractScannedResultCollector.java:78) > at > org.apache.carbondata.core.scan.collector.impl.DictionaryBasedResultCollector.fillMeasureData(DictionaryBasedResultCollector.java:158) > at > org.apache.carbondata.core.scan.collector.impl.DictionaryBasedResultCollector.collectData(DictionaryBasedResultCollector.java:115) > at > org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:51) > at > org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:32) > at > org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.getBatchResult(DetailQueryResultIterator.java:50) > at > org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.next(DetailQueryResultIterator.java:41) > at > org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.next(DetailQueryResultIterator.java:31) > at > org.apache.carbondata.core.scan.result.iterator.ChunkRowIterator.(ChunkRowIterator.java:41) > at > org.apache.carbondata.hive.CarbonHiveRecordReader.initialize(CarbonHiveRecordReader.java:84) > at > org.apache.carbondata.hive.CarbonHiveRecordReader.(CarbonHiveRecordReader.java:66) > at > org.apache.carbondata.hive.MapredCarbonInputFormat.getRecordReader(MapredCarbonInputFormat.java:68) > at > org.apache.hadoop.hive.ql.exec.FetchOperator$FetchInputFormatSplit.getRecordReader(FetchOperator.java:673) > at > org.apache.hadoop.hive.ql.exec.FetchOperator.getRecordReader(FetchOperator.java:323) > at > org.apache.hadoop.hive.ql.exec.FetchOperator.getNextRow(FetchOperator.java:445) > at > org.apache.hadoop.hive.ql.exec.FetchOperator.pushRow(FetchOperator.java:414) > at org.apache.hadoop.hive.ql.exec.FetchTask.fetch(FetchTask.java:140) > at org.apache.hadoop.hive.ql.Driver.getResults(Driver.java:1670) > at >
[GitHub] incubator-carbondata issue #709: [CARBONDATA-861] Improvements in query
Github user jackylk commented on the issue: https://github.com/apache/incubator-carbondata/pull/709 LGTM --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] incubator-carbondata pull request #783: [CARBONDATA-903] data load is not fa...
GitHub user mohammadshahidkhan opened a pull request: https://github.com/apache/incubator-carbondata/pull/783 [CARBONDATA-903] data load is not failing even though bad records exists in the data in case of unsafe sort or batch sort You can merge this pull request into a Git repository by running: $ git pull https://github.com/mohammadshahidkhan/incubator-carbondata dataload_termination_on_fail Alternatively you can review and apply these changes as the patch at: https://github.com/apache/incubator-carbondata/pull/783.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #783 commit 39d329a3ca6eda1d9e372a0dbabc804b5b7ad493 Author: mohammadshahidkhanDate: 2017-04-11T12:56:51Z [CARBONDATA-903] data load is not failing even though bad records exists in the data in case of unsafe sort or batch sort --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[jira] [Updated] (CARBONDATA-903) data load is not failing even though bad records exists in the data in case of unsafe sort or batch sort
[ https://issues.apache.org/jira/browse/CARBONDATA-903?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Mohammad Shahid Khan updated CARBONDATA-903: Summary: data load is not failing even though bad records exists in the data in case of unsafe sort or batch sort (was: data load is not failing even though bad records exists in the data) > data load is not failing even though bad records exists in the data in case > of unsafe sort or batch sort > > > Key: CARBONDATA-903 > URL: https://issues.apache.org/jira/browse/CARBONDATA-903 > Project: CarbonData > Issue Type: Bug >Reporter: Mohammad Shahid Khan >Assignee: Mohammad Shahid Khan >Priority: Critical > -- This message was sent by Atlassian JIRA (v6.3.15#6346)
[jira] [Commented] (CARBONDATA-903) data load is not failing even though bad records exists in the data
[ https://issues.apache.org/jira/browse/CARBONDATA-903?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15964327#comment-15964327 ] Mohammad Shahid Khan commented on CARBONDATA-903: - When carbon.load.use.batch.sort or enable.unsafe.sort or table bucket configured the data loading is not failing even in though the bad records exists in the data being loaded. > data load is not failing even though bad records exists in the data > --- > > Key: CARBONDATA-903 > URL: https://issues.apache.org/jira/browse/CARBONDATA-903 > Project: CarbonData > Issue Type: Bug >Reporter: Mohammad Shahid Khan >Assignee: Mohammad Shahid Khan >Priority: Critical > -- This message was sent by Atlassian JIRA (v6.3.15#6346)
[jira] [Assigned] (CARBONDATA-903) data load is not failing even though bad records exists in the data
[ https://issues.apache.org/jira/browse/CARBONDATA-903?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Mohammad Shahid Khan reassigned CARBONDATA-903: --- Assignee: Mohammad Shahid Khan > data load is not failing even though bad records exists in the data > --- > > Key: CARBONDATA-903 > URL: https://issues.apache.org/jira/browse/CARBONDATA-903 > Project: CarbonData > Issue Type: Bug >Reporter: Mohammad Shahid Khan >Assignee: Mohammad Shahid Khan >Priority: Critical > -- This message was sent by Atlassian JIRA (v6.3.15#6346)
[jira] [Created] (CARBONDATA-903) data load is not failing even though bad records exists in the data
Mohammad Shahid Khan created CARBONDATA-903: --- Summary: data load is not failing even though bad records exists in the data Key: CARBONDATA-903 URL: https://issues.apache.org/jira/browse/CARBONDATA-903 Project: CarbonData Issue Type: Bug Reporter: Mohammad Shahid Khan Priority: Critical -- This message was sent by Atlassian JIRA (v6.3.15#6346)
[jira] [Created] (CARBONDATA-902) NoClassDefFoundError for Decimal datatype during select queries
Neha Bhardwaj created CARBONDATA-902: Summary: NoClassDefFoundError for Decimal datatype during select queries Key: CARBONDATA-902 URL: https://issues.apache.org/jira/browse/CARBONDATA-902 Project: CarbonData Issue Type: Bug Components: data-query Environment: Spark 2.1, Hive 1.2.1 Reporter: Neha Bhardwaj Priority: Minor Attachments: testHive1.csv Decimal data type raises exception while selecting the data from the table in hive. Steps to reproduce: 1) In Spark Shell : a) Create Table - import org.apache.spark.sql.SparkSession import org.apache.spark.sql.CarbonSession._ val carbon = SparkSession.builder().config(sc.getConf).getOrCreateCarbonSession("hdfs://localhost:54310/opt/data") scala> carbon.sql(""" create table testHive1(id int,name string,dob timestamp,experience decimal,salary double,incentive bigint) stored by'carbondata' """).show b) Load Data - scala> carbon.sql(""" load data inpath 'hdfs://localhost:54310/Files/testHive1.csv' into table testHive1 """ ).show 2) In Hive : a) Add Jars - add jar /home/neha/incubator-carbondata/integration/hive/carbondata-hive-1.1.0-incubating-SNAPSHOT.jar; add jar /home/neha/incubator-carbondata/assembly/target/scala-2.11/carbondata_2.11-1.1.0-incubating-SNAPSHOT-shade-hadoop2.7.2.jar; b) Set Properties - set hive.mapred.supports.subdirectories=true; set mapreduce.input.fileinputformat.input.dir.recursive=true; c) Alter location - hive> alter table testHive1 set LOCATION 'hdfs://localhost:54310/opt/data/default/testhive1' ; d) Alter FileFormat - alter table testHive1 set FILEFORMAT INPUTFORMAT "org.apache.carbondata.hive.MapredCarbonInputFormat" OUTPUTFORMAT "org.apache.carbondata.hive.MapredCarbonOutputFormat" SERDE "org.apache.carbondata.hive.CarbonHiveSerDe"; e) Create Table - create table testHive1(id int,name string,dob timestamp,experience decimal,salary double,incentive bigint); f) Execute Queries - select * from testHive1; 3) Query : hive> select * from testHive1; Expected Output : ResultSet should display all the data present in the table. Result: Exception in thread "[main][partitionID:testhive1;queryID:8945394553892]" java.lang.NoClassDefFoundError: org/apache/spark/sql/types/Decimal at org.apache.carbondata.core.scan.collector.impl.AbstractScannedResultCollector.getMeasureData(AbstractScannedResultCollector.java:109) at org.apache.carbondata.core.scan.collector.impl.AbstractScannedResultCollector.fillMeasureData(AbstractScannedResultCollector.java:78) at org.apache.carbondata.core.scan.collector.impl.DictionaryBasedResultCollector.fillMeasureData(DictionaryBasedResultCollector.java:158) at org.apache.carbondata.core.scan.collector.impl.DictionaryBasedResultCollector.collectData(DictionaryBasedResultCollector.java:115) at org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:51) at org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:32) at org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.getBatchResult(DetailQueryResultIterator.java:50) at org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.next(DetailQueryResultIterator.java:41) at org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.next(DetailQueryResultIterator.java:31) at org.apache.carbondata.core.scan.result.iterator.ChunkRowIterator.(ChunkRowIterator.java:41) at org.apache.carbondata.hive.CarbonHiveRecordReader.initialize(CarbonHiveRecordReader.java:84) at org.apache.carbondata.hive.CarbonHiveRecordReader.(CarbonHiveRecordReader.java:66) at org.apache.carbondata.hive.MapredCarbonInputFormat.getRecordReader(MapredCarbonInputFormat.java:68) at org.apache.hadoop.hive.ql.exec.FetchOperator$FetchInputFormatSplit.getRecordReader(FetchOperator.java:673) at org.apache.hadoop.hive.ql.exec.FetchOperator.getRecordReader(FetchOperator.java:323) at org.apache.hadoop.hive.ql.exec.FetchOperator.getNextRow(FetchOperator.java:445) at org.apache.hadoop.hive.ql.exec.FetchOperator.pushRow(FetchOperator.java:414) at org.apache.hadoop.hive.ql.exec.FetchTask.fetch(FetchTask.java:140) at org.apache.hadoop.hive.ql.Driver.getResults(Driver.java:1670) at org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:233) at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:165) at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:376) at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:736) at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:681) at
[jira] [Commented] (CARBONDATA-688) Abnormal behaviour of double datatype when used in DICTIONARY_INCLUDE and filtering null values
[ https://issues.apache.org/jira/browse/CARBONDATA-688?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15964302#comment-15964302 ] Rahul Kumar commented on CARBONDATA-688: [~geetikagupta], This is working fine now . You can close it. > Abnormal behaviour of double datatype when used in DICTIONARY_INCLUDE and > filtering null values > --- > > Key: CARBONDATA-688 > URL: https://issues.apache.org/jira/browse/CARBONDATA-688 > Project: CarbonData > Issue Type: Bug > Components: data-query >Affects Versions: 1.1.0-incubating > Environment: Spark 2.1 >Reporter: Geetika Gupta >Priority: Minor > Attachments: 100_olap_C20.csv > > > I tried to create a table having double as a column and load null values into > that table. When I performed the select query on the table, it is displaying > wrong data. > Below are the commands used: > Create table : > create table Comp_VMALL_DICTIONARY_INCLUDE (imei string,deviceInformationId > int,MAC string,deviceColor string,device_backColor string,modelId > string,marketName string,AMSize string,ROMSize string,CUPAudit > string,CPIClocked string,series string,productionDate timestamp,bomCode > string,internalModels string, deliveryTime string, channelsId string, > channelsName string , deliveryAreaId string, deliveryCountry string, > deliveryProvince string, deliveryCity string,deliveryDistrict string, > deliveryStreet string, oxSingleNumber string, ActiveCheckTime string, > ActiveAreaId string, ActiveCountry string, ActiveProvince string, Activecity > string, ActiveDistrict string, ActiveStreet string, ActiveOperatorId string, > Active_releaseId string, Active_EMUIVersion string, Active_operaSysVersion > string, Active_BacVerNumber string, Active_BacFlashVer string, > Active_webUIVersion string, Active_webUITypeCarrVer > string,Active_webTypeDataVerNumber string, Active_operatorsVersion string, > Active_phonePADPartitionedVersions string, Latest_YEAR int, Latest_MONTH int, > Latest_DAY Decimal(30,10), Latest_HOUR string, Latest_areaId string, > Latest_country string, Latest_province string, Latest_city string, > Latest_district string, Latest_street string, Latest_releaseId string, > Latest_EMUIVersion string, Latest_operaSysVersion string, Latest_BacVerNumber > string, Latest_BacFlashVer string, Latest_webUIVersion string, > Latest_webUITypeCarrVer string, Latest_webTypeDataVerNumber string, > Latest_operatorsVersion string, Latest_phonePADPartitionedVersions string, > Latest_operatorId string, gamePointDescription string,gamePointId > double,contractNumber BigInt) STORED BY 'org.apache.carbondata.format' > TBLPROPERTIES('DICTIONARY_INCLUDE'='imei,deviceInformationId,productionDate,gamePointId,Latest_DAY,contractNumber'); > Load command: > LOAD DATA INPATH 'hdfs://localhost:54311/BabuStore/DATA/100_olap_C20.csv' > INTO table Comp_VMALL_DICTIONARY_INCLUDE options ('DELIMITER'=',', > 'QUOTECHAR'='"', > 'BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='imei,deviceInformationId,MAC,deviceColor,device_backColor,modelId,marketName,AMSize,ROMSize,CUPAudit,CPIClocked,series,productionDate,bomCode,internalModels,deliveryTime,channelsId,channelsName,deliveryAreaId,deliveryCountry,deliveryProvince,deliveryCity,deliveryDistrict,deliveryStreet,oxSingleNumber,contractNumber,ActiveCheckTime,ActiveAreaId,ActiveCountry,ActiveProvince,Activecity,ActiveDistrict,ActiveStreet,ActiveOperatorId,Active_releaseId,Active_EMUIVersion,Active_operaSysVersion,Active_BacVerNumber,Active_BacFlashVer,Active_webUIVersion,Active_webUITypeCarrVer,Active_webTypeDataVerNumber,Active_operatorsVersion,Active_phonePADPartitionedVersions,Latest_YEAR,Latest_MONTH,Latest_DAY,Latest_HOUR,Latest_areaId,Latest_country,Latest_province,Latest_city,Latest_district,Latest_street,Latest_releaseId,Latest_EMUIVersion,Latest_operaSysVersion,Latest_BacVerNumber,Latest_BacFlashVer,Latest_webUIVersion,Latest_webUITypeCarrVer,Latest_webTypeDataVerNumber,Latest_operatorsVersion,Latest_phonePADPartitionedVersions,Latest_operatorId,gamePointId,gamePointDescription'); > Select query: > select gamePointId from Comp_VMALL_DICTIONARY_INCLUDE where gamePointId IS > NOT NULL order by gamePointId; > select gamePointId from Comp_VMALL_DICTIONARY_INCLUDE where gamePointId is > NULL; > The first select command displays null values as well and the second command > displays no values. -- This message was sent by Atlassian JIRA (v6.3.15#6346)
[GitHub] incubator-carbondata pull request #781: [CARBONDATA-900] Is null query on a ...
GitHub user manishgupta88 reopened a pull request: https://github.com/apache/incubator-carbondata/pull/781 [CARBONDATA-900] Is null query on a newly added measure column is not returning proper results Analysis: When is null query is executed on newly added measure column, control goes to RowLevelFilterExecuterImpl class, where measure existence is checked. In case the measure is not found, bitset group is not getting populated with default values due to which that block is not returning any result. Solution: When query is on a restructured block where newly added column does not exist, create the default bitset group so that based on default value existence default bitset group is created and results are returned based on that. You can merge this pull request into a Git repository by running: $ git pull https://github.com/manishgupta88/incubator-carbondata alter_measure_query_fix Alternatively you can review and apply these changes as the patch at: https://github.com/apache/incubator-carbondata/pull/781.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #781 commit 1c68a23448e9a3a3a5fca9331125f7bd7aac11a4 Author: manishgupta88Date: 2017-04-11T11:24:03Z Problem: Is null query on a newly added measure column is not returning proper results. Analysis: When is null query is executed on newly added measure column, control goes to RowLevelFilterExecuterImpl class, where measure existence is checked. In case the measure is not found, bitset group is not getting populated with default values due to which that block is not returning any result. Solution: When query is on a restructured block where newly added column does not exist, create the default bitset group so that based on default value existence default bitset group is created and results are returned based on that. --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] incubator-carbondata pull request #781: [CARBONDATA-900] Is null query on a ...
Github user manishgupta88 closed the pull request at: https://github.com/apache/incubator-carbondata/pull/781 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[jira] [Updated] (CARBONDATA-901) Fix some spelling mistakes
[ https://issues.apache.org/jira/browse/CARBONDATA-901?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Cao Gaofei updated CARBONDATA-901: -- Request participants: (was: ) Summary: Fix some spelling mistakes (was: fix some spelling mistakes) > Fix some spelling mistakes > -- > > Key: CARBONDATA-901 > URL: https://issues.apache.org/jira/browse/CARBONDATA-901 > Project: CarbonData > Issue Type: Improvement > Components: core >Reporter: Cao Gaofei >Priority: Trivial > Time Spent: 20m > Remaining Estimate: 0h > > https://github.com/apache/incubator-carbondata/pull/782 -- This message was sent by Atlassian JIRA (v6.3.15#6346)
[GitHub] incubator-carbondata issue #780: [CARBONDATA-892] Fixed NullPointerException...
Github user CarbonDataQA commented on the issue: https://github.com/apache/incubator-carbondata/pull/780 Build Success with Spark 1.6.2, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/1569/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] incubator-carbondata issue #781: [CARBONDATA-900] Is null query on a newly a...
Github user CarbonDataQA commented on the issue: https://github.com/apache/incubator-carbondata/pull/781 Build Success with Spark 1.6.2, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/1568/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[jira] [Created] (CARBONDATA-900) Is null query on a newly added measure column is not returning proper results
Manish Gupta created CARBONDATA-900: --- Summary: Is null query on a newly added measure column is not returning proper results Key: CARBONDATA-900 URL: https://issues.apache.org/jira/browse/CARBONDATA-900 Project: CarbonData Issue Type: Bug Reporter: Manish Gupta Assignee: Manish Gupta Priority: Minor Fix For: 1.1.0-incubating When is null query is executed on newly added measure column, control goes to RowLevelFilterExecuterImpl class, where measure existence is checked. In case the measure is not found, bitset group is not getting populated with default values due to which that block is not returning any result. Below queries can be executed to reproduce the issue: CREATE TABLE uniqdata110 (CUST_ID int,CUST_NAME String) STORED BY 'carbondata' LOAD DATA INPATH '' into table uniqdata110 OPTIONS('BAD_RECORDS_LOGGER_ENABLE'='TRUE', 'BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME') ALTER TABLE uniqdata110 ADD COLUMNS (a6 int) LOAD DATA INPATH '' into table uniqdata110 OPTIONS('BAD_RECORDS_LOGGER_ENABLE'='TRUE', 'BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,a6') select * from uniqdata110 select * from uniqdata110 where a6 is null Data: 7,hello1 8,welcome1 bye,11 -- This message was sent by Atlassian JIRA (v6.3.15#6346)
[GitHub] incubator-carbondata pull request #779: [CARBONDATA-899] Added Support for D...
GitHub user bhavya411 opened a pull request: https://github.com/apache/incubator-carbondata/pull/779 [CARBONDATA-899] Added Support for DecimalType and Timestamp for spark-2.1 Added Support for DecimalType and Timestamp with Spark-2.1 Local Build is running fine. Manually Tested the functionality on Presto CLI You can merge this pull request into a Git repository by running: $ git pull https://github.com/bhavya411/incubator-carbondata CARBONDATA-899 Alternatively you can review and apply these changes as the patch at: https://github.com/apache/incubator-carbondata/pull/779.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #779 commit 2a1df96f100445b26287cb37549740cb45761990 Author: BhavyaDate: 2017-04-04T08:49:49Z Fixed Carbondata-848 commit ef9269b962b1f2332c25bc95a3af3d0aadc6b19e Author: Bhavya Date: 2017-04-04T10:41:20Z remove Presto module ' commit 30017786f948b0f88a385e54029dee00a85f5648 Author: Bhavya Date: 2017-04-11T09:46:39Z Added support for DecimalType and Fixed Date and time issues --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[jira] [Created] (CARBONDATA-899) Added Support for Decimal data type and Fixed the timestamp and date issues for Spark-2.1
Bhavya Aggarwal created CARBONDATA-899: -- Summary: Added Support for Decimal data type and Fixed the timestamp and date issues for Spark-2.1 Key: CARBONDATA-899 URL: https://issues.apache.org/jira/browse/CARBONDATA-899 Project: CarbonData Issue Type: Improvement Components: presto-integration Reporter: Bhavya Aggarwal Assignee: Bhavya Aggarwal Priority: Minor The Decimal Type correct support is added as well as issues related to timestamp and date are resolved. -- This message was sent by Atlassian JIRA (v6.3.15#6346)
[GitHub] incubator-carbondata issue #777: [CARBONDATA-863] Moved AlterTableAddColumnR...
Github user CarbonDataQA commented on the issue: https://github.com/apache/incubator-carbondata/pull/777 Build Success with Spark 1.6.2, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/1564/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] incubator-carbondata issue #778: [WIP] Added getAll dictionary to codegen of...
Github user CarbonDataQA commented on the issue: https://github.com/apache/incubator-carbondata/pull/778 Build Success with Spark 1.6.2, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/1563/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] incubator-carbondata issue #777: [CARBONDATA-863] Moved AlterTableAddColumnR...
Github user kunal642 commented on the issue: https://github.com/apache/incubator-carbondata/pull/777 retest this please --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] incubator-carbondata pull request #778: [WIP] Added getAll dictionary to cod...
GitHub user ravipesala opened a pull request: https://github.com/apache/incubator-carbondata/pull/778 [WIP] Added getAll dictionary to codegen of dictionary decoder. You can merge this pull request into a Git repository by running: $ git pull https://github.com/ravipesala/incubator-carbondata getAlldictionary-codegen Alternatively you can review and apply these changes as the patch at: https://github.com/apache/incubator-carbondata/pull/778.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #778 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] incubator-carbondata pull request #751: [CARBONDATA-816] Added Example for H...
Github user anubhav100 commented on a diff in the pull request: https://github.com/apache/incubator-carbondata/pull/751#discussion_r110853402 --- Diff: integration/hive/pom.xml --- @@ -64,6 +64,79 @@ compile +org.apache.spark +spark-network-common_2.10 +2.1.0 + + +org.apache.hadoop +hadoop-hdfs +2.7.3 --- End diff -- resolved --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] incubator-carbondata pull request #751: [CARBONDATA-816] Added Example for H...
Github user anubhav100 commented on a diff in the pull request: https://github.com/apache/incubator-carbondata/pull/751#discussion_r110853360 --- Diff: integration/hive/pom.xml --- @@ -64,6 +64,79 @@ compile +org.apache.spark +spark-network-common_2.10 +2.1.0 --- End diff -- resolved --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] incubator-carbondata pull request #751: [CARBONDATA-816] Added Example for H...
Github user anubhav100 commented on a diff in the pull request: https://github.com/apache/incubator-carbondata/pull/751#discussion_r110853283 --- Diff: integration/hive/pom.xml --- @@ -64,6 +64,79 @@ compile +org.apache.spark +spark-network-common_2.10 +2.1.0 + + +org.apache.hadoop +hadoop-hdfs +2.7.3 + + +ch.qos.logback +logback-classic + + +javax.servlet +servlet-api + + + + +org.apache.zookeeper +zookeeper +3.4.7 + + +jline +jline + + + + +org.apache.carbondata +carbondata-spark2 +${project.version} + + +org.apache.spark +spark-sql_${scala.binary.version} + + +org.apache.spark + spark-hive-thriftserver_${scala.binary.version} + + +org.apache.spark +spark-repl_${scala.binary.version} + + +org.apache.hadoop +hadoop-common +2.7.3 + + +org.apache.httpcomponents +httpclient +4.3.4 + + +org.apache.httpcomponents +httpcore +4.3-alpha1 + + +org.apache.hadoop --- End diff -- done --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] incubator-carbondata issue #777: [CARBONDATA-863] Moved AlterTableAddColumnR...
Github user CarbonDataQA commented on the issue: https://github.com/apache/incubator-carbondata/pull/777 Build Failed with Spark 1.6.2, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/1562/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] incubator-carbondata issue #777: [CARBONDATA-863] Moved AlterTableAddColumnR...
Github user CarbonDataQA commented on the issue: https://github.com/apache/incubator-carbondata/pull/777 Build Failed with Spark 1.6.2, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder/1561/ --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[jira] [Updated] (CARBONDATA-896) throws nullpointerException while insert data
[ https://issues.apache.org/jira/browse/CARBONDATA-896?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] sehriff updated CARBONDATA-896: --- Affects Version/s: 1.0.0-incubating > throws nullpointerException while insert data > - > > Key: CARBONDATA-896 > URL: https://issues.apache.org/jira/browse/CARBONDATA-896 > Project: CarbonData > Issue Type: Bug >Affects Versions: 1.0.0-incubating >Reporter: sehriff > > insert data into carbon table from hive using sql like: > cc.sql("insert into carbon.table_carbon select * from hivetable").show > and got the following error: > Job aborted due to stage failure: Task 0 in stage 26.3 failed 4 times, most > recent failure: Lost task 0.3 in stage 26.3 (TID 5628, HDD013): > java.lang.NullPointerException > at > org.apache.carbondata.core.datastore.filesystem.AbstractDFSCarbonFile.getLastModifiedTime(AbstractDFSCarbonFile.java:135) > at > org.apache.carbondata.core.datastore.filesystem.AbstractDFSCarbonFile.isFileModified(AbstractDFSCarbonFile.java:210) > at > org.apache.carbondata.core.cache.dictionary.AbstractDictionaryCache.isDictionaryMetaFileModified(AbstractDictionaryCache.java:119) > at > org.apache.carbondata.core.cache.dictionary.AbstractDictionaryCache.checkAndLoadDictionaryData(AbstractDictionaryCache.java:158) > at > org.apache.carbondata.core.cache.dictionary.ReverseDictionaryCache.getDictionary(ReverseDictionaryCache.java:174) > at > org.apache.carbondata.core.cache.dictionary.ReverseDictionaryCache.get(ReverseDictionaryCache.java:67) > at > org.apache.carbondata.core.cache.dictionary.ReverseDictionaryCache.get(ReverseDictionaryCache.java:38) > at > org.apache.carbondata.spark.load.CarbonLoaderUtil.getDictionary(CarbonLoaderUtil.java:463) > at > org.apache.carbondata.spark.load.CarbonLoaderUtil.getDictionary(CarbonLoaderUtil.java:469) > at > org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD$$anon$1.(CarbonGlobalDictionaryRDD.scala:413) > at > org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD.compute(CarbonGlobalDictionaryRDD.scala:342) > at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) > at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) > at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66) > at org.apache.spark.scheduler.Task.run(Task.scala:89) > at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) > at java.lang.Thread.run(Thread.java:745) -- This message was sent by Atlassian JIRA (v6.3.15#6346)
[GitHub] incubator-carbondata pull request #777: [CARBONDATA-863] Moved AlterTableAdd...
GitHub user kunal642 opened a pull request: https://github.com/apache/incubator-carbondata/pull/777 [CARBONDATA-863] Moved AlterTableAddColumnRDD call to AlterTableCommands You can merge this pull request into a Git repository by running: $ git pull https://github.com/kunal642/incubator-carbondata CARBONDATA-863 Alternatively you can review and apply these changes as the patch at: https://github.com/apache/incubator-carbondata/pull/777.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #777 commit d83cec14d02bf3ef828fc978c5a255f2c34ef40a Author: kunal642Date: 2017-04-11T08:45:10Z added AlterTableAddColumnRDD to AlterTableCommands --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[GitHub] incubator-carbondata issue #776: [CARBONDATA-897] Removed redundant fields f...
Github user CarbonDataQA commented on the issue: https://github.com/apache/incubator-carbondata/pull/776 Can one of the admins verify this patch? --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[jira] [Updated] (CARBONDATA-887) lazy rdd iterator for InsertInto
[ https://issues.apache.org/jira/browse/CARBONDATA-887?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Venkata Ramana G updated CARBONDATA-887: Priority: Minor (was: Major) > lazy rdd iterator for InsertInto > > > Key: CARBONDATA-887 > URL: https://issues.apache.org/jira/browse/CARBONDATA-887 > Project: CarbonData > Issue Type: Improvement >Reporter: QiangCai >Assignee: QiangCai >Priority: Minor > Fix For: 1.1.0-incubating > > Time Spent: 2.5h > Remaining Estimate: 0h > -- This message was sent by Atlassian JIRA (v6.3.15#6346)
[jira] [Resolved] (CARBONDATA-887) lazy rdd iterator for InsertInto
[ https://issues.apache.org/jira/browse/CARBONDATA-887?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Venkata Ramana G resolved CARBONDATA-887. - Resolution: Fixed Fix Version/s: 1.1.0-incubating > lazy rdd iterator for InsertInto > > > Key: CARBONDATA-887 > URL: https://issues.apache.org/jira/browse/CARBONDATA-887 > Project: CarbonData > Issue Type: Improvement >Reporter: QiangCai >Assignee: QiangCai >Priority: Minor > Fix For: 1.1.0-incubating > > Time Spent: 2.5h > Remaining Estimate: 0h > -- This message was sent by Atlassian JIRA (v6.3.15#6346)
[GitHub] incubator-carbondata pull request #776: [CARBONDATA-897] Removed redundant f...
GitHub user jatin9896 opened a pull request: https://github.com/apache/incubator-carbondata/pull/776 [CARBONDATA-897] Removed redundant fields from table in configuration-parameters.md You can merge this pull request into a Git repository by running: $ git pull https://github.com/jatin9896/incubator-carbondata docs/issuesInMdFileTable Alternatively you can review and apply these changes as the patch at: https://github.com/apache/incubator-carbondata/pull/776.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #776 commit 8e8135711c45e088c7b0cf1d47a6467ca6b62529 Author: jatinDate: 2017-04-11T08:08:10Z Removed redundant fields from table in configuration-parameters.md --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[jira] [Created] (CARBONDATA-898) When select query and alter table rename table is triggered concurrently, NullPointerException is getting thrown
Naresh P R created CARBONDATA-898: - Summary: When select query and alter table rename table is triggered concurrently, NullPointerException is getting thrown Key: CARBONDATA-898 URL: https://issues.apache.org/jira/browse/CARBONDATA-898 Project: CarbonData Issue Type: Bug Environment: Spark 2.1 Reporter: Naresh P R Assignee: Naresh P R Priority: Minor When user triggers select query and alter table rename table command concurrently, Select query is throwning NullPointerException if the files does not exist in hdfs. When dictionary file or schema file does not exist, File not found exception should be thrown -- This message was sent by Atlassian JIRA (v6.3.15#6346)
[GitHub] incubator-carbondata pull request #765: [CARBONDATA-887]lazy rdd iterator fo...
Github user asfgit closed the pull request at: https://github.com/apache/incubator-carbondata/pull/765 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---
[jira] [Created] (CARBONDATA-897) Redundant Fields Inside * **Global Dictionary Configurations** in Configuration-parameters.md
Pallavi Singh created CARBONDATA-897: Summary: Redundant Fields Inside * **Global Dictionary Configurations** in Configuration-parameters.md Key: CARBONDATA-897 URL: https://issues.apache.org/jira/browse/CARBONDATA-897 Project: CarbonData Issue Type: Bug Components: docs Reporter: Pallavi Singh Assignee: Pallavi Singh Priority: Minor Attachments: Configurations.png In the Configuration-parameters.md file under the table Global Dictionary Configurations the row for field high.cardinality.threshold has extra columns with redundant values in the md file. -- This message was sent by Atlassian JIRA (v6.3.15#6346)
[jira] [Created] (CARBONDATA-896) throws nullpointerException while insert data
sehriff created CARBONDATA-896: -- Summary: throws nullpointerException while insert data Key: CARBONDATA-896 URL: https://issues.apache.org/jira/browse/CARBONDATA-896 Project: CarbonData Issue Type: Bug Reporter: sehriff insert data into carbon table from hive using sql like: cc.sql("insert into carbon.table_carbon select * from hivetable").show and got the following error: Job aborted due to stage failure: Task 0 in stage 26.3 failed 4 times, most recent failure: Lost task 0.3 in stage 26.3 (TID 5628, HDD013): java.lang.NullPointerException at org.apache.carbondata.core.datastore.filesystem.AbstractDFSCarbonFile.getLastModifiedTime(AbstractDFSCarbonFile.java:135) at org.apache.carbondata.core.datastore.filesystem.AbstractDFSCarbonFile.isFileModified(AbstractDFSCarbonFile.java:210) at org.apache.carbondata.core.cache.dictionary.AbstractDictionaryCache.isDictionaryMetaFileModified(AbstractDictionaryCache.java:119) at org.apache.carbondata.core.cache.dictionary.AbstractDictionaryCache.checkAndLoadDictionaryData(AbstractDictionaryCache.java:158) at org.apache.carbondata.core.cache.dictionary.ReverseDictionaryCache.getDictionary(ReverseDictionaryCache.java:174) at org.apache.carbondata.core.cache.dictionary.ReverseDictionaryCache.get(ReverseDictionaryCache.java:67) at org.apache.carbondata.core.cache.dictionary.ReverseDictionaryCache.get(ReverseDictionaryCache.java:38) at org.apache.carbondata.spark.load.CarbonLoaderUtil.getDictionary(CarbonLoaderUtil.java:463) at org.apache.carbondata.spark.load.CarbonLoaderUtil.getDictionary(CarbonLoaderUtil.java:469) at org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD$$anon$1.(CarbonGlobalDictionaryRDD.scala:413) at org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD.compute(CarbonGlobalDictionaryRDD.scala:342) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66) at org.apache.spark.scheduler.Task.run(Task.scala:89) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745) -- This message was sent by Atlassian JIRA (v6.3.15#6346)
[GitHub] incubator-carbondata issue #765: [CARBONDATA-887]lazy rdd iterator for Inser...
Github user gvramana commented on the issue: https://github.com/apache/incubator-carbondata/pull/765 LGTM --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---