[GitHub] carbondata issue #2659: [CARBONDATA-2887] Fix complex filters on spark carbo...

2018-08-27 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2659
  
Build Success with Spark 2.2.1, Please check CI 
http://95.216.28.178:8080/job/ApacheCarbonPRBuilder1/39/



---


[GitHub] carbondata issue #2659: [CARBONDATA-2887] Fix complex filters on spark carbo...

2018-08-27 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2659
  
SDV Build Fail , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/6435/



---


[GitHub] carbondata issue #2623: [CARBONDATA-2844] [CARBONDATA-2865] Pass SK/AK to ex...

2018-08-27 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2623
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/8103/



---


[GitHub] carbondata issue #2654: [CARBONDATA-2896] Adaptive Encoding for Primitive da...

2018-08-27 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2654
  
SDV Build Fail , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/6434/



---


[jira] [Updated] (CARBONDATA-2896) Adaptive encoding for primitive data types

2018-08-27 Thread dhatchayani (JIRA)


 [ 
https://issues.apache.org/jira/browse/CARBONDATA-2896?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

dhatchayani updated CARBONDATA-2896:

Description: 
Currently Encoding and Decoding is present only for Dictionary, Measure 
Columns, but for no dictionary Primitive types encoding is *absent.*

*Encoding is a technique used to reduce the storage size and  after all these 
encoding, result will be compressed with snappy compression to further reduce 
the storage size.*

*With this feature, we support encoding on the no dictionary primitive data 
types also.*

> Adaptive encoding for primitive data types
> --
>
> Key: CARBONDATA-2896
> URL: https://issues.apache.org/jira/browse/CARBONDATA-2896
> Project: CarbonData
>  Issue Type: New Feature
>Reporter: dhatchayani
>Assignee: dhatchayani
>Priority: Major
>
> Currently Encoding and Decoding is present only for Dictionary, Measure 
> Columns, but for no dictionary Primitive types encoding is *absent.*
> *Encoding is a technique used to reduce the storage size and  after all these 
> encoding, result will be compressed with snappy compression to further reduce 
> the storage size.*
> *With this feature, we support encoding on the no dictionary primitive data 
> types also.*



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (CARBONDATA-2896) Adaptive encoding for primitive data types

2018-08-27 Thread dhatchayani (JIRA)


 [ 
https://issues.apache.org/jira/browse/CARBONDATA-2896?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

dhatchayani updated CARBONDATA-2896:

Description: 
Currently Encoding and Decoding is present only for Dictionary, Measure 
Columns, but for no dictionary Primitive types encoding is *absent.*

Encoding is a technique used to reduce the storage size and  after all these 
encoding, result will be compressed with snappy compression to further reduce 
the storage size.

With this feature, we support encoding on the no dictionary primitive data 
types also.

  was:
Currently Encoding and Decoding is present only for Dictionary, Measure 
Columns, but for no dictionary Primitive types encoding is *absent.*

*Encoding is a technique used to reduce the storage size and  after all these 
encoding, result will be compressed with snappy compression to further reduce 
the storage size.*

*With this feature, we support encoding on the no dictionary primitive data 
types also.*


> Adaptive encoding for primitive data types
> --
>
> Key: CARBONDATA-2896
> URL: https://issues.apache.org/jira/browse/CARBONDATA-2896
> Project: CarbonData
>  Issue Type: New Feature
>Reporter: dhatchayani
>Assignee: dhatchayani
>Priority: Major
>
> Currently Encoding and Decoding is present only for Dictionary, Measure 
> Columns, but for no dictionary Primitive types encoding is *absent.*
> Encoding is a technique used to reduce the storage size and  after all these 
> encoding, result will be compressed with snappy compression to further reduce 
> the storage size.
> With this feature, we support encoding on the no dictionary primitive data 
> types also.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] carbondata issue #2659: [CARBONDATA-2887] Fix complex filters on spark carbo...

2018-08-27 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2659
  
SDV Build Fail , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/6433/



---


[GitHub] carbondata issue #2628: [CARBONDATA-2851][CARBONDATA-2852] Support zstd as c...

2018-08-27 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2628
  
Build Failed with Spark 2.2.1, Please check CI 
http://95.216.28.178:8080/job/ApacheCarbonPRBuilder1/38/



---


[jira] [Created] (CARBONDATA-2896) Adaptive encoding for primitive data types

2018-08-27 Thread dhatchayani (JIRA)
dhatchayani created CARBONDATA-2896:
---

 Summary: Adaptive encoding for primitive data types
 Key: CARBONDATA-2896
 URL: https://issues.apache.org/jira/browse/CARBONDATA-2896
 Project: CarbonData
  Issue Type: New Feature
Reporter: dhatchayani
Assignee: dhatchayani






--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] carbondata issue #2623: [CARBONDATA-2844] [CARBONDATA-2865] Pass SK/AK to ex...

2018-08-27 Thread brijoobopanna
Github user brijoobopanna commented on the issue:

https://github.com/apache/carbondata/pull/2623
  
retest this please



---


[GitHub] carbondata issue #2628: [CARBONDATA-2851][CARBONDATA-2852] Support zstd as c...

2018-08-27 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2628
  
Build Failed  with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/8101/



---


[GitHub] carbondata issue #2628: [CARBONDATA-2851][CARBONDATA-2852] Support zstd as c...

2018-08-27 Thread xuchuanyin
Github user xuchuanyin commented on the issue:

https://github.com/apache/carbondata/pull/2628
  
retest this please


---


[GitHub] carbondata issue #2659: [CARBONDATA-2887] Fix complex filters on spark carbo...

2018-08-27 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2659
  
Build Failed with Spark 2.2.1, Please check CI 
http://95.216.28.178:8080/job/ApacheCarbonPRBuilder1/37/



---


[GitHub] carbondata issue #2659: [CARBONDATA-2887] Fix complex filters on spark carbo...

2018-08-27 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2659
  
Build Failed  with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/8100/



---


[GitHub] carbondata issue #2659: [CARBONDATA-2887] Fix complex filters on spark carbo...

2018-08-27 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2659
  
SDV Build Fail , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/6432/



---


[GitHub] carbondata issue #2659: [CARBONDATA-2887] Fix complex filters on spark carbo...

2018-08-27 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2659
  
SDV Build Fail , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/6431/



---


[GitHub] carbondata issue #2659: [CARBONDATA-2887] Fix complex filters on spark carbo...

2018-08-27 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2659
  
Build Failed with Spark 2.2.1, Please check CI 
http://95.216.28.178:8080/job/ApacheCarbonPRBuilder1/36/



---


[GitHub] carbondata issue #2628: [CARBONDATA-2851][CARBONDATA-2852] Support zstd as c...

2018-08-27 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2628
  
Build Failed with Spark 2.2.1, Please check CI 
http://95.216.28.178:8080/job/ApacheCarbonPRBuilder1/35/



---


[GitHub] carbondata issue #2659: [CARBONDATA-2887] Fix complex filters on spark carbo...

2018-08-27 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2659
  
SDV Build Fail , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/6430/



---


[GitHub] carbondata issue #2659: [CARBONDATA-2887] Fix complex filters on spark carbo...

2018-08-27 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2659
  
Build Failed  with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/8099/



---


[GitHub] carbondata issue #2659: [CARBONDATA-2887] Fix complex filters on spark carbo...

2018-08-27 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2659
  
Build Failed  with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/8098/



---


[GitHub] carbondata issue #2628: [CARBONDATA-2851][CARBONDATA-2852] Support zstd as c...

2018-08-27 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2628
  
Build Failed  with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/8097/



---


[GitHub] carbondata pull request #2662: [WIP][CARBONDATA-2889]Add decoder based fallb...

2018-08-27 Thread xuchuanyin
Github user xuchuanyin commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2662#discussion_r213161218
  
--- Diff: 
core/src/main/java/org/apache/carbondata/core/datastore/page/FallbackDecoderBasedColumnPageEncoder.java
 ---
@@ -0,0 +1,98 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements.  See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License.  You may obtain a copy of the License at
+ *
+ *http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+package org.apache.carbondata.core.datastore.page;
+
+import java.util.concurrent.Callable;
+
+import org.apache.carbondata.core.constants.CarbonCommonConstants;
+import org.apache.carbondata.core.datastore.TableSpec;
+import org.apache.carbondata.core.datastore.compression.CompressorFactory;
+import 
org.apache.carbondata.core.datastore.page.encoding.EncodedColumnPage;
+import org.apache.carbondata.core.keygenerator.KeyGenerator;
+import org.apache.carbondata.core.keygenerator.factory.KeyGeneratorFactory;
+import 
org.apache.carbondata.core.localdictionary.generator.LocalDictionaryGenerator;
+import org.apache.carbondata.core.metadata.datatype.DataType;
+import org.apache.carbondata.core.util.CarbonUtil;
+
+public class FallbackDecoderBasedColumnPageEncoder implements 
Callable {
+  /**
+   * actual local dictionary generated column page
+   */
+  private EncodedColumnPage encodedColumnPage;
+
+  /**
+   * actual index in the page
+   * this is required as in a blocklet few pages will be local dictionary
+   * encoded and few pages will be plain text encoding
+   * in this case local dictionary encoded page
+   */
+  private int pageIndex;
+
+  private LocalDictionaryGenerator localDictionaryGenerator;
+
+  public FallbackDecoderBasedColumnPageEncoder(EncodedColumnPage 
encodedColumnPage, int pageIndex,
+  LocalDictionaryGenerator localDictionaryGenerator) {
+this.encodedColumnPage = encodedColumnPage;
+this.pageIndex = pageIndex;
+this.localDictionaryGenerator = localDictionaryGenerator;
+  }
+
+  @Override public FallbackEncodedColumnPage call() throws Exception {
+
+// uncompress the encoded column page
+byte[] bytes = CompressorFactory.getInstance().getCompressor()
--- End diff --

emm, PR #2628 changed this. We should get the compressor from input 
configuration or from the metadata.


---


[GitHub] carbondata pull request #2662: [WIP][CARBONDATA-2889]Add decoder based fallb...

2018-08-27 Thread xuchuanyin
Github user xuchuanyin commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2662#discussion_r213160857
  
--- Diff: 
core/src/main/java/org/apache/carbondata/core/datastore/blocklet/EncodedBlocklet.java
 ---
@@ -87,19 +91,24 @@ private void addPageMetadata(EncodedTablePage 
encodedTablePage) {
* @param encodedTablePage
* encoded table page
*/
-  private void addEncodedMeasurePage(EncodedTablePage encodedTablePage) {
+  private void addEncodedMeasurePage(EncodedTablePage encodedTablePage,
+  Map localDictionaryGeneratorMap) {
 // for first page create new list
 if (null == encodedMeasureColumnPages) {
   encodedMeasureColumnPages = new ArrayList<>();
   // adding measure pages
   for (int i = 0; i < encodedTablePage.getNumMeasures(); i++) {
-BlockletEncodedColumnPage blockletEncodedColumnPage = new 
BlockletEncodedColumnPage(null);
-
blockletEncodedColumnPage.addEncodedColumnColumnPage(encodedTablePage.getMeasure(i));
+BlockletEncodedColumnPage blockletEncodedColumnPage = new 
BlockletEncodedColumnPage(null,
+Boolean.parseBoolean(CarbonProperties.getInstance()
--- End diff --

What if the configuration is changed during data loading? So that each 
column page will have different configuration, will this be OK?


---


[GitHub] carbondata issue #2628: [CARBONDATA-2851][CARBONDATA-2852] Support zstd as c...

2018-08-27 Thread xuchuanyin
Github user xuchuanyin commented on the issue:

https://github.com/apache/carbondata/pull/2628
  
retest this please


---


[GitHub] carbondata issue #2659: [CARBONDATA-2887] Fix complex filters on spark carbo...

2018-08-27 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2659
  
Build Failed with Spark 2.2.1, Please check CI 
http://95.216.28.178:8080/job/ApacheCarbonPRBuilder1/34/



---


[GitHub] carbondata issue #2664: [CARBONDATA-2895] Fix Query result mismatch with Bat...

2018-08-27 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2664
  
Build Success with Spark 2.2.1, Please check CI 
http://95.216.28.178:8080/job/ApacheCarbonPRBuilder1/33/



---


[GitHub] carbondata issue #2641: [CARBONDATA-2819] Fixed cannot drop preagg datamap o...

2018-08-27 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2641
  
Build Success with Spark 2.2.1, Please check CI 
http://95.216.28.178:8080/job/ApacheCarbonPRBuilder1/32/



---


[GitHub] carbondata issue #2663: [CARBONDATA-2894] Add support for complex map type t...

2018-08-27 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2663
  
Build Failed with Spark 2.2.1, Please check CI 
http://95.216.28.178:8080/job/ApacheCarbonPRBuilder1/31/



---


[GitHub] carbondata issue #2664: [CARBONDATA-2895] Fix Query result mismatch with Bat...

2018-08-27 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2664
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/8095/



---


[GitHub] carbondata issue #2659: [CARBONDATA-2887] Fix complex filters on spark carbo...

2018-08-27 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2659
  
SDV Build Fail , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/6429/



---


[GitHub] carbondata issue #2654: [WIP] Adaptive Encoding for Primitive data types

2018-08-27 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2654
  
Build Success with Spark 2.2.1, Please check CI 
http://95.216.28.178:8080/job/ApacheCarbonPRBuilder1/30/



---


[GitHub] carbondata issue #2659: [CARBONDATA-2887] Fix complex filters on spark carbo...

2018-08-27 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2659
  
Build Failed  with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/8096/



---


[GitHub] carbondata issue #2664: [CARBONDATA-2895] Fix Query result mismatch with Bat...

2018-08-27 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2664
  
SDV Build Fail , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/6428/



---


[GitHub] carbondata issue #2664: [CARBONDATA-2895] Fix Query result mismatch with Bat...

2018-08-27 Thread ajantha-bhat
Github user ajantha-bhat commented on the issue:

https://github.com/apache/carbondata/pull/2664
  
@jackylk , @ravipesala : Please review


---


[GitHub] carbondata pull request #2664: [CARBONDATA-2895] Fix Query result mismatch w...

2018-08-27 Thread ajantha-bhat
GitHub user ajantha-bhat opened a pull request:

https://github.com/apache/carbondata/pull/2664

[CARBONDATA-2895] Fix Query result mismatch with Batch-sort in save to disk 
(sort temp files) scenario.

**probelm:** Query result mismatch with Batch-sort in save to disk (sort 
temp files) scenario.

**scenario:**
a) Configure batchsort but give batch size more than 
UnsafeMemoryManager.INSTANCE.getUsableMemory().
b) Load data that is greater than batch size. Observe that 
unsafeMemoryManager save to disk happened as it cannot process one batch.  
c) so load happens in 2 batch. 
d) When query the results. There result data rows is more than expected 
data rows.


**root cause:**
For each batch, createSortDataRows() will be called.
Files saved to disk during sorting of previous batch was considered for 
this batch.


**solution:**
Files saved to disk during sorting of previous batch ,should not be 
considered for this batch.
Hence use batchID as rangeID field of sorttempfiles.
So getFilesToMergeSort() will select files of only this batch.

Be sure to do all of the following checklist to help us incorporate 
your contribution quickly and easily:

 - [ ] Any interfaces changed? NA
 
 - [ ] Any backward compatibility impacted? NA
 
 - [ ] Document update required? NA

 - [ ] Testing done. done
   
 - [ ] For large changes, please consider breaking it into sub-tasks under 
an umbrella JIRA.  NA



You can merge this pull request into a Git repository by running:

$ git pull https://github.com/ajantha-bhat/carbondata master_new

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/carbondata/pull/2664.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #2664


commit bad70a37508a2bad650aae2b150eecef59449a30
Author: ajantha-bhat 
Date:   2018-08-27T15:25:03Z

[CARBONDATA-2895] Fix Query result mismatch with Batch-sort in save to disk 
(sort temp files) scenario.

probelm: Query result mismatch with Batch-sort in save to disk (sort
temp files) scenario.

scenario:
a) Configure batchsort but give batch size more than
UnsafeMemoryManager.INSTANCE.getUsableMemory().
b) Load data that is greater than batch size. Observe that
unsafeMemoryManager save to disk happened as it cannot process one
batch.
c) so load happens in 2 batch.
d) When query the results. There result data rows is more than expected
data rows.

root cause:

For each batch, createSortDataRows() will be called.
Files saved to disk during sorting of previous batch was considered for
this batch.

solution:
Files saved to disk during sorting of previous batch ,should not be
considered for this batch.
Hence use batchID as rangeID field of sorttempfiles.
So getFilesToMergeSort() will select files of only this batch.




---


[jira] [Created] (CARBONDATA-2895) [Batch-sort]Query result mismatch with Batch-sort in save to disk (sort temp files) scenario.

2018-08-27 Thread Ajantha Bhat (JIRA)
Ajantha Bhat created CARBONDATA-2895:


 Summary: [Batch-sort]Query result mismatch with Batch-sort in save 
to disk (sort temp files) scenario.
 Key: CARBONDATA-2895
 URL: https://issues.apache.org/jira/browse/CARBONDATA-2895
 Project: CarbonData
  Issue Type: Bug
Reporter: Ajantha Bhat
Assignee: Ajantha Bhat


probelm: Query result mismatch with Batch-sort in save to disk (sort temp 
files) scenario.

scenario:
a) Configure batchsort but give batch size more than 
UnsafeMemoryManager.INSTANCE.getUsableMemory().
b) Load data that is greater than batch size. Observe that unsafeMemoryManager 
save to disk happened as it cannot process one batch.  
c) so load happens in 2 batch. 
d) When query the results. There result data rows is more than expected data 
rows.


root cause:

For each batch, createSortDataRows() will be called.
Files saved to disk during sorting of previous batch was considered for this 
batch.


solution:
Files saved to disk during sorting of previous batch ,should not be considered 
for this batch.
Hence use batchID as rangeID field of sorttempfiles.
So getFilesToMergeSort() will select files of only this batch.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] carbondata issue #2663: [CARBONDATA-2894] Add support for complex map type t...

2018-08-27 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2663
  
Build Failed  with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/8094/



---


[GitHub] carbondata issue #2628: [CARBONDATA-2851][CARBONDATA-2852] Support zstd as c...

2018-08-27 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2628
  
Build Failed with Spark 2.2.1, Please check CI 
http://95.216.28.178:8080/job/ApacheCarbonPRBuilder1/29/



---


[GitHub] carbondata issue #2641: [CARBONDATA-2819] Fixed cannot drop preagg datamap o...

2018-08-27 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2641
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/8092/



---


[GitHub] carbondata issue #2663: [CARBONDATA-2894] Add support for complex map type t...

2018-08-27 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2663
  
SDV Build Fail , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/6427/



---


[GitHub] carbondata issue #2623: [CARBONDATA-2844] [CARBONDATA-2865] Pass SK/AK to ex...

2018-08-27 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2623
  
Build Success with Spark 2.2.1, Please check CI 
http://95.216.28.178:8080/job/ApacheCarbonPRBuilder1/28/



---


[GitHub] carbondata issue #2663: [CARBONDATA-2894] Add support for complex map type t...

2018-08-27 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2663
  
SDV Build Fail , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/6426/



---


[GitHub] carbondata issue #2654: [WIP] Adaptive Encoding for Primitive data types

2018-08-27 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2654
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/8091/



---


[GitHub] carbondata pull request #2663: [CARBONDATA-2894] Add support for complex map...

2018-08-27 Thread manishgupta88
GitHub user manishgupta88 opened a pull request:

https://github.com/apache/carbondata/pull/2663

[CARBONDATA-2894] Add support for complex map type through spark carbon 
file format API

This PR supports loading querying complex map type through spark carbon 
file format API.

**Note: This PR is dependent on PR #2649** 

 - [ ] Any interfaces changed?
 No
 - [ ] Any backward compatibility impacted?
 No
 - [ ] Document update required?
No
 - [ ] Testing done
Added test cases   
 - [ ] For large changes, please consider breaking it into sub-tasks under 
an umbrella JIRA. 
NA


You can merge this pull request into a Git repository by running:

$ git pull https://github.com/manishgupta88/carbondata 
map_spark_carbon_file_support

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/carbondata/pull/2663.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #2663


commit e67bd0cb485c4bed15ce8ac3ef3be9b3a4f3798e
Author: manishgupta88 
Date:   2018-08-20T04:59:29Z

Added support for Avro Map type using SDK

commit 6db7f2a0d7c02406e0ecc9aa7ac69e2ec2e540a6
Author: manishgupta88 
Date:   2018-08-27T13:47:21Z

Add support for complex map type using spark carbon file format API




---


[jira] [Created] (CARBONDATA-2894) Add support for complex map type through spark carbon file format API

2018-08-27 Thread Manish Gupta (JIRA)
Manish Gupta created CARBONDATA-2894:


 Summary: Add support for complex map type through spark carbon 
file format API
 Key: CARBONDATA-2894
 URL: https://issues.apache.org/jira/browse/CARBONDATA-2894
 Project: CarbonData
  Issue Type: Sub-task
Reporter: Manish Gupta
Assignee: Manish Gupta






--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (CARBONDATA-2892) Data mismatch is seen in the Array-String and Array-Timestamp.

2018-08-27 Thread Prasanna Ravichandran (JIRA)


 [ 
https://issues.apache.org/jira/browse/CARBONDATA-2892?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanna Ravichandran updated CARBONDATA-2892:
--
Attachment: (was: Array.csv)

> Data mismatch is seen in the Array-String and Array-Timestamp.
> --
>
> Key: CARBONDATA-2892
> URL: https://issues.apache.org/jira/browse/CARBONDATA-2892
> Project: CarbonData
>  Issue Type: Bug
> Environment: 3 Node ANT.
>Reporter: Prasanna Ravichandran
>Priority: Major
> Attachments: Array.csv
>
>
> Data mismatch is seen in the Array-String and Array-Timestamp like mismatch 
> in data, order, date values. 
> *Test queries:*
> drop table if exists array_com_hive;
> create table array_com_hive (CUST_ID string, YEAR int, MONTH int, AGE int, 
> GENDER string, EDUCATED string, IS_MARRIED string, ARRAY_INT 
> array,ARRAY_STRING array,ARRAY_DATE array,CARD_COUNT 
> int,DEBIT_COUNT int, CREDIT_COUNT int, DEPOSIT double, HQ_DEPOSIT double) row 
> format delimited fields terminated by ',' collection items terminated by '$';
> load data local inpath '/opt/csv/complex/Array.csv' into table array_com_hive;
> drop table if exists array_com;
> create table Array_com (CUST_ID string, YEAR int, MONTH int, AGE int, GENDER 
> string, EDUCATED string, IS_MARRIED string, ARRAY_INT array,ARRAY_STRING 
> array,ARRAY_DATE array,CARD_COUNT int,DEBIT_COUNT int, 
> CREDIT_COUNT int, DEPOSIT double, HQ_DEPOSIT double) using carbon;
> insert into Array_com select * from array_com_hive;
> select * from array_com_hive order by CUST_ID ASC limit 3;
> select * from array_com order by CUST_ID ASC limit 3;
> *Expected result:*
> There should be no data mismatch and data in table should be same as it is in 
> CSV file.
> *Actual result:*
> Data mismatch is seen.
>  
>  
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (CARBONDATA-2892) Data mismatch is seen in the Array-String and Array-Timestamp.

2018-08-27 Thread Prasanna Ravichandran (JIRA)


 [ 
https://issues.apache.org/jira/browse/CARBONDATA-2892?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanna Ravichandran updated CARBONDATA-2892:
--
Attachment: Array.csv

> Data mismatch is seen in the Array-String and Array-Timestamp.
> --
>
> Key: CARBONDATA-2892
> URL: https://issues.apache.org/jira/browse/CARBONDATA-2892
> Project: CarbonData
>  Issue Type: Bug
> Environment: 3 Node ANT.
>Reporter: Prasanna Ravichandran
>Priority: Major
> Attachments: Array.csv
>
>
> Data mismatch is seen in the Array-String and Array-Timestamp like mismatch 
> in data, order, date values. 
> *Test queries:*
> drop table if exists array_com_hive;
> create table array_com_hive (CUST_ID string, YEAR int, MONTH int, AGE int, 
> GENDER string, EDUCATED string, IS_MARRIED string, ARRAY_INT 
> array,ARRAY_STRING array,ARRAY_DATE array,CARD_COUNT 
> int,DEBIT_COUNT int, CREDIT_COUNT int, DEPOSIT double, HQ_DEPOSIT double) row 
> format delimited fields terminated by ',' collection items terminated by '$';
> load data local inpath '/opt/csv/complex/Array.csv' into table array_com_hive;
> drop table if exists array_com;
> create table Array_com (CUST_ID string, YEAR int, MONTH int, AGE int, GENDER 
> string, EDUCATED string, IS_MARRIED string, ARRAY_INT array,ARRAY_STRING 
> array,ARRAY_DATE array,CARD_COUNT int,DEBIT_COUNT int, 
> CREDIT_COUNT int, DEPOSIT double, HQ_DEPOSIT double) using carbon;
> insert into Array_com select * from array_com_hive;
> select * from array_com_hive order by CUST_ID ASC limit 3;
> select * from array_com order by CUST_ID ASC limit 3;
> *Expected result:*
> There should be no data mismatch and data in table should be same as it is in 
> CSV file.
> *Actual result:*
> Data mismatch is seen.
>  
>  
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (CARBONDATA-2893) Job aborted during insert while loading the "Struct of Array" datatype values.

2018-08-27 Thread Prasanna Ravichandran (JIRA)


 [ 
https://issues.apache.org/jira/browse/CARBONDATA-2893?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanna Ravichandran updated CARBONDATA-2893:
--
Attachment: structofarray.csv

> Job aborted during insert while loading the "Struct of Array" datatype values.
> --
>
> Key: CARBONDATA-2893
> URL: https://issues.apache.org/jira/browse/CARBONDATA-2893
> Project: CarbonData
>  Issue Type: Bug
> Environment: 3 Node ANT.
>Reporter: Prasanna Ravichandran
>Priority: Major
> Attachments: structofarray.csv
>
>
> Job aborted during insert while loading the "Struct of Array" datatype values.
> *Test queries:*
> 0: jdbc:hive2:> drop table if exists STRUCT_OF_ARRAY_com_hive;
>  +--+-+
> |Result|
> +--+-+
>  +--+-+
>  No rows selected (0.026 seconds)
>  0: jdbc:hive2:> create table STRUCT_OF_ARRAY_com_hive (CUST_ID string, YEAR 
> int, MONTH int, AGE int, GENDER string, EDUCATED string, IS_MARRIED string, 
> STRUCT_OF_ARRAY struct,sal1: 
> array,state: array,date1: array>,CARD_COUNT 
> int,DEBIT_COUNT int, CREDIT_COUNT int, DEPOSIT float, HQ_DEPOSIT double) row 
> format delimited fields terminated by ',' collection items terminated by '$' 
> map keys terminated by '&';
>  +--+-+
> |Result|
> +--+-+
>  +--+-+
>  No rows selected (0.159 seconds)
>  0: jdbc:hive2:> load data local inpath '/opt/csv/complex/structofarray.csv' 
> into table STRUCT_OF_ARRAY_com_hive;
>  +--+-+
> |Result|
> +--+-+
>  +--+-+
>  No rows selected (0.217 seconds)
>  0: jdbc:hive2:> drop table if exists STRUCT_OF_ARRAY_com;
>  +--+-+
> |Result|
> +--+-+
>  +--+-+
>  No rows selected (0.03 seconds)
>  0: jdbc:hive2:> create table STRUCT_OF_ARRAY_com (CUST_ID string, YEAR int, 
> MONTH int, AGE int, GENDER string, EDUCATED string, IS_MARRIED string, 
> STRUCT_OF_ARRAY struct,sal1: 
> array,state: array,date1: array>,CARD_COUNT 
> int,DEBIT_COUNT int, CREDIT_COUNT int, DEPOSIT double, HQ_DEPOSIT double) 
> using carbon;
>  +--+-+
> |Result|
> +--+-+
>  +--+-+
>  No rows selected (0.099 seconds)
>  0: jdbc:hive2:> insert into STRUCT_OF_ARRAY_com select * from 
> STRUCT_OF_ARRAY_com_hive;
>  *Error: org.apache.spark.SparkException: Job aborted. (state=,code=0)*
>  
>  *Expected result:*
> Insert should be success.
> *Actual result:*
> Insert is showing job aborted.
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (CARBONDATA-2893) Job aborted during insert while loading the "Struct of Array" datatype values.

2018-08-27 Thread Prasanna Ravichandran (JIRA)


 [ 
https://issues.apache.org/jira/browse/CARBONDATA-2893?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanna Ravichandran updated CARBONDATA-2893:
--
Attachment: (was: structofarray.csv)

> Job aborted during insert while loading the "Struct of Array" datatype values.
> --
>
> Key: CARBONDATA-2893
> URL: https://issues.apache.org/jira/browse/CARBONDATA-2893
> Project: CarbonData
>  Issue Type: Bug
> Environment: 3 Node ANT.
>Reporter: Prasanna Ravichandran
>Priority: Major
> Attachments: structofarray.csv
>
>
> Job aborted during insert while loading the "Struct of Array" datatype values.
> *Test queries:*
> 0: jdbc:hive2:> drop table if exists STRUCT_OF_ARRAY_com_hive;
>  +--+-+
> |Result|
> +--+-+
>  +--+-+
>  No rows selected (0.026 seconds)
>  0: jdbc:hive2:> create table STRUCT_OF_ARRAY_com_hive (CUST_ID string, YEAR 
> int, MONTH int, AGE int, GENDER string, EDUCATED string, IS_MARRIED string, 
> STRUCT_OF_ARRAY struct,sal1: 
> array,state: array,date1: array>,CARD_COUNT 
> int,DEBIT_COUNT int, CREDIT_COUNT int, DEPOSIT float, HQ_DEPOSIT double) row 
> format delimited fields terminated by ',' collection items terminated by '$' 
> map keys terminated by '&';
>  +--+-+
> |Result|
> +--+-+
>  +--+-+
>  No rows selected (0.159 seconds)
>  0: jdbc:hive2:> load data local inpath '/opt/csv/complex/structofarray.csv' 
> into table STRUCT_OF_ARRAY_com_hive;
>  +--+-+
> |Result|
> +--+-+
>  +--+-+
>  No rows selected (0.217 seconds)
>  0: jdbc:hive2:> drop table if exists STRUCT_OF_ARRAY_com;
>  +--+-+
> |Result|
> +--+-+
>  +--+-+
>  No rows selected (0.03 seconds)
>  0: jdbc:hive2:> create table STRUCT_OF_ARRAY_com (CUST_ID string, YEAR int, 
> MONTH int, AGE int, GENDER string, EDUCATED string, IS_MARRIED string, 
> STRUCT_OF_ARRAY struct,sal1: 
> array,state: array,date1: array>,CARD_COUNT 
> int,DEBIT_COUNT int, CREDIT_COUNT int, DEPOSIT double, HQ_DEPOSIT double) 
> using carbon;
>  +--+-+
> |Result|
> +--+-+
>  +--+-+
>  No rows selected (0.099 seconds)
>  0: jdbc:hive2:> insert into STRUCT_OF_ARRAY_com select * from 
> STRUCT_OF_ARRAY_com_hive;
>  *Error: org.apache.spark.SparkException: Job aborted. (state=,code=0)*
>  
>  *Expected result:*
> Insert should be success.
> *Actual result:*
> Insert is showing job aborted.
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (CARBONDATA-2893) Job aborted during insert while loading the "Struct of Array" datatype values.

2018-08-27 Thread Prasanna Ravichandran (JIRA)


 [ 
https://issues.apache.org/jira/browse/CARBONDATA-2893?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanna Ravichandran updated CARBONDATA-2893:
--
Attachment: structofarray.csv

> Job aborted during insert while loading the "Struct of Array" datatype values.
> --
>
> Key: CARBONDATA-2893
> URL: https://issues.apache.org/jira/browse/CARBONDATA-2893
> Project: CarbonData
>  Issue Type: Bug
> Environment: 3 Node ANT.
>Reporter: Prasanna Ravichandran
>Priority: Major
> Attachments: structofarray.csv
>
>
> Job aborted during insert while loading the "Struct of Array" datatype values.
> *Test queries:*
> 0: jdbc:hive2:> drop table if exists STRUCT_OF_ARRAY_com_hive;
>  +--+-+
> |Result|
> +--+-+
>  +--+-+
>  No rows selected (0.026 seconds)
>  0: jdbc:hive2:> create table STRUCT_OF_ARRAY_com_hive (CUST_ID string, YEAR 
> int, MONTH int, AGE int, GENDER string, EDUCATED string, IS_MARRIED string, 
> STRUCT_OF_ARRAY struct,sal1: 
> array,state: array,date1: array>,CARD_COUNT 
> int,DEBIT_COUNT int, CREDIT_COUNT int, DEPOSIT float, HQ_DEPOSIT double) row 
> format delimited fields terminated by ',' collection items terminated by '$' 
> map keys terminated by '&';
>  +--+-+
> |Result|
> +--+-+
>  +--+-+
>  No rows selected (0.159 seconds)
>  0: jdbc:hive2:> load data local inpath '/opt/csv/complex/structofarray.csv' 
> into table STRUCT_OF_ARRAY_com_hive;
>  +--+-+
> |Result|
> +--+-+
>  +--+-+
>  No rows selected (0.217 seconds)
>  0: jdbc:hive2:> drop table if exists STRUCT_OF_ARRAY_com;
>  +--+-+
> |Result|
> +--+-+
>  +--+-+
>  No rows selected (0.03 seconds)
>  0: jdbc:hive2:> create table STRUCT_OF_ARRAY_com (CUST_ID string, YEAR int, 
> MONTH int, AGE int, GENDER string, EDUCATED string, IS_MARRIED string, 
> STRUCT_OF_ARRAY struct,sal1: 
> array,state: array,date1: array>,CARD_COUNT 
> int,DEBIT_COUNT int, CREDIT_COUNT int, DEPOSIT double, HQ_DEPOSIT double) 
> using carbon;
>  +--+-+
> |Result|
> +--+-+
>  +--+-+
>  No rows selected (0.099 seconds)
>  0: jdbc:hive2:> insert into STRUCT_OF_ARRAY_com select * from 
> STRUCT_OF_ARRAY_com_hive;
>  *Error: org.apache.spark.SparkException: Job aborted. (state=,code=0)*
>  
>  *Expected result:*
> Insert should be success.
> *Actual result:*
> Insert is showing job aborted.
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (CARBONDATA-2893) Job aborted during insert while loading the "Struct of Array" datatype values.

2018-08-27 Thread Prasanna Ravichandran (JIRA)


 [ 
https://issues.apache.org/jira/browse/CARBONDATA-2893?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanna Ravichandran updated CARBONDATA-2893:
--
Attachment: (was: arrayofstruct.csv)

> Job aborted during insert while loading the "Struct of Array" datatype values.
> --
>
> Key: CARBONDATA-2893
> URL: https://issues.apache.org/jira/browse/CARBONDATA-2893
> Project: CarbonData
>  Issue Type: Bug
> Environment: 3 Node ANT.
>Reporter: Prasanna Ravichandran
>Priority: Major
> Attachments: structofarray.csv
>
>
> Job aborted during insert while loading the "Struct of Array" datatype values.
> *Test queries:*
> 0: jdbc:hive2:> drop table if exists STRUCT_OF_ARRAY_com_hive;
>  +--+-+
> |Result|
> +--+-+
>  +--+-+
>  No rows selected (0.026 seconds)
>  0: jdbc:hive2:> create table STRUCT_OF_ARRAY_com_hive (CUST_ID string, YEAR 
> int, MONTH int, AGE int, GENDER string, EDUCATED string, IS_MARRIED string, 
> STRUCT_OF_ARRAY struct,sal1: 
> array,state: array,date1: array>,CARD_COUNT 
> int,DEBIT_COUNT int, CREDIT_COUNT int, DEPOSIT float, HQ_DEPOSIT double) row 
> format delimited fields terminated by ',' collection items terminated by '$' 
> map keys terminated by '&';
>  +--+-+
> |Result|
> +--+-+
>  +--+-+
>  No rows selected (0.159 seconds)
>  0: jdbc:hive2:> load data local inpath '/opt/csv/complex/structofarray.csv' 
> into table STRUCT_OF_ARRAY_com_hive;
>  +--+-+
> |Result|
> +--+-+
>  +--+-+
>  No rows selected (0.217 seconds)
>  0: jdbc:hive2:> drop table if exists STRUCT_OF_ARRAY_com;
>  +--+-+
> |Result|
> +--+-+
>  +--+-+
>  No rows selected (0.03 seconds)
>  0: jdbc:hive2:> create table STRUCT_OF_ARRAY_com (CUST_ID string, YEAR int, 
> MONTH int, AGE int, GENDER string, EDUCATED string, IS_MARRIED string, 
> STRUCT_OF_ARRAY struct,sal1: 
> array,state: array,date1: array>,CARD_COUNT 
> int,DEBIT_COUNT int, CREDIT_COUNT int, DEPOSIT double, HQ_DEPOSIT double) 
> using carbon;
>  +--+-+
> |Result|
> +--+-+
>  +--+-+
>  No rows selected (0.099 seconds)
>  0: jdbc:hive2:> insert into STRUCT_OF_ARRAY_com select * from 
> STRUCT_OF_ARRAY_com_hive;
>  *Error: org.apache.spark.SparkException: Job aborted. (state=,code=0)*
>  
>  *Expected result:*
> Insert should be success.
> *Actual result:*
> Insert is showing job aborted.
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (CARBONDATA-2893) Job aborted during insert while loading the "Struct of Array" datatype values.

2018-08-27 Thread Prasanna Ravichandran (JIRA)


 [ 
https://issues.apache.org/jira/browse/CARBONDATA-2893?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanna Ravichandran updated CARBONDATA-2893:
--
Description: 
Job aborted during insert while loading the "Struct of Array" datatype values.

*Test queries:*

0: jdbc:hive2:> drop table if exists STRUCT_OF_ARRAY_com_hive;
 +--+-+
|Result|

+--+-+
 +--+-+
 No rows selected (0.026 seconds)
 0: jdbc:hive2:> create table STRUCT_OF_ARRAY_com_hive (CUST_ID string, YEAR 
int, MONTH int, AGE int, GENDER string, EDUCATED string, IS_MARRIED string, 
STRUCT_OF_ARRAY struct,sal1: 
array,state: array,date1: array>,CARD_COUNT 
int,DEBIT_COUNT int, CREDIT_COUNT int, DEPOSIT float, HQ_DEPOSIT double) row 
format delimited fields terminated by ',' collection items terminated by '$' 
map keys terminated by '&';
 +--+-+
|Result|

+--+-+
 +--+-+
 No rows selected (0.159 seconds)
 0: jdbc:hive2:> load data local inpath '/opt/csv/complex/structofarray.csv' 
into table STRUCT_OF_ARRAY_com_hive;
 +--+-+
|Result|

+--+-+
 +--+-+
 No rows selected (0.217 seconds)
 0: jdbc:hive2:> drop table if exists STRUCT_OF_ARRAY_com;
 +--+-+
|Result|

+--+-+
 +--+-+
 No rows selected (0.03 seconds)
 0: jdbc:hive2:> create table STRUCT_OF_ARRAY_com (CUST_ID string, YEAR int, 
MONTH int, AGE int, GENDER string, EDUCATED string, IS_MARRIED string, 
STRUCT_OF_ARRAY struct,sal1: 
array,state: array,date1: array>,CARD_COUNT 
int,DEBIT_COUNT int, CREDIT_COUNT int, DEPOSIT double, HQ_DEPOSIT double) using 
carbon;
 +--+-+
|Result|

+--+-+
 +--+-+
 No rows selected (0.099 seconds)
 0: jdbc:hive2:> insert into STRUCT_OF_ARRAY_com select * from 
STRUCT_OF_ARRAY_com_hive;
 *Error: org.apache.spark.SparkException: Job aborted. (state=,code=0)*

 

 *Expected result:*

Insert should be success.

*Actual result:*

Insert is showing job aborted.

 

  was:
Job aborted during insert while loading the "Struct of Array" datatype values.

*Test queries:*

0: jdbc:hive2:> drop table if exists STRUCT_OF_ARRAY_com_hive;
+-+--+
| Result |
+-+--+
+-+--+
No rows selected (0.026 seconds)
0: jdbc:hive2:> create table STRUCT_OF_ARRAY_com_hive (CUST_ID string, YEAR 
int, MONTH int, AGE int, GENDER string, EDUCATED string, IS_MARRIED string, 
STRUCT_OF_ARRAY struct,sal1: 
array,state: array,date1: array>,CARD_COUNT 
int,DEBIT_COUNT int, CREDIT_COUNT int, DEPOSIT float, HQ_DEPOSIT double) row 
format delimited fields terminated by ',' collection items terminated by '$' 
map keys terminated by '&';
+-+--+
| Result |
+-+--+
+-+--+
No rows selected (0.159 seconds)
0: jdbc:hive2:> load data local inpath '/opt/csv/complex/structofarray.csv' 
into table STRUCT_OF_ARRAY_com_hive;
+-+--+
| Result |
+-+--+
+-+--+
No rows selected (0.217 seconds)
0: jdbc:hive2:> drop table if exists STRUCT_OF_ARRAY_com;
+-+--+
| Result |
+-+--+
+-+--+
No rows selected (0.03 seconds)
0: jdbc:hive2:> create table STRUCT_OF_ARRAY_com (CUST_ID string, YEAR int, 
MONTH int, AGE int, GENDER string, EDUCATED string, IS_MARRIED string, 
STRUCT_OF_ARRAY struct,sal1: 
array,state: array,date1: array>,CARD_COUNT 
int,DEBIT_COUNT int, CREDIT_COUNT int, DEPOSIT double, HQ_DEPOSIT double) using 
carbon;
+-+--+
| Result |
+-+--+
+-+--+
No rows selected (0.099 seconds)
0: jdbc:hive2:> insert into STRUCT_OF_ARRAY_com select * from 
STRUCT_OF_ARRAY_com_hive;
*Error: org.apache.spark.SparkException: Job aborted. (state=,code=0)*

 

 


> Job aborted during insert while loading the "Struct of Array" datatype values.
> --
>
> Key: CARBONDATA-2893
> URL: https://issues.apache.org/jira/browse/CARBONDATA-2893
> Project: CarbonData
>  Issue Type: Bug
> Environment: 3 Node ANT.
>Reporter: Prasanna Ravichandran
>Priority: Major
> Attachments: arrayofstruct.csv
>
>
> Job aborted during insert while loading the "Struct of Array" datatype values.
> *Test queries:*
> 0: jdbc:hive2:> drop table if exists STRUCT_OF_ARRAY_com_hive;
>  +--+-+
> |Result|
> +--+-+
>  +--+-+
>  No rows selected (0.026 seconds)
>  0: jdbc:hive2:> create table STRUCT_OF_ARRAY_com_hive (CUST_ID string, YEAR 
> int, MONTH int, AGE int, GENDER string, EDUCATED string, IS_MARRIED string, 
> STRUCT_OF_ARRAY struct,sal1: 
> array,state: array,date1: array>,CARD_COUNT 
> int,DEBIT_COUNT int, CREDIT_COUNT int, DEPOSIT float, HQ_DEPOSIT double) row 
> format delimited fields terminated by ',' collection items terminated by '$' 
> map keys terminated by '&';
>  +--+-+
> |Result|
> +--+-+
>  +--+-+
>  No rows selected (0.159 seconds)
>  0: 

[jira] [Created] (CARBONDATA-2893) Job aborted during insert while loading the "Struct of Array" datatype values.

2018-08-27 Thread Prasanna Ravichandran (JIRA)
Prasanna Ravichandran created CARBONDATA-2893:
-

 Summary: Job aborted during insert while loading the "Struct of 
Array" datatype values.
 Key: CARBONDATA-2893
 URL: https://issues.apache.org/jira/browse/CARBONDATA-2893
 Project: CarbonData
  Issue Type: Bug
 Environment: 3 Node ANT.
Reporter: Prasanna Ravichandran
 Attachments: arrayofstruct.csv

Job aborted during insert while loading the "Struct of Array" datatype values.

*Test queries:*

0: jdbc:hive2:> drop table if exists STRUCT_OF_ARRAY_com_hive;
+-+--+
| Result |
+-+--+
+-+--+
No rows selected (0.026 seconds)
0: jdbc:hive2:> create table STRUCT_OF_ARRAY_com_hive (CUST_ID string, YEAR 
int, MONTH int, AGE int, GENDER string, EDUCATED string, IS_MARRIED string, 
STRUCT_OF_ARRAY struct,sal1: 
array,state: array,date1: array>,CARD_COUNT 
int,DEBIT_COUNT int, CREDIT_COUNT int, DEPOSIT float, HQ_DEPOSIT double) row 
format delimited fields terminated by ',' collection items terminated by '$' 
map keys terminated by '&';
+-+--+
| Result |
+-+--+
+-+--+
No rows selected (0.159 seconds)
0: jdbc:hive2:> load data local inpath '/opt/csv/complex/structofarray.csv' 
into table STRUCT_OF_ARRAY_com_hive;
+-+--+
| Result |
+-+--+
+-+--+
No rows selected (0.217 seconds)
0: jdbc:hive2:> drop table if exists STRUCT_OF_ARRAY_com;
+-+--+
| Result |
+-+--+
+-+--+
No rows selected (0.03 seconds)
0: jdbc:hive2:> create table STRUCT_OF_ARRAY_com (CUST_ID string, YEAR int, 
MONTH int, AGE int, GENDER string, EDUCATED string, IS_MARRIED string, 
STRUCT_OF_ARRAY struct,sal1: 
array,state: array,date1: array>,CARD_COUNT 
int,DEBIT_COUNT int, CREDIT_COUNT int, DEPOSIT double, HQ_DEPOSIT double) using 
carbon;
+-+--+
| Result |
+-+--+
+-+--+
No rows selected (0.099 seconds)
0: jdbc:hive2:> insert into STRUCT_OF_ARRAY_com select * from 
STRUCT_OF_ARRAY_com_hive;
*Error: org.apache.spark.SparkException: Job aborted. (state=,code=0)*

 

 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] carbondata issue #2641: [CARBONDATA-2819] Fixed cannot drop preagg datamap o...

2018-08-27 Thread jackylk
Github user jackylk commented on the issue:

https://github.com/apache/carbondata/pull/2641
  
retest this please


---


[GitHub] carbondata issue #2654: [WIP] Adaptive Encoding for Primitive data types

2018-08-27 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2654
  
Build Success with Spark 2.2.1, Please check CI 
http://95.216.28.178:8080/job/ApacheCarbonPRBuilder1/27/



---


[jira] [Created] (CARBONDATA-2892) Data mismatch is seen in the Array-String and Array-Timestamp.

2018-08-27 Thread Prasanna Ravichandran (JIRA)
Prasanna Ravichandran created CARBONDATA-2892:
-

 Summary: Data mismatch is seen in the Array-String and 
Array-Timestamp.
 Key: CARBONDATA-2892
 URL: https://issues.apache.org/jira/browse/CARBONDATA-2892
 Project: CarbonData
  Issue Type: Bug
 Environment: 3 Node ANT.
Reporter: Prasanna Ravichandran
 Attachments: Array.csv

Data mismatch is seen in the Array-String and Array-Timestamp like mismatch in 
data, order, date values. 

*Test queries:*

drop table if exists array_com_hive;
create table array_com_hive (CUST_ID string, YEAR int, MONTH int, AGE int, 
GENDER string, EDUCATED string, IS_MARRIED string, ARRAY_INT 
array,ARRAY_STRING array,ARRAY_DATE array,CARD_COUNT 
int,DEBIT_COUNT int, CREDIT_COUNT int, DEPOSIT double, HQ_DEPOSIT double) row 
format delimited fields terminated by ',' collection items terminated by '$';
load data local inpath '/opt/csv/complex/Array.csv' into table array_com_hive;
drop table if exists array_com;
create table Array_com (CUST_ID string, YEAR int, MONTH int, AGE int, GENDER 
string, EDUCATED string, IS_MARRIED string, ARRAY_INT array,ARRAY_STRING 
array,ARRAY_DATE array,CARD_COUNT int,DEBIT_COUNT int, 
CREDIT_COUNT int, DEPOSIT double, HQ_DEPOSIT double) using carbon;
insert into Array_com select * from array_com_hive;
select * from array_com_hive order by CUST_ID ASC limit 3;
select * from array_com order by CUST_ID ASC limit 3;

*Expected result:*

There should be no data mismatch and data in table should be same as it is in 
CSV file.

*Actual result:*

Data mismatch is seen.

 

 

 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] carbondata issue #2654: [WIP] Adaptive Encoding for Primitive data types

2018-08-27 Thread dhatchayani
Github user dhatchayani commented on the issue:

https://github.com/apache/carbondata/pull/2654
  
retest this please


---


[GitHub] carbondata issue #2623: [CARBONDATA-2844] [CARBONDATA-2865] Pass SK/AK to ex...

2018-08-27 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2623
  
Build Failed  with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/8089/



---


[jira] [Created] (CARBONDATA-2891) Job aborted while loading long string 32k data into carbon table from hive

2018-08-27 Thread Rahul Singha (JIRA)
Rahul Singha created CARBONDATA-2891:


 Summary: Job aborted while loading long string 32k data into 
carbon table from hive
 Key: CARBONDATA-2891
 URL: https://issues.apache.org/jira/browse/CARBONDATA-2891
 Project: CarbonData
  Issue Type: Bug
  Components: data-load
Reporter: Rahul Singha


_*Steps:*_

 CREATE TABLE local1(id int, name string, description string,address string, 
note string) using carbon options('long_string_columns'='description,note');

CREATE TABLE local_hive(id int, name string, description string,address string, 
note string) ROW FORMAT DELIMITED FIELDS TERMINATED BY ',';

LOAD DATA local INPATH '/opt/csv/longStringData_100rec.csv' overwrite into 
table local_hive;

insert into local1 select * from local_hive;

_*Expected result:*_

Data should get loaded.

_*Actual Result:*_

Error: org.apache.spark.SparkException: Job aborted. (state=,code=0)



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] carbondata issue #2659: [CARBONDATA-2887] Fix complex filters on spark carbo...

2018-08-27 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2659
  
Build Failed with Spark 2.2.1, Please check CI 
http://95.216.28.178:8080/job/ApacheCarbonPRBuilder1/26/



---


[GitHub] carbondata issue #2654: [WIP] Adaptive Encoding for Primitive data types

2018-08-27 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2654
  
Build Failed  with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/8088/



---


[GitHub] carbondata issue #2628: [CARBONDATA-2851][CARBONDATA-2852] Support zstd as c...

2018-08-27 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2628
  
Build Failed  with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/8090/



---


[GitHub] carbondata issue #2628: [CARBONDATA-2851][CARBONDATA-2852] Support zstd as c...

2018-08-27 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2628
  
SDV Build Fail , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/6425/



---


[GitHub] carbondata issue #2659: [CARBONDATA-2887] Fix complex filters on spark carbo...

2018-08-27 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2659
  
Build Failed  with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/8087/



---


[GitHub] carbondata issue #2623: [CARBONDATA-2844] [CARBONDATA-2865] Pass SK/AK to ex...

2018-08-27 Thread kunal642
Github user kunal642 commented on the issue:

https://github.com/apache/carbondata/pull/2623
  
retest this please


---


[GitHub] carbondata issue #2623: [CARBONDATA-2844] [CARBONDATA-2865] Pass SK/AK to ex...

2018-08-27 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2623
  
Build Success with Spark 2.2.1, Please check CI 
http://95.216.28.178:8080/job/ApacheCarbonPRBuilder1/25/



---


[GitHub] carbondata issue #2654: [WIP] Adaptive Encoding for Primitive data types

2018-08-27 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2654
  
SDV Build Fail , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/6424/



---


[GitHub] carbondata issue #2659: [CARBONDATA-2887] Fix complex filters on spark carbo...

2018-08-27 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2659
  
SDV Build Fail , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/6423/



---


[GitHub] carbondata issue #2635: [CARBONDATA-2856][BloomDataMap] Fix bug in bloom ind...

2018-08-27 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2635
  
Build Success with Spark 2.2.1, Please check CI 
http://95.216.28.178:8080/job/ApacheCarbonPRBuilder1/24/



---


[jira] [Created] (CARBONDATA-2890) Use CarbonLoadModelBuilder instead of new CarbonLoadModel instance

2018-08-27 Thread xuchuanyin (JIRA)
xuchuanyin created CARBONDATA-2890:
--

 Summary: Use CarbonLoadModelBuilder instead of new CarbonLoadModel 
instance
 Key: CARBONDATA-2890
 URL: https://issues.apache.org/jira/browse/CARBONDATA-2890
 Project: CarbonData
  Issue Type: Sub-task
Reporter: xuchuanyin


Currently to get an instance of CarbonLoadModel, we can:
1. directly new an instance and set the member one by one;
2. or use the CarbonLoadModelBuilder to build an instance

However some of the members of CarbonLoadModel (such as ColumnCompressor, 
tableName) are required in the following procedure.

For the 1st method, these members may be forgotten to initialize. While for the 
2nd method, we can validate these members in the build method to ensure that 
these members are initialized.

So here I proposed to only use the CarbonLoadModelBuilder to instantiate a 
CarbonLoadModel.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] carbondata issue #2623: [CARBONDATA-2844] [CARBONDATA-2865] Pass SK/AK to ex...

2018-08-27 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2623
  
Build Failed  with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/8085/



---


[GitHub] carbondata issue #2659: [CARBONDATA-2887] Fix complex filters on spark carbo...

2018-08-27 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2659
  
SDV Build Fail , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/6422/



---


[GitHub] carbondata issue #2635: [CARBONDATA-2856][BloomDataMap] Fix bug in bloom ind...

2018-08-27 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2635
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/8084/



---


[GitHub] carbondata issue #2662: [WIP][CARBONDATA-2889]Add decoder based fallback mec...

2018-08-27 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2662
  
Build Success with Spark 2.2.1, Please check CI 
http://95.216.28.178:8080/job/ApacheCarbonPRBuilder1/23/



---


[GitHub] carbondata issue #2623: [CARBONDATA-2844] [CARBONDATA-2865] Pass SK/AK to ex...

2018-08-27 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2623
  
SDV Build Fail , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/6421/



---


[GitHub] carbondata pull request #2661: [CARBONDATA-2888] Support multi level subfold...

2018-08-27 Thread xuchuanyin
Github user xuchuanyin commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2661#discussion_r212897009
  
--- Diff: 
core/src/main/java/org/apache/carbondata/core/indexstore/blockletindex/SegmentIndexFileStore.java
 ---
@@ -338,6 +338,24 @@ private MergedBlockIndex 
readMergeBlockIndex(ThriftReader thriftReader) throws I
 });
   }
 
+  /**
+   * List all the index files of the segment.
+   *
+   * @param carbonFile directory
+   */
+  public static void getCarbonIndexFilesRecursively(CarbonFile carbonFile,
--- End diff --

Why not return the list instead of passing it as a parameter?


---


[GitHub] carbondata pull request #2661: [CARBONDATA-2888] Support multi level subfold...

2018-08-27 Thread xuchuanyin
Github user xuchuanyin commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2661#discussion_r212897508
  
--- Diff: 
core/src/main/java/org/apache/carbondata/core/metadata/schema/table/CarbonTable.java
 ---
@@ -268,6 +257,18 @@ public boolean accept(CarbonFile file) {
 return CarbonTable.buildFromTableInfo(tableInfoInfer);
   }
 
+  private static CarbonFile getFirstIndexFile(CarbonFile tablePath) {
+CarbonFile[] carbonFiles = tablePath.listFiles();
+for (CarbonFile carbonFile : carbonFiles) {
+  if (carbonFile.isDirectory()) {
+return getFirstIndexFile(carbonFile);
+  } else if 
(carbonFile.getName().endsWith(CarbonTablePath.INDEX_FILE_EXT)) {
+return carbonFile;
+  }
+}
+return null;
--- End diff --

why not just throw exception if the file does not exist?


---


[GitHub] carbondata issue #2635: [CARBONDATA-2856][BloomDataMap] Fix bug in bloom ind...

2018-08-27 Thread xuchuanyin
Github user xuchuanyin commented on the issue:

https://github.com/apache/carbondata/pull/2635
  
retest this please


---


[GitHub] carbondata issue #2623: [CARBONDATA-2844] Pass SK/AK to executor by serializ...

2018-08-27 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2623
  
Build Failed with Spark 2.2.1, Please check CI 
http://95.216.28.178:8080/job/ApacheCarbonPRBuilder1/22/



---


[GitHub] carbondata issue #2662: [WIP][CARBONDATA-2889]Add decoder based fallback mec...

2018-08-27 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2662
  
Build Failed  with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/8083/



---


[jira] [Assigned] (CARBONDATA-2885) Broadcast Issue and Small file distribution Issue

2018-08-27 Thread kumar vishal (JIRA)


 [ 
https://issues.apache.org/jira/browse/CARBONDATA-2885?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

kumar vishal reassigned CARBONDATA-2885:


Assignee: Babulal

> Broadcast Issue and Small file distribution Issue
> -
>
> Key: CARBONDATA-2885
> URL: https://issues.apache.org/jira/browse/CARBONDATA-2885
> Project: CarbonData
>  Issue Type: Bug
>Reporter: Babulal
>Assignee: Babulal
>Priority: Major
>
> Carbon Relation size is getting calculated wrongly ( always 0 ) for External 
> Table.
> Root Cause:- Because Tablestatus file is not present for external table
>  
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (CARBONDATA-2885) Broadcast Issue and Small file distribution Issue

2018-08-27 Thread kumar vishal (JIRA)


 [ 
https://issues.apache.org/jira/browse/CARBONDATA-2885?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

kumar vishal resolved CARBONDATA-2885.
--
Resolution: Fixed

> Broadcast Issue and Small file distribution Issue
> -
>
> Key: CARBONDATA-2885
> URL: https://issues.apache.org/jira/browse/CARBONDATA-2885
> Project: CarbonData
>  Issue Type: Bug
>Reporter: Babulal
>Assignee: Babulal
>Priority: Major
>
> Carbon Relation size is getting calculated wrongly ( always 0 ) for External 
> Table.
> Root Cause:- Because Tablestatus file is not present for external table
>  
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] carbondata pull request #2658: [Carbondata 2885]Broadcast Issue and Small fi...

2018-08-27 Thread asfgit
Github user asfgit closed the pull request at:

https://github.com/apache/carbondata/pull/2658


---


[GitHub] carbondata issue #2623: [CARBONDATA-2844] Pass SK/AK to executor by serializ...

2018-08-27 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2623
  
Build Failed  with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/8082/



---


[GitHub] carbondata issue #2662: [WIP][CARBONDATA-2889]Add decoder based fallback mec...

2018-08-27 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2662
  
SDV Build Fail , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/6420/



---


[GitHub] carbondata issue #2649: [CARBONDATA-2869] Add support for Avro Map data type...

2018-08-27 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2649
  
Build Success with Spark 2.2.1, Please check CI 
http://95.216.28.178:8080/job/ApacheCarbonPRBuilder1/21/



---


[GitHub] carbondata issue #2649: [CARBONDATA-2869] Add support for Avro Map data type...

2018-08-27 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2649
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/8080/



---


[jira] [Created] (CARBONDATA-2889) Support Decoder based fall back mechanism in Local Dictionary

2018-08-27 Thread Akash R Nilugal (JIRA)
Akash R Nilugal created CARBONDATA-2889:
---

 Summary: Support Decoder based fall back mechanism in Local 
Dictionary
 Key: CARBONDATA-2889
 URL: https://issues.apache.org/jira/browse/CARBONDATA-2889
 Project: CarbonData
  Issue Type: Sub-task
Reporter: Akash R Nilugal
Assignee: Akash R Nilugal


Currently, when the fallback is initiated for a column page in case of local 
dictionary, we are keeping both encoded data

and actual data in memory and then we form the new column page without 
dictionary encoding and then at last we free the Encoded Column Page.

Because of this offheap memory footprint increases.

 

We can reduce the offheap memory footprint. This can be done using decoder 
based fallback mechanism.

This means, no need to keep the actual data along with encoded data in encoded 
column page. We can keep only encoded data and to form a new column page, get 
the dictionary data from encoded column page by uncompressing and using 
dictionary data get the actual data using local dictionary generator and put it 
in new column page created and compress it again and give to consumer for 
writing blocklet. 

 

The above process may slow down the loading, but it will reduces the memory 
footprint. So we can give a property which will decide whether to take current 
fallback procedure or decoder based fallback mechanism dring fallback



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] carbondata pull request #2662: [CARBONDATA-2889]Add decoder based fallback m...

2018-08-27 Thread akashrn5
GitHub user akashrn5 opened a pull request:

https://github.com/apache/carbondata/pull/2662

[CARBONDATA-2889]Add decoder based fallback mechanism in local dictionary 
to reduce memory footprint



Be sure to do all of the following checklist to help us incorporate 
your contribution quickly and easily:

 - [ ] Any interfaces changed?
 
 - [ ] Any backward compatibility impacted?
 
 - [ ] Document update required?

 - [ ] Testing done
Please provide details on 
- Whether new unit test cases have been added or why no new tests 
are required?
- How it is tested? Please attach test report.
- Is it a performance related change? Please attach the performance 
test report.
- Any additional information to help reviewers in testing this 
change.
   
 - [ ] For large changes, please consider breaking it into sub-tasks under 
an umbrella JIRA. 



You can merge this pull request into a Git repository by running:

$ git pull https://github.com/akashrn5/incubator-carbondata fallback

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/carbondata/pull/2662.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #2662


commit 8729c673ed06ee16ed256270e31c495bd1568bfd
Author: akashrn5 
Date:   2018-08-20T04:59:26Z

Add decoder based fallback mechanism in local dictionary to reduce memory 
footprint




---


[GitHub] carbondata issue #2623: [CARBONDATA-2844] Pass SK/AK to executor by serializ...

2018-08-27 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2623
  
SDV Build Fail , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/6419/



---


[GitHub] carbondata issue #2623: [CARBONDATA-2844] Pass SK/AK to executor by serializ...

2018-08-27 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2623
  
SDV Build Fail , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/6418/



---