[jira] [Commented] (CARBONDATA-210) Support loading BZIP2 compressed CSV file

2016-10-18 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/CARBONDATA-210?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15587755#comment-15587755
 ] 

ASF GitHub Bot commented on CARBONDATA-210:
---

Github user asfgit closed the pull request at:

https://github.com/apache/incubator-carbondata/pull/151


> Support loading BZIP2 compressed CSV file
> -
>
> Key: CARBONDATA-210
> URL: https://issues.apache.org/jira/browse/CARBONDATA-210
> Project: CarbonData
>  Issue Type: Bug
>Reporter: Jacky Li
> Fix For: 0.2.0-incubating
>
>
> Support BZIP2 compressed CSV file, file extension name is bz2



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (CARBONDATA-324) Decimal and Bigint type columns contains Null, after load data

2016-10-18 Thread Harmeet Singh (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-324?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harmeet Singh updated CARBONDATA-324:
-
Description: 
Using Thrift server and Beeling client, i am trying to create a table and load 
the data from CSV. My tables contains BigInt and Decimal Column types, After 
load the data using Load Data command, The BigInt and Decimal Column contains 
Null Value. Bellow are the steps:

Step 1: 
> create database wednesday;
> use wednesday;
> CREATE TABLE one (id int, age iNt, name String, salary decimal, data bigInt, 
> weight double, dob timeStamp) STORED BY 'carbondata';

Step 2: 
Create a csv file which contains column values as below: 

id, age, name, salary, data, weight, dob
1, 54, james, 90, 292092, 34.2, 2016-05-04 22:55:00

Step 3: 
Load the data from CSV file as below: 
> LOAD DATA INPATH 'hdfs://localhost:54310/home/harmeet/sample3.csv' INTO TABLE 
> one;

Step 4: 
Select the data from table one, and BigInt and Decimal column contains Null 
value. 


> Decimal and Bigint type columns contains Null, after load data
> --
>
> Key: CARBONDATA-324
> URL: https://issues.apache.org/jira/browse/CARBONDATA-324
> Project: CarbonData
>  Issue Type: Bug
>Reporter: Harmeet Singh
>
> Using Thrift server and Beeling client, i am trying to create a table and 
> load the data from CSV. My tables contains BigInt and Decimal Column types, 
> After load the data using Load Data command, The BigInt and Decimal Column 
> contains Null Value. Bellow are the steps:
> Step 1: 
> > create database wednesday;
> > use wednesday;
> > CREATE TABLE one (id int, age iNt, name String, salary decimal, data 
> > bigInt, weight double, dob timeStamp) STORED BY 'carbondata';
> Step 2: 
> Create a csv file which contains column values as below: 
> id, age, name, salary, data, weight, dob
> 1, 54, james, 90, 292092, 34.2, 2016-05-04 22:55:00
> Step 3: 
> Load the data from CSV file as below: 
> > LOAD DATA INPATH 'hdfs://localhost:54310/home/harmeet/sample3.csv' INTO 
> > TABLE one;
> Step 4: 
> Select the data from table one, and BigInt and Decimal column contains Null 
> value. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CARBONDATA-324) Decimal and Bigint type columns contains Null, after load data

2016-10-18 Thread Harmeet Singh (JIRA)
Harmeet Singh created CARBONDATA-324:


 Summary: Decimal and Bigint type columns contains Null, after load 
data
 Key: CARBONDATA-324
 URL: https://issues.apache.org/jira/browse/CARBONDATA-324
 Project: CarbonData
  Issue Type: Bug
Reporter: Harmeet Singh






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CARBONDATA-323) Fix the load data local syntax

2016-10-18 Thread Fei Wang (JIRA)
Fei Wang created CARBONDATA-323:
---

 Summary: Fix the load data local syntax
 Key: CARBONDATA-323
 URL: https://issues.apache.org/jira/browse/CARBONDATA-323
 Project: CarbonData
  Issue Type: Bug
  Components: spark-integration
Affects Versions: 0.1.0-incubating
Reporter: Fei Wang
Assignee: Fei Wang
 Fix For: 0.2.0-incubating


carbon should not support load data local syntax, so fix it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (CARBONDATA-322) integrate spark 2.x

2016-10-18 Thread Fei Wang (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-322?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Fei Wang updated CARBONDATA-322:

Description: 
As spark 2.0 released. there are many nice features such as more efficient 
parser, vectorized execution, adaptive execution. 
It is good to integrate with spark 2.x

Another side now spark integration is heavy coupling with spark, we should 
redesign the spark integration, it should satisfy flowing requirement:

1. decoupled with spark, integrate according to spark datasource API(V2)
2. This integration should support vectorized carbon reader
3. Supoort write to carbondata from dadatrame
...


  was:
As spark 2.0 released. there are many nice features such as more efficient 
parser, vectorized execution, adaptive execution. It is good to integrate with 
spark 2.x

Another side now spark integration is heavy coupling with spark, we should 
redesign the spark integration, it should satisfy flowing requirement:

1. decoupled with spark, integrate according to spark datasource API(V2)
2. This integration should support vectorized carbon reader
3. Supoort write to carbondata from dadatrame
...



> integrate spark 2.x 
> 
>
> Key: CARBONDATA-322
> URL: https://issues.apache.org/jira/browse/CARBONDATA-322
> Project: CarbonData
>  Issue Type: Bug
>  Components: spark-integration
>Affects Versions: 0.2.0-incubating
>Reporter: Fei Wang
> Fix For: 0.3.0-incubating
>
>
> As spark 2.0 released. there are many nice features such as more efficient 
> parser, vectorized execution, adaptive execution. 
> It is good to integrate with spark 2.x
> Another side now spark integration is heavy coupling with spark, we should 
> redesign the spark integration, it should satisfy flowing requirement:
> 1. decoupled with spark, integrate according to spark datasource API(V2)
> 2. This integration should support vectorized carbon reader
> 3. Supoort write to carbondata from dadatrame
> ...



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (CARBONDATA-322) integrate spark 2.x

2016-10-18 Thread Fei Wang (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-322?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Fei Wang updated CARBONDATA-322:

Description: 
As spark 2.0 released. there are many nice features such as more efficient 
parser, vectorized execution, adaptive execution. 
It is good to integrate with spark 2.x

Another side now in carbondata, spark integration is heavy coupling with spark 
code and the code need clean, we should redesign the spark integration, it 
should satisfy flowing requirement:

1. decoupled with spark, integrate according to spark datasource API(V2)
2. This integration should support vectorized carbon reader
3. Supoort write to carbondata from dadatrame
...


  was:
As spark 2.0 released. there are many nice features such as more efficient 
parser, vectorized execution, adaptive execution. 
It is good to integrate with spark 2.x

Another side now spark integration is heavy coupling with spark, we should 
redesign the spark integration, it should satisfy flowing requirement:

1. decoupled with spark, integrate according to spark datasource API(V2)
2. This integration should support vectorized carbon reader
3. Supoort write to carbondata from dadatrame
...



> integrate spark 2.x 
> 
>
> Key: CARBONDATA-322
> URL: https://issues.apache.org/jira/browse/CARBONDATA-322
> Project: CarbonData
>  Issue Type: Bug
>  Components: spark-integration
>Affects Versions: 0.2.0-incubating
>Reporter: Fei Wang
> Fix For: 0.3.0-incubating
>
>
> As spark 2.0 released. there are many nice features such as more efficient 
> parser, vectorized execution, adaptive execution. 
> It is good to integrate with spark 2.x
> Another side now in carbondata, spark integration is heavy coupling with 
> spark code and the code need clean, we should redesign the spark integration, 
> it should satisfy flowing requirement:
> 1. decoupled with spark, integrate according to spark datasource API(V2)
> 2. This integration should support vectorized carbon reader
> 3. Supoort write to carbondata from dadatrame
> ...



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CARBONDATA-322) integrate spark 2.x

2016-10-18 Thread Fei Wang (JIRA)
Fei Wang created CARBONDATA-322:
---

 Summary: integrate spark 2.x 
 Key: CARBONDATA-322
 URL: https://issues.apache.org/jira/browse/CARBONDATA-322
 Project: CarbonData
  Issue Type: Bug
  Components: spark-integration
Affects Versions: 0.2.0-incubating
Reporter: Fei Wang
 Fix For: 0.3.0-incubating


As spark 2.0 released. there are many nice features such as more efficient 
parser, vectorized execution, adaptive execution. It is good to integrate with 
spark 2.x

Another side now spark integration is heavy coupling with spark, we should 
redesign the spark integration, it should satisfy flowing requirement:

1. decoupled with spark, integrate according to spark datasource API(V2)
2. This integration should support vectorized carbon reader
3. Supoort write to carbondata from dadatrame
...




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)