[jira] [Commented] (CARBONDATA-210) Support loading BZIP2 compressed CSV file
[ https://issues.apache.org/jira/browse/CARBONDATA-210?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15587755#comment-15587755 ] ASF GitHub Bot commented on CARBONDATA-210: --- Github user asfgit closed the pull request at: https://github.com/apache/incubator-carbondata/pull/151 > Support loading BZIP2 compressed CSV file > - > > Key: CARBONDATA-210 > URL: https://issues.apache.org/jira/browse/CARBONDATA-210 > Project: CarbonData > Issue Type: Bug >Reporter: Jacky Li > Fix For: 0.2.0-incubating > > > Support BZIP2 compressed CSV file, file extension name is bz2 -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (CARBONDATA-324) Decimal and Bigint type columns contains Null, after load data
[ https://issues.apache.org/jira/browse/CARBONDATA-324?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Harmeet Singh updated CARBONDATA-324: - Description: Using Thrift server and Beeling client, i am trying to create a table and load the data from CSV. My tables contains BigInt and Decimal Column types, After load the data using Load Data command, The BigInt and Decimal Column contains Null Value. Bellow are the steps: Step 1: > create database wednesday; > use wednesday; > CREATE TABLE one (id int, age iNt, name String, salary decimal, data bigInt, > weight double, dob timeStamp) STORED BY 'carbondata'; Step 2: Create a csv file which contains column values as below: id, age, name, salary, data, weight, dob 1, 54, james, 90, 292092, 34.2, 2016-05-04 22:55:00 Step 3: Load the data from CSV file as below: > LOAD DATA INPATH 'hdfs://localhost:54310/home/harmeet/sample3.csv' INTO TABLE > one; Step 4: Select the data from table one, and BigInt and Decimal column contains Null value. > Decimal and Bigint type columns contains Null, after load data > -- > > Key: CARBONDATA-324 > URL: https://issues.apache.org/jira/browse/CARBONDATA-324 > Project: CarbonData > Issue Type: Bug >Reporter: Harmeet Singh > > Using Thrift server and Beeling client, i am trying to create a table and > load the data from CSV. My tables contains BigInt and Decimal Column types, > After load the data using Load Data command, The BigInt and Decimal Column > contains Null Value. Bellow are the steps: > Step 1: > > create database wednesday; > > use wednesday; > > CREATE TABLE one (id int, age iNt, name String, salary decimal, data > > bigInt, weight double, dob timeStamp) STORED BY 'carbondata'; > Step 2: > Create a csv file which contains column values as below: > id, age, name, salary, data, weight, dob > 1, 54, james, 90, 292092, 34.2, 2016-05-04 22:55:00 > Step 3: > Load the data from CSV file as below: > > LOAD DATA INPATH 'hdfs://localhost:54310/home/harmeet/sample3.csv' INTO > > TABLE one; > Step 4: > Select the data from table one, and BigInt and Decimal column contains Null > value. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Created] (CARBONDATA-324) Decimal and Bigint type columns contains Null, after load data
Harmeet Singh created CARBONDATA-324: Summary: Decimal and Bigint type columns contains Null, after load data Key: CARBONDATA-324 URL: https://issues.apache.org/jira/browse/CARBONDATA-324 Project: CarbonData Issue Type: Bug Reporter: Harmeet Singh -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Created] (CARBONDATA-323) Fix the load data local syntax
Fei Wang created CARBONDATA-323: --- Summary: Fix the load data local syntax Key: CARBONDATA-323 URL: https://issues.apache.org/jira/browse/CARBONDATA-323 Project: CarbonData Issue Type: Bug Components: spark-integration Affects Versions: 0.1.0-incubating Reporter: Fei Wang Assignee: Fei Wang Fix For: 0.2.0-incubating carbon should not support load data local syntax, so fix it. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (CARBONDATA-322) integrate spark 2.x
[ https://issues.apache.org/jira/browse/CARBONDATA-322?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Fei Wang updated CARBONDATA-322: Description: As spark 2.0 released. there are many nice features such as more efficient parser, vectorized execution, adaptive execution. It is good to integrate with spark 2.x Another side now spark integration is heavy coupling with spark, we should redesign the spark integration, it should satisfy flowing requirement: 1. decoupled with spark, integrate according to spark datasource API(V2) 2. This integration should support vectorized carbon reader 3. Supoort write to carbondata from dadatrame ... was: As spark 2.0 released. there are many nice features such as more efficient parser, vectorized execution, adaptive execution. It is good to integrate with spark 2.x Another side now spark integration is heavy coupling with spark, we should redesign the spark integration, it should satisfy flowing requirement: 1. decoupled with spark, integrate according to spark datasource API(V2) 2. This integration should support vectorized carbon reader 3. Supoort write to carbondata from dadatrame ... > integrate spark 2.x > > > Key: CARBONDATA-322 > URL: https://issues.apache.org/jira/browse/CARBONDATA-322 > Project: CarbonData > Issue Type: Bug > Components: spark-integration >Affects Versions: 0.2.0-incubating >Reporter: Fei Wang > Fix For: 0.3.0-incubating > > > As spark 2.0 released. there are many nice features such as more efficient > parser, vectorized execution, adaptive execution. > It is good to integrate with spark 2.x > Another side now spark integration is heavy coupling with spark, we should > redesign the spark integration, it should satisfy flowing requirement: > 1. decoupled with spark, integrate according to spark datasource API(V2) > 2. This integration should support vectorized carbon reader > 3. Supoort write to carbondata from dadatrame > ... -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (CARBONDATA-322) integrate spark 2.x
[ https://issues.apache.org/jira/browse/CARBONDATA-322?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Fei Wang updated CARBONDATA-322: Description: As spark 2.0 released. there are many nice features such as more efficient parser, vectorized execution, adaptive execution. It is good to integrate with spark 2.x Another side now in carbondata, spark integration is heavy coupling with spark code and the code need clean, we should redesign the spark integration, it should satisfy flowing requirement: 1. decoupled with spark, integrate according to spark datasource API(V2) 2. This integration should support vectorized carbon reader 3. Supoort write to carbondata from dadatrame ... was: As spark 2.0 released. there are many nice features such as more efficient parser, vectorized execution, adaptive execution. It is good to integrate with spark 2.x Another side now spark integration is heavy coupling with spark, we should redesign the spark integration, it should satisfy flowing requirement: 1. decoupled with spark, integrate according to spark datasource API(V2) 2. This integration should support vectorized carbon reader 3. Supoort write to carbondata from dadatrame ... > integrate spark 2.x > > > Key: CARBONDATA-322 > URL: https://issues.apache.org/jira/browse/CARBONDATA-322 > Project: CarbonData > Issue Type: Bug > Components: spark-integration >Affects Versions: 0.2.0-incubating >Reporter: Fei Wang > Fix For: 0.3.0-incubating > > > As spark 2.0 released. there are many nice features such as more efficient > parser, vectorized execution, adaptive execution. > It is good to integrate with spark 2.x > Another side now in carbondata, spark integration is heavy coupling with > spark code and the code need clean, we should redesign the spark integration, > it should satisfy flowing requirement: > 1. decoupled with spark, integrate according to spark datasource API(V2) > 2. This integration should support vectorized carbon reader > 3. Supoort write to carbondata from dadatrame > ... -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Created] (CARBONDATA-322) integrate spark 2.x
Fei Wang created CARBONDATA-322: --- Summary: integrate spark 2.x Key: CARBONDATA-322 URL: https://issues.apache.org/jira/browse/CARBONDATA-322 Project: CarbonData Issue Type: Bug Components: spark-integration Affects Versions: 0.2.0-incubating Reporter: Fei Wang Fix For: 0.3.0-incubating As spark 2.0 released. there are many nice features such as more efficient parser, vectorized execution, adaptive execution. It is good to integrate with spark 2.x Another side now spark integration is heavy coupling with spark, we should redesign the spark integration, it should satisfy flowing requirement: 1. decoupled with spark, integrate according to spark datasource API(V2) 2. This integration should support vectorized carbon reader 3. Supoort write to carbondata from dadatrame ... -- This message was sent by Atlassian JIRA (v6.3.4#6332)