[jira] [Created] (CARBONDATA-2792) Create external table fails post schema restructure

2018-07-27 Thread Mohammad Shahid Khan (JIRA)
Mohammad Shahid Khan created CARBONDATA-2792:


 Summary: Create external table fails post schema restructure 
 Key: CARBONDATA-2792
 URL: https://issues.apache.org/jira/browse/CARBONDATA-2792
 Project: CarbonData
  Issue Type: Bug
Affects Versions: 1.4.0, 1.5.0, 1.4.2
Reporter: Mohammad Shahid Khan
Assignee: Mohammad Shahid Khan


Steps to reproduce.

sql("create table rstest1 (c1 string,c2 int) STORED BY 
'org.apache.carbondata.format'")
sql("Alter table rstest1 drop columns(c2)")
sql(
 "Alter table rstest1 add columns(c4 string) 
TBLPROPERTIES('DICTIONARY_EXCLUDE'='c4', " +
 "'DEFAULT.VALUE.c4'='def')")
sql(s"""CREATE EXTERNAL TABLE rsext STORED BY 'carbondata' LOCATION 
'$storeLocation/rstest1'""")



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (CARBONDATA-2749) In HDFS Empty tablestatus file is written during datalaod, iud or compaction when disk is full.

2018-07-17 Thread Mohammad Shahid Khan (JIRA)


 [ 
https://issues.apache.org/jira/browse/CARBONDATA-2749?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mohammad Shahid Khan reassigned CARBONDATA-2749:


Assignee: Mohammad Shahid Khan

> In HDFS Empty tablestatus file is written during datalaod, iud or compaction 
> when disk is full.
> ---
>
> Key: CARBONDATA-2749
> URL: https://issues.apache.org/jira/browse/CARBONDATA-2749
> Project: CarbonData
>  Issue Type: Bug
>Reporter: Mohammad Shahid Khan
>Assignee: Mohammad Shahid Khan
>Priority: Major
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> *Issue*: When a failure happens due to disk full during load, IUD or 
> Compaction, then while updating the tablestatus file, the tablestaus.tmp file 
> during atomic file operation remains empty, and in the finally block the 
> empty tablestaus.tmp file is getting renamed to the actual file. 
> This leads to empty tablestaus file.
> *Problem*: Once such problem happens the tablestatus file can not be 
> retrieved and the already loaded data can be used. 
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (CARBONDATA-2749) In HDFS Empty tablestatus file is written during datalaod, iud or compaction when disk is full.

2018-07-17 Thread Mohammad Shahid Khan (JIRA)
Mohammad Shahid Khan created CARBONDATA-2749:


 Summary: In HDFS Empty tablestatus file is written during 
datalaod, iud or compaction when disk is full.
 Key: CARBONDATA-2749
 URL: https://issues.apache.org/jira/browse/CARBONDATA-2749
 Project: CarbonData
  Issue Type: Bug
Reporter: Mohammad Shahid Khan


*Issue*: When a failure happens due to disk full during load, IUD or 
Compaction, then while updating the tablestatus file, the tablestaus.tmp file 
during atomic file operation remains empty, and in the finally block the empty 
tablestaus.tmp file is getting renamed to the actual file. 

This leads to empty tablestaus file.

*Problem*: Once such problem happens the tablestatus file can not be retrieved 
and the already loaded data can be used. 

 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (CARBONDATA-2729) Schema Compatibility problem between version 1.3.0 and 1.4.0

2018-07-11 Thread Mohammad Shahid Khan (JIRA)


 [ 
https://issues.apache.org/jira/browse/CARBONDATA-2729?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mohammad Shahid Khan updated CARBONDATA-2729:
-
Description: 
The table schema created with 1.3 is not compatible with 1.4.0.

Steps:

Do the below operation  on hive meta store.

1.  Create table and load data in 1.3 version.   
{code:java}
spark.sql("create database carbonversion_1_1");
spark.sql("use carbonversion_1_1");
spark.sql("""create table testIntType (c1 string,c2 int) STORED BY 
'org.apache.carbondata.format'""")
spark.sql("insert into testIntType select 'a',200"){code}
2. Switch to 1.4 version and create data map on the table created in step 1.
{code:java}
spark.sql("use carbonversion_1_3");
spark.sql("create datamap dm8 on table testinttype3 using 'preaggregate' as 
select c1, sum(c2) from testinttype3 group by c1")
{code}
 stack trace 
{code:java}
stack trace

Exception in thread "main" java.lang.NullPointerException
at 
org.apache.carbondata.core.metadata.converter.ThriftWrapperSchemaConverterImpl.fromWrapperToExternalSchemaEvolution(ThriftWrapperSchemaConverterImpl.java:87)
at 
org.apache.carbondata.core.metadata.converter.ThriftWrapperSchemaConverterImpl.fromWrapperToExternalTableSchema(ThriftWrapperSchemaConverterImpl.java:277)
at 
org.apache.carbondata.core.metadata.converter.ThriftWrapperSchemaConverterImpl.fromWrapperToExternalTableInfo(ThriftWrapperSchemaConverterImpl.java:310)
at 
org.apache.spark.sql.hive.CarbonHiveMetaStore.getThriftTableInfo(CarbonHiveMetaStore.scala:101){code}

  was:
The table schema created with 1.3 is not compatible with 1.4.0.

 stack trace 
{code:java}
stack trace

Exception in thread "main" java.lang.NullPointerException
at 
org.apache.carbondata.core.metadata.converter.ThriftWrapperSchemaConverterImpl.fromWrapperToExternalSchemaEvolution(ThriftWrapperSchemaConverterImpl.java:87)
at 
org.apache.carbondata.core.metadata.converter.ThriftWrapperSchemaConverterImpl.fromWrapperToExternalTableSchema(ThriftWrapperSchemaConverterImpl.java:277)
at 
org.apache.carbondata.core.metadata.converter.ThriftWrapperSchemaConverterImpl.fromWrapperToExternalTableInfo(ThriftWrapperSchemaConverterImpl.java:310)
at 
org.apache.spark.sql.hive.CarbonHiveMetaStore.getThriftTableInfo(CarbonHiveMetaStore.scala:101){code}


> Schema Compatibility problem between version 1.3.0 and 1.4.0
> 
>
> Key: CARBONDATA-2729
> URL: https://issues.apache.org/jira/browse/CARBONDATA-2729
> Project: CarbonData
>  Issue Type: Bug
>  Components: file-format
>Affects Versions: 1.4.0
>Reporter: Mohammad Shahid Khan
>Assignee: Mohammad Shahid Khan
>Priority: Major
> Fix For: 1.4.1
>
>
> The table schema created with 1.3 is not compatible with 1.4.0.
> Steps:
> Do the below operation  on hive meta store.
> 1.  Create table and load data in 1.3 version.   
> {code:java}
> spark.sql("create database carbonversion_1_1");
> spark.sql("use carbonversion_1_1");
> spark.sql("""create table testIntType (c1 string,c2 int) STORED BY 
> 'org.apache.carbondata.format'""")
> spark.sql("insert into testIntType select 'a',200"){code}
> 2. Switch to 1.4 version and create data map on the table created in step 1.
> {code:java}
> spark.sql("use carbonversion_1_3");
> spark.sql("create datamap dm8 on table testinttype3 using 'preaggregate' as 
> select c1, sum(c2) from testinttype3 group by c1")
> {code}
>  stack trace 
> {code:java}
> stack trace
> Exception in thread "main" java.lang.NullPointerException
> at 
> org.apache.carbondata.core.metadata.converter.ThriftWrapperSchemaConverterImpl.fromWrapperToExternalSchemaEvolution(ThriftWrapperSchemaConverterImpl.java:87)
> at 
> org.apache.carbondata.core.metadata.converter.ThriftWrapperSchemaConverterImpl.fromWrapperToExternalTableSchema(ThriftWrapperSchemaConverterImpl.java:277)
> at 
> org.apache.carbondata.core.metadata.converter.ThriftWrapperSchemaConverterImpl.fromWrapperToExternalTableInfo(ThriftWrapperSchemaConverterImpl.java:310)
> at 
> org.apache.spark.sql.hive.CarbonHiveMetaStore.getThriftTableInfo(CarbonHiveMetaStore.scala:101){code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (CARBONDATA-2729) Schema Compatibility problem between version 1.3.0 and 1.4.0

2018-07-11 Thread Mohammad Shahid Khan (JIRA)


 [ 
https://issues.apache.org/jira/browse/CARBONDATA-2729?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mohammad Shahid Khan updated CARBONDATA-2729:
-
Description: 
The table schema created with 1.3 is not compatible with 1.4.0.

 stack trace 
{code:java}
stack trace

Exception in thread "main" java.lang.NullPointerException
at 
org.apache.carbondata.core.metadata.converter.ThriftWrapperSchemaConverterImpl.fromWrapperToExternalSchemaEvolution(ThriftWrapperSchemaConverterImpl.java:87)
at 
org.apache.carbondata.core.metadata.converter.ThriftWrapperSchemaConverterImpl.fromWrapperToExternalTableSchema(ThriftWrapperSchemaConverterImpl.java:277)
at 
org.apache.carbondata.core.metadata.converter.ThriftWrapperSchemaConverterImpl.fromWrapperToExternalTableInfo(ThriftWrapperSchemaConverterImpl.java:310)
at 
org.apache.spark.sql.hive.CarbonHiveMetaStore.getThriftTableInfo(CarbonHiveMetaStore.scala:101){code}

  was:
The table schema created with 1.3 is not compatible with 1.4.0.

 


> Schema Compatibility problem between version 1.3.0 and 1.4.0
> 
>
> Key: CARBONDATA-2729
> URL: https://issues.apache.org/jira/browse/CARBONDATA-2729
> Project: CarbonData
>  Issue Type: Bug
>  Components: file-format
>Affects Versions: 1.4.0
>Reporter: Mohammad Shahid Khan
>Assignee: Mohammad Shahid Khan
>Priority: Major
> Fix For: 1.4.1
>
>
> The table schema created with 1.3 is not compatible with 1.4.0.
>  stack trace 
> {code:java}
> stack trace
> Exception in thread "main" java.lang.NullPointerException
> at 
> org.apache.carbondata.core.metadata.converter.ThriftWrapperSchemaConverterImpl.fromWrapperToExternalSchemaEvolution(ThriftWrapperSchemaConverterImpl.java:87)
> at 
> org.apache.carbondata.core.metadata.converter.ThriftWrapperSchemaConverterImpl.fromWrapperToExternalTableSchema(ThriftWrapperSchemaConverterImpl.java:277)
> at 
> org.apache.carbondata.core.metadata.converter.ThriftWrapperSchemaConverterImpl.fromWrapperToExternalTableInfo(ThriftWrapperSchemaConverterImpl.java:310)
> at 
> org.apache.spark.sql.hive.CarbonHiveMetaStore.getThriftTableInfo(CarbonHiveMetaStore.scala:101){code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (CARBONDATA-2729) Schema Compatibility problem between version 1.3.0 and 1.4.0

2018-07-11 Thread Mohammad Shahid Khan (JIRA)
Mohammad Shahid Khan created CARBONDATA-2729:


 Summary: Schema Compatibility problem between version 1.3.0 and 
1.4.0
 Key: CARBONDATA-2729
 URL: https://issues.apache.org/jira/browse/CARBONDATA-2729
 Project: CarbonData
  Issue Type: Bug
  Components: file-format
Affects Versions: 1.4.0
Reporter: Mohammad Shahid Khan
Assignee: Mohammad Shahid Khan
 Fix For: 1.4.1


The table schema created with 1.3 is not compatible with 1.4.0.

 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (CARBONDATA-2710) Refactor CarbonSparkSqlParser for better code reuse.

2018-07-09 Thread Mohammad Shahid Khan (JIRA)
Mohammad Shahid Khan created CARBONDATA-2710:


 Summary: Refactor CarbonSparkSqlParser for better code reuse.
 Key: CARBONDATA-2710
 URL: https://issues.apache.org/jira/browse/CARBONDATA-2710
 Project: CarbonData
  Issue Type: Improvement
Reporter: Mohammad Shahid Khan






--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (CARBONDATA-2674) Streaming with merge index enabled does not consider the merge index file while pruning.

2018-06-29 Thread Mohammad Shahid Khan (JIRA)


 [ 
https://issues.apache.org/jira/browse/CARBONDATA-2674?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mohammad Shahid Khan updated CARBONDATA-2674:
-
Summary: Streaming with merge index enabled does not consider the merge 
index file while pruning.   (was: Streaming with merge index enabled does not 
consider the index file while pruning.)

> Streaming with merge index enabled does not consider the merge index file 
> while pruning. 
> -
>
> Key: CARBONDATA-2674
> URL: https://issues.apache.org/jira/browse/CARBONDATA-2674
> Project: CarbonData
>  Issue Type: Bug
>Reporter: Mohammad Shahid Khan
>Priority: Major
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (CARBONDATA-2674) Streaming with merge index enabled does not consider the index file while pruning.

2018-06-29 Thread Mohammad Shahid Khan (JIRA)
Mohammad Shahid Khan created CARBONDATA-2674:


 Summary: Streaming with merge index enabled does not consider the 
index file while pruning.
 Key: CARBONDATA-2674
 URL: https://issues.apache.org/jira/browse/CARBONDATA-2674
 Project: CarbonData
  Issue Type: Bug
Reporter: Mohammad Shahid Khan






--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (CARBONDATA-2623) Add DataMap Pre and Pevent listener

2018-06-20 Thread Mohammad Shahid Khan (JIRA)
Mohammad Shahid Khan created CARBONDATA-2623:


 Summary: Add DataMap Pre and Pevent listener
 Key: CARBONDATA-2623
 URL: https://issues.apache.org/jira/browse/CARBONDATA-2623
 Project: CarbonData
  Issue Type: Improvement
Reporter: Mohammad Shahid Khan






--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (CARBONDATA-2621) Lock problem in index datamap

2018-06-20 Thread Mohammad Shahid Khan (JIRA)


 [ 
https://issues.apache.org/jira/browse/CARBONDATA-2621?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mohammad Shahid Khan updated CARBONDATA-2621:
-
Summary: Lock problem in index datamap  (was: Lock problem in for index 
datamap)

> Lock problem in index datamap
> -
>
> Key: CARBONDATA-2621
> URL: https://issues.apache.org/jira/browse/CARBONDATA-2621
> Project: CarbonData
>  Issue Type: Bug
>Affects Versions: 1.4.0
>Reporter: Mohammad Shahid Khan
>Assignee: Mohammad Shahid Khan
>Priority: Major
> Fix For: 1.5.0
>
>
> The locking for the index Datamap is not correct.
> The HDFS lock will not work properly, because the lock is getting created the 
> the local filesystem instead of HDFS. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (CARBONDATA-2621) Lock problem in for index datamap

2018-06-20 Thread Mohammad Shahid Khan (JIRA)
Mohammad Shahid Khan created CARBONDATA-2621:


 Summary: Lock problem in for index datamap
 Key: CARBONDATA-2621
 URL: https://issues.apache.org/jira/browse/CARBONDATA-2621
 Project: CarbonData
  Issue Type: Bug
Affects Versions: 1.4.0
Reporter: Mohammad Shahid Khan
Assignee: Mohammad Shahid Khan
 Fix For: 1.5.0


The locking for the index Datamap is not correct.

The HDFS lock will not work properly, because the lock is getting created the 

the local filesystem instead of HDFS. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (CARBONDATA-2436) Block pruning problem post the carbon schema restructure.

2018-05-04 Thread Mohammad Shahid Khan (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-2436?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mohammad Shahid Khan updated CARBONDATA-2436:
-
Summary: Block pruning problem post the carbon schema restructure.  (was: 
Blocklet pruning problem post the carbon schema restructure.)

> Block pruning problem post the carbon schema restructure.
> -
>
> Key: CARBONDATA-2436
> URL: https://issues.apache.org/jira/browse/CARBONDATA-2436
> Project: CarbonData
>  Issue Type: Bug
>Reporter: Mohammad Shahid Khan
>Priority: Major
>
> Currently datamap is pruning with segmentproperties from the 0th blcok of 
> BlockletDataMap is not 
> correct. As post restructure if the table is updated then all the block will 
> not have 
> symetric schema within the same segments.
> Fix: It must be ensured the block could be pruned with the same schema.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (CARBONDATA-2436) Blocklet pruning problem post the carbon schema restructure.

2018-05-04 Thread Mohammad Shahid Khan (JIRA)
Mohammad Shahid Khan created CARBONDATA-2436:


 Summary: Blocklet pruning problem post the carbon schema 
restructure.
 Key: CARBONDATA-2436
 URL: https://issues.apache.org/jira/browse/CARBONDATA-2436
 Project: CarbonData
  Issue Type: Bug
Reporter: Mohammad Shahid Khan


Currently datamap is pruning with segmentproperties from the 0th blcok of 
BlockletDataMap is not 
correct. As post restructure if the table is updated then all the block will 
not have 
symetric schema within the same segments.

Fix: It must be ensured the block could be pruned with the same schema.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (CARBONDATA-2410) Error message correction when column value length exceeds 320000 charactor

2018-04-27 Thread Mohammad Shahid Khan (JIRA)
Mohammad Shahid Khan created CARBONDATA-2410:


 Summary: Error message correction when column value length exceeds 
32 charactor
 Key: CARBONDATA-2410
 URL: https://issues.apache.org/jira/browse/CARBONDATA-2410
 Project: CarbonData
  Issue Type: Bug
Reporter: Mohammad Shahid Khan






--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (CARBONDATA-2289) If carbon merge index is enabled then after IUD operation if some blocks of a segment is deleted, then during query and IUD operation the driver is throwing FileNotF

2018-03-28 Thread Mohammad Shahid Khan (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-2289?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mohammad Shahid Khan updated CARBONDATA-2289:
-
Description: 
Root cause:

The carbonindexmerge file is containing deleted blocks (.*carbondata*) meta 
i.e. driver tries to load BlockletDataMap for deleted blocks and fails while 
populating BlockMetaInfo.

 Steps to reproduce:

sql("drop table if exists mitable")
sql("create table mitable(id int, issue date) stored by 'carbondata'")
sql("insert into table mitable select '1','2000-02-01'")
val table = CarbonMetadata.getInstance().getCarbonTable("default", "mitable")
new CarbonIndexFileMergeWriter()
 .mergeCarbonIndexFilesOfSegment("0", table.getTablePath, false)
sql("update mitable set(id)=(2) where issue = '2000-02-01'").show()
sql("clean files for table mitable")
sql("select * from mitable").show()

  was:
Root cause:

The carbonindexmerge file is containing deleted blocks (.*carbondata*) meta 
i.e. driver tries to load BlockletDataMap for deleted blocks and fails while 
populating BlockMetaInfo.

 


> If carbon merge index is enabled then after IUD operation if some blocks of a 
> segment is deleted, then during query and IUD operation the driver is 
> throwing FileNotFoundException while preparing BlockMetaInfo.
> -
>
> Key: CARBONDATA-2289
> URL: https://issues.apache.org/jira/browse/CARBONDATA-2289
> Project: CarbonData
>  Issue Type: Bug
>Reporter: Mohammad Shahid Khan
>Priority: Minor
>
> Root cause:
> The carbonindexmerge file is containing deleted blocks (.*carbondata*) meta 
> i.e. driver tries to load BlockletDataMap for deleted blocks and fails while 
> populating BlockMetaInfo.
>  Steps to reproduce:
> sql("drop table if exists mitable")
> sql("create table mitable(id int, issue date) stored by 'carbondata'")
> sql("insert into table mitable select '1','2000-02-01'")
> val table = CarbonMetadata.getInstance().getCarbonTable("default", "mitable")
> new CarbonIndexFileMergeWriter()
>  .mergeCarbonIndexFilesOfSegment("0", table.getTablePath, false)
> sql("update mitable set(id)=(2) where issue = '2000-02-01'").show()
> sql("clean files for table mitable")
> sql("select * from mitable").show()



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (CARBONDATA-2289) If carbon merge index is enabled then after IUD operation if some blocks of a segment is deleted, then during query and IUD operation the driver is throwing FileNotF

2018-03-28 Thread Mohammad Shahid Khan (JIRA)
Mohammad Shahid Khan created CARBONDATA-2289:


 Summary: If carbon merge index is enabled then after IUD operation 
if some blocks of a segment is deleted, then during query and IUD operation the 
driver is throwing FileNotFoundException while preparing BlockMetaInfo.
 Key: CARBONDATA-2289
 URL: https://issues.apache.org/jira/browse/CARBONDATA-2289
 Project: CarbonData
  Issue Type: Bug
Reporter: Mohammad Shahid Khan


Root cause:

The carbonindexmerge file is containing deleted blocks (.*carbondata*) meta 
i.e. driver tries to load BlockletDataMap for deleted blocks and fails while 
populating BlockMetaInfo.

 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (CARBONDATA-2285) spark integration code refactor

2018-03-26 Thread Mohammad Shahid Khan (JIRA)
Mohammad Shahid Khan created CARBONDATA-2285:


 Summary: spark integration code refactor
 Key: CARBONDATA-2285
 URL: https://issues.apache.org/jira/browse/CARBONDATA-2285
 Project: CarbonData
  Issue Type: Improvement
Reporter: Mohammad Shahid Khan
Assignee: Mohammad Shahid Khan


Spark integration code refactor to improve extensible and re-usability



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (CARBONDATA-1507) Dataload global sort fails with RpcTimeOutException: Futures timed out after [120 seconds]

2018-01-30 Thread Mohammad Shahid Khan (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-1507?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mohammad Shahid Khan resolved CARBONDATA-1507.
--
Resolution: Duplicate

> Dataload global sort fails with RpcTimeOutException: Futures timed out after 
> [120 seconds]
> --
>
> Key: CARBONDATA-1507
> URL: https://issues.apache.org/jira/browse/CARBONDATA-1507
> Project: CarbonData
>  Issue Type: Bug
>Reporter: Mohammad Shahid Khan
>Assignee: Mohammad Shahid Khan
>Priority: Major
>
> Dataload global sort fails with RpcTimeOutException: Futures timed out after 
> [120 seconds]



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (CARBONDATA-2056) Hadoop Configuration with access key and secret key should be passed while creating InputStream of distributed carbon file.

2018-01-18 Thread Mohammad Shahid Khan (JIRA)
Mohammad Shahid Khan created CARBONDATA-2056:


 Summary: Hadoop Configuration with access key and secret key 
should be passed while creating InputStream of distributed carbon file.
 Key: CARBONDATA-2056
 URL: https://issues.apache.org/jira/browse/CARBONDATA-2056
 Project: CarbonData
  Issue Type: Bug
Reporter: Mohammad Shahid Khan






--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (CARBONDATA-2049) CarbonCleanFilesCommand table path problem

2018-01-17 Thread Mohammad Shahid Khan (JIRA)
Mohammad Shahid Khan created CARBONDATA-2049:


 Summary: CarbonCleanFilesCommand table path problem 
 Key: CARBONDATA-2049
 URL: https://issues.apache.org/jira/browse/CARBONDATA-2049
 Project: CarbonData
  Issue Type: Bug
Reporter: Mohammad Shahid Khan
Assignee: Mohammad Shahid Khan


In CarbonCleanFilesCommand datbaseLocation is being passed instead of the 
tablePath in case of forceclean.

And in case of cleanGarbageData, storeLocation is being passed instead of the 
tablePath.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (CARBONDATA-2044) IDG update for CARBONDATA-2043 Configurable wait time for requesting executors and minimum registered executors ratio to continue the block distribution

2018-01-17 Thread Mohammad Shahid Khan (JIRA)
Mohammad Shahid Khan created CARBONDATA-2044:


 Summary: IDG update for CARBONDATA-2043  Configurable wait time 
for requesting executors and minimum registered executors ratio to continue the 
block distribution
 Key: CARBONDATA-2044
 URL: https://issues.apache.org/jira/browse/CARBONDATA-2044
 Project: CarbonData
  Issue Type: Sub-task
  Components: data-load
Reporter: Mohammad Shahid Khan
Assignee: Gururaj Shetty


Added below parameters:

*carbon.dynamicallocation.schedulertimeout :*                                   
                            Description: To configure wait time for executors 
to bring up.                                             default 5 sec, Min 5 
sec and max 15 sec.

*carbon.scheduler.minregisteredresourcesratio :*                                
                    Description: To configure minimum registered/started 
executors ratio to start block distribution.  default to 0.8min 0.1, max 1.0

Removed Parameters:

*carbon.max.executor.startup.timeout* – Remove it from IDG



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (CARBONDATA-2043) Configurable wait time for requesting executors and minimum registered executors ratio to continue the block distribution

2018-01-17 Thread Mohammad Shahid Khan (JIRA)
Mohammad Shahid Khan created CARBONDATA-2043:


 Summary: Configurable wait time for requesting executors and 
minimum registered executors ratio to continue the block distribution
 Key: CARBONDATA-2043
 URL: https://issues.apache.org/jira/browse/CARBONDATA-2043
 Project: CarbonData
  Issue Type: Improvement
Reporter: Mohammad Shahid Khan
Assignee: Mohammad Shahid Khan


* Configurable wait time for requesting executors and minimum registered 
executors ratio to continue the block distribution
- carbon.dynamicallocation.schedulertimeout : to configure wait time. defalt 
5sec, Min 5 sec and max 15 sec.
- carbon.scheduler.minregisteredresourcesratio : min 0.1, max 1.0 and default 
to 0.8 to configure minimum registered executors ratio.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (CARBONDATA-2009) REFRESH TABLE Limitation When HiveMetaStore is used

2018-01-10 Thread Mohammad Shahid Khan (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-2009?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mohammad Shahid Khan updated CARBONDATA-2009:
-
Description: Refresh table command will not register the carbon table if 
the old table is stored in the CarbonHiveMetastore  (was: Refresh table when 
spark.carbon.hive.schema.store is set to true ie when hive meta store is used.)

> REFRESH TABLE Limitation When HiveMetaStore is used
> ---
>
> Key: CARBONDATA-2009
> URL: https://issues.apache.org/jira/browse/CARBONDATA-2009
> Project: CarbonData
>  Issue Type: Sub-task
>Reporter: Mohammad Shahid Khan
>Assignee: Mohammad Arshad
>Priority: Minor
> Fix For: 1.3.0
>
>
> Refresh table command will not register the carbon table if the old table is 
> stored in the CarbonHiveMetastore



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Assigned] (CARBONDATA-2009) REFRESH TABLE Limitation When HiveMetaStore is used

2018-01-08 Thread Mohammad Shahid Khan (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-2009?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mohammad Shahid Khan reassigned CARBONDATA-2009:


Assignee: Mohammad Arshad

> REFRESH TABLE Limitation When HiveMetaStore is used
> ---
>
> Key: CARBONDATA-2009
> URL: https://issues.apache.org/jira/browse/CARBONDATA-2009
> Project: CarbonData
>  Issue Type: Sub-task
>Reporter: Mohammad Shahid Khan
>Assignee: Mohammad Arshad
>Priority: Minor
> Fix For: 1.3.0
>
>
> Refresh table when spark.carbon.hive.schema.store is set to true ie when hive 
> meta store is used.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (CARBONDATA-2009) REFRESH TABLE Limitation When HiveMetaStore is used

2018-01-08 Thread Mohammad Shahid Khan (JIRA)
Mohammad Shahid Khan created CARBONDATA-2009:


 Summary: REFRESH TABLE Limitation When HiveMetaStore is used
 Key: CARBONDATA-2009
 URL: https://issues.apache.org/jira/browse/CARBONDATA-2009
 Project: CarbonData
  Issue Type: Sub-task
Reporter: Mohammad Shahid Khan
Priority: Minor


Refresh table when spark.carbon.hive.schema.store is set to true ie when hive 
meta store is used.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (CARBONDATA-1955) Delta DataType calculation is incorrect for long type

2017-12-30 Thread Mohammad Shahid Khan (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-1955?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mohammad Shahid Khan updated CARBONDATA-1955:
-
Affects Version/s: 1.2.0

> Delta DataType calculation is incorrect for long type
> -
>
> Key: CARBONDATA-1955
> URL: https://issues.apache.org/jira/browse/CARBONDATA-1955
> Project: CarbonData
>  Issue Type: Bug
>Affects Versions: 1.2.0, 1.3.0
>Reporter: Mohammad Shahid Khan
>Assignee: Mohammad Shahid Khan
>




--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (CARBONDATA-1955) Delta DataType calculation is incorrect

2017-12-29 Thread Mohammad Shahid Khan (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-1955?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mohammad Shahid Khan updated CARBONDATA-1955:
-
Description: 



  was:

For deciding deltaDataType, adaptiveDataType should be passed to the calculate 
the deltaDataType otherwise it will end up with source type only.
DataType fitDelta(DataType dataType, Object max, Object min)
DataType  fitDeltaForDecimalType(DataType dataType, Object max, Object min, 
  DecimalConverterFactory.DecimalConverterType decimalConverterType)




> Delta DataType calculation is incorrect
> ---
>
> Key: CARBONDATA-1955
> URL: https://issues.apache.org/jira/browse/CARBONDATA-1955
> Project: CarbonData
>  Issue Type: Bug
>Affects Versions: 1.3.0
>Reporter: Mohammad Shahid Khan
>Assignee: Mohammad Shahid Khan
>




--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (CARBONDATA-1955) Delta DataType calculation is incorrect for long type

2017-12-29 Thread Mohammad Shahid Khan (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-1955?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mohammad Shahid Khan updated CARBONDATA-1955:
-
Summary: Delta DataType calculation is incorrect for long type  (was: Delta 
DataType calculation is incorrect)

> Delta DataType calculation is incorrect for long type
> -
>
> Key: CARBONDATA-1955
> URL: https://issues.apache.org/jira/browse/CARBONDATA-1955
> Project: CarbonData
>  Issue Type: Bug
>Affects Versions: 1.3.0
>Reporter: Mohammad Shahid Khan
>Assignee: Mohammad Shahid Khan
>




--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (CARBONDATA-1955) Delta DataType calculation is incorrect

2017-12-29 Thread Mohammad Shahid Khan (JIRA)
Mohammad Shahid Khan created CARBONDATA-1955:


 Summary: Delta DataType calculation is incorrect
 Key: CARBONDATA-1955
 URL: https://issues.apache.org/jira/browse/CARBONDATA-1955
 Project: CarbonData
  Issue Type: Bug
Affects Versions: 1.3.0
Reporter: Mohammad Shahid Khan
Assignee: Mohammad Shahid Khan



For deciding deltaDataType, adaptiveDataType should be passed to the calculate 
the deltaDataType otherwise it will end up with source type only.
DataType fitDelta(DataType dataType, Object max, Object min)
DataType  fitDeltaForDecimalType(DataType dataType, Object max, Object min, 
  DecimalConverterFactory.DecimalConverterType decimalConverterType)





--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (CARBONDATA-1948) Update help document for the change made for CARBONDATA-1929

2017-12-28 Thread Mohammad Shahid Khan (JIRA)
Mohammad Shahid Khan created CARBONDATA-1948:


 Summary: Update help document for the change made for 
CARBONDATA-1929
 Key: CARBONDATA-1948
 URL: https://issues.apache.org/jira/browse/CARBONDATA-1948
 Project: CarbonData
  Issue Type: Sub-task
Reporter: Mohammad Shahid Khan
Assignee: Gururaj Shetty
Priority: Minor


carbon.sort.file.write.buffer.size (minValue = 10 KB, maxValue=10MB, 
defaultValue =16 KB )
carbon.sort.intermediate.files.limit (minValue = 2, maxValue=50, defaultValue 
=20 )



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (CARBONDATA-1929) carbon property configuration validation

2017-12-21 Thread Mohammad Shahid Khan (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-1929?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mohammad Shahid Khan updated CARBONDATA-1929:
-
Description: 
 Issue 1 :
  In carbon.properties carbon.timestamp.format=testserwerwerwer12312312312312 
is configured and spark submit server is started using the command - 
bin/spark-submit --master yarn-client  --executor-memory 10G --executor-cores 5 
--driver-memory 5G --num-executors 3 --class 
org.apache.carbondata.spark.thriftserver.CarbonThriftServer --jars 
/srv/spark2.2Bigdata/install/spark/sparkJdbc/carbonlib/carbondata-common-1.3.0.0100.jar,/srv/spark2.2Bigdata/install/spark/sparkJdbc/carbonlib/carbondata-core-1.3.0.0100.jar,/srv/spark2.2Bigdata/install/spark/sparkJdbc/carbonlib/carbondata-format-1.3.0.0100.jar,/srv/spark2.2Bigdata/install/spark/sparkJdbc/carbonlib/carbondata-hadoop-1.3.0.0100.jar,/srv/spark2.2Bigdata/install/spark/sparkJdbc/carbonlib/carbondata-processing-1.3.0.0100.jar,/srv/spark2.2Bigdata/install/spark/sparkJdbc/carbonlib/carbondata-spark-common-1.3.0.0100.jar,/srv/spark2.2Bigdata/install/spark/sparkJdbc/carbonlib/carbondata-streaming-1.3.0.0100.jar
 
/srv/spark2.2Bigdata/install/spark/sparkJdbc/carbonlib/carbondata-spark2-1.3.0.0100.jar
 "hdfs://hacluster/carbonstore"

  From Beeline User tries to create a table 
Issue 2 :
 In carbon.properties the below properties is configured and spark submit 
server is started using the command 
 carbon.sort.file.write.buffer.size=test 
 carbon.sort.intermediate.files.limit=test

>From Beeline User creates a table 
User tries to load data.


Issue 3:
In carbon.properties the below properties is configured and spark submit server 
is started using the command 
carbon.block.meta.size.reserved.percentage=test 
>From beeline user creates table and tries to load data.


  was:
 Issue 1 :
  In carbon.properties carbon.timestamp.format=testserwerwerwer12312312312312 
is configured and spark submit server is started using the command - 
bin/spark-submit --master yarn-client  --executor-memory 10G --executor-cores 5 
--driver-memory 5G --num-executors 3 --class 
org.apache.carbondata.spark.thriftserver.CarbonThriftServer --jars 
/srv/spark2.2Bigdata/install/spark/sparkJdbc/carbonlib/carbondata-common-1.3.0.0100.jar,/srv/spark2.2Bigdata/install/spark/sparkJdbc/carbonlib/carbondata-core-1.3.0.0100.jar,/srv/spark2.2Bigdata/install/spark/sparkJdbc/carbonlib/carbondata-format-1.3.0.0100.jar,/srv/spark2.2Bigdata/install/spark/sparkJdbc/carbonlib/carbondata-hadoop-1.3.0.0100.jar,/srv/spark2.2Bigdata/install/spark/sparkJdbc/carbonlib/carbondata-processing-1.3.0.0100.jar,/srv/spark2.2Bigdata/install/spark/sparkJdbc/carbonlib/carbondata-spark-common-1.3.0.0100.jar,/srv/spark2.2Bigdata/install/spark/sparkJdbc/carbonlib/carbondata-streaming-1.3.0.0100.jar
 
/srv/spark2.2Bigdata/install/spark/sparkJdbc/carbonlib/carbondata-spark2-1.3.0.0100.jar
 "hdfs://hacluster/carbonstore"

  From Beeline User tries to create a table 
  Create table parsing fails as shown below in snap.
   
http://dts.huawei.com/net/dts/fckeditor/download.ashx?Path=sEDRdBxs6%2f6a0ax2Ss3KJMdclw69E%2f%2bmwFVkT%2f6WxsOF1uA6IA71DkJFN3iXJAG7vplZsvI%2bl7M%3d

Issue 2 :
 In carbon.properties the below properties is configured and spark submit 
server is started using the command 
 carbon.sort.file.write.buffer.size=test 
 carbon.sort.intermediate.files.limit=test

>From Beeline User creates a table 
User tries to load data.
Issue :Data loading fails as shown in snap.
http://dts.huawei.com/net/dts/fckeditor/download.ashx?Path=sEDRdBxs6%2f6a0ax2Ss3KJMdclw69E%2f%2bmwFVkT%2f6WxsOF1uA6IA71DjeZqxHmMLFdeVr460BZPL4%3d


Issue 3:
In carbon.properties the below properties is configured and spark submit server 
is started using the command 
carbon.block.meta.size.reserved.percentage=test 
>From beeline user creates table and tries to load data.
Issue :Data loading fails as shown in snap.
http://dts.huawei.com/net/dts/fckeditor/download.ashx?Path=sEDRdBxs6%2f6a0ax2Ss3KJMdclw69E%2f%2bmwFVkT%2f6WxsOF1uA6IA71Ds5xK0%2fNEhBDzhdFVtoc3sg%3d

 


> carbon property configuration validation
> 
>
> Key: CARBONDATA-1929
> URL: https://issues.apache.org/jira/browse/CARBONDATA-1929
> Project: CarbonData
>  Issue Type: Bug
>Reporter: Mohammad Shahid Khan
>
>  Issue 1 :
>   In carbon.properties carbon.timestamp.format=testserwerwerwer12312312312312 
> is configured and spark submit server is started using the command - 
> bin/spark-submit --master yarn-client  --executor-memory 10G --executor-cores 
> 5 --driver-memory 5G --num-executors 3 --class 
> org.apache.carbondata.spark.thriftserver.CarbonThriftServer --jars 
> 

[jira] [Created] (CARBONDATA-1929) carbon property configuration validation

2017-12-21 Thread Mohammad Shahid Khan (JIRA)
Mohammad Shahid Khan created CARBONDATA-1929:


 Summary: carbon property configuration validation
 Key: CARBONDATA-1929
 URL: https://issues.apache.org/jira/browse/CARBONDATA-1929
 Project: CarbonData
  Issue Type: Bug
Reporter: Mohammad Shahid Khan


 Issue 1 :
  In carbon.properties carbon.timestamp.format=testserwerwerwer12312312312312 
is configured and spark submit server is started using the command - 
bin/spark-submit --master yarn-client  --executor-memory 10G --executor-cores 5 
--driver-memory 5G --num-executors 3 --class 
org.apache.carbondata.spark.thriftserver.CarbonThriftServer --jars 
/srv/spark2.2Bigdata/install/spark/sparkJdbc/carbonlib/carbondata-common-1.3.0.0100.jar,/srv/spark2.2Bigdata/install/spark/sparkJdbc/carbonlib/carbondata-core-1.3.0.0100.jar,/srv/spark2.2Bigdata/install/spark/sparkJdbc/carbonlib/carbondata-format-1.3.0.0100.jar,/srv/spark2.2Bigdata/install/spark/sparkJdbc/carbonlib/carbondata-hadoop-1.3.0.0100.jar,/srv/spark2.2Bigdata/install/spark/sparkJdbc/carbonlib/carbondata-processing-1.3.0.0100.jar,/srv/spark2.2Bigdata/install/spark/sparkJdbc/carbonlib/carbondata-spark-common-1.3.0.0100.jar,/srv/spark2.2Bigdata/install/spark/sparkJdbc/carbonlib/carbondata-streaming-1.3.0.0100.jar
 
/srv/spark2.2Bigdata/install/spark/sparkJdbc/carbonlib/carbondata-spark2-1.3.0.0100.jar
 "hdfs://hacluster/carbonstore"

  From Beeline User tries to create a table 
  Create table parsing fails as shown below in snap.
   
http://dts.huawei.com/net/dts/fckeditor/download.ashx?Path=sEDRdBxs6%2f6a0ax2Ss3KJMdclw69E%2f%2bmwFVkT%2f6WxsOF1uA6IA71DkJFN3iXJAG7vplZsvI%2bl7M%3d

Issue 2 :
 In carbon.properties the below properties is configured and spark submit 
server is started using the command 
 carbon.sort.file.write.buffer.size=test 
 carbon.sort.intermediate.files.limit=test

>From Beeline User creates a table 
User tries to load data.
Issue :Data loading fails as shown in snap.
http://dts.huawei.com/net/dts/fckeditor/download.ashx?Path=sEDRdBxs6%2f6a0ax2Ss3KJMdclw69E%2f%2bmwFVkT%2f6WxsOF1uA6IA71DjeZqxHmMLFdeVr460BZPL4%3d


Issue 3:
In carbon.properties the below properties is configured and spark submit server 
is started using the command 
carbon.block.meta.size.reserved.percentage=test 
>From beeline user creates table and tries to load data.
Issue :Data loading fails as shown in snap.
http://dts.huawei.com/net/dts/fckeditor/download.ashx?Path=sEDRdBxs6%2f6a0ax2Ss3KJMdclw69E%2f%2bmwFVkT%2f6WxsOF1uA6IA71Ds5xK0%2fNEhBDzhdFVtoc3sg%3d

 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Resolved] (CARBONDATA-929) adding optional field number of blocklet in the carbonindex file

2017-12-20 Thread Mohammad Shahid Khan (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-929?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mohammad Shahid Khan resolved CARBONDATA-929.
-
Resolution: Won't Fix

Not required for current design. 
Please reopen if required in future.

> adding optional field number of blocklet in the carbonindex file 
> -
>
> Key: CARBONDATA-929
> URL: https://issues.apache.org/jira/browse/CARBONDATA-929
> Project: CarbonData
>  Issue Type: Improvement
>Reporter: Mohammad Shahid Khan
>Assignee: Mohammad Shahid Khan
>  Time Spent: 2h 50m
>  Remaining Estimate: 0h
>
> The num_blocklet needed to get the number of Blocklet within a block at the 
> driver
> {code}
>  /**
>  *  Block index information stored in index file for every block
>  */
> struct BlockIndex{
>   1: required i64 num_rows; // Total number of rows in this file
>   2: required string file_name; // Block file name
>   3: required i64 offset; // Offset of the footer
>   4: required carbondata.BlockletIndex block_index;   // Blocklet index
>   5: optional i32 num_blocklet; // total number of Blocklet
> }
> {code}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (CARBONDATA-1913) Global Sort data dataload fails for big with RPC timeout exception

2017-12-19 Thread Mohammad Shahid Khan (JIRA)
Mohammad Shahid Khan created CARBONDATA-1913:


 Summary: Global Sort data dataload fails for big with RPC timeout 
exception 
 Key: CARBONDATA-1913
 URL: https://issues.apache.org/jira/browse/CARBONDATA-1913
 Project: CarbonData
  Issue Type: Bug
Reporter: Mohammad Shahid Khan
Assignee: Mohammad Shahid Khan


When gloabl sort option is used for big data then for some times it fails with
RPC timeout after 120s. 

This is happening because the driver is not able to unpersist rdd cache with in 
120s.
The issue is happening due to rdd unpersist blocking call. Sometimes spark is 
not able 
to unppersist the rdd in default "spark.rpc.askTimeout" or 
"spark.network.timeout" time.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (CARBONDATA-1341) the data load not failing for bigdecimal datatype even though if action is fail and data is not in valid range

2017-12-19 Thread Mohammad Shahid Khan (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-1341?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mohammad Shahid Khan updated CARBONDATA-1341:
-
Description: 
Create table t1(c1 Decimal(3,2), c2 String) stored by 'carbondata'
For column c1, the expected allowed data should be 
2
2.1
2.11
Invalid data
2.111

Actual: But when we load / insert data 2.111 carbon is storing as null, when 
action is fail.
Expected : The load should fail if bad record action is fail.

> the data load not failing for bigdecimal datatype even though if action is 
> fail and data is not in valid range
> --
>
> Key: CARBONDATA-1341
> URL: https://issues.apache.org/jira/browse/CARBONDATA-1341
> Project: CarbonData
>  Issue Type: Bug
>Reporter: Mohammad Shahid Khan
>Priority: Minor
>
> Create table t1(c1 Decimal(3,2), c2 String) stored by 'carbondata'
> For column c1, the expected allowed data should be 
> 2
> 2.1
> 2.11
> Invalid data
> 2.111
> Actual: But when we load / insert data 2.111 carbon is storing as null, when 
> action is fail.
> Expected : The load should fail if bad record action is fail.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (CARBONDATA-1887) block pruning not happening is carbon for ShortType and SmallIntType columns

2017-12-12 Thread Mohammad Shahid Khan (JIRA)
Mohammad Shahid Khan created CARBONDATA-1887:


 Summary: block pruning not happening is carbon for ShortType and 
SmallIntType columns
 Key: CARBONDATA-1887
 URL: https://issues.apache.org/jira/browse/CARBONDATA-1887
 Project: CarbonData
  Issue Type: Bug
Reporter: Mohammad Shahid Khan
Assignee: Mohammad Shahid Khan




spark.sql(
  s"""
 | create table test_numeric_type(c1 int, c2 long, c3 smallint, c4 
bigint, c5 short) stored by 'carbondata'
   """.stripMargin).show()
spark.sql(
  s"""
 | insert into test_numeric_type select 
1,111,111,11,'2019-01-03 12:12:12'
   """.stripMargin).show()

spark.sql(
  s"""
 | insert into test_numeric_type select 
2,222,222,22,'2020-01-03 12:12:12'
   """.stripMargin).show()

spark.sql(
  s"""
 | insert into test_numeric_type select 
3,333,333,33,'2021-01-03 12:12:12'
   """.stripMargin).show()

spark.sql(
  s"""
 | select * from test_numeric_type where c5>
   """.stripMargin).show()

Only two blocks should be selected but all blocks are selected during query 
execution.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (CARBONDATA-1293) update on carbon data failed with carbon.update.persist.enable false

2017-12-07 Thread Mohammad Shahid Khan (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-1293?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mohammad Shahid Khan updated CARBONDATA-1293:
-
Priority: Major  (was: Critical)

> update on carbon data failed with carbon.update.persist.enable false
> 
>
> Key: CARBONDATA-1293
> URL: https://issues.apache.org/jira/browse/CARBONDATA-1293
> Project: CarbonData
>  Issue Type: Bug
>  Components: data-load
>Reporter: Mohammad Shahid Khan
>  Time Spent: 1.5h
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (CARBONDATA-1849) n Create pre aggregate table failled

2017-12-04 Thread Mohammad Shahid Khan (JIRA)
Mohammad Shahid Khan created CARBONDATA-1849:


 Summary: n Create pre aggregate table failled
 Key: CARBONDATA-1849
 URL: https://issues.apache.org/jira/browse/CARBONDATA-1849
 Project: CarbonData
  Issue Type: Bug
  Components: spark-integration
Reporter: Mohammad Shahid Khan


Create pre aggregate table is failing 
   spark.sql("""create table carbontable (c1 string,c2 int,c3 string,c5 string) 
STORED BY 'org.apache.carbondata.format'""")
spark.sql("insert into carbontable select 'a',1,'aa','aaa'")
spark. sql("insert into carbontable select 'b',1,'aa','aaa'")
spark.sql("insert into carbontable select 'a',10,'aa','aaa'")
spark.sql("create datamap preagg1 on table carbontable using 'preaggregate' 
as select c1,sum(c2) from carbontable group by c1")

Exception:
{code}
Exception in thread "main" 
org.apache.carbondata.processing.loading.exception.CarbonDataLoadingException: 
CSV header in DDL is not proper. Column names in schema and CSV header are not 
the same.
at 
org.apache.carbondata.spark.util.CommonUtil$.getCsvHeaderColumns(CommonUtil.scala:637)
at 
org.apache.carbondata.spark.util.DataLoadingUtil$.buildCarbonLoadModel(DataLoadingUtil.scala:312)
at 
org.apache.spark.sql.execution.command.management.CarbonLoadDataCommand.processData(CarbonLoadDataCommand.scala:133)
at 
org.apache.spark.sql.execution.command.DataCommand.run(package.scala:71)
at 
org.apache.spark.sql.execution.command.preaaggregate.CreatePreAggregateTableCommand.processData(CreatePreAggregateTableCommand.scala:139)
at 
org.apache.spark.sql.execution.command.datamap.CarbonCreateDataMapCommand.processData(CarbonCreateDataMapCommand.scala:80)
at 
org.apache.spark.sql.execution.command.AtomicRunnableCommand.run(package.scala:86)
at 
org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:58)
at 
org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:56)
at 
org.apache.spark.sql.execution.command.ExecutedCommandExec.doExecute(commands.scala:74)
at 
org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:114)
at 
org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:114)
at 
org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:135)
at 
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
at 
org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:132)
at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:113)
at 
org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:87)
at 
org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:87)
at org.apache.spark.sql.Dataset.(Dataset.scala:185)
at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:64)
at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:592)
at 
org.apache.carbondata.examples.CarbonSessionExample$.main(CarbonSessionExample.scala:35)
at 
org.apache.carbondata.examples.CarbonSessionExample.main(CarbonSessionExample.scala)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at com.intellij.rt.execution.application.AppMain.main(AppMain.java:147)
{code}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (CARBONDATA-1822) Support DDL to register the CarbonData table from existing carbon table data

2017-11-28 Thread Mohammad Shahid Khan (JIRA)
Mohammad Shahid Khan created CARBONDATA-1822:


 Summary: Support DDL to register the CarbonData table from 
existing carbon table data
 Key: CARBONDATA-1822
 URL: https://issues.apache.org/jira/browse/CARBONDATA-1822
 Project: CarbonData
  Issue Type: New Feature
  Components: spark-integration
Reporter: Mohammad Shahid Khan
Assignee: Mohammad Shahid Khan


For details please refer to the below link
http://apache-carbondata-dev-mailing-list-archive.1130556.n5.nabble.com/DDL-for-CarbonData-table-backup-and-recovery-new-feature-td27854.html



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (CARBONDATA-1249) Wrong order of columns in redirected csv for bad records

2017-11-21 Thread Mohammad Shahid Khan (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-1249?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mohammad Shahid Khan updated CARBONDATA-1249:
-
Summary: Wrong order of columns in redirected csv for bad records  (was: 
Wrong order of columns i n redirected csv for bad records)

> Wrong order of columns in redirected csv for bad records
> 
>
> Key: CARBONDATA-1249
> URL: https://issues.apache.org/jira/browse/CARBONDATA-1249
> Project: CarbonData
>  Issue Type: Bug
>Reporter: Mohammad Shahid Khan
>Assignee: Mohammad Shahid Khan
>Priority: Critical
>  Time Spent: 6h 10m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (CARBONDATA-1249) Wrong order of columns in redirected csv for bad records

2017-11-21 Thread Mohammad Shahid Khan (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-1249?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mohammad Shahid Khan updated CARBONDATA-1249:
-
Priority: Minor  (was: Critical)

> Wrong order of columns in redirected csv for bad records
> 
>
> Key: CARBONDATA-1249
> URL: https://issues.apache.org/jira/browse/CARBONDATA-1249
> Project: CarbonData
>  Issue Type: Bug
>Reporter: Mohammad Shahid Khan
>Assignee: Mohammad Shahid Khan
>Priority: Minor
>  Time Spent: 6h 10m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (CARBONDATA-1727) Dataload is successful even in case if the table is droped from other client.

2017-11-15 Thread Mohammad Shahid Khan (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-1727?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mohammad Shahid Khan updated CARBONDATA-1727:
-
Summary: Dataload is successful even in case if the table is droped from 
other client.  (was: Dataload is successful even in case if some user drops the 
table from other client.)

> Dataload is successful even in case if the table is droped from other client.
> -
>
> Key: CARBONDATA-1727
> URL: https://issues.apache.org/jira/browse/CARBONDATA-1727
> Project: CarbonData
>  Issue Type: Bug
>  Components: data-load, spark-integration
>Affects Versions: 1.2.0, 1.3.0
>Reporter: Mohammad Shahid Khan
>Assignee: Mohammad Shahid Khan
>Priority: Minor
>
> table drop has highest priority so even if on some table load operation is in 
> progress, 
> the table could be droped.
> If before finishing the load operation the table is dropped then the load 
> should fail.
> Steps:
> 1. Create table t1
> 2. Load data into t1 (big data that takes some time)
> 3. when load in progress, then drop the table
> Actual Result : The load is successful
> Expected Result : Final Load status should be fail.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (CARBONDATA-1727) Dataload is successful even in case if some user drops the table from other client.

2017-11-15 Thread Mohammad Shahid Khan (JIRA)
Mohammad Shahid Khan created CARBONDATA-1727:


 Summary: Dataload is successful even in case if some user drops 
the table from other client.
 Key: CARBONDATA-1727
 URL: https://issues.apache.org/jira/browse/CARBONDATA-1727
 Project: CarbonData
  Issue Type: Bug
  Components: data-load, spark-integration
Affects Versions: 1.2.0, 1.3.0
Reporter: Mohammad Shahid Khan
Assignee: Mohammad Shahid Khan
Priority: Minor


table drop has highest priority so even if on some table load operation is in 
progress, 
the table could be droped.
If before finishing the load operation the table is dropped then the load 
should fail.

Steps:
1. Create table t1
2. Load data into t1 (big data that takes some time)
3. when load in progress, then drop the table
Actual Result : The load is successful

Expected Result : Final Load status should be fail.




--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (CARBONDATA-1573) Support Database Location Configuration while Creating Database/ Support Creation of carbon Table in the database location

2017-10-26 Thread Mohammad Shahid Khan (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-1573?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mohammad Shahid Khan updated CARBONDATA-1573:
-
Summary: Support Database Location Configuration while Creating Database/ 
Support Creation of carbon Table in the database location  (was: Support 
Database Location Configuration while Creating Database)

> Support Database Location Configuration while Creating Database/ Support 
> Creation of carbon Table in the database location
> --
>
> Key: CARBONDATA-1573
> URL: https://issues.apache.org/jira/browse/CARBONDATA-1573
> Project: CarbonData
>  Issue Type: New Feature
>  Components: core, data-load, data-query, hadoop-integration, 
> presto-integration, spark-integration
>Reporter: Mohammad Shahid Khan
>Assignee: Mohammad Shahid Khan
>
> Support Creation of carbon table at the database location
> *Please refer to  for Design and discussion:*
> http://apache-carbondata-dev-mailing-list-archive.1130556.n5.nabble.com/DISCUSSION-Support-Database-Location-Configuration-while-Creating-Database-td23492.html



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Issue Comment Deleted] (CARBONDATA-1507) Dataload global sort fails with RpcTimeOutException: Futures timed out after [120 seconds]

2017-09-22 Thread Mohammad Shahid Khan (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-1507?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mohammad Shahid Khan updated CARBONDATA-1507:
-
Comment: was deleted

(was: This happening due to rdd unpersist  blocking calll. Sometime the RDD 
unpersist is not 
happening in  default "spark.rp.askTimeout" or "spark.network.timeout" time.
Solution, we may make the unpersist call as non blocking. And doing this will 
not have any functional impact 
as spark automatically monitors the cache usage on each node and drops out.)

> Dataload global sort fails with RpcTimeOutException: Futures timed out after 
> [120 seconds]
> --
>
> Key: CARBONDATA-1507
> URL: https://issues.apache.org/jira/browse/CARBONDATA-1507
> Project: CarbonData
>  Issue Type: Bug
>Reporter: Mohammad Shahid Khan
>Assignee: Mohammad Shahid Khan
>
> Dataload global sort fails with RpcTimeOutException: Futures timed out after 
> [120 seconds]



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Issue Comment Deleted] (CARBONDATA-1507) Dataload global sort fails with RpcTimeOutException: Futures timed out after [120 seconds]

2017-09-22 Thread Mohammad Shahid Khan (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-1507?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mohammad Shahid Khan updated CARBONDATA-1507:
-
Comment: was deleted

(was: This happening due to rdd unpersist  blocking calll. Sometime the RDD 
unpersist is not 
happening in  default "spark.rp.askTimeout" or "spark.network.timeout" time.
Solution, we may make the unpersist call as non blocking. And doing this will 
not have any functional impact 
as spark automatically monitors the cache usage on each node and drops out.)

> Dataload global sort fails with RpcTimeOutException: Futures timed out after 
> [120 seconds]
> --
>
> Key: CARBONDATA-1507
> URL: https://issues.apache.org/jira/browse/CARBONDATA-1507
> Project: CarbonData
>  Issue Type: Bug
>Reporter: Mohammad Shahid Khan
>Assignee: Mohammad Shahid Khan
>
> Dataload global sort fails with RpcTimeOutException: Futures timed out after 
> [120 seconds]



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (CARBONDATA-1507) Dataload global sort fails with RpcTimeOutException: Futures timed out after [120 seconds]

2017-09-22 Thread Mohammad Shahid Khan (JIRA)

[ 
https://issues.apache.org/jira/browse/CARBONDATA-1507?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16175990#comment-16175990
 ] 

Mohammad Shahid Khan commented on CARBONDATA-1507:
--

This happening due to rdd unpersist  blocking calll. Sometime the RDD unpersist 
is not 
happening in  default "spark.rp.askTimeout" or "spark.network.timeout" time.
Solution, we may make the unpersist call as non blocking. And doing this will 
not have any functional impact 
as spark automatically monitors the cache usage on each node and drops out.

> Dataload global sort fails with RpcTimeOutException: Futures timed out after 
> [120 seconds]
> --
>
> Key: CARBONDATA-1507
> URL: https://issues.apache.org/jira/browse/CARBONDATA-1507
> Project: CarbonData
>  Issue Type: Bug
>Reporter: Mohammad Shahid Khan
>Assignee: Mohammad Shahid Khan
>
> Dataload global sort fails with RpcTimeOutException: Futures timed out after 
> [120 seconds]



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (CARBONDATA-1507) Dataload global sort fails with RpcTimeOutException: Futures timed out after [120 seconds]

2017-09-22 Thread Mohammad Shahid Khan (JIRA)

[ 
https://issues.apache.org/jira/browse/CARBONDATA-1507?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16175989#comment-16175989
 ] 

Mohammad Shahid Khan commented on CARBONDATA-1507:
--

This happening due to rdd unpersist  blocking calll. Sometime the RDD unpersist 
is not 
happening in  default "spark.rp.askTimeout" or "spark.network.timeout" time.
Solution, we may make the unpersist call as non blocking. And doing this will 
not have any functional impact 
as spark automatically monitors the cache usage on each node and drops out.

> Dataload global sort fails with RpcTimeOutException: Futures timed out after 
> [120 seconds]
> --
>
> Key: CARBONDATA-1507
> URL: https://issues.apache.org/jira/browse/CARBONDATA-1507
> Project: CarbonData
>  Issue Type: Bug
>Reporter: Mohammad Shahid Khan
>Assignee: Mohammad Shahid Khan
>
> Dataload global sort fails with RpcTimeOutException: Futures timed out after 
> [120 seconds]



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (CARBONDATA-1507) Dataload global sort fails with RpcTimeOutException: Futures timed out after [120 seconds]

2017-09-22 Thread Mohammad Shahid Khan (JIRA)

[ 
https://issues.apache.org/jira/browse/CARBONDATA-1507?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16175988#comment-16175988
 ] 

Mohammad Shahid Khan commented on CARBONDATA-1507:
--

This happening due to rdd unpersist  blocking calll. Sometime the RDD unpersist 
is not 
happening in  default "spark.rp.askTimeout" or "spark.network.timeout" time.
Solution, we may make the unpersist call as non blocking. And doing this will 
not have any functional impact 
as spark automatically monitors the cache usage on each node and drops out.

> Dataload global sort fails with RpcTimeOutException: Futures timed out after 
> [120 seconds]
> --
>
> Key: CARBONDATA-1507
> URL: https://issues.apache.org/jira/browse/CARBONDATA-1507
> Project: CarbonData
>  Issue Type: Bug
>Reporter: Mohammad Shahid Khan
>Assignee: Mohammad Shahid Khan
>
> Dataload global sort fails with RpcTimeOutException: Futures timed out after 
> [120 seconds]



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (CARBONDATA-1507) Dataload global sort fails with RpcTimeOutException: Futures timed out after [120 seconds]

2017-09-22 Thread Mohammad Shahid Khan (JIRA)
Mohammad Shahid Khan created CARBONDATA-1507:


 Summary: Dataload global sort fails with RpcTimeOutException: 
Futures timed out after [120 seconds]
 Key: CARBONDATA-1507
 URL: https://issues.apache.org/jira/browse/CARBONDATA-1507
 Project: CarbonData
  Issue Type: Bug
Reporter: Mohammad Shahid Khan
Assignee: Mohammad Shahid Khan


Dataload global sort fails with RpcTimeOutException: Futures timed out after 
[120 seconds]






--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (CARBONDATA-1341) the data load not failing for bigdecimal datatype even though if action is fail and data is not in valid range

2017-07-29 Thread Mohammad Shahid Khan (JIRA)
Mohammad Shahid Khan created CARBONDATA-1341:


 Summary: the data load not failing for bigdecimal datatype even 
though if action is fail and data is not in valid range
 Key: CARBONDATA-1341
 URL: https://issues.apache.org/jira/browse/CARBONDATA-1341
 Project: CarbonData
  Issue Type: Bug
Reporter: Mohammad Shahid Khan
Priority: Minor






--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Assigned] (CARBONDATA-1309) During Global Sort the DataWriterProcessorStepImpl not getting shutdown even when datalaod is finished.

2017-07-18 Thread Mohammad Shahid Khan (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-1309?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mohammad Shahid Khan reassigned CARBONDATA-1309:


Assignee: Mohammad Shahid Khan

> During Global Sort the DataWriterProcessorStepImpl not getting shutdown even 
> when datalaod is finished.
> ---
>
> Key: CARBONDATA-1309
> URL: https://issues.apache.org/jira/browse/CARBONDATA-1309
> Project: CarbonData
>  Issue Type: Bug
>Reporter: Mohammad Shahid Khan
>Assignee: Mohammad Shahid Khan
>
> Global Sort is not closing the DataWriterProcessorStepImpl because of this 
> the thread started to print the number of records processed by the step at 
> the interval of 10 second is not getting shutdown.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (CARBONDATA-1309) During Global Sort the DataWriterProcessorStepImpl not getting shutdown even when datalaod is finished.

2017-07-18 Thread Mohammad Shahid Khan (JIRA)
Mohammad Shahid Khan created CARBONDATA-1309:


 Summary: During Global Sort the DataWriterProcessorStepImpl not 
getting shutdown even when datalaod is finished.
 Key: CARBONDATA-1309
 URL: https://issues.apache.org/jira/browse/CARBONDATA-1309
 Project: CarbonData
  Issue Type: Bug
Reporter: Mohammad Shahid Khan


Global Sort is not closing the DataWriterProcessorStepImpl because of this the 
thread started to print the number of records processed by the step at the 
interval of 10 second is not getting shutdown.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (CARBONDATA-1304) Support IUD with single_pass

2017-07-12 Thread Mohammad Shahid Khan (JIRA)
Mohammad Shahid Khan created CARBONDATA-1304:


 Summary: Support IUD with single_pass
 Key: CARBONDATA-1304
 URL: https://issues.apache.org/jira/browse/CARBONDATA-1304
 Project: CarbonData
  Issue Type: Bug
  Components: data-load, data-query
Reporter: Mohammad Shahid Khan
Assignee: Mohammad Shahid Khan






--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (CARBONDATA-1293) update on carbon data failed with carbon.update.persist.enable false

2017-07-11 Thread Mohammad Shahid Khan (JIRA)
Mohammad Shahid Khan created CARBONDATA-1293:


 Summary: update on carbon data failed with 
carbon.update.persist.enable false
 Key: CARBONDATA-1293
 URL: https://issues.apache.org/jira/browse/CARBONDATA-1293
 Project: CarbonData
  Issue Type: Bug
  Components: data-load
Reporter: Mohammad Shahid Khan
Priority: Critical






--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (CARBONDATA-1283) Carbon should continue with the default value if wrong value is set for the configurable parameter.

2017-07-10 Thread Mohammad Shahid Khan (JIRA)
Mohammad Shahid Khan created CARBONDATA-1283:


 Summary: Carbon should continue with the default value if wrong 
value is set for the configurable parameter.
 Key: CARBONDATA-1283
 URL: https://issues.apache.org/jira/browse/CARBONDATA-1283
 Project: CarbonData
  Issue Type: Bug
Reporter: Mohammad Shahid Khan


Carbon should continue with the default value if wrong value is set for the 
configurable parameters.
for example for below parameters if wrong value is configured then the 
executions is failing in the respective scenarios.
enable.unsafe.sort
carbon.custom.block.distribution
carbon.enable.vector.reader
carbon.lock.type
carbon.csv.read.buffersize.byte 






--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (CARBONDATA-1258) CarbonData should not allow loading Date Type values violating the boundary condition ("0001-01-01" through "9999-12-31")

2017-07-03 Thread Mohammad Shahid Khan (JIRA)
Mohammad Shahid Khan created CARBONDATA-1258:


 Summary: CarbonData should not allow loading Date Type values 
violating the boundary condition ("0001-01-01" through "-12-31")
 Key: CARBONDATA-1258
 URL: https://issues.apache.org/jira/browse/CARBONDATA-1258
 Project: CarbonData
  Issue Type: Bug
Reporter: Mohammad Shahid Khan
Assignee: Mohammad Shahid Khan
Priority: Critical


CarbonData should not allow loading Date Type values violating the boundary 
condition ("0001-01-01" through "-12-31")
For Date Type the value  X being loaded must be   X >= "0001-01-01" and X <= 
"-12-31".





--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (CARBONDATA-1252) Add BAD_RECORD_PATH option in Load options section in the Carbon Help doc

2017-06-30 Thread Mohammad Shahid Khan (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-1252?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mohammad Shahid Khan updated CARBONDATA-1252:
-
Summary: Add BAD_RECORD_PATH option in Load options section in the Carbon 
Help doc  (was: Add BAD_RECORD_PATH options in Load options in the Carbon Help 
doc)

> Add BAD_RECORD_PATH option in Load options section in the Carbon Help doc
> -
>
> Key: CARBONDATA-1252
> URL: https://issues.apache.org/jira/browse/CARBONDATA-1252
> Project: CarbonData
>  Issue Type: Sub-task
>Reporter: Mohammad Shahid Khan
>Priority: Minor
> Fix For: 1.2.0
>
>




--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (CARBONDATA-1252) Add BAD_RECORD_PATH options in Load options in the Carbon Help doc

2017-06-30 Thread Mohammad Shahid Khan (JIRA)
Mohammad Shahid Khan created CARBONDATA-1252:


 Summary: Add BAD_RECORD_PATH options in Load options in the Carbon 
Help doc
 Key: CARBONDATA-1252
 URL: https://issues.apache.org/jira/browse/CARBONDATA-1252
 Project: CarbonData
  Issue Type: Sub-task
Reporter: Mohammad Shahid Khan
Priority: Minor






--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (CARBONDATA-1117) Update SET & RESET command details in online help documentation

2017-06-30 Thread Mohammad Shahid Khan (JIRA)

[ 
https://issues.apache.org/jira/browse/CARBONDATA-1117?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16069604#comment-16069604
 ] 

Mohammad Shahid Khan commented on CARBONDATA-1117:
--

Please refer PR #https://github.com/apache/carbondata/pull/972 for details 
For information
Updated supported parameters
set -v command displays all session parameter list along with default value and 
usage doc
Example : set -v;
Supported parameters for dynamic set.
carbon.options.bad.records.logger.enable
carbon.options.bad.records.action
carbon.options.is.empty.data.bad.record
carbon.options.sort.scope
carbon.options.batch.sort.size.inmb
carbon.options.single.pass
carbon.options.bad.record.path
carbon.options.global.sort.partitions
enable.unsafe.sort
carbon.custom.block.distribution

> Update SET & RESET command details in online help documentation
> ---
>
> Key: CARBONDATA-1117
> URL: https://issues.apache.org/jira/browse/CARBONDATA-1117
> Project: CarbonData
>  Issue Type: Sub-task
>  Components: spark-integration
>Reporter: Manohar Vanam
>Assignee: Gururaj Shetty
>
> Update SET & RESET command details in online help documentation
> 1. update syntax
> 2. Examples with use cases



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (CARBONDATA-1249) Wrong order of columns i n redirected csv for bad records

2017-06-29 Thread Mohammad Shahid Khan (JIRA)
Mohammad Shahid Khan created CARBONDATA-1249:


 Summary: Wrong order of columns i n redirected csv for bad records
 Key: CARBONDATA-1249
 URL: https://issues.apache.org/jira/browse/CARBONDATA-1249
 Project: CarbonData
  Issue Type: Bug
Reporter: Mohammad Shahid Khan
Assignee: Mohammad Shahid Khan
Priority: Critical






--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (CARBONDATA-1218) In case of data-load failure the BadRecordsLogger.badRecordEntry map holding the task Status is not removing the task Entry.

2017-06-23 Thread Mohammad Shahid Khan (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-1218?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mohammad Shahid Khan updated CARBONDATA-1218:
-
Description: 
In case of data-load failure the BadRecordsLogger.badRecordEntry map holding 
the task Status is not removing the task Entry.
Because of this the next load is getting  failed even though the data being 
loaded has no bad records.

The map entry must be removed after load completion either success or fail.

  was:
In case of data-load failure the BadRecordsLogger.badRecordEntry map holding 
the task Status is not removing the task Entry.
Because of this the next load is getting  failed even though the data being 
loaded has no bad records.


> In case of data-load failure the BadRecordsLogger.badRecordEntry map holding 
> the task Status is not removing the task Entry.
> 
>
> Key: CARBONDATA-1218
> URL: https://issues.apache.org/jira/browse/CARBONDATA-1218
> Project: CarbonData
>  Issue Type: Bug
>Reporter: Mohammad Shahid Khan
>Assignee: Mohammad Shahid Khan
>
> In case of data-load failure the BadRecordsLogger.badRecordEntry map holding 
> the task Status is not removing the task Entry.
> Because of this the next load is getting  failed even though the data being 
> loaded has no bad records.
> The map entry must be removed after load completion either success or fail.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Assigned] (CARBONDATA-1218) In case of data-load failure the BadRecordsLogger.badRecordEntry map holding the task Status is not removing the task Entry.

2017-06-23 Thread Mohammad Shahid Khan (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-1218?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mohammad Shahid Khan reassigned CARBONDATA-1218:


Assignee: Mohammad Shahid Khan

> In case of data-load failure the BadRecordsLogger.badRecordEntry map holding 
> the task Status is not removing the task Entry.
> 
>
> Key: CARBONDATA-1218
> URL: https://issues.apache.org/jira/browse/CARBONDATA-1218
> Project: CarbonData
>  Issue Type: Bug
>Reporter: Mohammad Shahid Khan
>Assignee: Mohammad Shahid Khan
>
> In case of data-load failure the BadRecordsLogger.badRecordEntry map holding 
> the task Status is not removing the task Entry.
> Because of this the next load is getting  failed even though the data being 
> loaded has no bad records.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (CARBONDATA-1218) In case of data-load failure the BadRecordsLogger.badRecordEntry map holding the task Status is not removing the task Entry.

2017-06-23 Thread Mohammad Shahid Khan (JIRA)
Mohammad Shahid Khan created CARBONDATA-1218:


 Summary: In case of data-load failure the 
BadRecordsLogger.badRecordEntry map holding the task Status is not removing the 
task Entry.
 Key: CARBONDATA-1218
 URL: https://issues.apache.org/jira/browse/CARBONDATA-1218
 Project: CarbonData
  Issue Type: Bug
Reporter: Mohammad Shahid Khan


In case of data-load failure the BadRecordsLogger.badRecordEntry map holding 
the task Status is not removing the task Entry.
Because of this the next load is getting  failed even though the data being 
loaded has no bad records.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (CARBONDATA-1210) Exception should be thrown on bad record logger failure to write to log file or csv file.

2017-06-21 Thread Mohammad Shahid Khan (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-1210?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mohammad Shahid Khan updated CARBONDATA-1210:
-
Summary: Exception should be thrown on bad record logger failure to write 
to log file or csv file.  (was: Exception should be thrown on bad record write 
failure to log file or csv file.)

> Exception should be thrown on bad record logger failure to write to log file 
> or csv file.
> -
>
> Key: CARBONDATA-1210
> URL: https://issues.apache.org/jira/browse/CARBONDATA-1210
> Project: CarbonData
>  Issue Type: Bug
>Reporter: Mohammad Shahid Khan
>Priority: Minor
>




--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (CARBONDATA-1210) Exception should be thrown on bad record write failure to log file or csv file.

2017-06-21 Thread Mohammad Shahid Khan (JIRA)
Mohammad Shahid Khan created CARBONDATA-1210:


 Summary: Exception should be thrown on bad record write failure to 
log file or csv file.
 Key: CARBONDATA-1210
 URL: https://issues.apache.org/jira/browse/CARBONDATA-1210
 Project: CarbonData
  Issue Type: Bug
Reporter: Mohammad Shahid Khan
Priority: Minor






--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (CARBONDATA-1207) Resource leak problem in CarbonDictionaryWriter

2017-06-21 Thread Mohammad Shahid Khan (JIRA)
Mohammad Shahid Khan created CARBONDATA-1207:


 Summary: Resource leak problem in CarbonDictionaryWriter
 Key: CARBONDATA-1207
 URL: https://issues.apache.org/jira/browse/CARBONDATA-1207
 Project: CarbonData
  Issue Type: Bug
  Components: data-load
Reporter: Mohammad Shahid Khan
Assignee: Mohammad Shahid Khan


If in load during dictionary generation some exception happens during the 
dictionary writing or dictionary meta data writing, then stream will never be 
closed.
This will load DOS on the incremental load.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (CARBONDATA-1101) Avoid widening between wrapper classes

2017-05-29 Thread Mohammad Shahid Khan (JIRA)
Mohammad Shahid Khan created CARBONDATA-1101:


 Summary: Avoid widening between wrapper classes
 Key: CARBONDATA-1101
 URL: https://issues.apache.org/jira/browse/CARBONDATA-1101
 Project: CarbonData
  Issue Type: Bug
Reporter: Mohammad Shahid Khan
Assignee: Mohammad Shahid Khan


UnsafeIntermediateFileMerger.writeDataTofile does widening between Short, 
Integer, and Long wrapper classes.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[jira] [Assigned] (CARBONDATA-1077) ColumnDict and ALL_DICTIONARY_PATH must be used with SINGLE_PASS='true'

2017-05-24 Thread Mohammad Shahid Khan (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-1077?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mohammad Shahid Khan reassigned CARBONDATA-1077:


Assignee: Mohammad Shahid Khan

> ColumnDict  and ALL_DICTIONARY_PATH must be used with SINGLE_PASS='true'
> 
>
> Key: CARBONDATA-1077
> URL: https://issues.apache.org/jira/browse/CARBONDATA-1077
> Project: CarbonData
>  Issue Type: Bug
>Reporter: Mohammad Shahid Khan
>Assignee: Mohammad Shahid Khan
>Priority: Minor
>  Time Spent: 0.5h
>  Remaining Estimate: 0h
>
> Add validation that ColumnDict  and ALL_DICTIONARY_PATH must be used with 
> SINGLE_PASS='true'.
> Problem:
> If columndict is used without single path, and if the external dictionary 
> file 
> does not have all the values present in the actual raw data to be loaded then 
> load will not have proper data.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[jira] [Commented] (CARBONDATA-1049) avoid logging data into log file

2017-05-12 Thread Mohammad Shahid Khan (JIRA)

[ 
https://issues.apache.org/jira/browse/CARBONDATA-1049?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16007723#comment-16007723
 ] 

Mohammad Shahid Khan commented on CARBONDATA-1049:
--

avoid logging raw data into driver and executor log

> avoid logging data into log file
> 
>
> Key: CARBONDATA-1049
> URL: https://issues.apache.org/jira/browse/CARBONDATA-1049
> Project: CarbonData
>  Issue Type: Bug
>Reporter: Mohammad Shahid Khan
>




--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[jira] [Created] (CARBONDATA-1049) avoid logging data into log file

2017-05-12 Thread Mohammad Shahid Khan (JIRA)
Mohammad Shahid Khan created CARBONDATA-1049:


 Summary: avoid logging data into log file
 Key: CARBONDATA-1049
 URL: https://issues.apache.org/jira/browse/CARBONDATA-1049
 Project: CarbonData
  Issue Type: Bug
Reporter: Mohammad Shahid Khan






--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[jira] [Assigned] (CARBONDATA-991) add test case for validation the bad record log file creation at the configured bad store location

2017-04-26 Thread Mohammad Shahid Khan (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-991?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mohammad Shahid Khan reassigned CARBONDATA-991:
---

Assignee: Mohammad Shahid Khan

> add test case for validation the bad record log file creation at the 
> configured bad store location
> --
>
> Key: CARBONDATA-991
> URL: https://issues.apache.org/jira/browse/CARBONDATA-991
> Project: CarbonData
>  Issue Type: Test
>Reporter: Mohammad Shahid Khan
>Assignee: Mohammad Shahid Khan
>Priority: Trivial
>  Time Spent: 20m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[jira] [Created] (CARBONDATA-991) add test case for validation the bad record log file creation at the configured bad store location

2017-04-26 Thread Mohammad Shahid Khan (JIRA)
Mohammad Shahid Khan created CARBONDATA-991:
---

 Summary: add test case for validation the bad record log file 
creation at the configured bad store location
 Key: CARBONDATA-991
 URL: https://issues.apache.org/jira/browse/CARBONDATA-991
 Project: CarbonData
  Issue Type: Bug
Reporter: Mohammad Shahid Khan
Priority: Trivial






--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[jira] [Updated] (CARBONDATA-991) add test case for validation the bad record log file creation at the configured bad store location

2017-04-26 Thread Mohammad Shahid Khan (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-991?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mohammad Shahid Khan updated CARBONDATA-991:

Issue Type: Test  (was: Bug)

> add test case for validation the bad record log file creation at the 
> configured bad store location
> --
>
> Key: CARBONDATA-991
> URL: https://issues.apache.org/jira/browse/CARBONDATA-991
> Project: CarbonData
>  Issue Type: Test
>Reporter: Mohammad Shahid Khan
>Priority: Trivial
>




--
This message was sent by Atlassian JIRA
(v6.3.15#6346)