[ 
https://issues.apache.org/jira/browse/CARBONDATA-1914?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Venkata Ramana G resolved CARBONDATA-1914.
------------------------------------------
       Resolution: Fixed
         Assignee: sounak chakraborty
    Fix Version/s: 1.3.0

> Dictionary Cache Access Count Maintenance
> -----------------------------------------
>
>                 Key: CARBONDATA-1914
>                 URL: https://issues.apache.org/jira/browse/CARBONDATA-1914
>             Project: CarbonData
>          Issue Type: Bug
>            Reporter: sounak chakraborty
>            Assignee: sounak chakraborty
>             Fix For: 1.3.0
>
>          Time Spent: 3h 20m
>  Remaining Estimate: 0h
>
> When dictionary cache is being accessed the access count is being atomicaly 
> increased and after the access it has to be decremented. This access count 
> prevent it from removal from cache when cache overflow scenario comes into 
> picture. 
> There are some code places where access is incremented but not decremented 
> after usage is completed. This is a resource leak. 
> Test Case
> =======
> Precondition 
> ----------------
> (CarbonCommonConstants.CARBON_MAX_DRIVER_LRU_CACHE_SIZE, "1")
>  (CarbonCommonConstants.CARBON_MAX_EXECUTOR_LRU_CACHE_SIZE, "1")
> spark.sql("drop table if exists carbon_new6").show(200,false)
> spark.sql("drop table if exists carbon_new7").show(200,false)
> spark.sql("drop table if exists carbon_new8").show(200,false)
> //    spark.sql("CREATE TABLE carbon_new6 (CUST_ID INT,CUST_NAME 
> STRING,ACTIVE_EMUI_VERSION STRING, DOB TIMESTAMP, DOJ TIMESTAMP, 
> BIGINT_COLUMN1 BIGINT,BIGINT_COLUMN2 BIGINT,DECIMAL_COLUMN1 decimal(30,10), 
> DECIMAL_COLUMN2 DECIMAL(36,10),Double_COLUMN1 double, Double_COLUMN2 
> double,INTEGER_COLUMN1 INT) STORED BY 'org.apache.carbondata.format' 
> TBLPROPERTIES ('dictionary_include'='CUST_NAME')").show(200,false)
> //    spark.sql("LOAD DATA INPATH '/home/root1/data_2000.csv' INTO TABLE 
> carbon_new6 OPTIONS('DELIMITER'=',' , 
> 'QUOTECHAR'='\"','BAD_RECORDS_ACTION'='FORCE','SINGLE_PASS'='TRUE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,
>  INTEGER_COLUMN1')").show(200,false)
> spark.sql("CREATE TABLE carbon_new7 (CUST_ID INT,CUST_NAME 
> STRING,ACTIVE_EMUI_VERSION STRING, DOB TIMESTAMP, DOJ TIMESTAMP, 
> BIGINT_COLUMN1 BIGINT,BIGINT_COLUMN2 BIGINT,DECIMAL_COLUMN1 decimal(30,10), 
> DECIMAL_COLUMN2 DECIMAL(36,10),Double_COLUMN1 double, Double_COLUMN2 
> double,INTEGER_COLUMN1 INT) STORED BY 'org.apache.carbondata.format' 
> TBLPROPERTIES ('dictionary_include'='CUST_NAME, 
> ACTIVE_EMUI_VERSION,BIGINT_COLUMN1,Double_COLUMN1, 
> Double_COLUMN2')").show(200,false)
> spark.sql("LOAD DATA INPATH '/home/root1/data_2000.csv' INTO TABLE 
> carbon_new7 OPTIONS('DELIMITER'=',' , 
> 'QUOTECHAR'='\"','BAD_RECORDS_ACTION'='FORCE','SINGLE_PASS'='TRUE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,
>  INTEGER_COLUMN1')").show(200,false)
> spark.sql("LOAD DATA INPATH '/home/root1/data_2000.csv' INTO TABLE 
> carbon_new7 OPTIONS('DELIMITER'=',' , 
> 'QUOTECHAR'='\"','BAD_RECORDS_ACTION'='FORCE','SINGLE_PASS'='TRUE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,
>  INTEGER_COLUMN1')").show(200,false)
> spark.sql("CREATE TABLE carbon_new8 (CUST_ID INT,CUST_NAME 
> STRING,ACTIVE_EMUI_VERSION STRING, DOB TIMESTAMP, DOJ TIMESTAMP, 
> BIGINT_COLUMN1 BIGINT,BIGINT_COLUMN2 BIGINT,DECIMAL_COLUMN1 decimal(30,10), 
> DECIMAL_COLUMN2 DECIMAL(36,10),Double_COLUMN1 double, Double_COLUMN2 
> double,INTEGER_COLUMN1 INT) STORED BY 'org.apache.carbondata.format' 
> TBLPROPERTIES ('dictionary_include'='CUST_NAME, 
> ACTIVE_EMUI_VERSION,BIGINT_COLUMN1,Double_COLUMN1, 
> Double_COLUMN2')").show(200,false)
> spark.sql("LOAD DATA INPATH '/home/root1/data_2000.csv' INTO TABLE 
> carbon_new8 OPTIONS('DELIMITER'=',' , 
> 'QUOTECHAR'='\"','BAD_RECORDS_ACTION'='FORCE','SINGLE_PASS'='TRUE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,
>  INTEGER_COLUMN1')").show(200,false)
> spark.sql("LOAD DATA INPATH '/home/root1/data_2000.csv' INTO TABLE 
> carbon_new8 OPTIONS('DELIMITER'=',' , 
> 'QUOTECHAR'='\"','BAD_RECORDS_ACTION'='FORCE','SINGLE_PASS'='TRUE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,
>  INTEGER_COLUMN1')").show(200,false)



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

Reply via email to