[jira] [Created] (CARBONDATA-490) Unify all RDD in carbon-spark and carbon-spark2 module

2016-12-02 Thread Jacky Li (JIRA)
Jacky Li created CARBONDATA-490:
---

 Summary: Unify all RDD in carbon-spark and carbon-spark2 module
 Key: CARBONDATA-490
 URL: https://issues.apache.org/jira/browse/CARBONDATA-490
 Project: CarbonData
  Issue Type: Improvement
Reporter: Jacky Li


Currently there are duplicate RDD in carbon-spark and carbon-spark2 modules



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CARBONDATA-489) spark2 decimal issue

2016-12-02 Thread Fei Wang (JIRA)
Fei Wang created CARBONDATA-489:
---

 Summary: spark2 decimal issue
 Key: CARBONDATA-489
 URL: https://issues.apache.org/jira/browse/CARBONDATA-489
 Project: CarbonData
  Issue Type: Sub-task
  Components: spark-integration
Reporter: Fei Wang


create a table with decimal field and query it will throw error, do not support 
decimal(0, 0)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CARBONDATA-488) add InsertInto feature for spark2

2016-12-02 Thread QiangCai (JIRA)
QiangCai created CARBONDATA-488:
---

 Summary: add InsertInto feature for spark2
 Key: CARBONDATA-488
 URL: https://issues.apache.org/jira/browse/CARBONDATA-488
 Project: CarbonData
  Issue Type: New Feature
  Components: data-load
Affects Versions: 0.3.0-incubating
Reporter: QiangCai
Assignee: QiangCai
 Fix For: 0.3.0-incubating






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CARBONDATA-487) spark2 integration is not compiling

2016-12-02 Thread Jacky Li (JIRA)
Jacky Li created CARBONDATA-487:
---

 Summary: spark2 integration is not compiling
 Key: CARBONDATA-487
 URL: https://issues.apache.org/jira/browse/CARBONDATA-487
 Project: CarbonData
  Issue Type: Bug
Reporter: Jacky Li
 Fix For: 0.3.0-incubating


spark2 integration is not compiling



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CARBONDATA-486) Rreading dataframe concurrently will lead to wrong data

2016-12-02 Thread QiangCai (JIRA)
QiangCai created CARBONDATA-486:
---

 Summary: Rreading dataframe concurrently will lead to wrong data
 Key: CARBONDATA-486
 URL: https://issues.apache.org/jira/browse/CARBONDATA-486
 Project: CarbonData
  Issue Type: Bug
  Components: data-load
Affects Versions: 0.3.0-incubating
Reporter: QiangCai
Assignee: QiangCai
 Fix For: 0.3.0-incubating






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CARBONDATA-485) Refactor code for DataGraphExecuter

2016-12-02 Thread Prabhat Kashyap (JIRA)
Prabhat Kashyap created CARBONDATA-485:
--

 Summary: Refactor code for DataGraphExecuter
 Key: CARBONDATA-485
 URL: https://issues.apache.org/jira/browse/CARBONDATA-485
 Project: CarbonData
  Issue Type: Improvement
Reporter: Prabhat Kashyap
Priority: Trivial






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CARBONDATA-484) LRU cache for B-Tree to ensure to avoid out memory, when too many number of tables exits and all are not frequently used.

2016-12-02 Thread Mohammad Shahid Khan (JIRA)
Mohammad Shahid Khan created CARBONDATA-484:
---

 Summary: LRU cache for B-Tree to ensure to avoid out memory, when 
too many number of tables exits and all are not frequently used.
 Key: CARBONDATA-484
 URL: https://issues.apache.org/jira/browse/CARBONDATA-484
 Project: CarbonData
  Issue Type: Bug
Reporter: Mohammad Shahid Khan
Assignee: Mohammad Shahid Khan


*LRU Cache for B-Tree*
Problem:

CarbonData is maintaining two level of B-Tree cache, one at the driver level 
and another at executor level.  Currently CarbonData has the mechanism to 
invalidate the segments and blocks cache for the invalid table segments, but 
there is no eviction policy for the unused cached object. So the instance at 
which complete memory is utilized then the system will not be able to process 
any new requests.

Solution:

In the cache maintained at the driver level and at the executor there must be 
objects in cache currently not in use. Therefore system should have the 
mechanism to below mechanism.

1.   Set the max memory limit till which objects could be hold in the 
memory.

2.   When configured memory limit reached then identify the cached objects 
currently not in use so that the required memory could be freed without 
impacting the existing process.

3.   Eviction should be done only till the required memory is not meet.

For details please refer to attachments.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Re: B-Tree LRU cache (New Feature)

2016-12-02 Thread Venkata Gollamudi
Hi Shahid,

This solution, LRU cache for BTree is required to ensure to avoid out of
memory, when too many number of tables exists in store and all are not
frequently used.

Please raise an issue to track this feature.

Regards,
Ramana

On Wed, Nov 23, 2016 at 6:30 PM, mohdshahidkhan <
mohdshahidkhan1...@gmail.com> wrote:

> Please find Design document for B-Tree LRU cache
> https://drive.google.com/file/d/0B8sQb--59vO7bWxVeWs1ajBiMG8/view?usp=
> sharing
>
>
>
> --
> View this message in context: http://apache-carbondata-
> mailing-list-archive.1130556.n5.nabble.com/B-Tree-LRU-
> cache-New-Feature-tp2366p3130.html
> Sent from the Apache CarbonData Mailing List archive mailing list archive
> at Nabble.com.
>