[ 
https://issues.apache.org/jira/browse/YARN-3134?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Li Lu updated YARN-3134:
------------------------
    Attachment: YARN-3134-042715.patch

In this patch I addressed most of the review comments. Major changes include:
# Storing relationships between entities
# Addressing per-thread connection caching and close problems by using Guava 
loading cache
# Refactor collectors to allow writer injections upon creation
# Disable autocommit explicitly
# Using StringBuilders to replace StringBuffers

The only two points I've not addressed are:
# Make {{CONN_STRING}} configurable: I still think this is a little bit early 
for the POC stage. Adding this support is quick, though. 
# We need to redesign the way to store metrics for Phoenix. Explicit using 
HBase timestamp (PHOENIX-914) is not available in trunk. Meanwhile, YARN-3551 
is changing the object model and removing startTime and endTime that are part 
of the primary key for Phoenix. I've marked YARN-3551 as blocking this JIRA and 
will follow up. 

> [Storage implementation] Exploiting the option of using Phoenix to access 
> HBase backend
> ---------------------------------------------------------------------------------------
>
>                 Key: YARN-3134
>                 URL: https://issues.apache.org/jira/browse/YARN-3134
>             Project: Hadoop YARN
>          Issue Type: Sub-task
>          Components: timelineserver
>            Reporter: Zhijie Shen
>            Assignee: Li Lu
>         Attachments: YARN-3134-040915_poc.patch, YARN-3134-041015_poc.patch, 
> YARN-3134-041415_poc.patch, YARN-3134-042115.patch, YARN-3134-042715.patch, 
> YARN-3134DataSchema.pdf
>
>
> Quote the introduction on Phoenix web page:
> {code}
> Apache Phoenix is a relational database layer over HBase delivered as a 
> client-embedded JDBC driver targeting low latency queries over HBase data. 
> Apache Phoenix takes your SQL query, compiles it into a series of HBase 
> scans, and orchestrates the running of those scans to produce regular JDBC 
> result sets. The table metadata is stored in an HBase table and versioned, 
> such that snapshot queries over prior versions will automatically use the 
> correct schema. Direct use of the HBase API, along with coprocessors and 
> custom filters, results in performance on the order of milliseconds for small 
> queries, or seconds for tens of millions of rows.
> {code}
> It may simply our implementation read/write data from/to HBase, and can 
> easily build index and compose complex query.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to