[
https://issues.apache.org/jira/browse/SQOOP-1744?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14271730#comment-14271730
]
Joey Echeverria commented on SQOOP-1744:
----------------------------------------
[~vybs], I took a look at the design doc and overall it looks great. Here is
some more specific feedback:
# Option 2 is probably the right way to go, I don't see much value in the
overhead of a separate connector.
# I'd recommend using the URI to determine the dataset type rather than adding
an enum.
# There are two ways you can handle the additional configuration required
(column/key mapping and partitioning)
## Require that the dataset be pre-created. If the dataset doesn't exist, you
throw an error.
## Add a configuration validator to require the extra configs if the dataset
doesn't yet exist.
# What is required for Delta Fetch Merge (DFM)? Kite supports updating records
by key but I'm not sure what else you need.
# Is it ok for there to be a partial import on job failure? Kite writes to
HBase immediately, so you can't stage the results. If you needed to stage the
updates somehow, we'd have to add features to Kite to support that.
# Using Avro IDF sounds it would be great for Kite.
> Kite Connector Support : Read/Write data to HBase via Kite SDK
> --------------------------------------------------------------
>
> Key: SQOOP-1744
> URL: https://issues.apache.org/jira/browse/SQOOP-1744
> Project: Sqoop
> Issue Type: Bug
> Components: connectors
> Reporter: Qian Xu
> Assignee: Veena Basavaraj
> Fix For: 1.99.5
>
>
> Propose to read/ write data into HBase via the Kite SDK hbase module
> http://www.slideshare.net/HBaseCon/ecosystem-session-5
> A detailed design wiki to support basic read/ write and DFM is here
> https://cwiki.apache.org/confluence/display/SQOOP/Kite+Connector+Hbase+support
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)