[
https://issues.apache.org/jira/browse/HBASE-20429?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16592008#comment-16592008
]
Zach York commented on HBASE-20429:
-----------------------------------
If you're interested in HBASE-20723, you could also apply/review HBASE-20734
(the real fix for the issue), but I know you were tracking that as well.
I'm not sure what you mean by PUT-COPY, where is this happening? It's been a
while since I looked at the actual operations being called (and they might be
slightly different on our side). It seems with HBase, PUT-COPY's wouldn't need
to be used, but again I'm not sure what filesystem operation is being called
here that implements a put-copy under the hood.
> Support for mixed or write-heavy workloads on non-HDFS filesystems
> ------------------------------------------------------------------
>
> Key: HBASE-20429
> URL: https://issues.apache.org/jira/browse/HBASE-20429
> Project: HBase
> Issue Type: Umbrella
> Reporter: Andrew Purtell
> Priority: Major
>
> We can support reasonably well use cases on non-HDFS filesystems, like S3,
> where an external writer has loaded (and continues to load) HFiles via the
> bulk load mechanism, and then we serve out a read only workload at the HBase
> API.
> Mixed workloads or write-heavy workloads won't fare as well. In fact, data
> loss seems certain. It will depend in the specific filesystem, but all of the
> S3 backed Hadoop filesystems suffer from a couple of obvious problems,
> notably a lack of atomic rename.
> This umbrella will serve to collect some related ideas for consideration.
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)