[
https://issues.apache.org/jira/browse/HBASE-14004?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15042744#comment-15042744
]
Duo Zhang commented on HBASE-14004:
-----------------------------------
{quote}
Fixing HBase writing path that we should retry logging WAL in a new file rather
than rollback MemStore.
{quote}
To be clear, this means we will hold the {{WAL.sync}} request if there are some
entries have already been written out but not acked and never return until we
successfully write them out and get ack back. And if {{WAL.sync}} or
{{WAL.write}} fails(maybe due to queue full), we will still rollback MemStore
since we can confirm that the WAL entries have not been written out. Right?
And I think there is another task for us. Now the DFSOutputStream does not
provide a public method to get acked length. We can open a issue of HDFS
project and use reflection first in HBase. But there is still a problem that
{{hflush}} or {{hsync}} does not return the acked length which means get acked
length and {{hsync}} are two separated operations so it is hard to get the
exact acked length after calling {{hsync}}. Maybe we could get current total
write out bytes first(not acked length) and then call {{hsync}}, the acked
length after calling {{hsync}} must be larger than this value so it is safe to
use this value as "acked length". Any thoughts?
Thanks.
> [Replication] Inconsistency between Memstore and WAL may result in data in
> remote cluster that is not in the origin
> -------------------------------------------------------------------------------------------------------------------
>
> Key: HBASE-14004
> URL: https://issues.apache.org/jira/browse/HBASE-14004
> Project: HBase
> Issue Type: Bug
> Components: regionserver
> Reporter: He Liangliang
> Priority: Critical
> Labels: replication, wal
>
> Looks like the current write path can cause inconsistency between
> memstore/hfile and WAL which cause the slave cluster has more data than the
> master cluster.
> The simplified write path looks like:
> 1. insert record into Memstore
> 2. write record to WAL
> 3. sync WAL
> 4. rollback Memstore if 3 fails
> It's possible that the HDFS sync RPC call fails, but the data is already
> (may partially) transported to the DNs which finally get persisted. As a
> result, the handler will rollback the Memstore and the later flushed HFile
> will also skip this record.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)