[
https://issues.apache.org/jira/browse/HBASE-13396?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
stack updated HBASE-13396:
--------------------------
Fix Version/s: (was: 2.0.0)
Status: In Progress (was: Patch Available)
Probably still an issue. Patch does not apply. All changed in here now.
Unscheduling but leaving open because ... probably still an issue.
> Cleanup unclosed writers in later writer rolling
> ------------------------------------------------
>
> Key: HBASE-13396
> URL: https://issues.apache.org/jira/browse/HBASE-13396
> Project: HBase
> Issue Type: Bug
> Reporter: Liu Shaohui
> Assignee: Liu Shaohui
> Priority: Minor
> Attachments: HBASE-13396-v1.diff, HBASE-13396-v2.diff,
> HBASE-13396-v3.diff
>
>
> Currently, the default value of hbase.regionserver.logroll.errors.tolerated
> is 2, which means regionserver can tolerate two continuous failures of
> closing writers at most. Temporary problems of network or namenode may cause
> those failures. After those failures, the hdfs clients in RS may continue to
> renew the lease of the hlog of the writer and the namenode will not help to
> recover the lease of this hlog. So the last block of this hlog will be
> RBW(replica being written) state until the regionserver is down. Blocks in
> this state will block the datanode decommission and other operations in HDFS.
> So I think we need a mechanism to clean up those unclosed writers afterwards.
> A simple solution is to record those unclosed writers and attempt to close
> these writers until success.
> Discussions and suggestions are welcomed~ Thanks
--
This message was sent by Atlassian JIRA
(v6.4.14#64029)