[ 
https://issues.apache.org/jira/browse/HBASE-19954?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16369765#comment-16369765
 ] 

Mike Drob commented on HBASE-19954:
-----------------------------------

Right, so is it the same thing happening? We start and stop a DFS cluster with 
no region server start in between? I assume using a local file system is not an 
option for the intent if the test.

Splitting into multiple classes is fine. Don't need 3 though, just 2 is enough. 
Please add a comment to the one that doesn't use minihbase with a warning to 
developers to not add tests that try to stay region server there later.

> ShutdownHook should check whether shutdown hook is tracked by 
> ShutdownHookManager
> ---------------------------------------------------------------------------------
>
>                 Key: HBASE-19954
>                 URL: https://issues.apache.org/jira/browse/HBASE-19954
>             Project: HBase
>          Issue Type: Bug
>            Reporter: Ted Yu
>            Assignee: Ted Yu
>            Priority: Major
>             Fix For: 2.0.0-beta-2
>
>         Attachments: 19954.v1.txt, 19954.v2.txt, 19954.v3.txt
>
>
> Currently ShutdownHook#suppressHdfsShutdownHook() does the following:
> {code}
>       synchronized (fsShutdownHooks) {
>         boolean isFSCacheDisabled = 
> fs.getConf().getBoolean("fs.hdfs.impl.disable.cache", false);
>         if (!isFSCacheDisabled && 
> !fsShutdownHooks.containsKey(hdfsClientFinalizer)
>             && !ShutdownHookManager.deleteShutdownHook(hdfsClientFinalizer)) {
> {code}
> There is no check that ShutdownHookManager still tracks the shutdown hook, 
> leading to potential RuntimeException (as can be observed in hadoop3 Jenkins 
> job).



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

Reply via email to