[jira] [Updated] (HBASE-26420) Unexpected crash of meta RegionServer causes the cluster out of service

2021-11-03 Thread anonymous (Jira)


 [ 
https://issues.apache.org/jira/browse/HBASE-26420?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

anonymous updated HBASE-26420:
--
Attachment: hbase-root-master-C3HM1.log

> Unexpected crash of meta RegionServer causes the cluster out of service
> ---
>
> Key: HBASE-26420
> URL: https://issues.apache.org/jira/browse/HBASE-26420
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.7.1
>Reporter: anonymous
>Priority: Major
> Attachments: hbase-root-master-C3HM1.log
>
>
> We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, 
> C3RS1, C3RS2, C3RS3. 
> We use an external ZooKeeper cluster which is a pseudo-distributed cluster:
> {code:java}
>   
> hbase.zookeeper.quorum
> C3hb-zk
>   
>   
> hbase.zookeeper.property.clientPort
> 11181
>   
> {code}
> For other HBase options, we use the default settings. The buggy scenario is 
> as follows:
> 1. Start the cluster, C3HM1 becomes the active master;
> 2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" 
> on ZooKeeper;
> {code:java}
> [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.createNonSequential(RecoverableZooKeeper.java:665),
>  
> org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.create(RecoverableZooKeeper.java:644),
>  org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndWatch(ZKUtil.java:1182), 
> org.apache.hadoop.hbase.zookeeper.MetaTableLocator.setMetaLocation(MetaTableLocator.java:464),
>  
> org.apache.hadoop.hbase.regionserver.HRegionServer.postOpenDeployTasks(HRegionServer.java:2182),
>  
> org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler$PostOpenDeployTasksThread.run(OpenRegionHandler.java:329)]
> {code}
> 3. The meta server is still not online after 10 minutes. The data of znode 
> "/hbase/master" is C3HM1.
> While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, 
> everything works fine. And the bug does not appear on HBase-2.4.5.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HBASE-26420) Unexpected crash of meta RegionServer causes the cluster out of service

2021-11-03 Thread anonymous (Jira)


 [ 
https://issues.apache.org/jira/browse/HBASE-26420?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

anonymous updated HBASE-26420:
--
Description: 
We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, 
C3RS1, C3RS2, C3RS3. 
We use an external ZooKeeper cluster which is a pseudo-distributed cluster:

{code:java}
  
hbase.zookeeper.quorum
C3hb-zk
  
  
hbase.zookeeper.property.clientPort
11181
  
{code}

For other HBase options, we use the default settings. The buggy scenario is as 
follows:

1. Start the cluster, C3HM1 becomes the active master;
2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" on 
ZooKeeper;

{code:java}
[org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.createNonSequential(RecoverableZooKeeper.java:665),
 
org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.create(RecoverableZooKeeper.java:644),
 org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndWatch(ZKUtil.java:1182), 
org.apache.hadoop.hbase.zookeeper.MetaTableLocator.setMetaLocation(MetaTableLocator.java:464),
 
org.apache.hadoop.hbase.regionserver.HRegionServer.postOpenDeployTasks(HRegionServer.java:2182),
 
org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler$PostOpenDeployTasksThread.run(OpenRegionHandler.java:329)]
{code}

3. The meta server is still not online after 10 minutes. The data of znode 
"/hbase/master" is C3HM1.

While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, 
everything works fine. And the bug does not appear on HBase-2.4.5.

  was:
We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, 
C3RS1, C3RS2, C3RS3. 
We use an external ZooKeeper cluster which is a pseudo-distributed cluster:

{code:java}
  
hbase.zookeeper.quorum
C3hb-zk
  
  
hbase.zookeeper.property.clientPort
11181
  
{code}

For other HBase options, we use the default settings. The buggy scenario is as 
follows:

1. Start the cluster, C3HM1 becomes the active master;
2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" on 
ZooKeeper;

{code:java}
[org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.createNonSequential(RecoverableZooKeeper.java:665),
 
org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.create(RecoverableZooKeeper.java:644),
 org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndWatch(ZKUtil.java:1182), 
org.apache.hadoop.hbase.zookeeper.MetaTableLocator.setMetaLocation(MetaTableLocator.java:464),
 
org.apache.hadoop.hbase.regionserver.HRegionServer.postOpenDeployTasks(HRegionServer.java:2182),
 
org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler$PostOpenDeployTasksThread.run(OpenRegionHandler.java:329)]
{code}

3. The meta server is still not online after 10 minutes.

While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, 
everything works fine. And the bug does not appear on HBase-2.4.5.


> Unexpected crash of meta RegionServer causes the cluster out of service
> ---
>
> Key: HBASE-26420
> URL: https://issues.apache.org/jira/browse/HBASE-26420
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.7.1
>Reporter: anonymous
>Priority: Major
>
> We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, 
> C3RS1, C3RS2, C3RS3. 
> We use an external ZooKeeper cluster which is a pseudo-distributed cluster:
> {code:java}
>   
> hbase.zookeeper.quorum
> C3hb-zk
>   
>   
> hbase.zookeeper.property.clientPort
> 11181
>   
> {code}
> For other HBase options, we use the default settings. The buggy scenario is 
> as follows:
> 1. Start the cluster, C3HM1 becomes the active master;
> 2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" 
> on ZooKeeper;
> {code:java}
> [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.createNonSequential(RecoverableZooKeeper.java:665),
>  
> org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.create(RecoverableZooKeeper.java:644),
>  org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndWatch(ZKUtil.java:1182), 
> org.apache.hadoop.hbase.zookeeper.MetaTableLocator.setMetaLocation(MetaTableLocator.java:464),
>  
> org.apache.hadoop.hbase.regionserver.HRegionServer.postOpenDeployTasks(HRegionServer.java:2182),
>  
> org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler$PostOpenDeployTasksThread.run(OpenRegionHandler.java:329)]
> {code}
> 3. The meta server is still not online after 10 minutes. The data of znode 
> "/hbase/master" is C3HM1.
> While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, 
> everything works fine. And the bug does not appear on HBase-2.4.5.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HBASE-26420) Unexpected crash of meta RegionServer causes the cluster out of service

2021-11-03 Thread anonymous (Jira)


 [ 
https://issues.apache.org/jira/browse/HBASE-26420?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

anonymous updated HBASE-26420:
--
Description: 
We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, 
C3RS1, C3RS2, C3RS3. 
We use an external ZooKeeper cluster which is a pseudo-distributed cluster:

{code:java}
  
hbase.zookeeper.quorum
C3hb-zk
  
  
hbase.zookeeper.property.clientPort
11181
  
{code}

For other HBase options, we use the default settings. The buggy scenario is as 
follows:

1. Start the cluster, C3HM1 becomes the active master;
2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" on 
ZooKeeper;

{code:java}
[org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.createNonSequential(RecoverableZooKeeper.java:665),
 
org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.create(RecoverableZooKeeper.java:644),
 org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndWatch(ZKUtil.java:1182), 
org.apache.hadoop.hbase.zookeeper.MetaTableLocator.setMetaLocation(MetaTableLocator.java:464),
 
org.apache.hadoop.hbase.regionserver.HRegionServer.postOpenDeployTasks(HRegionServer.java:2182),
 
org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler$PostOpenDeployTasksThread.run(OpenRegionHandler.java:329)]
{code}

3. The meta server is still not online after 10 minutes.

While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, 
everything works fine. And the bug does not appear on HBase-2.4.5.

  was:
We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, 
C3RS1, C3RS2, C3RS3. 
We use an external ZooKeeper cluster which is a pseudo-distributed cluster:

{code:java}
  
hbase.zookeeper.quorum
C3hb-zk
  
  
hbase.zookeeper.property.clientPort
11181
  
{code}

For other HBase options, we use the default settings. The buggy scenario is as 
follows:

1. Start the cluster, C3HM1 becomes the active master;
2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" on 
ZooKeeper;

{code:java}
[org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.createNonSequential(RecoverableZooKeeper.java:665),
 
org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.create(RecoverableZooKeeper.java:644),
 org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndWatch(ZKUtil.java:1182), 
org.apache.hadoop.hbase.zookeeper.MetaTableLocator.setMetaLocation(MetaTableLocator.java:464),
 
org.apache.hadoop.hbase.regionserver.HRegionServer.postOpenDeployTasks(HRegionServer.java:2182),
 
org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler$PostOpenDeployTasksThread.run(OpenRegionHandler.java:329)]
{code}

3. The meta server is still not online after 10 minutes.

While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, 
everything works fine. And the bug disappears on HBase-2.4.5.


> Unexpected crash of meta RegionServer causes the cluster out of service
> ---
>
> Key: HBASE-26420
> URL: https://issues.apache.org/jira/browse/HBASE-26420
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.7.1
>Reporter: anonymous
>Priority: Major
>
> We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, 
> C3RS1, C3RS2, C3RS3. 
> We use an external ZooKeeper cluster which is a pseudo-distributed cluster:
> {code:java}
>   
> hbase.zookeeper.quorum
> C3hb-zk
>   
>   
> hbase.zookeeper.property.clientPort
> 11181
>   
> {code}
> For other HBase options, we use the default settings. The buggy scenario is 
> as follows:
> 1. Start the cluster, C3HM1 becomes the active master;
> 2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" 
> on ZooKeeper;
> {code:java}
> [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.createNonSequential(RecoverableZooKeeper.java:665),
>  
> org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.create(RecoverableZooKeeper.java:644),
>  org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndWatch(ZKUtil.java:1182), 
> org.apache.hadoop.hbase.zookeeper.MetaTableLocator.setMetaLocation(MetaTableLocator.java:464),
>  
> org.apache.hadoop.hbase.regionserver.HRegionServer.postOpenDeployTasks(HRegionServer.java:2182),
>  
> org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler$PostOpenDeployTasksThread.run(OpenRegionHandler.java:329)]
> {code}
> 3. The meta server is still not online after 10 minutes.
> While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, 
> everything works fine. And the bug does not appear on HBase-2.4.5.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HBASE-26420) Unexpected crash of meta RegionServer causes the cluster out of service

2021-11-03 Thread anonymous (Jira)


 [ 
https://issues.apache.org/jira/browse/HBASE-26420?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

anonymous updated HBASE-26420:
--
Description: 
We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, 
C3RS1, C3RS2, C3RS3. 
We use an external ZooKeeper cluster which is a pseudo-distributed cluster:

{code:java}
  
hbase.zookeeper.quorum
C3hb-zk
  
  
hbase.zookeeper.property.clientPort
11181
  
{code}

For other HBase options, we use the default settings. The buggy scenario is as 
follows:

1. Start the cluster, C3HM1 becomes the active master;
2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" on 
ZooKeeper;

{code:java}
[org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.createNonSequential(RecoverableZooKeeper.java:665),
 
org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.create(RecoverableZooKeeper.java:644),
 org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndWatch(ZKUtil.java:1182), 
org.apache.hadoop.hbase.zookeeper.MetaTableLocator.setMetaLocation(MetaTableLocator.java:464),
 
org.apache.hadoop.hbase.regionserver.HRegionServer.postOpenDeployTasks(HRegionServer.java:2182),
 
org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler$PostOpenDeployTasksThread.run(OpenRegionHandler.java:329)]
{code}

3. The meta server is still not online after 10 minutes.

While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, 
everything works fine. And the bug disappears on HBase-2.4.5.

  was:
We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, 
C3RS1, C3RS2, C3RS3. 
We use an external ZooKeeper cluster which is a pseudo-distributed cluster:

{code:java}
  
hbase.zookeeper.quorum
C3hb-zk
  
  
hbase.zookeeper.property.clientPort
11181
  
{code}

For other HBase options, we use the default settings. The buggy scenario is as 
follows:

1. Start the cluster, C3HM1 becomes the active master;
2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" on 
ZooKeeper;

{code:java}
[org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.createNonSequential(RecoverableZooKeeper.java:665),
 
org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.create(RecoverableZooKeeper.java:644),
 org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndWatch(ZKUtil.java:1182), 
org.apache.hadoop.hbase.zookeeper.MetaTableLocator.setMetaLocation(MetaTableLocator.java:464),
 
org.apache.hadoop.hbase.regionserver.HRegionServer.postOpenDeployTasks(HRegionServer.java:2182),
 
org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler$PostOpenDeployTasksThread.run(OpenRegionHandler.java:329)]
{code}

3. The meta server has not been online after 10 minutes.

While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, 
everything works fine. And the bug disappears on HBase-2.4.5.


> Unexpected crash of meta RegionServer causes the cluster out of service
> ---
>
> Key: HBASE-26420
> URL: https://issues.apache.org/jira/browse/HBASE-26420
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.7.1
>Reporter: anonymous
>Priority: Major
>
> We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, 
> C3RS1, C3RS2, C3RS3. 
> We use an external ZooKeeper cluster which is a pseudo-distributed cluster:
> {code:java}
>   
> hbase.zookeeper.quorum
> C3hb-zk
>   
>   
> hbase.zookeeper.property.clientPort
> 11181
>   
> {code}
> For other HBase options, we use the default settings. The buggy scenario is 
> as follows:
> 1. Start the cluster, C3HM1 becomes the active master;
> 2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" 
> on ZooKeeper;
> {code:java}
> [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.createNonSequential(RecoverableZooKeeper.java:665),
>  
> org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.create(RecoverableZooKeeper.java:644),
>  org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndWatch(ZKUtil.java:1182), 
> org.apache.hadoop.hbase.zookeeper.MetaTableLocator.setMetaLocation(MetaTableLocator.java:464),
>  
> org.apache.hadoop.hbase.regionserver.HRegionServer.postOpenDeployTasks(HRegionServer.java:2182),
>  
> org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler$PostOpenDeployTasksThread.run(OpenRegionHandler.java:329)]
> {code}
> 3. The meta server is still not online after 10 minutes.
> While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, 
> everything works fine. And the bug disappears on HBase-2.4.5.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HBASE-26370) Unexpected crash fails the truncate operation and makes a disabled table enabled

2021-11-03 Thread anonymous (Jira)


 [ 
https://issues.apache.org/jira/browse/HBASE-26370?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

anonymous updated HBASE-26370:
--
Description: 
There is a table {{mytable}} in a HBase cluster with two HMasters and two 
RegionServers.

1. Current active HMaster is HMaster1;
2. Client disabled the table {{mytable}};
3. {{mytable}} is in disabled state;
4. Client request to truncate {{mytable}};
5. HMaster1 tries to truncate {{mytable}}. For the TruncateTableState 
{{TRUNCATE_TABLE_ASSIGN_REGIONS}}, HMaster1 assigns the regions of the table 
and mark the table as Enabling;
6. HMaster1 crashes;
7. HMaster2 becomes active;
8. The truncate operation fails due to {{TableNotDisabledException}}, and 
{{mytable}} was left as enabled.

In HMaster2, we got following info:

{code:java}
INFO [ProcedureExecutor-12] procedure2.ProcedureExecutor: Rolledback procedure 
TruncateTableProcedure (table=mytable preserveSplits=true) id=9 owner=myhbase 
state=ROLLEDBACK exec-time=4.6030 sec 
exception=org.apache.hadoop.hbase.TableNotDisabledException: mytable
{code}


  was:
There is a table {{mytable}} in a HBase cluster with two HMasters and two 
RegionServers.

1. Current active HMaster is HMaster1;
2. Client disabled the table {{mytable}};
3. {{mytable}} is in disabled state;
4. Client request to truncate {{mytable}};
5. HMaster1 tries to truncate {{mytable}}. For the TruncateTableState 
{{TRUNCATE_TABLE_ASSIGN_REGIONS}}, HMaster1 assigns regions of the table and 
mark the table as Enabling;
6. HMaster1 crashes;
7. HMaster2 becomes active;
8. The truncate operation fails due to {{TableNotDisabledException}}, and 
{{mytable}} was left as enabled.

In HMaster2, we got following info:

{code:java}
INFO [ProcedureExecutor-12] procedure2.ProcedureExecutor: Rolledback procedure 
TruncateTableProcedure (table=mytable preserveSplits=true) id=9 owner=myhbase 
state=ROLLEDBACK exec-time=4.6030 sec 
exception=org.apache.hadoop.hbase.TableNotDisabledException: mytable
{code}



> Unexpected crash fails the truncate operation and makes a disabled table 
> enabled
> 
>
> Key: HBASE-26370
> URL: https://issues.apache.org/jira/browse/HBASE-26370
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.7.1
>Reporter: anonymous
>Priority: Major
>
> There is a table {{mytable}} in a HBase cluster with two HMasters and two 
> RegionServers.
> 1. Current active HMaster is HMaster1;
> 2. Client disabled the table {{mytable}};
> 3. {{mytable}} is in disabled state;
> 4. Client request to truncate {{mytable}};
> 5. HMaster1 tries to truncate {{mytable}}. For the TruncateTableState 
> {{TRUNCATE_TABLE_ASSIGN_REGIONS}}, HMaster1 assigns the regions of the table 
> and mark the table as Enabling;
> 6. HMaster1 crashes;
> 7. HMaster2 becomes active;
> 8. The truncate operation fails due to {{TableNotDisabledException}}, and 
> {{mytable}} was left as enabled.
> In HMaster2, we got following info:
> {code:java}
> INFO [ProcedureExecutor-12] procedure2.ProcedureExecutor: Rolledback 
> procedure TruncateTableProcedure (table=mytable preserveSplits=true) id=9 
> owner=myhbase state=ROLLEDBACK exec-time=4.6030 sec 
> exception=org.apache.hadoop.hbase.TableNotDisabledException: mytable
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HBASE-26370) Unexpected crash fails the truncate operation and makes a disabled table enabled

2021-11-03 Thread anonymous (Jira)


 [ 
https://issues.apache.org/jira/browse/HBASE-26370?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

anonymous updated HBASE-26370:
--
Description: 
There is a table {{mytable}} in a HBase cluster with two HMasters and two 
RegionServers.

1. Current active HMaster is HMaster1;
2. Client disabled the table {{mytable}};
3. {{mytable}} is in disabled state;
4. Client request to truncate {{mytable}};
5. HMaster1 tries to truncate {{mytable}}. For the TruncateTableState 
{{TRUNCATE_TABLE_ASSIGN_REGIONS}}, HMaster1 assigns regions of the table and 
mark the table as Enabling;
6. HMaster1 crashes;
7. HMaster2 becomes active;
8. The truncate operation fails due to {{TableNotDisabledException}}, and 
{{mytable}} was left as enabled.

In HMaster2, we got following info:

{code:java}
INFO [ProcedureExecutor-12] procedure2.ProcedureExecutor: Rolledback procedure 
TruncateTableProcedure (table=mytable preserveSplits=true) id=9 owner=myhbase 
state=ROLLEDBACK exec-time=4.6030 sec 
exception=org.apache.hadoop.hbase.TableNotDisabledException: mytable
{code}


  was:
There is a table {{mytable}} in a HBase cluster with two HMasters and two 
RegionServers.

1. Current active HMaster is HMaster1;
2. Client disabled the table {{mytable}};
3. {{mytable}} is in disabled state;
4. Client request to truncate {{mytable}};
5. HMaster1 tries to truncate {{mytable}}. For the TruncateTableState 
{{TRUNCATE_TABLE_ASSIGN_REGIONS}}, HMaster1 tries to assign regions of the 
table and mark the table as Enabling;
6. HMaster1 crashes;
7. HMaster2 becomes active;
8. The truncate operation fails due to {{TableNotDisabledException}}, and 
{{mytable}} was left as enabled.

In HMaster2, we got following info:

{code:java}
INFO [ProcedureExecutor-12] procedure2.ProcedureExecutor: Rolledback procedure 
TruncateTableProcedure (table=mytable preserveSplits=true) id=9 owner=myhbase 
state=ROLLEDBACK exec-time=4.6030 sec 
exception=org.apache.hadoop.hbase.TableNotDisabledException: mytable
{code}



> Unexpected crash fails the truncate operation and makes a disabled table 
> enabled
> 
>
> Key: HBASE-26370
> URL: https://issues.apache.org/jira/browse/HBASE-26370
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.7.1
>Reporter: anonymous
>Priority: Major
>
> There is a table {{mytable}} in a HBase cluster with two HMasters and two 
> RegionServers.
> 1. Current active HMaster is HMaster1;
> 2. Client disabled the table {{mytable}};
> 3. {{mytable}} is in disabled state;
> 4. Client request to truncate {{mytable}};
> 5. HMaster1 tries to truncate {{mytable}}. For the TruncateTableState 
> {{TRUNCATE_TABLE_ASSIGN_REGIONS}}, HMaster1 assigns regions of the table and 
> mark the table as Enabling;
> 6. HMaster1 crashes;
> 7. HMaster2 becomes active;
> 8. The truncate operation fails due to {{TableNotDisabledException}}, and 
> {{mytable}} was left as enabled.
> In HMaster2, we got following info:
> {code:java}
> INFO [ProcedureExecutor-12] procedure2.ProcedureExecutor: Rolledback 
> procedure TruncateTableProcedure (table=mytable preserveSplits=true) id=9 
> owner=myhbase state=ROLLEDBACK exec-time=4.6030 sec 
> exception=org.apache.hadoop.hbase.TableNotDisabledException: mytable
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HBASE-26391) Unexpected crash in the process of deleting a table causes stale data in ZooKeeper

2021-11-03 Thread anonymous (Jira)


 [ 
https://issues.apache.org/jira/browse/HBASE-26391?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

anonymous updated HBASE-26391:
--
Description: 
There is a table {{mytable}} in a HBase cluster with two HMasters and two 
RegionServers.

1. Current active HMaster is HMaster1;
2. Client request to delete the table {{mytable}};
3. HMaster1 deletes znode in ZooKeeper: /hbase/table/mytable

{code:java}
[org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.delete(RecoverableZooKeeper.java:238),
 
org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1367),
 
org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1359),
 
org.apache.hadoop.hbase.zookeeper.ZKTableStateManager.setDeletedTable(ZKTableStateManager.java:244),
 
org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.deleteAssignmentState(DeleteTableProcedure.java:383),
 
org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:136),
 
org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:57),
 
org.apache.hadoop.hbase.procedure2.StateMachineProcedure.execute(StateMachineProcedure.java:139),
 org.apache.hadoop.hbase.procedure2.Procedure.doExecute(Procedure.java:499), 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execProcedure(ProcedureExecutor.java:1166),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:954),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:907),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.access$400(ProcedureExecutor.java:76),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java:481),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java)]
{code}

4. HMaster1 crashes before deleting the znode: /hbase/table-lock/mytable

{code:java}
[org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.delete(RecoverableZooKeeper.java:238),
 org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNode(ZKUtil.java:1341), 
org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNode(ZKUtil.java:1330), 
org.apache.hadoop.hbase.master.TableLockManager$ZKTableLockManager.tableDeleted(TableLockManager.java:442),
 
org.apache.hadoop.hbase.master.procedure.MasterProcedureScheduler.markTableAsDeleted(MasterProcedureScheduler.java:895),
 
org.apache.hadoop.hbase.master.procedure.MasterProcedureScheduler.completionCleanup(MasterProcedureScheduler.java:261),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.procedureFinished(ProcedureExecutor.java:1415),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:974),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:907),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.access$400(ProcedureExecutor.java:76),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java:481),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java)]
{code}

5. HMaster2 becomes active
6. The table {{mytablle}} was deleted, but the znode 
"/hbase/table-lock/mytable" was left on ZooKeeper




  was:
There is a table {{mytable}} in a HBase cluster with two HMasters and two 
RegionServers.

1. Current active HMaster is HMaster1;
2. Client request to delete the table {{mytable}};
3. HMaster1 tries to delete znode in ZooKeeper: /hbase/table/mytable

{code:java}
[org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.delete(RecoverableZooKeeper.java:238),
 
org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1367),
 
org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1359),
 
org.apache.hadoop.hbase.zookeeper.ZKTableStateManager.setDeletedTable(ZKTableStateManager.java:244),
 
org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.deleteAssignmentState(DeleteTableProcedure.java:383),
 
org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:136),
 
org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:57),
 
org.apache.hadoop.hbase.procedure2.StateMachineProcedure.execute(StateMachineProcedure.java:139),
 org.apache.hadoop.hbase.procedure2.Procedure.doExecute(Procedure.java:499), 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execProcedure(ProcedureExecutor.java:1166),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:954),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:907),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.access$400(ProcedureExecutor.java:76),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java:481),
 

[jira] [Updated] (HBASE-26391) Unexpected crash in the process of deleting a table causes stale data in ZooKeeper

2021-11-03 Thread anonymous (Jira)


 [ 
https://issues.apache.org/jira/browse/HBASE-26391?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

anonymous updated HBASE-26391:
--
Description: 
There is a table {{mytable}} in a HBase cluster with two HMasters and two 
RegionServers.

1. Current active HMaster is HMaster1;
2. Client request to delete the table {{mytable}};
3. HMaster1 deletes znode in ZooKeeper: /hbase/table/mytable

{code:java}
[org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.delete(RecoverableZooKeeper.java:238),
 
org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1367),
 
org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1359),
 
org.apache.hadoop.hbase.zookeeper.ZKTableStateManager.setDeletedTable(ZKTableStateManager.java:244),
 
org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.deleteAssignmentState(DeleteTableProcedure.java:383),
 
org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:136),
 
org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:57),
 
org.apache.hadoop.hbase.procedure2.StateMachineProcedure.execute(StateMachineProcedure.java:139),
 org.apache.hadoop.hbase.procedure2.Procedure.doExecute(Procedure.java:499), 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execProcedure(ProcedureExecutor.java:1166),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:954),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:907),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.access$400(ProcedureExecutor.java:76),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java:481),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java)]
{code}

4. HMaster1 crashes before deleting the znode: /hbase/table-lock/mytable

{code:java}
[org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.delete(RecoverableZooKeeper.java:238),
 org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNode(ZKUtil.java:1341), 
org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNode(ZKUtil.java:1330), 
org.apache.hadoop.hbase.master.TableLockManager$ZKTableLockManager.tableDeleted(TableLockManager.java:442),
 
org.apache.hadoop.hbase.master.procedure.MasterProcedureScheduler.markTableAsDeleted(MasterProcedureScheduler.java:895),
 
org.apache.hadoop.hbase.master.procedure.MasterProcedureScheduler.completionCleanup(MasterProcedureScheduler.java:261),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.procedureFinished(ProcedureExecutor.java:1415),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:974),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:907),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.access$400(ProcedureExecutor.java:76),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java:481),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java)]
{code}

5. HMaster2 becomes active
6. The table {{mytable}} was deleted, but the znode "/hbase/table-lock/mytable" 
was left on ZooKeeper




  was:
There is a table {{mytable}} in a HBase cluster with two HMasters and two 
RegionServers.

1. Current active HMaster is HMaster1;
2. Client request to delete the table {{mytable}};
3. HMaster1 deletes znode in ZooKeeper: /hbase/table/mytable

{code:java}
[org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.delete(RecoverableZooKeeper.java:238),
 
org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1367),
 
org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1359),
 
org.apache.hadoop.hbase.zookeeper.ZKTableStateManager.setDeletedTable(ZKTableStateManager.java:244),
 
org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.deleteAssignmentState(DeleteTableProcedure.java:383),
 
org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:136),
 
org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:57),
 
org.apache.hadoop.hbase.procedure2.StateMachineProcedure.execute(StateMachineProcedure.java:139),
 org.apache.hadoop.hbase.procedure2.Procedure.doExecute(Procedure.java:499), 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execProcedure(ProcedureExecutor.java:1166),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:954),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:907),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.access$400(ProcedureExecutor.java:76),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java:481),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java)]
{code}


[jira] [Updated] (HBASE-26420) Unexpected crash of meta RegionServer causes the cluster out of service

2021-11-03 Thread anonymous (Jira)


 [ 
https://issues.apache.org/jira/browse/HBASE-26420?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

anonymous updated HBASE-26420:
--
Description: 
We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, 
C3RS1, C3RS2, C3RS3. 
We use an external ZooKeeper cluster which is a pseudo-distributed cluster:

{code:java}
  
hbase.zookeeper.quorum
C3hb-zk
  
  
hbase.zookeeper.property.clientPort
11181
  
{code}

For other HBase options, we use the default settings. The buggy scenario is as 
follows:

1. Start the cluster, C3HM1 becomes the active master;
2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" on 
ZooKeeper;

{code:java}
[org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.createNonSequential(RecoverableZooKeeper.java:665),
 
org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.create(RecoverableZooKeeper.java:644),
 org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndWatch(ZKUtil.java:1182), 
org.apache.hadoop.hbase.zookeeper.MetaTableLocator.setMetaLocation(MetaTableLocator.java:464),
 
org.apache.hadoop.hbase.regionserver.HRegionServer.postOpenDeployTasks(HRegionServer.java:2182),
 
org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler$PostOpenDeployTasksThread.run(OpenRegionHandler.java:329)]
{code}

3. The meta server has not been online after 10 minutes.

While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, 
everything works fine. And the bug disappears on HBase-2.4.5.

  was:
We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, 
C3RS1, C3RS2, C3RS3. 
We use an external ZooKeeper cluster which is a pseudo-distributed cluster:

{code:java}
  
hbase.zookeeper.quorum
C3hb-zk
  
  
hbase.zookeeper.property.clientPort
11181
  
{code}

For other HBase options, we use the default settings. The buggy scenario is as 
follows:

1. Start the cluster;
2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" on 
ZooKeeper;

{code:java}
[org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.createNonSequential(RecoverableZooKeeper.java:665),
 
org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.create(RecoverableZooKeeper.java:644),
 org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndWatch(ZKUtil.java:1182), 
org.apache.hadoop.hbase.zookeeper.MetaTableLocator.setMetaLocation(MetaTableLocator.java:464),
 
org.apache.hadoop.hbase.regionserver.HRegionServer.postOpenDeployTasks(HRegionServer.java:2182),
 
org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler$PostOpenDeployTasksThread.run(OpenRegionHandler.java:329)]
{code}

3. The meta server has not been online after 10 minutes.

While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, 
everything works fine. And the bug disappears on HBase-2.4.5.


> Unexpected crash of meta RegionServer causes the cluster out of service
> ---
>
> Key: HBASE-26420
> URL: https://issues.apache.org/jira/browse/HBASE-26420
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.7.1
>Reporter: anonymous
>Priority: Major
>
> We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, 
> C3RS1, C3RS2, C3RS3. 
> We use an external ZooKeeper cluster which is a pseudo-distributed cluster:
> {code:java}
>   
> hbase.zookeeper.quorum
> C3hb-zk
>   
>   
> hbase.zookeeper.property.clientPort
> 11181
>   
> {code}
> For other HBase options, we use the default settings. The buggy scenario is 
> as follows:
> 1. Start the cluster, C3HM1 becomes the active master;
> 2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" 
> on ZooKeeper;
> {code:java}
> [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.createNonSequential(RecoverableZooKeeper.java:665),
>  
> org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.create(RecoverableZooKeeper.java:644),
>  org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndWatch(ZKUtil.java:1182), 
> org.apache.hadoop.hbase.zookeeper.MetaTableLocator.setMetaLocation(MetaTableLocator.java:464),
>  
> org.apache.hadoop.hbase.regionserver.HRegionServer.postOpenDeployTasks(HRegionServer.java:2182),
>  
> org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler$PostOpenDeployTasksThread.run(OpenRegionHandler.java:329)]
> {code}
> 3. The meta server has not been online after 10 minutes.
> While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, 
> everything works fine. And the bug disappears on HBase-2.4.5.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HBASE-26420) Unexpected crash of meta RegionServer causes the cluster out of service

2021-11-03 Thread anonymous (Jira)


 [ 
https://issues.apache.org/jira/browse/HBASE-26420?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

anonymous updated HBASE-26420:
--
Description: 
We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, 
C3RS1, C3RS2, C3RS3. 
We use an external ZooKeeper cluster which is a pseudo-distributed cluster:

{code:java}
  
hbase.zookeeper.quorum
C3hb-zk
  
  
hbase.zookeeper.property.clientPort
11181
  
{code}

For other HBase options, we use the default settings. The buggy scenario is as 
follows:

1. Start the cluster;
2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" on 
ZooKeeper;

{code:java}
[org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.createNonSequential(RecoverableZooKeeper.java:665),
 
org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.create(RecoverableZooKeeper.java:644),
 org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndWatch(ZKUtil.java:1182), 
org.apache.hadoop.hbase.zookeeper.MetaTableLocator.setMetaLocation(MetaTableLocator.java:464),
 
org.apache.hadoop.hbase.regionserver.HRegionServer.postOpenDeployTasks(HRegionServer.java:2182),
 
org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler$PostOpenDeployTasksThread.run(OpenRegionHandler.java:329)]
{code}

3. The meta server has not been online after 10 minutes.

While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, 
everything works fine. And the bug disappears on HBase-2.4.5.

  was:
We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, 
C3RS1, C3RS2, C3RS3. 
We use an external ZooKeeper cluster which is a pseudo-distributed cluster:

{code:java}
  
hbase.zookeeper.quorum
C3hb-zk
  
  
hbase.zookeeper.property.clientPort
11181
  
{code}

For other HBase options, we use the default settings. The buggy scenario is as 
follows:

1. Start the cluster;
2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" on 
ZooKeeper;

{code:java}
[org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.createNonSequential(RecoverableZooKeeper.java:665),
 
org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.create(RecoverableZooKeeper.java:644),
 org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndWatch(ZKUtil.java:1182), 
org.apache.hadoop.hbase.zookeeper.MetaTableLocator.setMetaLocation(MetaTableLocator.java:464),
 
org.apache.hadoop.hbase.regionserver.HRegionServer.postOpenDeployTasks(HRegionServer.java:2182),
 
org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler$PostOpenDeployTasksThread.run(OpenRegionHandler.java:329)]
{code}

3. The meta server has not been online after 10 minutes.

While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, 
everything works fine.


> Unexpected crash of meta RegionServer causes the cluster out of service
> ---
>
> Key: HBASE-26420
> URL: https://issues.apache.org/jira/browse/HBASE-26420
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.7.1
>Reporter: anonymous
>Priority: Major
>
> We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, 
> C3RS1, C3RS2, C3RS3. 
> We use an external ZooKeeper cluster which is a pseudo-distributed cluster:
> {code:java}
>   
> hbase.zookeeper.quorum
> C3hb-zk
>   
>   
> hbase.zookeeper.property.clientPort
> 11181
>   
> {code}
> For other HBase options, we use the default settings. The buggy scenario is 
> as follows:
> 1. Start the cluster;
> 2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" 
> on ZooKeeper;
> {code:java}
> [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.createNonSequential(RecoverableZooKeeper.java:665),
>  
> org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.create(RecoverableZooKeeper.java:644),
>  org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndWatch(ZKUtil.java:1182), 
> org.apache.hadoop.hbase.zookeeper.MetaTableLocator.setMetaLocation(MetaTableLocator.java:464),
>  
> org.apache.hadoop.hbase.regionserver.HRegionServer.postOpenDeployTasks(HRegionServer.java:2182),
>  
> org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler$PostOpenDeployTasksThread.run(OpenRegionHandler.java:329)]
> {code}
> 3. The meta server has not been online after 10 minutes.
> While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, 
> everything works fine. And the bug disappears on HBase-2.4.5.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HBASE-26420) Unexpected crash of meta RegionServer causes the cluster out of service

2021-11-03 Thread anonymous (Jira)


 [ 
https://issues.apache.org/jira/browse/HBASE-26420?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

anonymous updated HBASE-26420:
--
Description: 
We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, 
C3RS1, C3RS2, C3RS3. 
We use an external ZooKeeper cluster which is a pseudo-distributed cluster:

{code:java}
  
hbase.zookeeper.quorum
C3hb-zk
  
  
hbase.zookeeper.property.clientPort
11181
  
{code}

For other HBase options, we use the default settings. The buggy scenario is as 
follows:

1. Start the cluster;
2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" on 
ZooKeeper;

{code:java}
[org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.createNonSequential(RecoverableZooKeeper.java:665),
 
org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.create(RecoverableZooKeeper.java:644),
 org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndWatch(ZKUtil.java:1182), 
org.apache.hadoop.hbase.zookeeper.MetaTableLocator.setMetaLocation(MetaTableLocator.java:464),
 
org.apache.hadoop.hbase.regionserver.HRegionServer.postOpenDeployTasks(HRegionServer.java:2182),
 
org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler$PostOpenDeployTasksThread.run(OpenRegionHandler.java:329)]
{code}

3. The meta server has not been online after 10 minutes.

While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, 
everything works fine.

  was:
We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, 
C3RS1, C3RS2, C3RS3. 
We use an external ZooKeeper cluster which is a pseudo-distributed cluster:

{code:java}
  
hbase.zookeeper.quorum
C3hb-zk
  
  
hbase.zookeeper.property.clientPort
11181
  
{code}
For other HBase options, we use the default settings. The buggy scenario is as 
follows:

1. Start the cluster;
2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" on 
ZooKeeper;

{code:java}
[org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.createNonSequential(RecoverableZooKeeper.java:665),
 
org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.create(RecoverableZooKeeper.java:644),
 org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndWatch(ZKUtil.java:1182), 
org.apache.hadoop.hbase.zookeeper.MetaTableLocator.setMetaLocation(MetaTableLocator.java:464),
 
org.apache.hadoop.hbase.regionserver.HRegionServer.postOpenDeployTasks(HRegionServer.java:2182),
 
org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler$PostOpenDeployTasksThread.run(OpenRegionHandler.java:329)]
{code}

3. The meta server has not been online after 10 minutes.

While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, 
everything works fine.


> Unexpected crash of meta RegionServer causes the cluster out of service
> ---
>
> Key: HBASE-26420
> URL: https://issues.apache.org/jira/browse/HBASE-26420
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.7.1
>Reporter: anonymous
>Priority: Major
>
> We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, 
> C3RS1, C3RS2, C3RS3. 
> We use an external ZooKeeper cluster which is a pseudo-distributed cluster:
> {code:java}
>   
> hbase.zookeeper.quorum
> C3hb-zk
>   
>   
> hbase.zookeeper.property.clientPort
> 11181
>   
> {code}
> For other HBase options, we use the default settings. The buggy scenario is 
> as follows:
> 1. Start the cluster;
> 2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" 
> on ZooKeeper;
> {code:java}
> [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.createNonSequential(RecoverableZooKeeper.java:665),
>  
> org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.create(RecoverableZooKeeper.java:644),
>  org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndWatch(ZKUtil.java:1182), 
> org.apache.hadoop.hbase.zookeeper.MetaTableLocator.setMetaLocation(MetaTableLocator.java:464),
>  
> org.apache.hadoop.hbase.regionserver.HRegionServer.postOpenDeployTasks(HRegionServer.java:2182),
>  
> org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler$PostOpenDeployTasksThread.run(OpenRegionHandler.java:329)]
> {code}
> 3. The meta server has not been online after 10 minutes.
> While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, 
> everything works fine.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HBASE-26420) Unexpected crash of meta RegionServer causes the cluster out of service

2021-11-03 Thread anonymous (Jira)


 [ 
https://issues.apache.org/jira/browse/HBASE-26420?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

anonymous updated HBASE-26420:
--
Description: 
We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, 
C3RS1, C3RS2, C3RS3. 
We use an external ZooKeeper cluster which is a pseudo-distributed cluster:

{code:java}
  
hbase.zookeeper.quorum
C3hb-zk
  
  
hbase.zookeeper.property.clientPort
11181
  
{code}
For other HBase options, we use the default settings. The buggy scenario is as 
follows:

1. Start the cluster;
2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" on 
ZooKeeper;

{code:java}
[org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.createNonSequential(RecoverableZooKeeper.java:665),
 
org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.create(RecoverableZooKeeper.java:644),
 org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndWatch(ZKUtil.java:1182), 
org.apache.hadoop.hbase.zookeeper.MetaTableLocator.setMetaLocation(MetaTableLocator.java:464),
 
org.apache.hadoop.hbase.regionserver.HRegionServer.postOpenDeployTasks(HRegionServer.java:2182),
 
org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler$PostOpenDeployTasksThread.run(OpenRegionHandler.java:329)]
{code}

3. The meta server has not been online after 10 minutes.

While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, 
everything works fine.

  was:
We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, 
C3RS1, C3RS2, C3RS3. 
We use an external ZooKeeper cluster which is a pseudo-distributed cluster:

{code:java}
  
hbase.zookeeper.quorum
C3hb-zk
  
  
hbase.zookeeper.property.clientPort
11181
  
{code}
For other HBase options, we use the default settings. The buggy scenario is as 
follows:

1. Start the cluster;
2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" on 
ZooKeeper;

{code:java}
// Some comments here

{code}

3. The meta server has not been online after 10 minutes.

While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, 
everything works fine.


> Unexpected crash of meta RegionServer causes the cluster out of service
> ---
>
> Key: HBASE-26420
> URL: https://issues.apache.org/jira/browse/HBASE-26420
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.7.1
>Reporter: anonymous
>Priority: Major
>
> We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, 
> C3RS1, C3RS2, C3RS3. 
> We use an external ZooKeeper cluster which is a pseudo-distributed cluster:
> {code:java}
>   
> hbase.zookeeper.quorum
> C3hb-zk
>   
>   
> hbase.zookeeper.property.clientPort
> 11181
>   
> {code}
> For other HBase options, we use the default settings. The buggy scenario is 
> as follows:
> 1. Start the cluster;
> 2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" 
> on ZooKeeper;
> {code:java}
> [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.createNonSequential(RecoverableZooKeeper.java:665),
>  
> org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.create(RecoverableZooKeeper.java:644),
>  org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndWatch(ZKUtil.java:1182), 
> org.apache.hadoop.hbase.zookeeper.MetaTableLocator.setMetaLocation(MetaTableLocator.java:464),
>  
> org.apache.hadoop.hbase.regionserver.HRegionServer.postOpenDeployTasks(HRegionServer.java:2182),
>  
> org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler$PostOpenDeployTasksThread.run(OpenRegionHandler.java:329)]
> {code}
> 3. The meta server has not been online after 10 minutes.
> While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, 
> everything works fine.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HBASE-26420) Unexpected crash of meta RegionServer causes the cluster out of service

2021-11-03 Thread anonymous (Jira)


 [ 
https://issues.apache.org/jira/browse/HBASE-26420?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

anonymous updated HBASE-26420:
--
Description: 
We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, 
C3RS1, C3RS2, C3RS3. 
We use an external ZooKeeper cluster which is a pseudo-distributed cluster:

{code:java}
  
hbase.zookeeper.quorum
C3hb-zk
  
  
hbase.zookeeper.property.clientPort
11181
  
{code}
For other HBase options, we use the default settings. The buggy scenario is as 
follows:

1. Start the cluster;
2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" on 
ZooKeeper;

{code:java}
// Some comments here
public String getFoo()
{
return foo;
}
{code}

3. The meta server has not been online after 10 minutes.

While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, 
everything works fine.

> Unexpected crash of meta RegionServer causes the cluster out of service
> ---
>
> Key: HBASE-26420
> URL: https://issues.apache.org/jira/browse/HBASE-26420
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.7.1
>Reporter: anonymous
>Priority: Major
>
> We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, 
> C3RS1, C3RS2, C3RS3. 
> We use an external ZooKeeper cluster which is a pseudo-distributed cluster:
> {code:java}
>   
> hbase.zookeeper.quorum
> C3hb-zk
>   
>   
> hbase.zookeeper.property.clientPort
> 11181
>   
> {code}
> For other HBase options, we use the default settings. The buggy scenario is 
> as follows:
> 1. Start the cluster;
> 2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" 
> on ZooKeeper;
> {code:java}
> // Some comments here
> public String getFoo()
> {
> return foo;
> }
> {code}
> 3. The meta server has not been online after 10 minutes.
> While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, 
> everything works fine.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HBASE-26420) Unexpected crash of meta RegionServer causes the cluster out of service

2021-11-03 Thread anonymous (Jira)


 [ 
https://issues.apache.org/jira/browse/HBASE-26420?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

anonymous updated HBASE-26420:
--
Description: 
We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, 
C3RS1, C3RS2, C3RS3. 
We use an external ZooKeeper cluster which is a pseudo-distributed cluster:

{code:java}
  
hbase.zookeeper.quorum
C3hb-zk
  
  
hbase.zookeeper.property.clientPort
11181
  
{code}
For other HBase options, we use the default settings. The buggy scenario is as 
follows:

1. Start the cluster;
2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" on 
ZooKeeper;

{code:java}
// Some comments here

{code}

3. The meta server has not been online after 10 minutes.

While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, 
everything works fine.

  was:
We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, 
C3RS1, C3RS2, C3RS3. 
We use an external ZooKeeper cluster which is a pseudo-distributed cluster:

{code:java}
  
hbase.zookeeper.quorum
C3hb-zk
  
  
hbase.zookeeper.property.clientPort
11181
  
{code}
For other HBase options, we use the default settings. The buggy scenario is as 
follows:

1. Start the cluster;
2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" on 
ZooKeeper;

{code:java}
// Some comments here
public String getFoo()
{
return foo;
}
{code}

3. The meta server has not been online after 10 minutes.

While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, 
everything works fine.


> Unexpected crash of meta RegionServer causes the cluster out of service
> ---
>
> Key: HBASE-26420
> URL: https://issues.apache.org/jira/browse/HBASE-26420
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.7.1
>Reporter: anonymous
>Priority: Major
>
> We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, 
> C3RS1, C3RS2, C3RS3. 
> We use an external ZooKeeper cluster which is a pseudo-distributed cluster:
> {code:java}
>   
> hbase.zookeeper.quorum
> C3hb-zk
>   
>   
> hbase.zookeeper.property.clientPort
> 11181
>   
> {code}
> For other HBase options, we use the default settings. The buggy scenario is 
> as follows:
> 1. Start the cluster;
> 2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" 
> on ZooKeeper;
> {code:java}
> // Some comments here
> {code}
> 3. The meta server has not been online after 10 minutes.
> While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, 
> everything works fine.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HBASE-26420) Unexpected crash of meta RegionServer causes the cluster out of service

2021-11-03 Thread anonymous (Jira)


 [ 
https://issues.apache.org/jira/browse/HBASE-26420?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

anonymous updated HBASE-26420:
--
Summary: Unexpected crash of meta RegionServer causes the cluster out of 
service  (was: Unexpected crash of HMaster causes the cluster out of service)

> Unexpected crash of meta RegionServer causes the cluster out of service
> ---
>
> Key: HBASE-26420
> URL: https://issues.apache.org/jira/browse/HBASE-26420
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.7.1
>Reporter: anonymous
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (HBASE-26420) Unexpected crash of HMaster causes the cluster out of service

2021-11-03 Thread anonymous (Jira)
anonymous created HBASE-26420:
-

 Summary: Unexpected crash of HMaster causes the cluster out of 
service
 Key: HBASE-26420
 URL: https://issues.apache.org/jira/browse/HBASE-26420
 Project: HBase
  Issue Type: Bug
Affects Versions: 1.7.1
Reporter: anonymous






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HBASE-26391) Unexpected crash in the process of deleting a table causes stale data

2021-10-22 Thread anonymous (Jira)


 [ 
https://issues.apache.org/jira/browse/HBASE-26391?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

anonymous updated HBASE-26391:
--
Summary: Unexpected crash in the process of deleting a table causes stale 
data  (was: Unexpected crash in the process of deleting table causes stale data)

> Unexpected crash in the process of deleting a table causes stale data
> -
>
> Key: HBASE-26391
> URL: https://issues.apache.org/jira/browse/HBASE-26391
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.7.1
>Reporter: anonymous
>Priority: Major
>
> There is a table {{mytable}} in a HBase cluster with two HMasters and two 
> RegionServers.
> 1. Current active HMaster is HMaster1;
> 2. Client request to delete the table {{mytable}};
> 3. HMaster1 tries to delete znode in ZooKeeper: /hbase/table/mytable
> {code:java}
> [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.delete(RecoverableZooKeeper.java:238),
>  
> org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1367),
>  
> org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1359),
>  
> org.apache.hadoop.hbase.zookeeper.ZKTableStateManager.setDeletedTable(ZKTableStateManager.java:244),
>  
> org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.deleteAssignmentState(DeleteTableProcedure.java:383),
>  
> org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:136),
>  
> org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:57),
>  
> org.apache.hadoop.hbase.procedure2.StateMachineProcedure.execute(StateMachineProcedure.java:139),
>  org.apache.hadoop.hbase.procedure2.Procedure.doExecute(Procedure.java:499), 
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execProcedure(ProcedureExecutor.java:1166),
>  
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:954),
>  
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:907),
>  
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor.access$400(ProcedureExecutor.java:76),
>  
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java:481),
>  
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java)]
> {code}
> 4. HMaster1 crashes before deleting the znode: /hbase/table-lock/mytable
> {code:java}
> [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.delete(RecoverableZooKeeper.java:238),
>  org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNode(ZKUtil.java:1341), 
> org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNode(ZKUtil.java:1330), 
> org.apache.hadoop.hbase.master.TableLockManager$ZKTableLockManager.tableDeleted(TableLockManager.java:442),
>  
> org.apache.hadoop.hbase.master.procedure.MasterProcedureScheduler.markTableAsDeleted(MasterProcedureScheduler.java:895),
>  
> org.apache.hadoop.hbase.master.procedure.MasterProcedureScheduler.completionCleanup(MasterProcedureScheduler.java:261),
>  
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor.procedureFinished(ProcedureExecutor.java:1415),
>  
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:974),
>  
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:907),
>  
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor.access$400(ProcedureExecutor.java:76),
>  
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java:481),
>  
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java)]
> {code}
> 5. HMaster2 becomes active
> 6. The table {{mytablle}} was deleted, but the znode 
> "/hbase/table-lock/mytable" was left on ZooKeeper



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HBASE-26391) Unexpected crash in the process of deleting a table causes stale data in ZooKeeper

2021-10-22 Thread anonymous (Jira)


 [ 
https://issues.apache.org/jira/browse/HBASE-26391?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

anonymous updated HBASE-26391:
--
Summary: Unexpected crash in the process of deleting a table causes stale 
data in ZooKeeper  (was: Unexpected crash in the process of deleting a table 
causes stale data)

> Unexpected crash in the process of deleting a table causes stale data in 
> ZooKeeper
> --
>
> Key: HBASE-26391
> URL: https://issues.apache.org/jira/browse/HBASE-26391
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.7.1
>Reporter: anonymous
>Priority: Major
>
> There is a table {{mytable}} in a HBase cluster with two HMasters and two 
> RegionServers.
> 1. Current active HMaster is HMaster1;
> 2. Client request to delete the table {{mytable}};
> 3. HMaster1 tries to delete znode in ZooKeeper: /hbase/table/mytable
> {code:java}
> [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.delete(RecoverableZooKeeper.java:238),
>  
> org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1367),
>  
> org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1359),
>  
> org.apache.hadoop.hbase.zookeeper.ZKTableStateManager.setDeletedTable(ZKTableStateManager.java:244),
>  
> org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.deleteAssignmentState(DeleteTableProcedure.java:383),
>  
> org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:136),
>  
> org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:57),
>  
> org.apache.hadoop.hbase.procedure2.StateMachineProcedure.execute(StateMachineProcedure.java:139),
>  org.apache.hadoop.hbase.procedure2.Procedure.doExecute(Procedure.java:499), 
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execProcedure(ProcedureExecutor.java:1166),
>  
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:954),
>  
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:907),
>  
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor.access$400(ProcedureExecutor.java:76),
>  
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java:481),
>  
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java)]
> {code}
> 4. HMaster1 crashes before deleting the znode: /hbase/table-lock/mytable
> {code:java}
> [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.delete(RecoverableZooKeeper.java:238),
>  org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNode(ZKUtil.java:1341), 
> org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNode(ZKUtil.java:1330), 
> org.apache.hadoop.hbase.master.TableLockManager$ZKTableLockManager.tableDeleted(TableLockManager.java:442),
>  
> org.apache.hadoop.hbase.master.procedure.MasterProcedureScheduler.markTableAsDeleted(MasterProcedureScheduler.java:895),
>  
> org.apache.hadoop.hbase.master.procedure.MasterProcedureScheduler.completionCleanup(MasterProcedureScheduler.java:261),
>  
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor.procedureFinished(ProcedureExecutor.java:1415),
>  
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:974),
>  
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:907),
>  
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor.access$400(ProcedureExecutor.java:76),
>  
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java:481),
>  
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java)]
> {code}
> 5. HMaster2 becomes active
> 6. The table {{mytablle}} was deleted, but the znode 
> "/hbase/table-lock/mytable" was left on ZooKeeper



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HBASE-26391) Unexpected crash in the process of deleting table causes stale data

2021-10-22 Thread anonymous (Jira)


 [ 
https://issues.apache.org/jira/browse/HBASE-26391?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

anonymous updated HBASE-26391:
--
Description: 
There is a table {{mytable}} in a HBase cluster with two HMasters and two 
RegionServers.

1. Current active HMaster is HMaster1;
2. Client request to delete the table {{mytable}};
3. HMaster1 tries to delete znode in ZooKeeper: /hbase/table/mytable

{code:java}
[org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.delete(RecoverableZooKeeper.java:238),
 
org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1367),
 
org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1359),
 
org.apache.hadoop.hbase.zookeeper.ZKTableStateManager.setDeletedTable(ZKTableStateManager.java:244),
 
org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.deleteAssignmentState(DeleteTableProcedure.java:383),
 
org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:136),
 
org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:57),
 
org.apache.hadoop.hbase.procedure2.StateMachineProcedure.execute(StateMachineProcedure.java:139),
 org.apache.hadoop.hbase.procedure2.Procedure.doExecute(Procedure.java:499), 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execProcedure(ProcedureExecutor.java:1166),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:954),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:907),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.access$400(ProcedureExecutor.java:76),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java:481),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java)]
{code}

4. HMaster1 crashes before deleting the znode: /hbase/table-lock/mytable

{code:java}
[org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.delete(RecoverableZooKeeper.java:238),
 org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNode(ZKUtil.java:1341), 
org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNode(ZKUtil.java:1330), 
org.apache.hadoop.hbase.master.TableLockManager$ZKTableLockManager.tableDeleted(TableLockManager.java:442),
 
org.apache.hadoop.hbase.master.procedure.MasterProcedureScheduler.markTableAsDeleted(MasterProcedureScheduler.java:895),
 
org.apache.hadoop.hbase.master.procedure.MasterProcedureScheduler.completionCleanup(MasterProcedureScheduler.java:261),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.procedureFinished(ProcedureExecutor.java:1415),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:974),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:907),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.access$400(ProcedureExecutor.java:76),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java:481),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java)]
{code}

5. HMaster2 becomes active
6. The table {{mytablle}} was deleted, but the znode 
"/hbase/table-lock/mytable" was left on ZooKeeper




  was:
There is a table {{mytable}} in a HBase cluster with two HMasters and two 
RegionServers.

1. Current active HMaster is HMaster1;
2. Client request to delete the table {{mytable}};
3. HMaster1 tries to delete znode in ZooKeeper: /hbase/table/mytable

{code:java}
[org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.delete(RecoverableZooKeeper.java:238),
 
org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1367),
 
org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1359),
 
org.apache.hadoop.hbase.zookeeper.ZKTableStateManager.setDeletedTable(ZKTableStateManager.java:244),
 
org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.deleteAssignmentState(DeleteTableProcedure.java:383),
 
org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:136),
 
org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:57),
 
org.apache.hadoop.hbase.procedure2.StateMachineProcedure.execute(StateMachineProcedure.java:139),
 org.apache.hadoop.hbase.procedure2.Procedure.doExecute(Procedure.java:499), 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execProcedure(ProcedureExecutor.java:1166),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:954),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:907),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.access$400(ProcedureExecutor.java:76),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java:481),
 

[jira] [Updated] (HBASE-26391) Unexpected crash in the process of deleting table causes stale data

2021-10-22 Thread anonymous (Jira)


 [ 
https://issues.apache.org/jira/browse/HBASE-26391?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

anonymous updated HBASE-26391:
--
Description: 
There is a table {{mytable}} in a HBase cluster with two HMasters and two 
RegionServers.

1. Current active HMaster is HMaster1;
2. Client request to delete the table {{mytable}};
3. HMaster1 tries to delete znode in ZooKeeper: /hbase/table/mytable

{code:java}
[org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.delete(RecoverableZooKeeper.java:238),
 
org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1367),
 
org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1359),
 
org.apache.hadoop.hbase.zookeeper.ZKTableStateManager.setDeletedTable(ZKTableStateManager.java:244),
 
org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.deleteAssignmentState(DeleteTableProcedure.java:383),
 
org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:136),
 
org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:57),
 
org.apache.hadoop.hbase.procedure2.StateMachineProcedure.execute(StateMachineProcedure.java:139),
 org.apache.hadoop.hbase.procedure2.Procedure.doExecute(Procedure.java:499), 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execProcedure(ProcedureExecutor.java:1166),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:954),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:907),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.access$400(ProcedureExecutor.java:76),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java:481),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java)]
{code}

4. HMaster1 crashes before deleting the znode: /hbase/table-lock/mytable

{code:java}
[org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.delete(RecoverableZooKeeper.java:238),
 org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNode(ZKUtil.java:1341), 
org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNode(ZKUtil.java:1330), 
org.apache.hadoop.hbase.master.TableLockManager$ZKTableLockManager.tableDeleted(TableLockManager.java:442),
 
org.apache.hadoop.hbase.master.procedure.MasterProcedureScheduler.markTableAsDeleted(MasterProcedureScheduler.java:895),
 
org.apache.hadoop.hbase.master.procedure.MasterProcedureScheduler.completionCleanup(MasterProcedureScheduler.java:261),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.procedureFinished(ProcedureExecutor.java:1415),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:974),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:907),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.access$400(ProcedureExecutor.java:76),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java:481),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java)]
{code}

5. HMaster2 becomes active
6. znode "/hbase/table-lock/mytable" was left on ZooKeeper




  was:
There is a table {{mytable}} in a HBase cluster with two HMasters and two 
RegionServers.

1. Current active HMaster is HMaster1;
2. Client request to delete the table {{mytable}};
3. HMaster1 tries to delete znode in ZooKeeper: /hbase/table/mytable

{[org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.delete(RecoverableZooKeeper.java:238),
 
org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1367),
 
org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1359),
 
org.apache.hadoop.hbase.zookeeper.ZKTableStateManager.setDeletedTable(ZKTableStateManager.java:244),
 
org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.deleteAssignmentState(DeleteTableProcedure.java:383),
 
org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:136),
 
org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:57),
 
org.apache.hadoop.hbase.procedure2.StateMachineProcedure.execute(StateMachineProcedure.java:139),
 org.apache.hadoop.hbase.procedure2.Procedure.doExecute(Procedure.java:499), 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execProcedure(ProcedureExecutor.java:1166),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:954),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:907),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.access$400(ProcedureExecutor.java:76),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java:481),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java)]}






> Unexpected crash in the process of 

[jira] [Updated] (HBASE-26391) Unexpected crash in the process of deleting table causes stale data

2021-10-22 Thread anonymous (Jira)


 [ 
https://issues.apache.org/jira/browse/HBASE-26391?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

anonymous updated HBASE-26391:
--
Description: 
There is a table {{mytable}} in a HBase cluster with two HMasters and two 
RegionServers.

1. Current active HMaster is HMaster1;
2. Client request to delete the table {{mytable}};
3. HMaster1 tries to delete znode in ZooKeeper: /hbase/table/mytable

{[org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.delete(RecoverableZooKeeper.java:238),
 
org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1367),
 
org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1359),
 
org.apache.hadoop.hbase.zookeeper.ZKTableStateManager.setDeletedTable(ZKTableStateManager.java:244),
 
org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.deleteAssignmentState(DeleteTableProcedure.java:383),
 
org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:136),
 
org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:57),
 
org.apache.hadoop.hbase.procedure2.StateMachineProcedure.execute(StateMachineProcedure.java:139),
 org.apache.hadoop.hbase.procedure2.Procedure.doExecute(Procedure.java:499), 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execProcedure(ProcedureExecutor.java:1166),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:954),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:907),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor.access$400(ProcedureExecutor.java:76),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java:481),
 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java)]}





> Unexpected crash in the process of deleting table causes stale data
> ---
>
> Key: HBASE-26391
> URL: https://issues.apache.org/jira/browse/HBASE-26391
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.7.1
>Reporter: anonymous
>Priority: Major
>
> There is a table {{mytable}} in a HBase cluster with two HMasters and two 
> RegionServers.
> 1. Current active HMaster is HMaster1;
> 2. Client request to delete the table {{mytable}};
> 3. HMaster1 tries to delete znode in ZooKeeper: /hbase/table/mytable
> {[org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.delete(RecoverableZooKeeper.java:238),
>  
> org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1367),
>  
> org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1359),
>  
> org.apache.hadoop.hbase.zookeeper.ZKTableStateManager.setDeletedTable(ZKTableStateManager.java:244),
>  
> org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.deleteAssignmentState(DeleteTableProcedure.java:383),
>  
> org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:136),
>  
> org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:57),
>  
> org.apache.hadoop.hbase.procedure2.StateMachineProcedure.execute(StateMachineProcedure.java:139),
>  org.apache.hadoop.hbase.procedure2.Procedure.doExecute(Procedure.java:499), 
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execProcedure(ProcedureExecutor.java:1166),
>  
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:954),
>  
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:907),
>  
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor.access$400(ProcedureExecutor.java:76),
>  
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java:481),
>  
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java)]}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (HBASE-26391) Unexpected crash in the process of deleting table causes stale data

2021-10-22 Thread anonymous (Jira)
anonymous created HBASE-26391:
-

 Summary: Unexpected crash in the process of deleting table causes 
stale data
 Key: HBASE-26391
 URL: https://issues.apache.org/jira/browse/HBASE-26391
 Project: HBase
  Issue Type: Bug
Affects Versions: 1.7.1
Reporter: anonymous






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HBASE-26370) Unexpected crash fails the truncate operation and makes a disabled table enabled

2021-10-19 Thread anonymous (Jira)


 [ 
https://issues.apache.org/jira/browse/HBASE-26370?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

anonymous updated HBASE-26370:
--
Description: 
There is a table {{mytable}} in a HBase cluster with two HMasters and two 
RegionServers.

1. Current active HMaster is HMaster1;
2. Client disabled the table {{mytable}};
3. {{mytable}} is in disabled state;
4. Client request to truncate {{mytable}};
5. HMaster1 tries to truncate {{mytable}}. For the TruncateTableState 
{{TRUNCATE_TABLE_ASSIGN_REGIONS}}, HMaster1 tries to assign regions of the 
table and mark the table as Enabling;
6. HMaster1 crashes;
7. HMaster2 becomes active;
8. The truncate operation fails due to {{TableNotDisabledException}}, and 
{{mytable}} was left as enabled.

In HMaster2, we got following info:

{code:java}
INFO [ProcedureExecutor-12] procedure2.ProcedureExecutor: Rolledback procedure 
TruncateTableProcedure (table=mytable preserveSplits=true) id=9 owner=myhbase 
state=ROLLEDBACK exec-time=4.6030 sec 
exception=org.apache.hadoop.hbase.TableNotDisabledException: mytable
{code}


  was:
There is a table {{mytable}} in a HBase cluster with two HMasters and two 
RegionServers.

1. Current active HMaster is HMaster1;
2. Client disabled the table {{mytable}};
3. {{mytable}} is in disabled state;
4. Client request to truncate {{mytable}};
5. HMaster1 tries to truncate {{mytable}}. For the TruncateTableState 
{{TRUNCATE_TABLE_ASSIGN_REGIONS}}, HMaster1 tries to assign regions of the 
table and mark the table as Enabling;
6. HMaster1 crashes;
7. HMaster2 becomes active;
8. The truncate operation fails due to {{TableNotDisabledException}}, and 
{{mytable}} was left as enabled.


> Unexpected crash fails the truncate operation and makes a disabled table 
> enabled
> 
>
> Key: HBASE-26370
> URL: https://issues.apache.org/jira/browse/HBASE-26370
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.7.1
>Reporter: anonymous
>Priority: Major
>
> There is a table {{mytable}} in a HBase cluster with two HMasters and two 
> RegionServers.
> 1. Current active HMaster is HMaster1;
> 2. Client disabled the table {{mytable}};
> 3. {{mytable}} is in disabled state;
> 4. Client request to truncate {{mytable}};
> 5. HMaster1 tries to truncate {{mytable}}. For the TruncateTableState 
> {{TRUNCATE_TABLE_ASSIGN_REGIONS}}, HMaster1 tries to assign regions of the 
> table and mark the table as Enabling;
> 6. HMaster1 crashes;
> 7. HMaster2 becomes active;
> 8. The truncate operation fails due to {{TableNotDisabledException}}, and 
> {{mytable}} was left as enabled.
> In HMaster2, we got following info:
> {code:java}
> INFO [ProcedureExecutor-12] procedure2.ProcedureExecutor: Rolledback 
> procedure TruncateTableProcedure (table=mytable preserveSplits=true) id=9 
> owner=myhbase state=ROLLEDBACK exec-time=4.6030 sec 
> exception=org.apache.hadoop.hbase.TableNotDisabledException: mytable
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HBASE-26370) Unexpected crash fails the truncate operation and makes a disabled table enabled

2021-10-19 Thread anonymous (Jira)


 [ 
https://issues.apache.org/jira/browse/HBASE-26370?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

anonymous updated HBASE-26370:
--
Summary: Unexpected crash fails the truncate operation and makes a disabled 
table enabled  (was: Unexpected crash makes a disabled table enabled)

> Unexpected crash fails the truncate operation and makes a disabled table 
> enabled
> 
>
> Key: HBASE-26370
> URL: https://issues.apache.org/jira/browse/HBASE-26370
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.7.1
>Reporter: anonymous
>Priority: Major
>
> There is a table {{mytable}} in a HBase cluster with two HMasters and two 
> RegionServers.
> 1. Current active HMaster is HMaster1;
> 2. Client disabled the table {{mytable}};
> 3. {{mytable}} is in disabled state;
> 4. Client request to truncate {{mytable}};
> 5. HMaster1 tries to truncate {{mytable}}. For the TruncateTableState 
> {{TRUNCATE_TABLE_ASSIGN_REGIONS}}, HMaster1 tries to assign regions of the 
> table and mark the table as Enabling;
> 6. HMaster1 crashes;
> 7. HMaster2 becomes active;
> 8. The truncate operation fails due to {{TableNotDisabledException}}, and 
> {{mytable}} was left as enabled.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HBASE-26370) Unexpected crash makes a disabled table enabled

2021-10-19 Thread anonymous (Jira)


 [ 
https://issues.apache.org/jira/browse/HBASE-26370?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

anonymous updated HBASE-26370:
--
Summary: Unexpected crash makes a disabled table enabled  (was: Unexpected 
crash makes a disabled table be enabled)

> Unexpected crash makes a disabled table enabled
> ---
>
> Key: HBASE-26370
> URL: https://issues.apache.org/jira/browse/HBASE-26370
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.7.1
>Reporter: anonymous
>Priority: Major
>
> There is a table {{mytable}} in a HBase cluster with two HMasters and two 
> RegionServers.
> 1. Current active HMaster is HMaster1;
> 2. Client disabled the table {{mytable}};
> 3. {{mytable}} is in disabled state;
> 4. Client request to truncate {{mytable}};
> 5. HMaster1 tries to truncate {{mytable}}. For the TruncateTableState 
> {{TRUNCATE_TABLE_ASSIGN_REGIONS}}, HMaster1 tries to assign regions of the 
> table and mark the table as Enabling;
> 6. HMaster1 crashes;
> 7. HMaster2 becomes active;
> 8. The truncate operation fails due to {{TableNotDisabledException}}, and 
> {{mytable}} was left as enabled.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HBASE-26370) Unexpected crash makes a disabled table be enabled

2021-10-19 Thread anonymous (Jira)


 [ 
https://issues.apache.org/jira/browse/HBASE-26370?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

anonymous updated HBASE-26370:
--
Summary: Unexpected crash makes a disabled table be enabled  (was: 
Unexpected crash make a disabled table be enabled)

> Unexpected crash makes a disabled table be enabled
> --
>
> Key: HBASE-26370
> URL: https://issues.apache.org/jira/browse/HBASE-26370
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.7.1
>Reporter: anonymous
>Priority: Major
>
> There is a table {{mytable}} in a HBase cluster with two HMasters and two 
> RegionServers.
> 1. Current active HMaster is HMaster1;
> 2. Client disabled the table {{mytable}};
> 3. {{mytable}} is in disabled state;
> 4. Client request to truncate {{mytable}};
> 5. HMaster1 tries to truncate {{mytable}}. For the TruncateTableState 
> {{TRUNCATE_TABLE_ASSIGN_REGIONS}}, HMaster1 tries to assign regions of the 
> table and mark the table as Enabling;
> 6. HMaster1 crashes;
> 7. HMaster2 becomes active;
> 8. The truncate operation fails due to {{TableNotDisabledException}}, and 
> {{mytable}} was left as enabled.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (HBASE-26370) Unexpected crash make a disabled table be enabled

2021-10-19 Thread anonymous (Jira)
anonymous created HBASE-26370:
-

 Summary: Unexpected crash make a disabled table be enabled
 Key: HBASE-26370
 URL: https://issues.apache.org/jira/browse/HBASE-26370
 Project: HBase
  Issue Type: Bug
Affects Versions: 1.7.1
Reporter: anonymous


There is a table {{mytable}} in a HBase cluster with two HMasters and two 
RegionServers.

1. Current active HMaster is HMaster1;
2. Client disabled the table {{mytable}};
3. {{mytable}} is in disabled state;
4. Client request to truncate {{mytable}};
5. HMaster1 tries to truncate {{mytable}}. For the TruncateTableState 
{{TRUNCATE_TABLE_ASSIGN_REGIONS}}, HMaster1 tries to assign regions of the 
table and mark the table as Enabling;
6. HMaster1 crashes;
7. HMaster2 becomes active;
8. The truncate operation fails due to {{TableNotDisabledException}}, and 
{{mytable}} was left as enabled.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HBASE-19400) Add missing security checks in MasterRpcServices

2018-02-23 Thread Anonymous (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-19400?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anonymous updated HBASE-19400:
--
Status: Patch Available  (was: Reopened)

> Add missing security checks in MasterRpcServices
> 
>
> Key: HBASE-19400
> URL: https://issues.apache.org/jira/browse/HBASE-19400
> Project: HBase
>  Issue Type: Sub-task
>Affects Versions: 2.0.0-beta-1
>Reporter: Balazs Meszaros
>Assignee: Appy
>Priority: Major
> Fix For: 2.0.0-beta-2
>
> Attachments: HBASE-19400.branch-1.001.patch, 
> HBASE-19400.master.001.patch, HBASE-19400.master.002.patch, 
> HBASE-19400.master.003.patch, HBASE-19400.master.004.patch, 
> HBASE-19400.master.004.patch, HBASE-19400.master.005.patch, 
> HBASE-19400.master.006.patch, HBASE-19400.master.007.patch, 
> HBASE-19400.master.007.patch
>
>
> The following RPC methods in MasterRpcServices do not have ACL check for 
> ADMIN rights.
> - normalize
> - setNormalizerRunning
> - runCatalogScan
> - enableCatalogJanitor
> - runCleanerChore
> - setCleanerChoreRunning
> - execMasterService
> - execProcedure
> - execProcedureWithRet



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-18230) Generated LICENSE file includes unsubstituted Velocity variables

2017-06-26 Thread Anonymous (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-18230?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anonymous updated HBASE-18230:
--
Status: Patch Available  (was: Reopened)

> Generated LICENSE file includes unsubstituted Velocity variables
> 
>
> Key: HBASE-18230
> URL: https://issues.apache.org/jira/browse/HBASE-18230
> Project: HBase
>  Issue Type: Bug
>  Components: build
>Affects Versions: 2.0.0-alpha-1
>Reporter: Mike Drob
>Assignee: Mike Drob
> Fix For: 3.0.0, 1.4.0, 1.3.2, 1.2.7, 2.0.0-alpha-2, 1.1.12
>
> Attachments: HBASE-18230.addendum.patch, HBASE-18230.patch
>
>
> From the release vote:
> {quote}
> we have a ton of places where we have velocity variables instead of
> copyright years, but IIRC that's a problem on branch-1 right now too.
> {quote}
> This is referring to lines like these:
> {noformat}
>   * javax.annotation-api, ${dep.licenses[0].comments}
>   * javax.servlet-api, ${dep.licenses[0].comments}
>   * jetty-schemas, ${dep.licenses[0].comments}
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HBASE-15291) FileSystem not closed in secure bulkLoad

2016-02-28 Thread Anonymous (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-15291?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anonymous updated HBASE-15291:
--
Affects Version/s: 1.0.2
   0.98.16.1
   Status: Patch Available  (was: Reopened)

> FileSystem not closed in secure bulkLoad
> 
>
> Key: HBASE-15291
> URL: https://issues.apache.org/jira/browse/HBASE-15291
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 0.98.16.1, 1.0.2
>Reporter: Yong Zhang
>Assignee: Yong Zhang
> Fix For: 2.0.0, 1.3.0, 1.2.1, 0.98.18
>
> Attachments: HBASE-15291.001.patch, HBASE-15291.002.patch, 
> HBASE-15291.addendum, patch
>
>
> FileSystem not closed in secure bulkLoad after bulkLoad  finish, it will 
> cause memory used more and more if too many bulkLoad .



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HBASE-14497) Reverse Scan threw StackOverflow caused by readPt checking

2015-10-08 Thread Anonymous (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-14497?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anonymous updated HBASE-14497:
--
Status: Patch Available  (was: Reopened)

> Reverse Scan threw StackOverflow caused by readPt checking
> --
>
> Key: HBASE-14497
> URL: https://issues.apache.org/jira/browse/HBASE-14497
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 0.98.14, 2.0.0, 1.3.0
>Reporter: Yerui Sun
>Assignee: Yerui Sun
> Fix For: 2.0.0, 1.3.0
>
> Attachments: 14497-branch-1-v6.patch, 14497-master-v6.patch, 
> HBASE-14497-0.98-v6.patch, HBASE-14497-0.98.patch, 
> HBASE-14497-branch-1-v2.patch, HBASE-14497-branch-1-v3.patch, 
> HBASE-14497-branch-1-v6.patch, HBASE-14497-branch-1.patch, 
> HBASE-14497-master-v2.patch, HBASE-14497-master-v3.patch, 
> HBASE-14497-master-v3.patch, HBASE-14497-master-v4.patch, 
> HBASE-14497-master-v5.patch, HBASE-14497-master.patch
>
>
> I met stack overflow error in StoreFileScanner.seekToPreviousRow using 
> reversed scan. I searched and founded HBASE-14155, but it seems to be a 
> different reason.
> The seekToPreviousRow will fetch the row which closest before, and compare 
> mvcc to the readPt, which acquired when scanner created. If the row's mvcc is 
> bigger than readPt, an recursive call of seekToPreviousRow will invoked, to 
> find the next closest before row.
> Considering we created a scanner for reversed scan, and some data with 
> smaller rows was written and flushed, before calling scanner next. When 
> seekToPreviousRow was invoked, it would call itself recursively, until all 
> rows which written after scanner created were iterated. The depth of 
> recursive calling stack depends on the count of rows, the stack overflow 
> error will be threw if the count of rows is large, like 1.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HBASE-13965) Stochastic Load Balancer JMX Metrics

2015-08-05 Thread Anonymous (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-13965?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anonymous updated HBASE-13965:
--
Status: Patch Available  (was: Reopened)

 Stochastic Load Balancer JMX Metrics
 

 Key: HBASE-13965
 URL: https://issues.apache.org/jira/browse/HBASE-13965
 Project: HBase
  Issue Type: Improvement
  Components: Balancer, metrics
Reporter: Lei Chen
Assignee: Lei Chen
 Fix For: 2.0.0

 Attachments: 13965-addendum.txt, HBASE-13965-branch-1.patch, 
 HBASE-13965-v10.patch, HBASE-13965-v11.patch, HBASE-13965-v3.patch, 
 HBASE-13965-v4.patch, HBASE-13965-v5.patch, HBASE-13965-v6.patch, 
 HBASE-13965-v7.patch, HBASE-13965-v8.patch, HBASE-13965-v9.patch, 
 HBASE-13965_v2.patch, HBase-13965-JConsole.png, HBase-13965-v1.patch, 
 stochasticloadbalancerclasses_v2.png


 Today’s default HBase load balancer (the Stochastic load balancer) is cost 
 function based. The cost function weights are tunable but no visibility into 
 those cost function results is directly provided.
 A driving example is a cluster we have been tuning which has skewed rack size 
 (one rack has half the nodes of the other few racks). We are tuning the 
 cluster for uniform response time from all region servers with the ability to 
 tolerate a rack failure. Balancing LocalityCost, RegionReplicaRack Cost and 
 RegionCountSkew Cost is difficult without a way to attribute each cost 
 function’s contribution to overall cost. 
 What this jira proposes is to provide visibility via JMX into each cost 
 function of the stochastic load balancer, as well as the overall cost of the 
 balancing plan.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HBASE-13082) Coarsen StoreScanner locks to RegionScanner

2015-03-08 Thread Anonymous (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-13082?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anonymous updated HBASE-13082:
--
Status: Patch Available  (was: Reopened)

 Coarsen StoreScanner locks to RegionScanner
 ---

 Key: HBASE-13082
 URL: https://issues.apache.org/jira/browse/HBASE-13082
 Project: HBase
  Issue Type: Bug
Reporter: Lars Hofhansl
Assignee: Lars Hofhansl
 Fix For: 2.0.0, 1.1.0

 Attachments: 13082-test.txt, 13082-v2.txt, 13082-v3.txt, 13082.txt, 
 13082.txt, gc.png, gc.png, gc.png, hits.png, next.png, next.png


 Continuing where HBASE-10015 left of.
 We can avoid locking (and memory fencing) inside StoreScanner by deferring to 
 the lock already held by the RegionScanner.
 In tests this shows quite a scan improvement and reduced CPU (the fences make 
 the cores wait for memory fetches).
 There are some drawbacks too:
 * All calls to RegionScanner need to be remain synchronized
 * Implementors of coprocessors need to be diligent in following the locking 
 contract. For example Phoenix does not lock RegionScanner.nextRaw() and 
 required in the documentation (not picking on Phoenix, this one is my fault 
 as I told them it's OK)
 * possible starving of flushes and compaction with heavy read load. 
 RegionScanner operations would keep getting the locks and the 
 flushes/compactions would not be able finalize the set of files.
 I'll have a patch soon.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HBASE-5075) regionserver crashed and failover

2012-11-14 Thread Anonymous (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-5075?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anonymous updated HBASE-5075:
-

Affects Version/s: (was: 0.92.1)
   0.90.5
 Hadoop Flags: Incompatible change
   Status: Patch Available  (was: Reopened)

 regionserver crashed and failover
 -

 Key: HBASE-5075
 URL: https://issues.apache.org/jira/browse/HBASE-5075
 Project: HBase
  Issue Type: Improvement
  Components: monitoring, regionserver, Replication, Zookeeper
Affects Versions: 0.90.5
Reporter: zhiyuan.dai
 Fix For: 0.90.5

 Attachments: Degion of Failure Detection.pdf, HBase-5075-shell.patch, 
 HBase-5075-src.patch


 regionserver crashed,it is too long time to notify hmaster.when hmaster know 
 regionserver's shutdown,it is long time to fetch the hlog's lease.
 hbase is a online db, availability is very important.
 i have a idea to improve availability, monitor node to check regionserver's 
 pid.if this pid not exsits,i think the rs down,i will delete the znode,and 
 force close the hlog file.
 so the period maybe 100ms.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HBASE-5673) The OOM problem of IPC client call cause all handle block

2012-09-14 Thread Anonymous (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-5673?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anonymous updated HBASE-5673:
-

 Tags: 1
Affects Version/s: (was: 0.90.6)
 Release Note: 1
 Hadoop Flags: Incompatible change  (was: Reviewed)
   Status: Patch Available  (was: Reopened)

 The OOM problem of IPC client call  cause all handle block
 --

 Key: HBASE-5673
 URL: https://issues.apache.org/jira/browse/HBASE-5673
 Project: HBase
  Issue Type: Bug
 Environment: 0.90.6
Reporter: xufeng
Assignee: xufeng
  Labels: 
 Fix For: 0.92.3

 Attachments: HBASE-5673-90.patch, HBASE-5673-90-V2.patch


 if HBaseClient meet unable to create new native thread exception, the call 
 will never complete because it be lost in calls queue.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira