[jira] [Updated] (HBASE-26420) Unexpected crash of meta RegionServer causes the cluster out of service
[ https://issues.apache.org/jira/browse/HBASE-26420?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] anonymous updated HBASE-26420: -- Attachment: hbase-root-master-C3HM1.log > Unexpected crash of meta RegionServer causes the cluster out of service > --- > > Key: HBASE-26420 > URL: https://issues.apache.org/jira/browse/HBASE-26420 > Project: HBase > Issue Type: Bug >Affects Versions: 1.7.1 >Reporter: anonymous >Priority: Major > Attachments: hbase-root-master-C3HM1.log > > > We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, > C3RS1, C3RS2, C3RS3. > We use an external ZooKeeper cluster which is a pseudo-distributed cluster: > {code:java} > > hbase.zookeeper.quorum > C3hb-zk > > > hbase.zookeeper.property.clientPort > 11181 > > {code} > For other HBase options, we use the default settings. The buggy scenario is > as follows: > 1. Start the cluster, C3HM1 becomes the active master; > 2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" > on ZooKeeper; > {code:java} > [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.createNonSequential(RecoverableZooKeeper.java:665), > > org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.create(RecoverableZooKeeper.java:644), > org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndWatch(ZKUtil.java:1182), > org.apache.hadoop.hbase.zookeeper.MetaTableLocator.setMetaLocation(MetaTableLocator.java:464), > > org.apache.hadoop.hbase.regionserver.HRegionServer.postOpenDeployTasks(HRegionServer.java:2182), > > org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler$PostOpenDeployTasksThread.run(OpenRegionHandler.java:329)] > {code} > 3. The meta server is still not online after 10 minutes. The data of znode > "/hbase/master" is C3HM1. > While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, > everything works fine. And the bug does not appear on HBase-2.4.5. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (HBASE-26420) Unexpected crash of meta RegionServer causes the cluster out of service
[ https://issues.apache.org/jira/browse/HBASE-26420?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] anonymous updated HBASE-26420: -- Description: We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, C3RS1, C3RS2, C3RS3. We use an external ZooKeeper cluster which is a pseudo-distributed cluster: {code:java} hbase.zookeeper.quorum C3hb-zk hbase.zookeeper.property.clientPort 11181 {code} For other HBase options, we use the default settings. The buggy scenario is as follows: 1. Start the cluster, C3HM1 becomes the active master; 2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" on ZooKeeper; {code:java} [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.createNonSequential(RecoverableZooKeeper.java:665), org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.create(RecoverableZooKeeper.java:644), org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndWatch(ZKUtil.java:1182), org.apache.hadoop.hbase.zookeeper.MetaTableLocator.setMetaLocation(MetaTableLocator.java:464), org.apache.hadoop.hbase.regionserver.HRegionServer.postOpenDeployTasks(HRegionServer.java:2182), org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler$PostOpenDeployTasksThread.run(OpenRegionHandler.java:329)] {code} 3. The meta server is still not online after 10 minutes. The data of znode "/hbase/master" is C3HM1. While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, everything works fine. And the bug does not appear on HBase-2.4.5. was: We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, C3RS1, C3RS2, C3RS3. We use an external ZooKeeper cluster which is a pseudo-distributed cluster: {code:java} hbase.zookeeper.quorum C3hb-zk hbase.zookeeper.property.clientPort 11181 {code} For other HBase options, we use the default settings. The buggy scenario is as follows: 1. Start the cluster, C3HM1 becomes the active master; 2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" on ZooKeeper; {code:java} [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.createNonSequential(RecoverableZooKeeper.java:665), org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.create(RecoverableZooKeeper.java:644), org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndWatch(ZKUtil.java:1182), org.apache.hadoop.hbase.zookeeper.MetaTableLocator.setMetaLocation(MetaTableLocator.java:464), org.apache.hadoop.hbase.regionserver.HRegionServer.postOpenDeployTasks(HRegionServer.java:2182), org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler$PostOpenDeployTasksThread.run(OpenRegionHandler.java:329)] {code} 3. The meta server is still not online after 10 minutes. While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, everything works fine. And the bug does not appear on HBase-2.4.5. > Unexpected crash of meta RegionServer causes the cluster out of service > --- > > Key: HBASE-26420 > URL: https://issues.apache.org/jira/browse/HBASE-26420 > Project: HBase > Issue Type: Bug >Affects Versions: 1.7.1 >Reporter: anonymous >Priority: Major > > We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, > C3RS1, C3RS2, C3RS3. > We use an external ZooKeeper cluster which is a pseudo-distributed cluster: > {code:java} > > hbase.zookeeper.quorum > C3hb-zk > > > hbase.zookeeper.property.clientPort > 11181 > > {code} > For other HBase options, we use the default settings. The buggy scenario is > as follows: > 1. Start the cluster, C3HM1 becomes the active master; > 2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" > on ZooKeeper; > {code:java} > [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.createNonSequential(RecoverableZooKeeper.java:665), > > org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.create(RecoverableZooKeeper.java:644), > org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndWatch(ZKUtil.java:1182), > org.apache.hadoop.hbase.zookeeper.MetaTableLocator.setMetaLocation(MetaTableLocator.java:464), > > org.apache.hadoop.hbase.regionserver.HRegionServer.postOpenDeployTasks(HRegionServer.java:2182), > > org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler$PostOpenDeployTasksThread.run(OpenRegionHandler.java:329)] > {code} > 3. The meta server is still not online after 10 minutes. The data of znode > "/hbase/master" is C3HM1. > While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, > everything works fine. And the bug does not appear on HBase-2.4.5. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (HBASE-26420) Unexpected crash of meta RegionServer causes the cluster out of service
[ https://issues.apache.org/jira/browse/HBASE-26420?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] anonymous updated HBASE-26420: -- Description: We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, C3RS1, C3RS2, C3RS3. We use an external ZooKeeper cluster which is a pseudo-distributed cluster: {code:java} hbase.zookeeper.quorum C3hb-zk hbase.zookeeper.property.clientPort 11181 {code} For other HBase options, we use the default settings. The buggy scenario is as follows: 1. Start the cluster, C3HM1 becomes the active master; 2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" on ZooKeeper; {code:java} [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.createNonSequential(RecoverableZooKeeper.java:665), org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.create(RecoverableZooKeeper.java:644), org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndWatch(ZKUtil.java:1182), org.apache.hadoop.hbase.zookeeper.MetaTableLocator.setMetaLocation(MetaTableLocator.java:464), org.apache.hadoop.hbase.regionserver.HRegionServer.postOpenDeployTasks(HRegionServer.java:2182), org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler$PostOpenDeployTasksThread.run(OpenRegionHandler.java:329)] {code} 3. The meta server is still not online after 10 minutes. While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, everything works fine. And the bug does not appear on HBase-2.4.5. was: We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, C3RS1, C3RS2, C3RS3. We use an external ZooKeeper cluster which is a pseudo-distributed cluster: {code:java} hbase.zookeeper.quorum C3hb-zk hbase.zookeeper.property.clientPort 11181 {code} For other HBase options, we use the default settings. The buggy scenario is as follows: 1. Start the cluster, C3HM1 becomes the active master; 2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" on ZooKeeper; {code:java} [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.createNonSequential(RecoverableZooKeeper.java:665), org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.create(RecoverableZooKeeper.java:644), org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndWatch(ZKUtil.java:1182), org.apache.hadoop.hbase.zookeeper.MetaTableLocator.setMetaLocation(MetaTableLocator.java:464), org.apache.hadoop.hbase.regionserver.HRegionServer.postOpenDeployTasks(HRegionServer.java:2182), org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler$PostOpenDeployTasksThread.run(OpenRegionHandler.java:329)] {code} 3. The meta server is still not online after 10 minutes. While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, everything works fine. And the bug disappears on HBase-2.4.5. > Unexpected crash of meta RegionServer causes the cluster out of service > --- > > Key: HBASE-26420 > URL: https://issues.apache.org/jira/browse/HBASE-26420 > Project: HBase > Issue Type: Bug >Affects Versions: 1.7.1 >Reporter: anonymous >Priority: Major > > We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, > C3RS1, C3RS2, C3RS3. > We use an external ZooKeeper cluster which is a pseudo-distributed cluster: > {code:java} > > hbase.zookeeper.quorum > C3hb-zk > > > hbase.zookeeper.property.clientPort > 11181 > > {code} > For other HBase options, we use the default settings. The buggy scenario is > as follows: > 1. Start the cluster, C3HM1 becomes the active master; > 2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" > on ZooKeeper; > {code:java} > [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.createNonSequential(RecoverableZooKeeper.java:665), > > org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.create(RecoverableZooKeeper.java:644), > org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndWatch(ZKUtil.java:1182), > org.apache.hadoop.hbase.zookeeper.MetaTableLocator.setMetaLocation(MetaTableLocator.java:464), > > org.apache.hadoop.hbase.regionserver.HRegionServer.postOpenDeployTasks(HRegionServer.java:2182), > > org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler$PostOpenDeployTasksThread.run(OpenRegionHandler.java:329)] > {code} > 3. The meta server is still not online after 10 minutes. > While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, > everything works fine. And the bug does not appear on HBase-2.4.5. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (HBASE-26420) Unexpected crash of meta RegionServer causes the cluster out of service
[ https://issues.apache.org/jira/browse/HBASE-26420?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] anonymous updated HBASE-26420: -- Description: We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, C3RS1, C3RS2, C3RS3. We use an external ZooKeeper cluster which is a pseudo-distributed cluster: {code:java} hbase.zookeeper.quorum C3hb-zk hbase.zookeeper.property.clientPort 11181 {code} For other HBase options, we use the default settings. The buggy scenario is as follows: 1. Start the cluster, C3HM1 becomes the active master; 2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" on ZooKeeper; {code:java} [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.createNonSequential(RecoverableZooKeeper.java:665), org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.create(RecoverableZooKeeper.java:644), org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndWatch(ZKUtil.java:1182), org.apache.hadoop.hbase.zookeeper.MetaTableLocator.setMetaLocation(MetaTableLocator.java:464), org.apache.hadoop.hbase.regionserver.HRegionServer.postOpenDeployTasks(HRegionServer.java:2182), org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler$PostOpenDeployTasksThread.run(OpenRegionHandler.java:329)] {code} 3. The meta server is still not online after 10 minutes. While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, everything works fine. And the bug disappears on HBase-2.4.5. was: We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, C3RS1, C3RS2, C3RS3. We use an external ZooKeeper cluster which is a pseudo-distributed cluster: {code:java} hbase.zookeeper.quorum C3hb-zk hbase.zookeeper.property.clientPort 11181 {code} For other HBase options, we use the default settings. The buggy scenario is as follows: 1. Start the cluster, C3HM1 becomes the active master; 2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" on ZooKeeper; {code:java} [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.createNonSequential(RecoverableZooKeeper.java:665), org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.create(RecoverableZooKeeper.java:644), org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndWatch(ZKUtil.java:1182), org.apache.hadoop.hbase.zookeeper.MetaTableLocator.setMetaLocation(MetaTableLocator.java:464), org.apache.hadoop.hbase.regionserver.HRegionServer.postOpenDeployTasks(HRegionServer.java:2182), org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler$PostOpenDeployTasksThread.run(OpenRegionHandler.java:329)] {code} 3. The meta server has not been online after 10 minutes. While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, everything works fine. And the bug disappears on HBase-2.4.5. > Unexpected crash of meta RegionServer causes the cluster out of service > --- > > Key: HBASE-26420 > URL: https://issues.apache.org/jira/browse/HBASE-26420 > Project: HBase > Issue Type: Bug >Affects Versions: 1.7.1 >Reporter: anonymous >Priority: Major > > We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, > C3RS1, C3RS2, C3RS3. > We use an external ZooKeeper cluster which is a pseudo-distributed cluster: > {code:java} > > hbase.zookeeper.quorum > C3hb-zk > > > hbase.zookeeper.property.clientPort > 11181 > > {code} > For other HBase options, we use the default settings. The buggy scenario is > as follows: > 1. Start the cluster, C3HM1 becomes the active master; > 2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" > on ZooKeeper; > {code:java} > [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.createNonSequential(RecoverableZooKeeper.java:665), > > org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.create(RecoverableZooKeeper.java:644), > org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndWatch(ZKUtil.java:1182), > org.apache.hadoop.hbase.zookeeper.MetaTableLocator.setMetaLocation(MetaTableLocator.java:464), > > org.apache.hadoop.hbase.regionserver.HRegionServer.postOpenDeployTasks(HRegionServer.java:2182), > > org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler$PostOpenDeployTasksThread.run(OpenRegionHandler.java:329)] > {code} > 3. The meta server is still not online after 10 minutes. > While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, > everything works fine. And the bug disappears on HBase-2.4.5. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (HBASE-26370) Unexpected crash fails the truncate operation and makes a disabled table enabled
[ https://issues.apache.org/jira/browse/HBASE-26370?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] anonymous updated HBASE-26370: -- Description: There is a table {{mytable}} in a HBase cluster with two HMasters and two RegionServers. 1. Current active HMaster is HMaster1; 2. Client disabled the table {{mytable}}; 3. {{mytable}} is in disabled state; 4. Client request to truncate {{mytable}}; 5. HMaster1 tries to truncate {{mytable}}. For the TruncateTableState {{TRUNCATE_TABLE_ASSIGN_REGIONS}}, HMaster1 assigns the regions of the table and mark the table as Enabling; 6. HMaster1 crashes; 7. HMaster2 becomes active; 8. The truncate operation fails due to {{TableNotDisabledException}}, and {{mytable}} was left as enabled. In HMaster2, we got following info: {code:java} INFO [ProcedureExecutor-12] procedure2.ProcedureExecutor: Rolledback procedure TruncateTableProcedure (table=mytable preserveSplits=true) id=9 owner=myhbase state=ROLLEDBACK exec-time=4.6030 sec exception=org.apache.hadoop.hbase.TableNotDisabledException: mytable {code} was: There is a table {{mytable}} in a HBase cluster with two HMasters and two RegionServers. 1. Current active HMaster is HMaster1; 2. Client disabled the table {{mytable}}; 3. {{mytable}} is in disabled state; 4. Client request to truncate {{mytable}}; 5. HMaster1 tries to truncate {{mytable}}. For the TruncateTableState {{TRUNCATE_TABLE_ASSIGN_REGIONS}}, HMaster1 assigns regions of the table and mark the table as Enabling; 6. HMaster1 crashes; 7. HMaster2 becomes active; 8. The truncate operation fails due to {{TableNotDisabledException}}, and {{mytable}} was left as enabled. In HMaster2, we got following info: {code:java} INFO [ProcedureExecutor-12] procedure2.ProcedureExecutor: Rolledback procedure TruncateTableProcedure (table=mytable preserveSplits=true) id=9 owner=myhbase state=ROLLEDBACK exec-time=4.6030 sec exception=org.apache.hadoop.hbase.TableNotDisabledException: mytable {code} > Unexpected crash fails the truncate operation and makes a disabled table > enabled > > > Key: HBASE-26370 > URL: https://issues.apache.org/jira/browse/HBASE-26370 > Project: HBase > Issue Type: Bug >Affects Versions: 1.7.1 >Reporter: anonymous >Priority: Major > > There is a table {{mytable}} in a HBase cluster with two HMasters and two > RegionServers. > 1. Current active HMaster is HMaster1; > 2. Client disabled the table {{mytable}}; > 3. {{mytable}} is in disabled state; > 4. Client request to truncate {{mytable}}; > 5. HMaster1 tries to truncate {{mytable}}. For the TruncateTableState > {{TRUNCATE_TABLE_ASSIGN_REGIONS}}, HMaster1 assigns the regions of the table > and mark the table as Enabling; > 6. HMaster1 crashes; > 7. HMaster2 becomes active; > 8. The truncate operation fails due to {{TableNotDisabledException}}, and > {{mytable}} was left as enabled. > In HMaster2, we got following info: > {code:java} > INFO [ProcedureExecutor-12] procedure2.ProcedureExecutor: Rolledback > procedure TruncateTableProcedure (table=mytable preserveSplits=true) id=9 > owner=myhbase state=ROLLEDBACK exec-time=4.6030 sec > exception=org.apache.hadoop.hbase.TableNotDisabledException: mytable > {code} -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (HBASE-26370) Unexpected crash fails the truncate operation and makes a disabled table enabled
[ https://issues.apache.org/jira/browse/HBASE-26370?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] anonymous updated HBASE-26370: -- Description: There is a table {{mytable}} in a HBase cluster with two HMasters and two RegionServers. 1. Current active HMaster is HMaster1; 2. Client disabled the table {{mytable}}; 3. {{mytable}} is in disabled state; 4. Client request to truncate {{mytable}}; 5. HMaster1 tries to truncate {{mytable}}. For the TruncateTableState {{TRUNCATE_TABLE_ASSIGN_REGIONS}}, HMaster1 assigns regions of the table and mark the table as Enabling; 6. HMaster1 crashes; 7. HMaster2 becomes active; 8. The truncate operation fails due to {{TableNotDisabledException}}, and {{mytable}} was left as enabled. In HMaster2, we got following info: {code:java} INFO [ProcedureExecutor-12] procedure2.ProcedureExecutor: Rolledback procedure TruncateTableProcedure (table=mytable preserveSplits=true) id=9 owner=myhbase state=ROLLEDBACK exec-time=4.6030 sec exception=org.apache.hadoop.hbase.TableNotDisabledException: mytable {code} was: There is a table {{mytable}} in a HBase cluster with two HMasters and two RegionServers. 1. Current active HMaster is HMaster1; 2. Client disabled the table {{mytable}}; 3. {{mytable}} is in disabled state; 4. Client request to truncate {{mytable}}; 5. HMaster1 tries to truncate {{mytable}}. For the TruncateTableState {{TRUNCATE_TABLE_ASSIGN_REGIONS}}, HMaster1 tries to assign regions of the table and mark the table as Enabling; 6. HMaster1 crashes; 7. HMaster2 becomes active; 8. The truncate operation fails due to {{TableNotDisabledException}}, and {{mytable}} was left as enabled. In HMaster2, we got following info: {code:java} INFO [ProcedureExecutor-12] procedure2.ProcedureExecutor: Rolledback procedure TruncateTableProcedure (table=mytable preserveSplits=true) id=9 owner=myhbase state=ROLLEDBACK exec-time=4.6030 sec exception=org.apache.hadoop.hbase.TableNotDisabledException: mytable {code} > Unexpected crash fails the truncate operation and makes a disabled table > enabled > > > Key: HBASE-26370 > URL: https://issues.apache.org/jira/browse/HBASE-26370 > Project: HBase > Issue Type: Bug >Affects Versions: 1.7.1 >Reporter: anonymous >Priority: Major > > There is a table {{mytable}} in a HBase cluster with two HMasters and two > RegionServers. > 1. Current active HMaster is HMaster1; > 2. Client disabled the table {{mytable}}; > 3. {{mytable}} is in disabled state; > 4. Client request to truncate {{mytable}}; > 5. HMaster1 tries to truncate {{mytable}}. For the TruncateTableState > {{TRUNCATE_TABLE_ASSIGN_REGIONS}}, HMaster1 assigns regions of the table and > mark the table as Enabling; > 6. HMaster1 crashes; > 7. HMaster2 becomes active; > 8. The truncate operation fails due to {{TableNotDisabledException}}, and > {{mytable}} was left as enabled. > In HMaster2, we got following info: > {code:java} > INFO [ProcedureExecutor-12] procedure2.ProcedureExecutor: Rolledback > procedure TruncateTableProcedure (table=mytable preserveSplits=true) id=9 > owner=myhbase state=ROLLEDBACK exec-time=4.6030 sec > exception=org.apache.hadoop.hbase.TableNotDisabledException: mytable > {code} -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (HBASE-26391) Unexpected crash in the process of deleting a table causes stale data in ZooKeeper
[ https://issues.apache.org/jira/browse/HBASE-26391?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] anonymous updated HBASE-26391: -- Description: There is a table {{mytable}} in a HBase cluster with two HMasters and two RegionServers. 1. Current active HMaster is HMaster1; 2. Client request to delete the table {{mytable}}; 3. HMaster1 deletes znode in ZooKeeper: /hbase/table/mytable {code:java} [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.delete(RecoverableZooKeeper.java:238), org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1367), org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1359), org.apache.hadoop.hbase.zookeeper.ZKTableStateManager.setDeletedTable(ZKTableStateManager.java:244), org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.deleteAssignmentState(DeleteTableProcedure.java:383), org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:136), org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:57), org.apache.hadoop.hbase.procedure2.StateMachineProcedure.execute(StateMachineProcedure.java:139), org.apache.hadoop.hbase.procedure2.Procedure.doExecute(Procedure.java:499), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execProcedure(ProcedureExecutor.java:1166), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:954), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:907), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.access$400(ProcedureExecutor.java:76), org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java:481), org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java)] {code} 4. HMaster1 crashes before deleting the znode: /hbase/table-lock/mytable {code:java} [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.delete(RecoverableZooKeeper.java:238), org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNode(ZKUtil.java:1341), org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNode(ZKUtil.java:1330), org.apache.hadoop.hbase.master.TableLockManager$ZKTableLockManager.tableDeleted(TableLockManager.java:442), org.apache.hadoop.hbase.master.procedure.MasterProcedureScheduler.markTableAsDeleted(MasterProcedureScheduler.java:895), org.apache.hadoop.hbase.master.procedure.MasterProcedureScheduler.completionCleanup(MasterProcedureScheduler.java:261), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.procedureFinished(ProcedureExecutor.java:1415), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:974), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:907), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.access$400(ProcedureExecutor.java:76), org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java:481), org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java)] {code} 5. HMaster2 becomes active 6. The table {{mytablle}} was deleted, but the znode "/hbase/table-lock/mytable" was left on ZooKeeper was: There is a table {{mytable}} in a HBase cluster with two HMasters and two RegionServers. 1. Current active HMaster is HMaster1; 2. Client request to delete the table {{mytable}}; 3. HMaster1 tries to delete znode in ZooKeeper: /hbase/table/mytable {code:java} [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.delete(RecoverableZooKeeper.java:238), org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1367), org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1359), org.apache.hadoop.hbase.zookeeper.ZKTableStateManager.setDeletedTable(ZKTableStateManager.java:244), org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.deleteAssignmentState(DeleteTableProcedure.java:383), org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:136), org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:57), org.apache.hadoop.hbase.procedure2.StateMachineProcedure.execute(StateMachineProcedure.java:139), org.apache.hadoop.hbase.procedure2.Procedure.doExecute(Procedure.java:499), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execProcedure(ProcedureExecutor.java:1166), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:954), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:907), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.access$400(ProcedureExecutor.java:76), org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java:481),
[jira] [Updated] (HBASE-26391) Unexpected crash in the process of deleting a table causes stale data in ZooKeeper
[ https://issues.apache.org/jira/browse/HBASE-26391?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] anonymous updated HBASE-26391: -- Description: There is a table {{mytable}} in a HBase cluster with two HMasters and two RegionServers. 1. Current active HMaster is HMaster1; 2. Client request to delete the table {{mytable}}; 3. HMaster1 deletes znode in ZooKeeper: /hbase/table/mytable {code:java} [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.delete(RecoverableZooKeeper.java:238), org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1367), org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1359), org.apache.hadoop.hbase.zookeeper.ZKTableStateManager.setDeletedTable(ZKTableStateManager.java:244), org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.deleteAssignmentState(DeleteTableProcedure.java:383), org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:136), org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:57), org.apache.hadoop.hbase.procedure2.StateMachineProcedure.execute(StateMachineProcedure.java:139), org.apache.hadoop.hbase.procedure2.Procedure.doExecute(Procedure.java:499), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execProcedure(ProcedureExecutor.java:1166), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:954), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:907), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.access$400(ProcedureExecutor.java:76), org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java:481), org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java)] {code} 4. HMaster1 crashes before deleting the znode: /hbase/table-lock/mytable {code:java} [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.delete(RecoverableZooKeeper.java:238), org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNode(ZKUtil.java:1341), org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNode(ZKUtil.java:1330), org.apache.hadoop.hbase.master.TableLockManager$ZKTableLockManager.tableDeleted(TableLockManager.java:442), org.apache.hadoop.hbase.master.procedure.MasterProcedureScheduler.markTableAsDeleted(MasterProcedureScheduler.java:895), org.apache.hadoop.hbase.master.procedure.MasterProcedureScheduler.completionCleanup(MasterProcedureScheduler.java:261), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.procedureFinished(ProcedureExecutor.java:1415), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:974), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:907), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.access$400(ProcedureExecutor.java:76), org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java:481), org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java)] {code} 5. HMaster2 becomes active 6. The table {{mytable}} was deleted, but the znode "/hbase/table-lock/mytable" was left on ZooKeeper was: There is a table {{mytable}} in a HBase cluster with two HMasters and two RegionServers. 1. Current active HMaster is HMaster1; 2. Client request to delete the table {{mytable}}; 3. HMaster1 deletes znode in ZooKeeper: /hbase/table/mytable {code:java} [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.delete(RecoverableZooKeeper.java:238), org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1367), org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1359), org.apache.hadoop.hbase.zookeeper.ZKTableStateManager.setDeletedTable(ZKTableStateManager.java:244), org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.deleteAssignmentState(DeleteTableProcedure.java:383), org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:136), org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:57), org.apache.hadoop.hbase.procedure2.StateMachineProcedure.execute(StateMachineProcedure.java:139), org.apache.hadoop.hbase.procedure2.Procedure.doExecute(Procedure.java:499), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execProcedure(ProcedureExecutor.java:1166), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:954), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:907), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.access$400(ProcedureExecutor.java:76), org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java:481), org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java)] {code}
[jira] [Updated] (HBASE-26420) Unexpected crash of meta RegionServer causes the cluster out of service
[ https://issues.apache.org/jira/browse/HBASE-26420?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] anonymous updated HBASE-26420: -- Description: We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, C3RS1, C3RS2, C3RS3. We use an external ZooKeeper cluster which is a pseudo-distributed cluster: {code:java} hbase.zookeeper.quorum C3hb-zk hbase.zookeeper.property.clientPort 11181 {code} For other HBase options, we use the default settings. The buggy scenario is as follows: 1. Start the cluster, C3HM1 becomes the active master; 2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" on ZooKeeper; {code:java} [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.createNonSequential(RecoverableZooKeeper.java:665), org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.create(RecoverableZooKeeper.java:644), org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndWatch(ZKUtil.java:1182), org.apache.hadoop.hbase.zookeeper.MetaTableLocator.setMetaLocation(MetaTableLocator.java:464), org.apache.hadoop.hbase.regionserver.HRegionServer.postOpenDeployTasks(HRegionServer.java:2182), org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler$PostOpenDeployTasksThread.run(OpenRegionHandler.java:329)] {code} 3. The meta server has not been online after 10 minutes. While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, everything works fine. And the bug disappears on HBase-2.4.5. was: We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, C3RS1, C3RS2, C3RS3. We use an external ZooKeeper cluster which is a pseudo-distributed cluster: {code:java} hbase.zookeeper.quorum C3hb-zk hbase.zookeeper.property.clientPort 11181 {code} For other HBase options, we use the default settings. The buggy scenario is as follows: 1. Start the cluster; 2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" on ZooKeeper; {code:java} [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.createNonSequential(RecoverableZooKeeper.java:665), org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.create(RecoverableZooKeeper.java:644), org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndWatch(ZKUtil.java:1182), org.apache.hadoop.hbase.zookeeper.MetaTableLocator.setMetaLocation(MetaTableLocator.java:464), org.apache.hadoop.hbase.regionserver.HRegionServer.postOpenDeployTasks(HRegionServer.java:2182), org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler$PostOpenDeployTasksThread.run(OpenRegionHandler.java:329)] {code} 3. The meta server has not been online after 10 minutes. While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, everything works fine. And the bug disappears on HBase-2.4.5. > Unexpected crash of meta RegionServer causes the cluster out of service > --- > > Key: HBASE-26420 > URL: https://issues.apache.org/jira/browse/HBASE-26420 > Project: HBase > Issue Type: Bug >Affects Versions: 1.7.1 >Reporter: anonymous >Priority: Major > > We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, > C3RS1, C3RS2, C3RS3. > We use an external ZooKeeper cluster which is a pseudo-distributed cluster: > {code:java} > > hbase.zookeeper.quorum > C3hb-zk > > > hbase.zookeeper.property.clientPort > 11181 > > {code} > For other HBase options, we use the default settings. The buggy scenario is > as follows: > 1. Start the cluster, C3HM1 becomes the active master; > 2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" > on ZooKeeper; > {code:java} > [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.createNonSequential(RecoverableZooKeeper.java:665), > > org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.create(RecoverableZooKeeper.java:644), > org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndWatch(ZKUtil.java:1182), > org.apache.hadoop.hbase.zookeeper.MetaTableLocator.setMetaLocation(MetaTableLocator.java:464), > > org.apache.hadoop.hbase.regionserver.HRegionServer.postOpenDeployTasks(HRegionServer.java:2182), > > org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler$PostOpenDeployTasksThread.run(OpenRegionHandler.java:329)] > {code} > 3. The meta server has not been online after 10 minutes. > While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, > everything works fine. And the bug disappears on HBase-2.4.5. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (HBASE-26420) Unexpected crash of meta RegionServer causes the cluster out of service
[ https://issues.apache.org/jira/browse/HBASE-26420?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] anonymous updated HBASE-26420: -- Description: We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, C3RS1, C3RS2, C3RS3. We use an external ZooKeeper cluster which is a pseudo-distributed cluster: {code:java} hbase.zookeeper.quorum C3hb-zk hbase.zookeeper.property.clientPort 11181 {code} For other HBase options, we use the default settings. The buggy scenario is as follows: 1. Start the cluster; 2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" on ZooKeeper; {code:java} [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.createNonSequential(RecoverableZooKeeper.java:665), org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.create(RecoverableZooKeeper.java:644), org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndWatch(ZKUtil.java:1182), org.apache.hadoop.hbase.zookeeper.MetaTableLocator.setMetaLocation(MetaTableLocator.java:464), org.apache.hadoop.hbase.regionserver.HRegionServer.postOpenDeployTasks(HRegionServer.java:2182), org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler$PostOpenDeployTasksThread.run(OpenRegionHandler.java:329)] {code} 3. The meta server has not been online after 10 minutes. While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, everything works fine. And the bug disappears on HBase-2.4.5. was: We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, C3RS1, C3RS2, C3RS3. We use an external ZooKeeper cluster which is a pseudo-distributed cluster: {code:java} hbase.zookeeper.quorum C3hb-zk hbase.zookeeper.property.clientPort 11181 {code} For other HBase options, we use the default settings. The buggy scenario is as follows: 1. Start the cluster; 2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" on ZooKeeper; {code:java} [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.createNonSequential(RecoverableZooKeeper.java:665), org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.create(RecoverableZooKeeper.java:644), org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndWatch(ZKUtil.java:1182), org.apache.hadoop.hbase.zookeeper.MetaTableLocator.setMetaLocation(MetaTableLocator.java:464), org.apache.hadoop.hbase.regionserver.HRegionServer.postOpenDeployTasks(HRegionServer.java:2182), org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler$PostOpenDeployTasksThread.run(OpenRegionHandler.java:329)] {code} 3. The meta server has not been online after 10 minutes. While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, everything works fine. > Unexpected crash of meta RegionServer causes the cluster out of service > --- > > Key: HBASE-26420 > URL: https://issues.apache.org/jira/browse/HBASE-26420 > Project: HBase > Issue Type: Bug >Affects Versions: 1.7.1 >Reporter: anonymous >Priority: Major > > We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, > C3RS1, C3RS2, C3RS3. > We use an external ZooKeeper cluster which is a pseudo-distributed cluster: > {code:java} > > hbase.zookeeper.quorum > C3hb-zk > > > hbase.zookeeper.property.clientPort > 11181 > > {code} > For other HBase options, we use the default settings. The buggy scenario is > as follows: > 1. Start the cluster; > 2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" > on ZooKeeper; > {code:java} > [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.createNonSequential(RecoverableZooKeeper.java:665), > > org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.create(RecoverableZooKeeper.java:644), > org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndWatch(ZKUtil.java:1182), > org.apache.hadoop.hbase.zookeeper.MetaTableLocator.setMetaLocation(MetaTableLocator.java:464), > > org.apache.hadoop.hbase.regionserver.HRegionServer.postOpenDeployTasks(HRegionServer.java:2182), > > org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler$PostOpenDeployTasksThread.run(OpenRegionHandler.java:329)] > {code} > 3. The meta server has not been online after 10 minutes. > While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, > everything works fine. And the bug disappears on HBase-2.4.5. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (HBASE-26420) Unexpected crash of meta RegionServer causes the cluster out of service
[ https://issues.apache.org/jira/browse/HBASE-26420?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] anonymous updated HBASE-26420: -- Description: We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, C3RS1, C3RS2, C3RS3. We use an external ZooKeeper cluster which is a pseudo-distributed cluster: {code:java} hbase.zookeeper.quorum C3hb-zk hbase.zookeeper.property.clientPort 11181 {code} For other HBase options, we use the default settings. The buggy scenario is as follows: 1. Start the cluster; 2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" on ZooKeeper; {code:java} [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.createNonSequential(RecoverableZooKeeper.java:665), org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.create(RecoverableZooKeeper.java:644), org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndWatch(ZKUtil.java:1182), org.apache.hadoop.hbase.zookeeper.MetaTableLocator.setMetaLocation(MetaTableLocator.java:464), org.apache.hadoop.hbase.regionserver.HRegionServer.postOpenDeployTasks(HRegionServer.java:2182), org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler$PostOpenDeployTasksThread.run(OpenRegionHandler.java:329)] {code} 3. The meta server has not been online after 10 minutes. While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, everything works fine. was: We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, C3RS1, C3RS2, C3RS3. We use an external ZooKeeper cluster which is a pseudo-distributed cluster: {code:java} hbase.zookeeper.quorum C3hb-zk hbase.zookeeper.property.clientPort 11181 {code} For other HBase options, we use the default settings. The buggy scenario is as follows: 1. Start the cluster; 2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" on ZooKeeper; {code:java} [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.createNonSequential(RecoverableZooKeeper.java:665), org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.create(RecoverableZooKeeper.java:644), org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndWatch(ZKUtil.java:1182), org.apache.hadoop.hbase.zookeeper.MetaTableLocator.setMetaLocation(MetaTableLocator.java:464), org.apache.hadoop.hbase.regionserver.HRegionServer.postOpenDeployTasks(HRegionServer.java:2182), org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler$PostOpenDeployTasksThread.run(OpenRegionHandler.java:329)] {code} 3. The meta server has not been online after 10 minutes. While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, everything works fine. > Unexpected crash of meta RegionServer causes the cluster out of service > --- > > Key: HBASE-26420 > URL: https://issues.apache.org/jira/browse/HBASE-26420 > Project: HBase > Issue Type: Bug >Affects Versions: 1.7.1 >Reporter: anonymous >Priority: Major > > We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, > C3RS1, C3RS2, C3RS3. > We use an external ZooKeeper cluster which is a pseudo-distributed cluster: > {code:java} > > hbase.zookeeper.quorum > C3hb-zk > > > hbase.zookeeper.property.clientPort > 11181 > > {code} > For other HBase options, we use the default settings. The buggy scenario is > as follows: > 1. Start the cluster; > 2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" > on ZooKeeper; > {code:java} > [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.createNonSequential(RecoverableZooKeeper.java:665), > > org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.create(RecoverableZooKeeper.java:644), > org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndWatch(ZKUtil.java:1182), > org.apache.hadoop.hbase.zookeeper.MetaTableLocator.setMetaLocation(MetaTableLocator.java:464), > > org.apache.hadoop.hbase.regionserver.HRegionServer.postOpenDeployTasks(HRegionServer.java:2182), > > org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler$PostOpenDeployTasksThread.run(OpenRegionHandler.java:329)] > {code} > 3. The meta server has not been online after 10 minutes. > While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, > everything works fine. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (HBASE-26420) Unexpected crash of meta RegionServer causes the cluster out of service
[ https://issues.apache.org/jira/browse/HBASE-26420?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] anonymous updated HBASE-26420: -- Description: We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, C3RS1, C3RS2, C3RS3. We use an external ZooKeeper cluster which is a pseudo-distributed cluster: {code:java} hbase.zookeeper.quorum C3hb-zk hbase.zookeeper.property.clientPort 11181 {code} For other HBase options, we use the default settings. The buggy scenario is as follows: 1. Start the cluster; 2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" on ZooKeeper; {code:java} [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.createNonSequential(RecoverableZooKeeper.java:665), org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.create(RecoverableZooKeeper.java:644), org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndWatch(ZKUtil.java:1182), org.apache.hadoop.hbase.zookeeper.MetaTableLocator.setMetaLocation(MetaTableLocator.java:464), org.apache.hadoop.hbase.regionserver.HRegionServer.postOpenDeployTasks(HRegionServer.java:2182), org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler$PostOpenDeployTasksThread.run(OpenRegionHandler.java:329)] {code} 3. The meta server has not been online after 10 minutes. While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, everything works fine. was: We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, C3RS1, C3RS2, C3RS3. We use an external ZooKeeper cluster which is a pseudo-distributed cluster: {code:java} hbase.zookeeper.quorum C3hb-zk hbase.zookeeper.property.clientPort 11181 {code} For other HBase options, we use the default settings. The buggy scenario is as follows: 1. Start the cluster; 2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" on ZooKeeper; {code:java} // Some comments here {code} 3. The meta server has not been online after 10 minutes. While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, everything works fine. > Unexpected crash of meta RegionServer causes the cluster out of service > --- > > Key: HBASE-26420 > URL: https://issues.apache.org/jira/browse/HBASE-26420 > Project: HBase > Issue Type: Bug >Affects Versions: 1.7.1 >Reporter: anonymous >Priority: Major > > We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, > C3RS1, C3RS2, C3RS3. > We use an external ZooKeeper cluster which is a pseudo-distributed cluster: > {code:java} > > hbase.zookeeper.quorum > C3hb-zk > > > hbase.zookeeper.property.clientPort > 11181 > > {code} > For other HBase options, we use the default settings. The buggy scenario is > as follows: > 1. Start the cluster; > 2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" > on ZooKeeper; > {code:java} > [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.createNonSequential(RecoverableZooKeeper.java:665), > > org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.create(RecoverableZooKeeper.java:644), > org.apache.hadoop.hbase.zookeeper.ZKUtil.createAndWatch(ZKUtil.java:1182), > org.apache.hadoop.hbase.zookeeper.MetaTableLocator.setMetaLocation(MetaTableLocator.java:464), > > org.apache.hadoop.hbase.regionserver.HRegionServer.postOpenDeployTasks(HRegionServer.java:2182), > > org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler$PostOpenDeployTasksThread.run(OpenRegionHandler.java:329)] > {code} > 3. The meta server has not been online after 10 minutes. > While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, > everything works fine. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (HBASE-26420) Unexpected crash of meta RegionServer causes the cluster out of service
[ https://issues.apache.org/jira/browse/HBASE-26420?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] anonymous updated HBASE-26420: -- Description: We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, C3RS1, C3RS2, C3RS3. We use an external ZooKeeper cluster which is a pseudo-distributed cluster: {code:java} hbase.zookeeper.quorum C3hb-zk hbase.zookeeper.property.clientPort 11181 {code} For other HBase options, we use the default settings. The buggy scenario is as follows: 1. Start the cluster; 2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" on ZooKeeper; {code:java} // Some comments here public String getFoo() { return foo; } {code} 3. The meta server has not been online after 10 minutes. While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, everything works fine. > Unexpected crash of meta RegionServer causes the cluster out of service > --- > > Key: HBASE-26420 > URL: https://issues.apache.org/jira/browse/HBASE-26420 > Project: HBase > Issue Type: Bug >Affects Versions: 1.7.1 >Reporter: anonymous >Priority: Major > > We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, > C3RS1, C3RS2, C3RS3. > We use an external ZooKeeper cluster which is a pseudo-distributed cluster: > {code:java} > > hbase.zookeeper.quorum > C3hb-zk > > > hbase.zookeeper.property.clientPort > 11181 > > {code} > For other HBase options, we use the default settings. The buggy scenario is > as follows: > 1. Start the cluster; > 2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" > on ZooKeeper; > {code:java} > // Some comments here > public String getFoo() > { > return foo; > } > {code} > 3. The meta server has not been online after 10 minutes. > While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, > everything works fine. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (HBASE-26420) Unexpected crash of meta RegionServer causes the cluster out of service
[ https://issues.apache.org/jira/browse/HBASE-26420?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] anonymous updated HBASE-26420: -- Description: We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, C3RS1, C3RS2, C3RS3. We use an external ZooKeeper cluster which is a pseudo-distributed cluster: {code:java} hbase.zookeeper.quorum C3hb-zk hbase.zookeeper.property.clientPort 11181 {code} For other HBase options, we use the default settings. The buggy scenario is as follows: 1. Start the cluster; 2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" on ZooKeeper; {code:java} // Some comments here {code} 3. The meta server has not been online after 10 minutes. While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, everything works fine. was: We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, C3RS1, C3RS2, C3RS3. We use an external ZooKeeper cluster which is a pseudo-distributed cluster: {code:java} hbase.zookeeper.quorum C3hb-zk hbase.zookeeper.property.clientPort 11181 {code} For other HBase options, we use the default settings. The buggy scenario is as follows: 1. Start the cluster; 2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" on ZooKeeper; {code:java} // Some comments here public String getFoo() { return foo; } {code} 3. The meta server has not been online after 10 minutes. While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, everything works fine. > Unexpected crash of meta RegionServer causes the cluster out of service > --- > > Key: HBASE-26420 > URL: https://issues.apache.org/jira/browse/HBASE-26420 > Project: HBase > Issue Type: Bug >Affects Versions: 1.7.1 >Reporter: anonymous >Priority: Major > > We have a cluster of two HMasters, C3HM1 and C3HM2, and three RegionServers, > C3RS1, C3RS2, C3RS3. > We use an external ZooKeeper cluster which is a pseudo-distributed cluster: > {code:java} > > hbase.zookeeper.quorum > C3hb-zk > > > hbase.zookeeper.property.clientPort > 11181 > > {code} > For other HBase options, we use the default settings. The buggy scenario is > as follows: > 1. Start the cluster; > 2. C3RS2 crashes right before creating the znode "/hbase/meta-region-server" > on ZooKeeper; > {code:java} > // Some comments here > {code} > 3. The meta server has not been online after 10 minutes. > While the C3RS2 crashes after creating the "/hbase/meta-region-server" znode, > everything works fine. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (HBASE-26420) Unexpected crash of meta RegionServer causes the cluster out of service
[ https://issues.apache.org/jira/browse/HBASE-26420?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] anonymous updated HBASE-26420: -- Summary: Unexpected crash of meta RegionServer causes the cluster out of service (was: Unexpected crash of HMaster causes the cluster out of service) > Unexpected crash of meta RegionServer causes the cluster out of service > --- > > Key: HBASE-26420 > URL: https://issues.apache.org/jira/browse/HBASE-26420 > Project: HBase > Issue Type: Bug >Affects Versions: 1.7.1 >Reporter: anonymous >Priority: Major > -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Created] (HBASE-26420) Unexpected crash of HMaster causes the cluster out of service
anonymous created HBASE-26420: - Summary: Unexpected crash of HMaster causes the cluster out of service Key: HBASE-26420 URL: https://issues.apache.org/jira/browse/HBASE-26420 Project: HBase Issue Type: Bug Affects Versions: 1.7.1 Reporter: anonymous -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (HBASE-26391) Unexpected crash in the process of deleting a table causes stale data
[ https://issues.apache.org/jira/browse/HBASE-26391?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] anonymous updated HBASE-26391: -- Summary: Unexpected crash in the process of deleting a table causes stale data (was: Unexpected crash in the process of deleting table causes stale data) > Unexpected crash in the process of deleting a table causes stale data > - > > Key: HBASE-26391 > URL: https://issues.apache.org/jira/browse/HBASE-26391 > Project: HBase > Issue Type: Bug >Affects Versions: 1.7.1 >Reporter: anonymous >Priority: Major > > There is a table {{mytable}} in a HBase cluster with two HMasters and two > RegionServers. > 1. Current active HMaster is HMaster1; > 2. Client request to delete the table {{mytable}}; > 3. HMaster1 tries to delete znode in ZooKeeper: /hbase/table/mytable > {code:java} > [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.delete(RecoverableZooKeeper.java:238), > > org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1367), > > org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1359), > > org.apache.hadoop.hbase.zookeeper.ZKTableStateManager.setDeletedTable(ZKTableStateManager.java:244), > > org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.deleteAssignmentState(DeleteTableProcedure.java:383), > > org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:136), > > org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:57), > > org.apache.hadoop.hbase.procedure2.StateMachineProcedure.execute(StateMachineProcedure.java:139), > org.apache.hadoop.hbase.procedure2.Procedure.doExecute(Procedure.java:499), > org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execProcedure(ProcedureExecutor.java:1166), > > org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:954), > > org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:907), > > org.apache.hadoop.hbase.procedure2.ProcedureExecutor.access$400(ProcedureExecutor.java:76), > > org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java:481), > > org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java)] > {code} > 4. HMaster1 crashes before deleting the znode: /hbase/table-lock/mytable > {code:java} > [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.delete(RecoverableZooKeeper.java:238), > org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNode(ZKUtil.java:1341), > org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNode(ZKUtil.java:1330), > org.apache.hadoop.hbase.master.TableLockManager$ZKTableLockManager.tableDeleted(TableLockManager.java:442), > > org.apache.hadoop.hbase.master.procedure.MasterProcedureScheduler.markTableAsDeleted(MasterProcedureScheduler.java:895), > > org.apache.hadoop.hbase.master.procedure.MasterProcedureScheduler.completionCleanup(MasterProcedureScheduler.java:261), > > org.apache.hadoop.hbase.procedure2.ProcedureExecutor.procedureFinished(ProcedureExecutor.java:1415), > > org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:974), > > org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:907), > > org.apache.hadoop.hbase.procedure2.ProcedureExecutor.access$400(ProcedureExecutor.java:76), > > org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java:481), > > org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java)] > {code} > 5. HMaster2 becomes active > 6. The table {{mytablle}} was deleted, but the znode > "/hbase/table-lock/mytable" was left on ZooKeeper -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (HBASE-26391) Unexpected crash in the process of deleting a table causes stale data in ZooKeeper
[ https://issues.apache.org/jira/browse/HBASE-26391?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] anonymous updated HBASE-26391: -- Summary: Unexpected crash in the process of deleting a table causes stale data in ZooKeeper (was: Unexpected crash in the process of deleting a table causes stale data) > Unexpected crash in the process of deleting a table causes stale data in > ZooKeeper > -- > > Key: HBASE-26391 > URL: https://issues.apache.org/jira/browse/HBASE-26391 > Project: HBase > Issue Type: Bug >Affects Versions: 1.7.1 >Reporter: anonymous >Priority: Major > > There is a table {{mytable}} in a HBase cluster with two HMasters and two > RegionServers. > 1. Current active HMaster is HMaster1; > 2. Client request to delete the table {{mytable}}; > 3. HMaster1 tries to delete znode in ZooKeeper: /hbase/table/mytable > {code:java} > [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.delete(RecoverableZooKeeper.java:238), > > org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1367), > > org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1359), > > org.apache.hadoop.hbase.zookeeper.ZKTableStateManager.setDeletedTable(ZKTableStateManager.java:244), > > org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.deleteAssignmentState(DeleteTableProcedure.java:383), > > org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:136), > > org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:57), > > org.apache.hadoop.hbase.procedure2.StateMachineProcedure.execute(StateMachineProcedure.java:139), > org.apache.hadoop.hbase.procedure2.Procedure.doExecute(Procedure.java:499), > org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execProcedure(ProcedureExecutor.java:1166), > > org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:954), > > org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:907), > > org.apache.hadoop.hbase.procedure2.ProcedureExecutor.access$400(ProcedureExecutor.java:76), > > org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java:481), > > org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java)] > {code} > 4. HMaster1 crashes before deleting the znode: /hbase/table-lock/mytable > {code:java} > [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.delete(RecoverableZooKeeper.java:238), > org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNode(ZKUtil.java:1341), > org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNode(ZKUtil.java:1330), > org.apache.hadoop.hbase.master.TableLockManager$ZKTableLockManager.tableDeleted(TableLockManager.java:442), > > org.apache.hadoop.hbase.master.procedure.MasterProcedureScheduler.markTableAsDeleted(MasterProcedureScheduler.java:895), > > org.apache.hadoop.hbase.master.procedure.MasterProcedureScheduler.completionCleanup(MasterProcedureScheduler.java:261), > > org.apache.hadoop.hbase.procedure2.ProcedureExecutor.procedureFinished(ProcedureExecutor.java:1415), > > org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:974), > > org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:907), > > org.apache.hadoop.hbase.procedure2.ProcedureExecutor.access$400(ProcedureExecutor.java:76), > > org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java:481), > > org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java)] > {code} > 5. HMaster2 becomes active > 6. The table {{mytablle}} was deleted, but the znode > "/hbase/table-lock/mytable" was left on ZooKeeper -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (HBASE-26391) Unexpected crash in the process of deleting table causes stale data
[ https://issues.apache.org/jira/browse/HBASE-26391?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] anonymous updated HBASE-26391: -- Description: There is a table {{mytable}} in a HBase cluster with two HMasters and two RegionServers. 1. Current active HMaster is HMaster1; 2. Client request to delete the table {{mytable}}; 3. HMaster1 tries to delete znode in ZooKeeper: /hbase/table/mytable {code:java} [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.delete(RecoverableZooKeeper.java:238), org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1367), org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1359), org.apache.hadoop.hbase.zookeeper.ZKTableStateManager.setDeletedTable(ZKTableStateManager.java:244), org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.deleteAssignmentState(DeleteTableProcedure.java:383), org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:136), org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:57), org.apache.hadoop.hbase.procedure2.StateMachineProcedure.execute(StateMachineProcedure.java:139), org.apache.hadoop.hbase.procedure2.Procedure.doExecute(Procedure.java:499), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execProcedure(ProcedureExecutor.java:1166), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:954), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:907), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.access$400(ProcedureExecutor.java:76), org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java:481), org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java)] {code} 4. HMaster1 crashes before deleting the znode: /hbase/table-lock/mytable {code:java} [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.delete(RecoverableZooKeeper.java:238), org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNode(ZKUtil.java:1341), org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNode(ZKUtil.java:1330), org.apache.hadoop.hbase.master.TableLockManager$ZKTableLockManager.tableDeleted(TableLockManager.java:442), org.apache.hadoop.hbase.master.procedure.MasterProcedureScheduler.markTableAsDeleted(MasterProcedureScheduler.java:895), org.apache.hadoop.hbase.master.procedure.MasterProcedureScheduler.completionCleanup(MasterProcedureScheduler.java:261), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.procedureFinished(ProcedureExecutor.java:1415), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:974), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:907), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.access$400(ProcedureExecutor.java:76), org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java:481), org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java)] {code} 5. HMaster2 becomes active 6. The table {{mytablle}} was deleted, but the znode "/hbase/table-lock/mytable" was left on ZooKeeper was: There is a table {{mytable}} in a HBase cluster with two HMasters and two RegionServers. 1. Current active HMaster is HMaster1; 2. Client request to delete the table {{mytable}}; 3. HMaster1 tries to delete znode in ZooKeeper: /hbase/table/mytable {code:java} [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.delete(RecoverableZooKeeper.java:238), org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1367), org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1359), org.apache.hadoop.hbase.zookeeper.ZKTableStateManager.setDeletedTable(ZKTableStateManager.java:244), org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.deleteAssignmentState(DeleteTableProcedure.java:383), org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:136), org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:57), org.apache.hadoop.hbase.procedure2.StateMachineProcedure.execute(StateMachineProcedure.java:139), org.apache.hadoop.hbase.procedure2.Procedure.doExecute(Procedure.java:499), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execProcedure(ProcedureExecutor.java:1166), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:954), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:907), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.access$400(ProcedureExecutor.java:76), org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java:481),
[jira] [Updated] (HBASE-26391) Unexpected crash in the process of deleting table causes stale data
[ https://issues.apache.org/jira/browse/HBASE-26391?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] anonymous updated HBASE-26391: -- Description: There is a table {{mytable}} in a HBase cluster with two HMasters and two RegionServers. 1. Current active HMaster is HMaster1; 2. Client request to delete the table {{mytable}}; 3. HMaster1 tries to delete znode in ZooKeeper: /hbase/table/mytable {code:java} [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.delete(RecoverableZooKeeper.java:238), org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1367), org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1359), org.apache.hadoop.hbase.zookeeper.ZKTableStateManager.setDeletedTable(ZKTableStateManager.java:244), org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.deleteAssignmentState(DeleteTableProcedure.java:383), org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:136), org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:57), org.apache.hadoop.hbase.procedure2.StateMachineProcedure.execute(StateMachineProcedure.java:139), org.apache.hadoop.hbase.procedure2.Procedure.doExecute(Procedure.java:499), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execProcedure(ProcedureExecutor.java:1166), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:954), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:907), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.access$400(ProcedureExecutor.java:76), org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java:481), org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java)] {code} 4. HMaster1 crashes before deleting the znode: /hbase/table-lock/mytable {code:java} [org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.delete(RecoverableZooKeeper.java:238), org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNode(ZKUtil.java:1341), org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNode(ZKUtil.java:1330), org.apache.hadoop.hbase.master.TableLockManager$ZKTableLockManager.tableDeleted(TableLockManager.java:442), org.apache.hadoop.hbase.master.procedure.MasterProcedureScheduler.markTableAsDeleted(MasterProcedureScheduler.java:895), org.apache.hadoop.hbase.master.procedure.MasterProcedureScheduler.completionCleanup(MasterProcedureScheduler.java:261), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.procedureFinished(ProcedureExecutor.java:1415), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:974), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:907), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.access$400(ProcedureExecutor.java:76), org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java:481), org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java)] {code} 5. HMaster2 becomes active 6. znode "/hbase/table-lock/mytable" was left on ZooKeeper was: There is a table {{mytable}} in a HBase cluster with two HMasters and two RegionServers. 1. Current active HMaster is HMaster1; 2. Client request to delete the table {{mytable}}; 3. HMaster1 tries to delete znode in ZooKeeper: /hbase/table/mytable {[org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.delete(RecoverableZooKeeper.java:238), org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1367), org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1359), org.apache.hadoop.hbase.zookeeper.ZKTableStateManager.setDeletedTable(ZKTableStateManager.java:244), org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.deleteAssignmentState(DeleteTableProcedure.java:383), org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:136), org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:57), org.apache.hadoop.hbase.procedure2.StateMachineProcedure.execute(StateMachineProcedure.java:139), org.apache.hadoop.hbase.procedure2.Procedure.doExecute(Procedure.java:499), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execProcedure(ProcedureExecutor.java:1166), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:954), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:907), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.access$400(ProcedureExecutor.java:76), org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java:481), org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java)]} > Unexpected crash in the process of
[jira] [Updated] (HBASE-26391) Unexpected crash in the process of deleting table causes stale data
[ https://issues.apache.org/jira/browse/HBASE-26391?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] anonymous updated HBASE-26391: -- Description: There is a table {{mytable}} in a HBase cluster with two HMasters and two RegionServers. 1. Current active HMaster is HMaster1; 2. Client request to delete the table {{mytable}}; 3. HMaster1 tries to delete znode in ZooKeeper: /hbase/table/mytable {[org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.delete(RecoverableZooKeeper.java:238), org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1367), org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1359), org.apache.hadoop.hbase.zookeeper.ZKTableStateManager.setDeletedTable(ZKTableStateManager.java:244), org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.deleteAssignmentState(DeleteTableProcedure.java:383), org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:136), org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:57), org.apache.hadoop.hbase.procedure2.StateMachineProcedure.execute(StateMachineProcedure.java:139), org.apache.hadoop.hbase.procedure2.Procedure.doExecute(Procedure.java:499), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execProcedure(ProcedureExecutor.java:1166), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:954), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:907), org.apache.hadoop.hbase.procedure2.ProcedureExecutor.access$400(ProcedureExecutor.java:76), org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java:481), org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java)]} > Unexpected crash in the process of deleting table causes stale data > --- > > Key: HBASE-26391 > URL: https://issues.apache.org/jira/browse/HBASE-26391 > Project: HBase > Issue Type: Bug >Affects Versions: 1.7.1 >Reporter: anonymous >Priority: Major > > There is a table {{mytable}} in a HBase cluster with two HMasters and two > RegionServers. > 1. Current active HMaster is HMaster1; > 2. Client request to delete the table {{mytable}}; > 3. HMaster1 tries to delete znode in ZooKeeper: /hbase/table/mytable > {[org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.delete(RecoverableZooKeeper.java:238), > > org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1367), > > org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeFailSilent(ZKUtil.java:1359), > > org.apache.hadoop.hbase.zookeeper.ZKTableStateManager.setDeletedTable(ZKTableStateManager.java:244), > > org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.deleteAssignmentState(DeleteTableProcedure.java:383), > > org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:136), > > org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:57), > > org.apache.hadoop.hbase.procedure2.StateMachineProcedure.execute(StateMachineProcedure.java:139), > org.apache.hadoop.hbase.procedure2.Procedure.doExecute(Procedure.java:499), > org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execProcedure(ProcedureExecutor.java:1166), > > org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:954), > > org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:907), > > org.apache.hadoop.hbase.procedure2.ProcedureExecutor.access$400(ProcedureExecutor.java:76), > > org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java:481), > > org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java)]} -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Created] (HBASE-26391) Unexpected crash in the process of deleting table causes stale data
anonymous created HBASE-26391: - Summary: Unexpected crash in the process of deleting table causes stale data Key: HBASE-26391 URL: https://issues.apache.org/jira/browse/HBASE-26391 Project: HBase Issue Type: Bug Affects Versions: 1.7.1 Reporter: anonymous -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (HBASE-26370) Unexpected crash fails the truncate operation and makes a disabled table enabled
[ https://issues.apache.org/jira/browse/HBASE-26370?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] anonymous updated HBASE-26370: -- Description: There is a table {{mytable}} in a HBase cluster with two HMasters and two RegionServers. 1. Current active HMaster is HMaster1; 2. Client disabled the table {{mytable}}; 3. {{mytable}} is in disabled state; 4. Client request to truncate {{mytable}}; 5. HMaster1 tries to truncate {{mytable}}. For the TruncateTableState {{TRUNCATE_TABLE_ASSIGN_REGIONS}}, HMaster1 tries to assign regions of the table and mark the table as Enabling; 6. HMaster1 crashes; 7. HMaster2 becomes active; 8. The truncate operation fails due to {{TableNotDisabledException}}, and {{mytable}} was left as enabled. In HMaster2, we got following info: {code:java} INFO [ProcedureExecutor-12] procedure2.ProcedureExecutor: Rolledback procedure TruncateTableProcedure (table=mytable preserveSplits=true) id=9 owner=myhbase state=ROLLEDBACK exec-time=4.6030 sec exception=org.apache.hadoop.hbase.TableNotDisabledException: mytable {code} was: There is a table {{mytable}} in a HBase cluster with two HMasters and two RegionServers. 1. Current active HMaster is HMaster1; 2. Client disabled the table {{mytable}}; 3. {{mytable}} is in disabled state; 4. Client request to truncate {{mytable}}; 5. HMaster1 tries to truncate {{mytable}}. For the TruncateTableState {{TRUNCATE_TABLE_ASSIGN_REGIONS}}, HMaster1 tries to assign regions of the table and mark the table as Enabling; 6. HMaster1 crashes; 7. HMaster2 becomes active; 8. The truncate operation fails due to {{TableNotDisabledException}}, and {{mytable}} was left as enabled. > Unexpected crash fails the truncate operation and makes a disabled table > enabled > > > Key: HBASE-26370 > URL: https://issues.apache.org/jira/browse/HBASE-26370 > Project: HBase > Issue Type: Bug >Affects Versions: 1.7.1 >Reporter: anonymous >Priority: Major > > There is a table {{mytable}} in a HBase cluster with two HMasters and two > RegionServers. > 1. Current active HMaster is HMaster1; > 2. Client disabled the table {{mytable}}; > 3. {{mytable}} is in disabled state; > 4. Client request to truncate {{mytable}}; > 5. HMaster1 tries to truncate {{mytable}}. For the TruncateTableState > {{TRUNCATE_TABLE_ASSIGN_REGIONS}}, HMaster1 tries to assign regions of the > table and mark the table as Enabling; > 6. HMaster1 crashes; > 7. HMaster2 becomes active; > 8. The truncate operation fails due to {{TableNotDisabledException}}, and > {{mytable}} was left as enabled. > In HMaster2, we got following info: > {code:java} > INFO [ProcedureExecutor-12] procedure2.ProcedureExecutor: Rolledback > procedure TruncateTableProcedure (table=mytable preserveSplits=true) id=9 > owner=myhbase state=ROLLEDBACK exec-time=4.6030 sec > exception=org.apache.hadoop.hbase.TableNotDisabledException: mytable > {code} -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (HBASE-26370) Unexpected crash fails the truncate operation and makes a disabled table enabled
[ https://issues.apache.org/jira/browse/HBASE-26370?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] anonymous updated HBASE-26370: -- Summary: Unexpected crash fails the truncate operation and makes a disabled table enabled (was: Unexpected crash makes a disabled table enabled) > Unexpected crash fails the truncate operation and makes a disabled table > enabled > > > Key: HBASE-26370 > URL: https://issues.apache.org/jira/browse/HBASE-26370 > Project: HBase > Issue Type: Bug >Affects Versions: 1.7.1 >Reporter: anonymous >Priority: Major > > There is a table {{mytable}} in a HBase cluster with two HMasters and two > RegionServers. > 1. Current active HMaster is HMaster1; > 2. Client disabled the table {{mytable}}; > 3. {{mytable}} is in disabled state; > 4. Client request to truncate {{mytable}}; > 5. HMaster1 tries to truncate {{mytable}}. For the TruncateTableState > {{TRUNCATE_TABLE_ASSIGN_REGIONS}}, HMaster1 tries to assign regions of the > table and mark the table as Enabling; > 6. HMaster1 crashes; > 7. HMaster2 becomes active; > 8. The truncate operation fails due to {{TableNotDisabledException}}, and > {{mytable}} was left as enabled. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (HBASE-26370) Unexpected crash makes a disabled table enabled
[ https://issues.apache.org/jira/browse/HBASE-26370?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] anonymous updated HBASE-26370: -- Summary: Unexpected crash makes a disabled table enabled (was: Unexpected crash makes a disabled table be enabled) > Unexpected crash makes a disabled table enabled > --- > > Key: HBASE-26370 > URL: https://issues.apache.org/jira/browse/HBASE-26370 > Project: HBase > Issue Type: Bug >Affects Versions: 1.7.1 >Reporter: anonymous >Priority: Major > > There is a table {{mytable}} in a HBase cluster with two HMasters and two > RegionServers. > 1. Current active HMaster is HMaster1; > 2. Client disabled the table {{mytable}}; > 3. {{mytable}} is in disabled state; > 4. Client request to truncate {{mytable}}; > 5. HMaster1 tries to truncate {{mytable}}. For the TruncateTableState > {{TRUNCATE_TABLE_ASSIGN_REGIONS}}, HMaster1 tries to assign regions of the > table and mark the table as Enabling; > 6. HMaster1 crashes; > 7. HMaster2 becomes active; > 8. The truncate operation fails due to {{TableNotDisabledException}}, and > {{mytable}} was left as enabled. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (HBASE-26370) Unexpected crash makes a disabled table be enabled
[ https://issues.apache.org/jira/browse/HBASE-26370?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] anonymous updated HBASE-26370: -- Summary: Unexpected crash makes a disabled table be enabled (was: Unexpected crash make a disabled table be enabled) > Unexpected crash makes a disabled table be enabled > -- > > Key: HBASE-26370 > URL: https://issues.apache.org/jira/browse/HBASE-26370 > Project: HBase > Issue Type: Bug >Affects Versions: 1.7.1 >Reporter: anonymous >Priority: Major > > There is a table {{mytable}} in a HBase cluster with two HMasters and two > RegionServers. > 1. Current active HMaster is HMaster1; > 2. Client disabled the table {{mytable}}; > 3. {{mytable}} is in disabled state; > 4. Client request to truncate {{mytable}}; > 5. HMaster1 tries to truncate {{mytable}}. For the TruncateTableState > {{TRUNCATE_TABLE_ASSIGN_REGIONS}}, HMaster1 tries to assign regions of the > table and mark the table as Enabling; > 6. HMaster1 crashes; > 7. HMaster2 becomes active; > 8. The truncate operation fails due to {{TableNotDisabledException}}, and > {{mytable}} was left as enabled. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Created] (HBASE-26370) Unexpected crash make a disabled table be enabled
anonymous created HBASE-26370: - Summary: Unexpected crash make a disabled table be enabled Key: HBASE-26370 URL: https://issues.apache.org/jira/browse/HBASE-26370 Project: HBase Issue Type: Bug Affects Versions: 1.7.1 Reporter: anonymous There is a table {{mytable}} in a HBase cluster with two HMasters and two RegionServers. 1. Current active HMaster is HMaster1; 2. Client disabled the table {{mytable}}; 3. {{mytable}} is in disabled state; 4. Client request to truncate {{mytable}}; 5. HMaster1 tries to truncate {{mytable}}. For the TruncateTableState {{TRUNCATE_TABLE_ASSIGN_REGIONS}}, HMaster1 tries to assign regions of the table and mark the table as Enabling; 6. HMaster1 crashes; 7. HMaster2 becomes active; 8. The truncate operation fails due to {{TableNotDisabledException}}, and {{mytable}} was left as enabled. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (HBASE-19400) Add missing security checks in MasterRpcServices
[ https://issues.apache.org/jira/browse/HBASE-19400?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Anonymous updated HBASE-19400: -- Status: Patch Available (was: Reopened) > Add missing security checks in MasterRpcServices > > > Key: HBASE-19400 > URL: https://issues.apache.org/jira/browse/HBASE-19400 > Project: HBase > Issue Type: Sub-task >Affects Versions: 2.0.0-beta-1 >Reporter: Balazs Meszaros >Assignee: Appy >Priority: Major > Fix For: 2.0.0-beta-2 > > Attachments: HBASE-19400.branch-1.001.patch, > HBASE-19400.master.001.patch, HBASE-19400.master.002.patch, > HBASE-19400.master.003.patch, HBASE-19400.master.004.patch, > HBASE-19400.master.004.patch, HBASE-19400.master.005.patch, > HBASE-19400.master.006.patch, HBASE-19400.master.007.patch, > HBASE-19400.master.007.patch > > > The following RPC methods in MasterRpcServices do not have ACL check for > ADMIN rights. > - normalize > - setNormalizerRunning > - runCatalogScan > - enableCatalogJanitor > - runCleanerChore > - setCleanerChoreRunning > - execMasterService > - execProcedure > - execProcedureWithRet -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HBASE-18230) Generated LICENSE file includes unsubstituted Velocity variables
[ https://issues.apache.org/jira/browse/HBASE-18230?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Anonymous updated HBASE-18230: -- Status: Patch Available (was: Reopened) > Generated LICENSE file includes unsubstituted Velocity variables > > > Key: HBASE-18230 > URL: https://issues.apache.org/jira/browse/HBASE-18230 > Project: HBase > Issue Type: Bug > Components: build >Affects Versions: 2.0.0-alpha-1 >Reporter: Mike Drob >Assignee: Mike Drob > Fix For: 3.0.0, 1.4.0, 1.3.2, 1.2.7, 2.0.0-alpha-2, 1.1.12 > > Attachments: HBASE-18230.addendum.patch, HBASE-18230.patch > > > From the release vote: > {quote} > we have a ton of places where we have velocity variables instead of > copyright years, but IIRC that's a problem on branch-1 right now too. > {quote} > This is referring to lines like these: > {noformat} > * javax.annotation-api, ${dep.licenses[0].comments} > * javax.servlet-api, ${dep.licenses[0].comments} > * jetty-schemas, ${dep.licenses[0].comments} > {noformat} -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[jira] [Updated] (HBASE-15291) FileSystem not closed in secure bulkLoad
[ https://issues.apache.org/jira/browse/HBASE-15291?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Anonymous updated HBASE-15291: -- Affects Version/s: 1.0.2 0.98.16.1 Status: Patch Available (was: Reopened) > FileSystem not closed in secure bulkLoad > > > Key: HBASE-15291 > URL: https://issues.apache.org/jira/browse/HBASE-15291 > Project: HBase > Issue Type: Bug >Affects Versions: 0.98.16.1, 1.0.2 >Reporter: Yong Zhang >Assignee: Yong Zhang > Fix For: 2.0.0, 1.3.0, 1.2.1, 0.98.18 > > Attachments: HBASE-15291.001.patch, HBASE-15291.002.patch, > HBASE-15291.addendum, patch > > > FileSystem not closed in secure bulkLoad after bulkLoad finish, it will > cause memory used more and more if too many bulkLoad . -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HBASE-14497) Reverse Scan threw StackOverflow caused by readPt checking
[ https://issues.apache.org/jira/browse/HBASE-14497?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Anonymous updated HBASE-14497: -- Status: Patch Available (was: Reopened) > Reverse Scan threw StackOverflow caused by readPt checking > -- > > Key: HBASE-14497 > URL: https://issues.apache.org/jira/browse/HBASE-14497 > Project: HBase > Issue Type: Bug >Affects Versions: 0.98.14, 2.0.0, 1.3.0 >Reporter: Yerui Sun >Assignee: Yerui Sun > Fix For: 2.0.0, 1.3.0 > > Attachments: 14497-branch-1-v6.patch, 14497-master-v6.patch, > HBASE-14497-0.98-v6.patch, HBASE-14497-0.98.patch, > HBASE-14497-branch-1-v2.patch, HBASE-14497-branch-1-v3.patch, > HBASE-14497-branch-1-v6.patch, HBASE-14497-branch-1.patch, > HBASE-14497-master-v2.patch, HBASE-14497-master-v3.patch, > HBASE-14497-master-v3.patch, HBASE-14497-master-v4.patch, > HBASE-14497-master-v5.patch, HBASE-14497-master.patch > > > I met stack overflow error in StoreFileScanner.seekToPreviousRow using > reversed scan. I searched and founded HBASE-14155, but it seems to be a > different reason. > The seekToPreviousRow will fetch the row which closest before, and compare > mvcc to the readPt, which acquired when scanner created. If the row's mvcc is > bigger than readPt, an recursive call of seekToPreviousRow will invoked, to > find the next closest before row. > Considering we created a scanner for reversed scan, and some data with > smaller rows was written and flushed, before calling scanner next. When > seekToPreviousRow was invoked, it would call itself recursively, until all > rows which written after scanner created were iterated. The depth of > recursive calling stack depends on the count of rows, the stack overflow > error will be threw if the count of rows is large, like 1. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HBASE-13965) Stochastic Load Balancer JMX Metrics
[ https://issues.apache.org/jira/browse/HBASE-13965?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Anonymous updated HBASE-13965: -- Status: Patch Available (was: Reopened) Stochastic Load Balancer JMX Metrics Key: HBASE-13965 URL: https://issues.apache.org/jira/browse/HBASE-13965 Project: HBase Issue Type: Improvement Components: Balancer, metrics Reporter: Lei Chen Assignee: Lei Chen Fix For: 2.0.0 Attachments: 13965-addendum.txt, HBASE-13965-branch-1.patch, HBASE-13965-v10.patch, HBASE-13965-v11.patch, HBASE-13965-v3.patch, HBASE-13965-v4.patch, HBASE-13965-v5.patch, HBASE-13965-v6.patch, HBASE-13965-v7.patch, HBASE-13965-v8.patch, HBASE-13965-v9.patch, HBASE-13965_v2.patch, HBase-13965-JConsole.png, HBase-13965-v1.patch, stochasticloadbalancerclasses_v2.png Today’s default HBase load balancer (the Stochastic load balancer) is cost function based. The cost function weights are tunable but no visibility into those cost function results is directly provided. A driving example is a cluster we have been tuning which has skewed rack size (one rack has half the nodes of the other few racks). We are tuning the cluster for uniform response time from all region servers with the ability to tolerate a rack failure. Balancing LocalityCost, RegionReplicaRack Cost and RegionCountSkew Cost is difficult without a way to attribute each cost function’s contribution to overall cost. What this jira proposes is to provide visibility via JMX into each cost function of the stochastic load balancer, as well as the overall cost of the balancing plan. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HBASE-13082) Coarsen StoreScanner locks to RegionScanner
[ https://issues.apache.org/jira/browse/HBASE-13082?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Anonymous updated HBASE-13082: -- Status: Patch Available (was: Reopened) Coarsen StoreScanner locks to RegionScanner --- Key: HBASE-13082 URL: https://issues.apache.org/jira/browse/HBASE-13082 Project: HBase Issue Type: Bug Reporter: Lars Hofhansl Assignee: Lars Hofhansl Fix For: 2.0.0, 1.1.0 Attachments: 13082-test.txt, 13082-v2.txt, 13082-v3.txt, 13082.txt, 13082.txt, gc.png, gc.png, gc.png, hits.png, next.png, next.png Continuing where HBASE-10015 left of. We can avoid locking (and memory fencing) inside StoreScanner by deferring to the lock already held by the RegionScanner. In tests this shows quite a scan improvement and reduced CPU (the fences make the cores wait for memory fetches). There are some drawbacks too: * All calls to RegionScanner need to be remain synchronized * Implementors of coprocessors need to be diligent in following the locking contract. For example Phoenix does not lock RegionScanner.nextRaw() and required in the documentation (not picking on Phoenix, this one is my fault as I told them it's OK) * possible starving of flushes and compaction with heavy read load. RegionScanner operations would keep getting the locks and the flushes/compactions would not be able finalize the set of files. I'll have a patch soon. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HBASE-5075) regionserver crashed and failover
[ https://issues.apache.org/jira/browse/HBASE-5075?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Anonymous updated HBASE-5075: - Affects Version/s: (was: 0.92.1) 0.90.5 Hadoop Flags: Incompatible change Status: Patch Available (was: Reopened) regionserver crashed and failover - Key: HBASE-5075 URL: https://issues.apache.org/jira/browse/HBASE-5075 Project: HBase Issue Type: Improvement Components: monitoring, regionserver, Replication, Zookeeper Affects Versions: 0.90.5 Reporter: zhiyuan.dai Fix For: 0.90.5 Attachments: Degion of Failure Detection.pdf, HBase-5075-shell.patch, HBase-5075-src.patch regionserver crashed,it is too long time to notify hmaster.when hmaster know regionserver's shutdown,it is long time to fetch the hlog's lease. hbase is a online db, availability is very important. i have a idea to improve availability, monitor node to check regionserver's pid.if this pid not exsits,i think the rs down,i will delete the znode,and force close the hlog file. so the period maybe 100ms. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HBASE-5673) The OOM problem of IPC client call cause all handle block
[ https://issues.apache.org/jira/browse/HBASE-5673?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Anonymous updated HBASE-5673: - Tags: 1 Affects Version/s: (was: 0.90.6) Release Note: 1 Hadoop Flags: Incompatible change (was: Reviewed) Status: Patch Available (was: Reopened) The OOM problem of IPC client call cause all handle block -- Key: HBASE-5673 URL: https://issues.apache.org/jira/browse/HBASE-5673 Project: HBase Issue Type: Bug Environment: 0.90.6 Reporter: xufeng Assignee: xufeng Labels: Fix For: 0.92.3 Attachments: HBASE-5673-90.patch, HBASE-5673-90-V2.patch if HBaseClient meet unable to create new native thread exception, the call will never complete because it be lost in calls queue. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira