[jira] [Commented] (HBASE-16853) Regions are assigned to Region Servers in /hbase/draining after HBase Master failover
[ https://issues.apache.org/jira/browse/HBASE-16853?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15665193#comment-15665193 ] Hudson commented on HBASE-16853: ABORTED: Integrated in Jenkins build HBase-0.98-matrix #415 (See [https://builds.apache.org/job/HBase-0.98-matrix/415/]) HBASE-16853 Regions are assigned to Region Servers in /hbase/draining (apurtell: rev dba43b62823cbaa663cf0c2f7b7e4dcd668bdbce) * (edit) hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestAssignmentListener.java * (edit) hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java * (edit) hbase-server/src/main/java/org/apache/hadoop/hbase/zookeeper/DrainingServerTracker.java > Regions are assigned to Region Servers in /hbase/draining after HBase Master > failover > - > > Key: HBASE-16853 > URL: https://issues.apache.org/jira/browse/HBASE-16853 > Project: HBase > Issue Type: Bug > Components: Balancer, Region Assignment >Affects Versions: 2.0.0, 1.3.0 >Reporter: David Pope >Assignee: David Pope > Fix For: 2.0.0, 1.3.0, 1.4.0, 0.98.24 > > Attachments: 16853.v2.txt, HBASE-16853.branch-1.3-v1.patch, > HBASE-16853.branch-1.3-v2.patch > > > h2. Problem > If there are Region Servers registered as "draining", they will continue to > have "draining" znodes after a HMaster failover; however, the balancer will > assign regions to them. > h2. How to reproduce (on hbase master): > # Add regionserver to /hbase/draining: {{bin/hbase-jruby > bin/draining_servers.rb add server1:16205}} > # Unload the regionserver: {{bin/hbase-jruby bin/region_mover.rb unload > server1:16205}} > # Kill the Active HMaster and failover to the Backup HMaster > # Run the balancer: {{hbase shell <<< "balancer"}} > # Notice regions get assigned on new Active Master to Region Servers in > /hbase/draining > h2. Root Cause > The Backup HMaster initializes the {{DrainingServerTracker}} before the > Region Servers are registered as "online" with the {{ServerManager}}. As a > result, the {{ServerManager.drainingServers}} isn't populated with existing > Region Servers in draining when we have an HMaster failover. > E.g., > # We have a region server in draining: {{server1,16205,1000}} > # The {{RegionServerTracker}} starts up and adds a ZK watcher on the Znode > for this RegionServer: {{/hbase/rs/server1,16205,1000}} > # The {{DrainingServerTracker}} starts and processes each Znode under > {{/hbase/draining}}, but the Region Server isn't registered as "online" so it > isn't added to the {{ServerManager.drainingServers}} list. > # The Region Server is added to the {{DrainingServerTracker.drainingServers}} > list. > # The Region Server's Znode watcher is triggered and the ZK watcher is > restarted. > # The Region Server is registered with {{ServerManager}} as "online". > *END STATE:* The Region Server has a Znode in {{/hbase/draining}}, but it is > registered as "online" and the Balancer will start assigning regions to it. > {code} > $ bin/hbase-jruby bin/draining_servers.rb list > [1] server1,16205,1000 > $ grep server1,16205,1000 logs/master-server1.log > 2016-10-14 16:02:47,713 DEBUG [server1:16001.activeMasterManager] > zookeeper.ZKUtil: master:16001-0x157c56adc810014, quorum=localhost:2181, > baseZNode=/hbase Set watcher on existing znode=/hbase/rs/server1,16205,1000 > [2] 2016-10-14 16:02:47,722 DEBUG [server1:16001.activeMasterManager] > zookeeper.RegionServerTracker: Added tracking of RS > /hbase/rs/server1,16205,1000 > 2016-10-14 16:02:47,730 DEBUG [server1:16001.activeMasterManager] > zookeeper.ZKUtil: master:16001-0x157c56adc810014, quorum=localhost:2181, > baseZNode=/hbase Set watcher on existing > znode=/hbase/draining/server1,16205,1000 > [3] 2016-10-14 16:02:47,731 WARN [server1:16001.activeMasterManager] > master.ServerManager: Server server1,16205,1000 is not currently online. > Ignoring request to add it to draining list. > [4] 2016-10-14 16:02:47,731 INFO [server1:16001.activeMasterManager] > zookeeper.DrainingServerTracker: Draining RS node created, adding to list > [server1,16205,1000] > 2016-10-14 16:02:47,971 DEBUG [main-EventThread] zookeeper.ZKUtil: > master:16001-0x157c56adc810014, quorum=localhost:2181, baseZNode=/hbase Set > watcher on existing > znode=/hbase/rs/dev6918.prn2.facebook.com,16205,1476486047114 > [5] 2016-10-14 16:02:47,976 DEBUG [main-EventThread] > zookeeper.RegionServerTracker: Added tracking of RS > /hbase/rs/server1,16205,1000 > [6] 2016-10-14 16:02:52,084 INFO > [RpcServer.FifoWFPBQ.default.handler=29,queue=2,port=16001] > master.ServerManager: Registering server=server1,16205,1000 > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-16853) Regions are assigned to Region Servers in /hbase/draining after HBase Master failover
[ https://issues.apache.org/jira/browse/HBASE-16853?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15659967#comment-15659967 ] Hudson commented on HBASE-16853: FAILURE: Integrated in Jenkins build HBase-0.98-on-Hadoop-1.1 #1288 (See [https://builds.apache.org/job/HBase-0.98-on-Hadoop-1.1/1288/]) HBASE-16853 Regions are assigned to Region Servers in /hbase/draining (apurtell: rev dba43b62823cbaa663cf0c2f7b7e4dcd668bdbce) * (edit) hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestAssignmentListener.java * (edit) hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java * (edit) hbase-server/src/main/java/org/apache/hadoop/hbase/zookeeper/DrainingServerTracker.java > Regions are assigned to Region Servers in /hbase/draining after HBase Master > failover > - > > Key: HBASE-16853 > URL: https://issues.apache.org/jira/browse/HBASE-16853 > Project: HBase > Issue Type: Bug > Components: Balancer, Region Assignment >Affects Versions: 2.0.0, 1.3.0 >Reporter: David Pope >Assignee: David Pope > Fix For: 2.0.0, 1.3.0, 1.4.0, 0.98.24 > > Attachments: 16853.v2.txt, HBASE-16853.branch-1.3-v1.patch, > HBASE-16853.branch-1.3-v2.patch > > > h2. Problem > If there are Region Servers registered as "draining", they will continue to > have "draining" znodes after a HMaster failover; however, the balancer will > assign regions to them. > h2. How to reproduce (on hbase master): > # Add regionserver to /hbase/draining: {{bin/hbase-jruby > bin/draining_servers.rb add server1:16205}} > # Unload the regionserver: {{bin/hbase-jruby bin/region_mover.rb unload > server1:16205}} > # Kill the Active HMaster and failover to the Backup HMaster > # Run the balancer: {{hbase shell <<< "balancer"}} > # Notice regions get assigned on new Active Master to Region Servers in > /hbase/draining > h2. Root Cause > The Backup HMaster initializes the {{DrainingServerTracker}} before the > Region Servers are registered as "online" with the {{ServerManager}}. As a > result, the {{ServerManager.drainingServers}} isn't populated with existing > Region Servers in draining when we have an HMaster failover. > E.g., > # We have a region server in draining: {{server1,16205,1000}} > # The {{RegionServerTracker}} starts up and adds a ZK watcher on the Znode > for this RegionServer: {{/hbase/rs/server1,16205,1000}} > # The {{DrainingServerTracker}} starts and processes each Znode under > {{/hbase/draining}}, but the Region Server isn't registered as "online" so it > isn't added to the {{ServerManager.drainingServers}} list. > # The Region Server is added to the {{DrainingServerTracker.drainingServers}} > list. > # The Region Server's Znode watcher is triggered and the ZK watcher is > restarted. > # The Region Server is registered with {{ServerManager}} as "online". > *END STATE:* The Region Server has a Znode in {{/hbase/draining}}, but it is > registered as "online" and the Balancer will start assigning regions to it. > {code} > $ bin/hbase-jruby bin/draining_servers.rb list > [1] server1,16205,1000 > $ grep server1,16205,1000 logs/master-server1.log > 2016-10-14 16:02:47,713 DEBUG [server1:16001.activeMasterManager] > zookeeper.ZKUtil: master:16001-0x157c56adc810014, quorum=localhost:2181, > baseZNode=/hbase Set watcher on existing znode=/hbase/rs/server1,16205,1000 > [2] 2016-10-14 16:02:47,722 DEBUG [server1:16001.activeMasterManager] > zookeeper.RegionServerTracker: Added tracking of RS > /hbase/rs/server1,16205,1000 > 2016-10-14 16:02:47,730 DEBUG [server1:16001.activeMasterManager] > zookeeper.ZKUtil: master:16001-0x157c56adc810014, quorum=localhost:2181, > baseZNode=/hbase Set watcher on existing > znode=/hbase/draining/server1,16205,1000 > [3] 2016-10-14 16:02:47,731 WARN [server1:16001.activeMasterManager] > master.ServerManager: Server server1,16205,1000 is not currently online. > Ignoring request to add it to draining list. > [4] 2016-10-14 16:02:47,731 INFO [server1:16001.activeMasterManager] > zookeeper.DrainingServerTracker: Draining RS node created, adding to list > [server1,16205,1000] > 2016-10-14 16:02:47,971 DEBUG [main-EventThread] zookeeper.ZKUtil: > master:16001-0x157c56adc810014, quorum=localhost:2181, baseZNode=/hbase Set > watcher on existing > znode=/hbase/rs/dev6918.prn2.facebook.com,16205,1476486047114 > [5] 2016-10-14 16:02:47,976 DEBUG [main-EventThread] > zookeeper.RegionServerTracker: Added tracking of RS > /hbase/rs/server1,16205,1000 > [6] 2016-10-14 16:02:52,084 INFO > [RpcServer.FifoWFPBQ.default.handler=29,queue=2,port=16001] > master.ServerManager: Registering server=server1,16205,1000 > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-16853) Regions are assigned to Region Servers in /hbase/draining after HBase Master failover
[ https://issues.apache.org/jira/browse/HBASE-16853?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15581703#comment-15581703 ] Hudson commented on HBASE-16853: SUCCESS: Integrated in Jenkins build HBase-Trunk_matrix #1801 (See [https://builds.apache.org/job/HBase-Trunk_matrix/1801/]) HBASE-16853 Regions are assigned to Region Servers in /hbase/draining (tedyu: rev 109db38b6ad091b23593ee46b1e919136aed7886) * (edit) hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestAssignmentListener.java * (edit) hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java * (edit) hbase-server/src/main/java/org/apache/hadoop/hbase/zookeeper/DrainingServerTracker.java > Regions are assigned to Region Servers in /hbase/draining after HBase Master > failover > - > > Key: HBASE-16853 > URL: https://issues.apache.org/jira/browse/HBASE-16853 > Project: HBase > Issue Type: Bug > Components: Balancer, Region Assignment >Affects Versions: 2.0.0, 1.3.0 >Reporter: David Pope >Assignee: David Pope > Fix For: 2.0.0, 1.3.0, 1.4.0 > > Attachments: 16853.v2.txt, HBASE-16853.branch-1.3-v1.patch, > HBASE-16853.branch-1.3-v2.patch > > > h2. Problem > If there are Region Servers registered as "draining", they will continue to > have "draining" znodes after a HMaster failover; however, the balancer will > assign regions to them. > h2. How to reproduce (on hbase master): > # Add regionserver to /hbase/draining: {{bin/hbase-jruby > bin/draining_servers.rb add server1:16205}} > # Unload the regionserver: {{bin/hbase-jruby bin/region_mover.rb unload > server1:16205}} > # Kill the Active HMaster and failover to the Backup HMaster > # Run the balancer: {{hbase shell <<< "balancer"}} > # Notice regions get assigned on new Active Master to Region Servers in > /hbase/draining > h2. Root Cause > The Backup HMaster initializes the {{DrainingServerTracker}} before the > Region Servers are registered as "online" with the {{ServerManager}}. As a > result, the {{ServerManager.drainingServers}} isn't populated with existing > Region Servers in draining when we have an HMaster failover. > E.g., > # We have a region server in draining: {{server1,16205,1000}} > # The {{RegionServerTracker}} starts up and adds a ZK watcher on the Znode > for this RegionServer: {{/hbase/rs/server1,16205,1000}} > # The {{DrainingServerTracker}} starts and processes each Znode under > {{/hbase/draining}}, but the Region Server isn't registered as "online" so it > isn't added to the {{ServerManager.drainingServers}} list. > # The Region Server is added to the {{DrainingServerTracker.drainingServers}} > list. > # The Region Server's Znode watcher is triggered and the ZK watcher is > restarted. > # The Region Server is registered with {{ServerManager}} as "online". > *END STATE:* The Region Server has a Znode in {{/hbase/draining}}, but it is > registered as "online" and the Balancer will start assigning regions to it. > {code} > $ bin/hbase-jruby bin/draining_servers.rb list > [1] server1,16205,1000 > $ grep server1,16205,1000 logs/master-server1.log > 2016-10-14 16:02:47,713 DEBUG [server1:16001.activeMasterManager] > zookeeper.ZKUtil: master:16001-0x157c56adc810014, quorum=localhost:2181, > baseZNode=/hbase Set watcher on existing znode=/hbase/rs/server1,16205,1000 > [2] 2016-10-14 16:02:47,722 DEBUG [server1:16001.activeMasterManager] > zookeeper.RegionServerTracker: Added tracking of RS > /hbase/rs/server1,16205,1000 > 2016-10-14 16:02:47,730 DEBUG [server1:16001.activeMasterManager] > zookeeper.ZKUtil: master:16001-0x157c56adc810014, quorum=localhost:2181, > baseZNode=/hbase Set watcher on existing > znode=/hbase/draining/server1,16205,1000 > [3] 2016-10-14 16:02:47,731 WARN [server1:16001.activeMasterManager] > master.ServerManager: Server server1,16205,1000 is not currently online. > Ignoring request to add it to draining list. > [4] 2016-10-14 16:02:47,731 INFO [server1:16001.activeMasterManager] > zookeeper.DrainingServerTracker: Draining RS node created, adding to list > [server1,16205,1000] > 2016-10-14 16:02:47,971 DEBUG [main-EventThread] zookeeper.ZKUtil: > master:16001-0x157c56adc810014, quorum=localhost:2181, baseZNode=/hbase Set > watcher on existing > znode=/hbase/rs/dev6918.prn2.facebook.com,16205,1476486047114 > [5] 2016-10-14 16:02:47,976 DEBUG [main-EventThread] > zookeeper.RegionServerTracker: Added tracking of RS > /hbase/rs/server1,16205,1000 > [6] 2016-10-14 16:02:52,084 INFO > [RpcServer.FifoWFPBQ.default.handler=29,queue=2,port=16001] > master.ServerManager: Registering server=server1,16205,1000 > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-16853) Regions are assigned to Region Servers in /hbase/draining after HBase Master failover
[ https://issues.apache.org/jira/browse/HBASE-16853?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15581642#comment-15581642 ] Hudson commented on HBASE-16853: SUCCESS: Integrated in Jenkins build HBase-1.3-JDK7 #44 (See [https://builds.apache.org/job/HBase-1.3-JDK7/44/]) HBASE-16853 Regions are assigned to Region Servers in /hbase/draining (tedyu: rev d38310aa4d97c0be0e2d1f5113630510a02f7ade) * (edit) hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestAssignmentListener.java * (edit) hbase-server/src/main/java/org/apache/hadoop/hbase/zookeeper/DrainingServerTracker.java * (edit) hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java > Regions are assigned to Region Servers in /hbase/draining after HBase Master > failover > - > > Key: HBASE-16853 > URL: https://issues.apache.org/jira/browse/HBASE-16853 > Project: HBase > Issue Type: Bug > Components: Balancer, Region Assignment >Affects Versions: 2.0.0, 1.3.0 >Reporter: David Pope >Assignee: David Pope > Fix For: 2.0.0, 1.3.0, 1.4.0 > > Attachments: 16853.v2.txt, HBASE-16853.branch-1.3-v1.patch, > HBASE-16853.branch-1.3-v2.patch > > > h2. Problem > If there are Region Servers registered as "draining", they will continue to > have "draining" znodes after a HMaster failover; however, the balancer will > assign regions to them. > h2. How to reproduce (on hbase master): > # Add regionserver to /hbase/draining: {{bin/hbase-jruby > bin/draining_servers.rb add server1:16205}} > # Unload the regionserver: {{bin/hbase-jruby bin/region_mover.rb unload > server1:16205}} > # Kill the Active HMaster and failover to the Backup HMaster > # Run the balancer: {{hbase shell <<< "balancer"}} > # Notice regions get assigned on new Active Master to Region Servers in > /hbase/draining > h2. Root Cause > The Backup HMaster initializes the {{DrainingServerTracker}} before the > Region Servers are registered as "online" with the {{ServerManager}}. As a > result, the {{ServerManager.drainingServers}} isn't populated with existing > Region Servers in draining when we have an HMaster failover. > E.g., > # We have a region server in draining: {{server1,16205,1000}} > # The {{RegionServerTracker}} starts up and adds a ZK watcher on the Znode > for this RegionServer: {{/hbase/rs/server1,16205,1000}} > # The {{DrainingServerTracker}} starts and processes each Znode under > {{/hbase/draining}}, but the Region Server isn't registered as "online" so it > isn't added to the {{ServerManager.drainingServers}} list. > # The Region Server is added to the {{DrainingServerTracker.drainingServers}} > list. > # The Region Server's Znode watcher is triggered and the ZK watcher is > restarted. > # The Region Server is registered with {{ServerManager}} as "online". > *END STATE:* The Region Server has a Znode in {{/hbase/draining}}, but it is > registered as "online" and the Balancer will start assigning regions to it. > {code} > $ bin/hbase-jruby bin/draining_servers.rb list > [1] server1,16205,1000 > $ grep server1,16205,1000 logs/master-server1.log > 2016-10-14 16:02:47,713 DEBUG [server1:16001.activeMasterManager] > zookeeper.ZKUtil: master:16001-0x157c56adc810014, quorum=localhost:2181, > baseZNode=/hbase Set watcher on existing znode=/hbase/rs/server1,16205,1000 > [2] 2016-10-14 16:02:47,722 DEBUG [server1:16001.activeMasterManager] > zookeeper.RegionServerTracker: Added tracking of RS > /hbase/rs/server1,16205,1000 > 2016-10-14 16:02:47,730 DEBUG [server1:16001.activeMasterManager] > zookeeper.ZKUtil: master:16001-0x157c56adc810014, quorum=localhost:2181, > baseZNode=/hbase Set watcher on existing > znode=/hbase/draining/server1,16205,1000 > [3] 2016-10-14 16:02:47,731 WARN [server1:16001.activeMasterManager] > master.ServerManager: Server server1,16205,1000 is not currently online. > Ignoring request to add it to draining list. > [4] 2016-10-14 16:02:47,731 INFO [server1:16001.activeMasterManager] > zookeeper.DrainingServerTracker: Draining RS node created, adding to list > [server1,16205,1000] > 2016-10-14 16:02:47,971 DEBUG [main-EventThread] zookeeper.ZKUtil: > master:16001-0x157c56adc810014, quorum=localhost:2181, baseZNode=/hbase Set > watcher on existing > znode=/hbase/rs/dev6918.prn2.facebook.com,16205,1476486047114 > [5] 2016-10-14 16:02:47,976 DEBUG [main-EventThread] > zookeeper.RegionServerTracker: Added tracking of RS > /hbase/rs/server1,16205,1000 > [6] 2016-10-14 16:02:52,084 INFO > [RpcServer.FifoWFPBQ.default.handler=29,queue=2,port=16001] > master.ServerManager: Registering server=server1,16205,1000 > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-16853) Regions are assigned to Region Servers in /hbase/draining after HBase Master failover
[ https://issues.apache.org/jira/browse/HBASE-16853?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15580993#comment-15580993 ] Hadoop QA commented on HBASE-16853: --- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 13s {color} | {color:blue} Docker mode activated. {color} | | {color:blue}0{color} | {color:blue} patch {color} | {color:blue} 0m 3s {color} | {color:blue} The patch file was not named according to hbase's naming conventions. Please see https://yetus.apache.org/documentation/0.3.0/precommit-patchnames for instructions. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s {color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s {color} | {color:green} The patch appears to include 1 new or modified test files. {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 2m 57s {color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 33s {color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 44s {color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 13s {color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 39s {color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 25s {color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 44s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 34s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 34s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 45s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 13s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s {color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} hadoopcheck {color} | {color:green} 26m 34s {color} | {color:green} Patch does not cause any errors with Hadoop 2.4.0 2.4.1 2.5.0 2.5.1 2.5.2 2.6.1 2.6.2 2.6.3 2.7.1. {color} | | {color:green}+1{color} | {color:green} hbaseprotoc {color} | {color:green} 0m 12s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 47s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 25s {color} | {color:green} the patch passed {color} | | {color:red}-1{color} | {color:red} unit {color} | {color:red} 79m 12s {color} | {color:red} hbase-server in the patch failed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 14s {color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 117m 41s {color} | {color:black} {color} | \\ \\ || Reason || Tests || | Timed out junit tests | org.apache.hadoop.hbase.security.visibility.TestVisibilityLabelsWithDeletes | | | org.apache.hadoop.hbase.filter.TestMultiRowRangeFilter | \\ \\ || Subsystem || Report/Notes || | Docker | Client=1.11.2 Server=1.11.2 Image:yetus/hbase:7bda515 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12833663/16853.v2.txt | | JIRA Issue | HBASE-16853 | | Optional Tests | asflicense javac javadoc unit findbugs hadoopcheck hbaseanti checkstyle compile | | uname | Linux dbd83b6f41e1 3.13.0-95-generic #142-Ubuntu SMP Fri Aug 12 17:00:09 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /home/jenkins/jenkins-slave/workspace/PreCommit-HBASE-Build/component/dev-support/hbase-personality.sh | | git revision | master / 0f38415 | | Default Java | 1.8.0_101 | | findbugs | v3.0.0 | | unit | https://builds.apache.org/job/PreCommit-HBASE-Build/4042/artifact/patchprocess/patch-unit-hbase-server.txt | | unit test logs | https://builds.apache.org/job/PreCommit-HBASE-Build/4042/artifact/patchprocess/patch-unit-hbase-server.txt | | Test Results | https://builds.apache.org/job/PreCommit-HBASE-Build/4042/testReport/ | | modules | C: hbase-server U: hbase-server | | Console
[jira] [Commented] (HBASE-16853) Regions are assigned to Region Servers in /hbase/draining after HBase Master failover
[ https://issues.apache.org/jira/browse/HBASE-16853?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15580931#comment-15580931 ] Heng Chen commented on HBASE-16853: --- +1 > Regions are assigned to Region Servers in /hbase/draining after HBase Master > failover > - > > Key: HBASE-16853 > URL: https://issues.apache.org/jira/browse/HBASE-16853 > Project: HBase > Issue Type: Bug > Components: Balancer, Region Assignment >Affects Versions: 2.0.0, 1.3.0 >Reporter: David Pope >Assignee: David Pope > Fix For: 2.0.0, 1.3.0, 1.4.0 > > Attachments: 16853.v2.txt, HBASE-16853.branch-1.3-v1.patch, > HBASE-16853.branch-1.3-v2.patch > > > h2. Problem > If there are Region Servers registered as "draining", they will continue to > have "draining" znodes after a HMaster failover; however, the balancer will > assign regions to them. > h2. How to reproduce (on hbase master): > # Add regionserver to /hbase/draining: {{bin/hbase-jruby > bin/draining_servers.rb add server1:16205}} > # Unload the regionserver: {{bin/hbase-jruby bin/region_mover.rb unload > server1:16205}} > # Kill the Active HMaster and failover to the Backup HMaster > # Run the balancer: {{hbase shell <<< "balancer"}} > # Notice regions get assigned on new Active Master to Region Servers in > /hbase/draining > h2. Root Cause > The Backup HMaster initializes the {{DrainingServerTracker}} before the > Region Servers are registered as "online" with the {{ServerManager}}. As a > result, the {{ServerManager.drainingServers}} isn't populated with existing > Region Servers in draining when we have an HMaster failover. > E.g., > # We have a region server in draining: {{server1,16205,1000}} > # The {{RegionServerTracker}} starts up and adds a ZK watcher on the Znode > for this RegionServer: {{/hbase/rs/server1,16205,1000}} > # The {{DrainingServerTracker}} starts and processes each Znode under > {{/hbase/draining}}, but the Region Server isn't registered as "online" so it > isn't added to the {{ServerManager.drainingServers}} list. > # The Region Server is added to the {{DrainingServerTracker.drainingServers}} > list. > # The Region Server's Znode watcher is triggered and the ZK watcher is > restarted. > # The Region Server is registered with {{ServerManager}} as "online". > *END STATE:* The Region Server has a Znode in {{/hbase/draining}}, but it is > registered as "online" and the Balancer will start assigning regions to it. > {code} > $ bin/hbase-jruby bin/draining_servers.rb list > [1] server1,16205,1000 > $ grep server1,16205,1000 logs/master-server1.log > 2016-10-14 16:02:47,713 DEBUG [server1:16001.activeMasterManager] > zookeeper.ZKUtil: master:16001-0x157c56adc810014, quorum=localhost:2181, > baseZNode=/hbase Set watcher on existing znode=/hbase/rs/server1,16205,1000 > [2] 2016-10-14 16:02:47,722 DEBUG [server1:16001.activeMasterManager] > zookeeper.RegionServerTracker: Added tracking of RS > /hbase/rs/server1,16205,1000 > 2016-10-14 16:02:47,730 DEBUG [server1:16001.activeMasterManager] > zookeeper.ZKUtil: master:16001-0x157c56adc810014, quorum=localhost:2181, > baseZNode=/hbase Set watcher on existing > znode=/hbase/draining/server1,16205,1000 > [3] 2016-10-14 16:02:47,731 WARN [server1:16001.activeMasterManager] > master.ServerManager: Server server1,16205,1000 is not currently online. > Ignoring request to add it to draining list. > [4] 2016-10-14 16:02:47,731 INFO [server1:16001.activeMasterManager] > zookeeper.DrainingServerTracker: Draining RS node created, adding to list > [server1,16205,1000] > 2016-10-14 16:02:47,971 DEBUG [main-EventThread] zookeeper.ZKUtil: > master:16001-0x157c56adc810014, quorum=localhost:2181, baseZNode=/hbase Set > watcher on existing > znode=/hbase/rs/dev6918.prn2.facebook.com,16205,1476486047114 > [5] 2016-10-14 16:02:47,976 DEBUG [main-EventThread] > zookeeper.RegionServerTracker: Added tracking of RS > /hbase/rs/server1,16205,1000 > [6] 2016-10-14 16:02:52,084 INFO > [RpcServer.FifoWFPBQ.default.handler=29,queue=2,port=16001] > master.ServerManager: Registering server=server1,16205,1000 > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-16853) Regions are assigned to Region Servers in /hbase/draining after HBase Master failover
[ https://issues.apache.org/jira/browse/HBASE-16853?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15580933#comment-15580933 ] Ted Yu commented on HBASE-16853: Integrated to branch-1.3 and branch-1 Waiting for QA on master patch. > Regions are assigned to Region Servers in /hbase/draining after HBase Master > failover > - > > Key: HBASE-16853 > URL: https://issues.apache.org/jira/browse/HBASE-16853 > Project: HBase > Issue Type: Bug > Components: Balancer, Region Assignment >Affects Versions: 2.0.0, 1.3.0 >Reporter: David Pope >Assignee: David Pope > Fix For: 2.0.0, 1.3.0, 1.4.0 > > Attachments: 16853.v2.txt, HBASE-16853.branch-1.3-v1.patch, > HBASE-16853.branch-1.3-v2.patch > > > h2. Problem > If there are Region Servers registered as "draining", they will continue to > have "draining" znodes after a HMaster failover; however, the balancer will > assign regions to them. > h2. How to reproduce (on hbase master): > # Add regionserver to /hbase/draining: {{bin/hbase-jruby > bin/draining_servers.rb add server1:16205}} > # Unload the regionserver: {{bin/hbase-jruby bin/region_mover.rb unload > server1:16205}} > # Kill the Active HMaster and failover to the Backup HMaster > # Run the balancer: {{hbase shell <<< "balancer"}} > # Notice regions get assigned on new Active Master to Region Servers in > /hbase/draining > h2. Root Cause > The Backup HMaster initializes the {{DrainingServerTracker}} before the > Region Servers are registered as "online" with the {{ServerManager}}. As a > result, the {{ServerManager.drainingServers}} isn't populated with existing > Region Servers in draining when we have an HMaster failover. > E.g., > # We have a region server in draining: {{server1,16205,1000}} > # The {{RegionServerTracker}} starts up and adds a ZK watcher on the Znode > for this RegionServer: {{/hbase/rs/server1,16205,1000}} > # The {{DrainingServerTracker}} starts and processes each Znode under > {{/hbase/draining}}, but the Region Server isn't registered as "online" so it > isn't added to the {{ServerManager.drainingServers}} list. > # The Region Server is added to the {{DrainingServerTracker.drainingServers}} > list. > # The Region Server's Znode watcher is triggered and the ZK watcher is > restarted. > # The Region Server is registered with {{ServerManager}} as "online". > *END STATE:* The Region Server has a Znode in {{/hbase/draining}}, but it is > registered as "online" and the Balancer will start assigning regions to it. > {code} > $ bin/hbase-jruby bin/draining_servers.rb list > [1] server1,16205,1000 > $ grep server1,16205,1000 logs/master-server1.log > 2016-10-14 16:02:47,713 DEBUG [server1:16001.activeMasterManager] > zookeeper.ZKUtil: master:16001-0x157c56adc810014, quorum=localhost:2181, > baseZNode=/hbase Set watcher on existing znode=/hbase/rs/server1,16205,1000 > [2] 2016-10-14 16:02:47,722 DEBUG [server1:16001.activeMasterManager] > zookeeper.RegionServerTracker: Added tracking of RS > /hbase/rs/server1,16205,1000 > 2016-10-14 16:02:47,730 DEBUG [server1:16001.activeMasterManager] > zookeeper.ZKUtil: master:16001-0x157c56adc810014, quorum=localhost:2181, > baseZNode=/hbase Set watcher on existing > znode=/hbase/draining/server1,16205,1000 > [3] 2016-10-14 16:02:47,731 WARN [server1:16001.activeMasterManager] > master.ServerManager: Server server1,16205,1000 is not currently online. > Ignoring request to add it to draining list. > [4] 2016-10-14 16:02:47,731 INFO [server1:16001.activeMasterManager] > zookeeper.DrainingServerTracker: Draining RS node created, adding to list > [server1,16205,1000] > 2016-10-14 16:02:47,971 DEBUG [main-EventThread] zookeeper.ZKUtil: > master:16001-0x157c56adc810014, quorum=localhost:2181, baseZNode=/hbase Set > watcher on existing > znode=/hbase/rs/dev6918.prn2.facebook.com,16205,1476486047114 > [5] 2016-10-14 16:02:47,976 DEBUG [main-EventThread] > zookeeper.RegionServerTracker: Added tracking of RS > /hbase/rs/server1,16205,1000 > [6] 2016-10-14 16:02:52,084 INFO > [RpcServer.FifoWFPBQ.default.handler=29,queue=2,port=16001] > master.ServerManager: Registering server=server1,16205,1000 > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-16853) Regions are assigned to Region Servers in /hbase/draining after HBase Master failover
[ https://issues.apache.org/jira/browse/HBASE-16853?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15580883#comment-15580883 ] David Pope commented on HBASE-16853: Thanks, Ted. That's a patch for master, correct? Do you want any others, such as branch-1? > Regions are assigned to Region Servers in /hbase/draining after HBase Master > failover > - > > Key: HBASE-16853 > URL: https://issues.apache.org/jira/browse/HBASE-16853 > Project: HBase > Issue Type: Bug > Components: Balancer, Region Assignment >Affects Versions: 2.0.0, 1.3.0 >Reporter: David Pope >Assignee: David Pope > Fix For: 2.0.0, 1.3.0, 1.4.0 > > Attachments: 16853.v2.txt, HBASE-16853.branch-1.3-v1.patch, > HBASE-16853.branch-1.3-v2.patch > > > h2. Problem > If there are Region Servers registered as "draining", they will continue to > have "draining" znodes after a HMaster failover; however, the balancer will > assign regions to them. > h2. How to reproduce (on hbase master): > # Add regionserver to /hbase/draining: {{bin/hbase-jruby > bin/draining_servers.rb add server1:16205}} > # Unload the regionserver: {{bin/hbase-jruby bin/region_mover.rb unload > server1:16205}} > # Kill the Active HMaster and failover to the Backup HMaster > # Run the balancer: {{hbase shell <<< "balancer"}} > # Notice regions get assigned on new Active Master to Region Servers in > /hbase/draining > h2. Root Cause > The Backup HMaster initializes the {{DrainingServerTracker}} before the > Region Servers are registered as "online" with the {{ServerManager}}. As a > result, the {{ServerManager.drainingServers}} isn't populated with existing > Region Servers in draining when we have an HMaster failover. > E.g., > # We have a region server in draining: {{server1,16205,1000}} > # The {{RegionServerTracker}} starts up and adds a ZK watcher on the Znode > for this RegionServer: {{/hbase/rs/server1,16205,1000}} > # The {{DrainingServerTracker}} starts and processes each Znode under > {{/hbase/draining}}, but the Region Server isn't registered as "online" so it > isn't added to the {{ServerManager.drainingServers}} list. > # The Region Server is added to the {{DrainingServerTracker.drainingServers}} > list. > # The Region Server's Znode watcher is triggered and the ZK watcher is > restarted. > # The Region Server is registered with {{ServerManager}} as "online". > *END STATE:* The Region Server has a Znode in {{/hbase/draining}}, but it is > registered as "online" and the Balancer will start assigning regions to it. > {code} > $ bin/hbase-jruby bin/draining_servers.rb list > [1] server1,16205,1000 > $ grep server1,16205,1000 logs/master-server1.log > 2016-10-14 16:02:47,713 DEBUG [server1:16001.activeMasterManager] > zookeeper.ZKUtil: master:16001-0x157c56adc810014, quorum=localhost:2181, > baseZNode=/hbase Set watcher on existing znode=/hbase/rs/server1,16205,1000 > [2] 2016-10-14 16:02:47,722 DEBUG [server1:16001.activeMasterManager] > zookeeper.RegionServerTracker: Added tracking of RS > /hbase/rs/server1,16205,1000 > 2016-10-14 16:02:47,730 DEBUG [server1:16001.activeMasterManager] > zookeeper.ZKUtil: master:16001-0x157c56adc810014, quorum=localhost:2181, > baseZNode=/hbase Set watcher on existing > znode=/hbase/draining/server1,16205,1000 > [3] 2016-10-14 16:02:47,731 WARN [server1:16001.activeMasterManager] > master.ServerManager: Server server1,16205,1000 is not currently online. > Ignoring request to add it to draining list. > [4] 2016-10-14 16:02:47,731 INFO [server1:16001.activeMasterManager] > zookeeper.DrainingServerTracker: Draining RS node created, adding to list > [server1,16205,1000] > 2016-10-14 16:02:47,971 DEBUG [main-EventThread] zookeeper.ZKUtil: > master:16001-0x157c56adc810014, quorum=localhost:2181, baseZNode=/hbase Set > watcher on existing > znode=/hbase/rs/dev6918.prn2.facebook.com,16205,1476486047114 > [5] 2016-10-14 16:02:47,976 DEBUG [main-EventThread] > zookeeper.RegionServerTracker: Added tracking of RS > /hbase/rs/server1,16205,1000 > [6] 2016-10-14 16:02:52,084 INFO > [RpcServer.FifoWFPBQ.default.handler=29,queue=2,port=16001] > master.ServerManager: Registering server=server1,16205,1000 > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-16853) Regions are assigned to Region Servers in /hbase/draining after HBase Master failover
[ https://issues.apache.org/jira/browse/HBASE-16853?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15580709#comment-15580709 ] Hadoop QA commented on HBASE-16853: --- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 16s {color} | {color:blue} Docker mode activated. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s {color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s {color} | {color:green} The patch appears to include 1 new or modified test files. {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 1m 49s {color} | {color:green} branch-1 passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 30s {color} | {color:green} branch-1 passed with JDK v1.8.0_101 {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 35s {color} | {color:green} branch-1 passed with JDK v1.7.0_80 {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 55s {color} | {color:green} branch-1 passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 16s {color} | {color:green} branch-1 passed {color} | | {color:red}-1{color} | {color:red} findbugs {color} | {color:red} 1m 53s {color} | {color:red} hbase-server in branch-1 has 1 extant Findbugs warnings. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 25s {color} | {color:green} branch-1 passed with JDK v1.8.0_101 {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 33s {color} | {color:green} branch-1 passed with JDK v1.7.0_80 {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 45s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 30s {color} | {color:green} the patch passed with JDK v1.8.0_101 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 30s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 33s {color} | {color:green} the patch passed with JDK v1.7.0_80 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 33s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 56s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 16s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s {color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} hadoopcheck {color} | {color:green} 16m 24s {color} | {color:green} The patch does not cause any errors with Hadoop 2.4.0 2.4.1 2.5.0 2.5.1 2.5.2 2.6.1 2.6.2 2.6.3 2.7.1. {color} | | {color:green}+1{color} | {color:green} hbaseprotoc {color} | {color:green} 0m 15s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 2m 5s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 24s {color} | {color:green} the patch passed with JDK v1.8.0_101 {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 33s {color} | {color:green} the patch passed with JDK v1.7.0_80 {color} | | {color:red}-1{color} | {color:red} unit {color} | {color:red} 72m 36s {color} | {color:red} hbase-server in the patch failed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 15s {color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 103m 12s {color} | {color:black} {color} | \\ \\ || Reason || Tests || | Failed junit tests | hadoop.hbase.mapreduce.TestMultiTableSnapshotInputFormat | | Timed out junit tests | org.apache.hadoop.hbase.TestHBaseTestingUtility | | | org.apache.hadoop.hbase.regionserver.TestClusterId | | | org.apache.hadoop.hbase.quotas.TestQuotaAdmin | | | org.apache.hadoop.hbase.filter.TestFuzzyRowFilterEndToEnd | \\ \\ || Subsystem || Report/Notes || | Docker | Client=1.11.2 Server=1.11.2 Image:yetus/hbase:b2c5d84 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12833648/HBASE-16853.branch-1.3-v2.patch | | JIRA Issue | HBASE-16853 | | Optional Tests | asflicense javac javadoc unit findbugs hadoopcheck hbaseanti
[jira] [Commented] (HBASE-16853) Regions are assigned to Region Servers in /hbase/draining after HBase Master failover
[ https://issues.apache.org/jira/browse/HBASE-16853?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15580034#comment-15580034 ] Ted Yu commented on HBASE-16853: lgtm Can you add the new test to TestAssignmentListener (which uses ServerListener) ? Please attach patch for master. > Regions are assigned to Region Servers in /hbase/draining after HBase Master > failover > - > > Key: HBASE-16853 > URL: https://issues.apache.org/jira/browse/HBASE-16853 > Project: HBase > Issue Type: Bug > Components: Balancer, Region Assignment >Affects Versions: 2.0.0, 1.3.0 >Reporter: David Pope >Assignee: David Pope > Attachments: HBASE-16853.branch-1.3-v1.patch > > > h2. Problem > If there are Region Servers registered as "draining", they will continue to > have "draining" znodes after a HMaster failover; however, the balancer will > assign regions to them. > h2. How to reproduce (on hbase master): > # Add regionserver to /hbase/draining: {{bin/hbase-jruby > bin/draining_servers.rb add server1:16205}} > # Unload the regionserver: {{bin/hbase-jruby bin/region_mover.rb unload > server1:16205}} > # Kill the Active HMaster and failover to the Backup HMaster > # Run the balancer: {{hbase shell <<< "balancer"}} > # Notice regions get assigned on new Active Master to Region Servers in > /hbase/draining > h2. Root Cause > The Backup HMaster initializes the {{DrainingServerTracker}} before the > Region Servers are registered as "online" with the {{ServerManager}}. As a > result, the {{ServerManager.drainingServers}} isn't populated with existing > Region Servers in draining when we have an HMaster failover. > E.g., > # We have a region server in draining: {{server1,16205,1000}} > # The {{RegionServerTracker}} starts up and adds a ZK watcher on the Znode > for this RegionServer: {{/hbase/rs/server1,16205,1000}} > # The {{DrainingServerTracker}} starts and processes each Znode under > {{/hbase/draining}}, but the Region Server isn't registered as "online" so it > isn't added to the {{ServerManager.drainingServers}} list. > # The Region Server is added to the {{DrainingServerTracker.drainingServers}} > list. > # The Region Server's Znode watcher is triggered and the ZK watcher is > restarted. > # The Region Server is registered with {{ServerManager}} as "online". > *END STATE:* The Region Server has a Znode in {{/hbase/draining}}, but it is > registered as "online" and the Balancer will start assigning regions to it. > {code} > $ bin/hbase-jruby bin/draining_servers.rb list > [1] server1,16205,1000 > $ grep server1,16205,1000 logs/master-server1.log > 2016-10-14 16:02:47,713 DEBUG [server1:16001.activeMasterManager] > zookeeper.ZKUtil: master:16001-0x157c56adc810014, quorum=localhost:2181, > baseZNode=/hbase Set watcher on existing znode=/hbase/rs/server1,16205,1000 > [2] 2016-10-14 16:02:47,722 DEBUG [server1:16001.activeMasterManager] > zookeeper.RegionServerTracker: Added tracking of RS > /hbase/rs/server1,16205,1000 > 2016-10-14 16:02:47,730 DEBUG [server1:16001.activeMasterManager] > zookeeper.ZKUtil: master:16001-0x157c56adc810014, quorum=localhost:2181, > baseZNode=/hbase Set watcher on existing > znode=/hbase/draining/server1,16205,1000 > [3] 2016-10-14 16:02:47,731 WARN [server1:16001.activeMasterManager] > master.ServerManager: Server server1,16205,1000 is not currently online. > Ignoring request to add it to draining list. > [4] 2016-10-14 16:02:47,731 INFO [server1:16001.activeMasterManager] > zookeeper.DrainingServerTracker: Draining RS node created, adding to list > [server1,16205,1000] > 2016-10-14 16:02:47,971 DEBUG [main-EventThread] zookeeper.ZKUtil: > master:16001-0x157c56adc810014, quorum=localhost:2181, baseZNode=/hbase Set > watcher on existing > znode=/hbase/rs/dev6918.prn2.facebook.com,16205,1476486047114 > [5] 2016-10-14 16:02:47,976 DEBUG [main-EventThread] > zookeeper.RegionServerTracker: Added tracking of RS > /hbase/rs/server1,16205,1000 > [6] 2016-10-14 16:02:52,084 INFO > [RpcServer.FifoWFPBQ.default.handler=29,queue=2,port=16001] > master.ServerManager: Registering server=server1,16205,1000 > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-16853) Regions are assigned to Region Servers in /hbase/draining after HBase Master failover
[ https://issues.apache.org/jira/browse/HBASE-16853?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15579860#comment-15579860 ] Hadoop QA commented on HBASE-16853: --- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 16s {color} | {color:blue} Docker mode activated. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s {color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s {color} | {color:green} The patch appears to include 1 new or modified test files. {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 2m 1s {color} | {color:green} branch-1 passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 30s {color} | {color:green} branch-1 passed with JDK v1.8.0_101 {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 34s {color} | {color:green} branch-1 passed with JDK v1.7.0_80 {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 56s {color} | {color:green} branch-1 passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 16s {color} | {color:green} branch-1 passed {color} | | {color:red}-1{color} | {color:red} findbugs {color} | {color:red} 1m 54s {color} | {color:red} hbase-server in branch-1 has 1 extant Findbugs warnings. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 24s {color} | {color:green} branch-1 passed with JDK v1.8.0_101 {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 32s {color} | {color:green} branch-1 passed with JDK v1.7.0_80 {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 46s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 30s {color} | {color:green} the patch passed with JDK v1.8.0_101 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 30s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 33s {color} | {color:green} the patch passed with JDK v1.7.0_80 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 33s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 56s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 16s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s {color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} hadoopcheck {color} | {color:green} 16m 29s {color} | {color:green} The patch does not cause any errors with Hadoop 2.4.0 2.4.1 2.5.0 2.5.1 2.5.2 2.6.1 2.6.2 2.6.3 2.7.1. {color} | | {color:green}+1{color} | {color:green} hbaseprotoc {color} | {color:green} 0m 14s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 2m 7s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 25s {color} | {color:green} the patch passed with JDK v1.8.0_101 {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 33s {color} | {color:green} the patch passed with JDK v1.7.0_80 {color} | | {color:red}-1{color} | {color:red} unit {color} | {color:red} 72m 20s {color} | {color:red} hbase-server in the patch failed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 15s {color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 103m 15s {color} | {color:black} {color} | \\ \\ || Reason || Tests || | Timed out junit tests | org.apache.hadoop.hbase.filter.TestScanRowPrefix | | | org.apache.hadoop.hbase.security.visibility.TestVisibilityLabelsWithDeletes | \\ \\ || Subsystem || Report/Notes || | Docker | Client=1.11.2 Server=1.11.2 Image:yetus/hbase:b2c5d84 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12833637/HBASE-16853.branch-1.3-v1.patch | | JIRA Issue | HBASE-16853 | | Optional Tests | asflicense javac javadoc unit findbugs hadoopcheck hbaseanti checkstyle compile | | uname | Linux abcf67c6de6a 3.13.0-95-generic #142-Ubuntu SMP Fri Aug 12 17:00:09 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | |