[jira] [Commented] (HBASE-16853) Regions are assigned to Region Servers in /hbase/draining after HBase Master failover

2016-11-14 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-16853?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15665193#comment-15665193
 ] 

Hudson commented on HBASE-16853:


ABORTED: Integrated in Jenkins build HBase-0.98-matrix #415 (See 
[https://builds.apache.org/job/HBase-0.98-matrix/415/])
HBASE-16853 Regions are assigned to Region Servers in /hbase/draining 
(apurtell: rev dba43b62823cbaa663cf0c2f7b7e4dcd668bdbce)
* (edit) 
hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestAssignmentListener.java
* (edit) 
hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java
* (edit) 
hbase-server/src/main/java/org/apache/hadoop/hbase/zookeeper/DrainingServerTracker.java


> Regions are assigned to Region Servers in /hbase/draining after HBase Master 
> failover
> -
>
> Key: HBASE-16853
> URL: https://issues.apache.org/jira/browse/HBASE-16853
> Project: HBase
>  Issue Type: Bug
>  Components: Balancer, Region Assignment
>Affects Versions: 2.0.0, 1.3.0
>Reporter: David Pope
>Assignee: David Pope
> Fix For: 2.0.0, 1.3.0, 1.4.0, 0.98.24
>
> Attachments: 16853.v2.txt, HBASE-16853.branch-1.3-v1.patch, 
> HBASE-16853.branch-1.3-v2.patch
>
>
> h2. Problem
> If there are Region Servers registered as "draining", they will continue to 
> have "draining" znodes after a HMaster failover; however, the balancer will 
> assign regions to them.
> h2. How to reproduce (on hbase master):
> # Add regionserver to /hbase/draining: {{bin/hbase-jruby 
> bin/draining_servers.rb add server1:16205}}
> # Unload the regionserver:  {{bin/hbase-jruby bin/region_mover.rb unload 
> server1:16205}}
> # Kill the Active HMaster and failover to the Backup HMaster
> # Run the balancer: {{hbase shell <<< "balancer"}}
> # Notice regions get assigned on new Active Master to Region Servers in 
> /hbase/draining
> h2. Root Cause
> The Backup HMaster initializes the {{DrainingServerTracker}} before the 
> Region Servers are registered as "online" with the {{ServerManager}}.  As a 
> result, the {{ServerManager.drainingServers}} isn't populated with existing 
> Region Servers in draining when we have an HMaster failover.
> E.g., 
> # We have a region server in draining: {{server1,16205,1000}}
> # The {{RegionServerTracker}} starts up and adds a ZK watcher on the Znode 
> for this RegionServer: {{/hbase/rs/server1,16205,1000}}
> # The {{DrainingServerTracker}} starts and processes each Znode under 
> {{/hbase/draining}}, but the Region Server isn't registered as "online" so it 
> isn't added to the {{ServerManager.drainingServers}} list.
> # The Region Server is added to the {{DrainingServerTracker.drainingServers}} 
> list.
> # The Region Server's Znode watcher is triggered and the ZK watcher is 
> restarted.
> # The Region Server is registered with {{ServerManager}} as "online".
> *END STATE:* The Region Server has a Znode in {{/hbase/draining}}, but it is 
> registered as "online" and the Balancer will start assigning regions to it.
> {code}
> $ bin/hbase-jruby bin/draining_servers.rb list
> [1] server1,16205,1000
> $ grep server1,16205,1000 logs/master-server1.log
> 2016-10-14 16:02:47,713 DEBUG [server1:16001.activeMasterManager] 
> zookeeper.ZKUtil: master:16001-0x157c56adc810014, quorum=localhost:2181, 
> baseZNode=/hbase Set watcher on existing znode=/hbase/rs/server1,16205,1000
> [2] 2016-10-14 16:02:47,722 DEBUG [server1:16001.activeMasterManager] 
> zookeeper.RegionServerTracker: Added tracking of RS 
> /hbase/rs/server1,16205,1000
> 2016-10-14 16:02:47,730 DEBUG [server1:16001.activeMasterManager] 
> zookeeper.ZKUtil: master:16001-0x157c56adc810014, quorum=localhost:2181, 
> baseZNode=/hbase Set watcher on existing 
> znode=/hbase/draining/server1,16205,1000
> [3] 2016-10-14 16:02:47,731 WARN  [server1:16001.activeMasterManager] 
> master.ServerManager: Server server1,16205,1000 is not currently online. 
> Ignoring request to add it to draining list.
> [4] 2016-10-14 16:02:47,731 INFO  [server1:16001.activeMasterManager] 
> zookeeper.DrainingServerTracker: Draining RS node created, adding to list 
> [server1,16205,1000]
> 2016-10-14 16:02:47,971 DEBUG [main-EventThread] zookeeper.ZKUtil: 
> master:16001-0x157c56adc810014, quorum=localhost:2181, baseZNode=/hbase Set 
> watcher on existing 
> znode=/hbase/rs/dev6918.prn2.facebook.com,16205,1476486047114
> [5] 2016-10-14 16:02:47,976 DEBUG [main-EventThread] 
> zookeeper.RegionServerTracker: Added tracking of RS 
> /hbase/rs/server1,16205,1000
> [6] 2016-10-14 16:02:52,084 INFO  
> [RpcServer.FifoWFPBQ.default.handler=29,queue=2,port=16001] 
> master.ServerManager: Registering server=server1,16205,1000
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HBASE-16853) Regions are assigned to Region Servers in /hbase/draining after HBase Master failover

2016-11-12 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-16853?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15659967#comment-15659967
 ] 

Hudson commented on HBASE-16853:


FAILURE: Integrated in Jenkins build HBase-0.98-on-Hadoop-1.1 #1288 (See 
[https://builds.apache.org/job/HBase-0.98-on-Hadoop-1.1/1288/])
HBASE-16853 Regions are assigned to Region Servers in /hbase/draining 
(apurtell: rev dba43b62823cbaa663cf0c2f7b7e4dcd668bdbce)
* (edit) 
hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestAssignmentListener.java
* (edit) 
hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java
* (edit) 
hbase-server/src/main/java/org/apache/hadoop/hbase/zookeeper/DrainingServerTracker.java


> Regions are assigned to Region Servers in /hbase/draining after HBase Master 
> failover
> -
>
> Key: HBASE-16853
> URL: https://issues.apache.org/jira/browse/HBASE-16853
> Project: HBase
>  Issue Type: Bug
>  Components: Balancer, Region Assignment
>Affects Versions: 2.0.0, 1.3.0
>Reporter: David Pope
>Assignee: David Pope
> Fix For: 2.0.0, 1.3.0, 1.4.0, 0.98.24
>
> Attachments: 16853.v2.txt, HBASE-16853.branch-1.3-v1.patch, 
> HBASE-16853.branch-1.3-v2.patch
>
>
> h2. Problem
> If there are Region Servers registered as "draining", they will continue to 
> have "draining" znodes after a HMaster failover; however, the balancer will 
> assign regions to them.
> h2. How to reproduce (on hbase master):
> # Add regionserver to /hbase/draining: {{bin/hbase-jruby 
> bin/draining_servers.rb add server1:16205}}
> # Unload the regionserver:  {{bin/hbase-jruby bin/region_mover.rb unload 
> server1:16205}}
> # Kill the Active HMaster and failover to the Backup HMaster
> # Run the balancer: {{hbase shell <<< "balancer"}}
> # Notice regions get assigned on new Active Master to Region Servers in 
> /hbase/draining
> h2. Root Cause
> The Backup HMaster initializes the {{DrainingServerTracker}} before the 
> Region Servers are registered as "online" with the {{ServerManager}}.  As a 
> result, the {{ServerManager.drainingServers}} isn't populated with existing 
> Region Servers in draining when we have an HMaster failover.
> E.g., 
> # We have a region server in draining: {{server1,16205,1000}}
> # The {{RegionServerTracker}} starts up and adds a ZK watcher on the Znode 
> for this RegionServer: {{/hbase/rs/server1,16205,1000}}
> # The {{DrainingServerTracker}} starts and processes each Znode under 
> {{/hbase/draining}}, but the Region Server isn't registered as "online" so it 
> isn't added to the {{ServerManager.drainingServers}} list.
> # The Region Server is added to the {{DrainingServerTracker.drainingServers}} 
> list.
> # The Region Server's Znode watcher is triggered and the ZK watcher is 
> restarted.
> # The Region Server is registered with {{ServerManager}} as "online".
> *END STATE:* The Region Server has a Znode in {{/hbase/draining}}, but it is 
> registered as "online" and the Balancer will start assigning regions to it.
> {code}
> $ bin/hbase-jruby bin/draining_servers.rb list
> [1] server1,16205,1000
> $ grep server1,16205,1000 logs/master-server1.log
> 2016-10-14 16:02:47,713 DEBUG [server1:16001.activeMasterManager] 
> zookeeper.ZKUtil: master:16001-0x157c56adc810014, quorum=localhost:2181, 
> baseZNode=/hbase Set watcher on existing znode=/hbase/rs/server1,16205,1000
> [2] 2016-10-14 16:02:47,722 DEBUG [server1:16001.activeMasterManager] 
> zookeeper.RegionServerTracker: Added tracking of RS 
> /hbase/rs/server1,16205,1000
> 2016-10-14 16:02:47,730 DEBUG [server1:16001.activeMasterManager] 
> zookeeper.ZKUtil: master:16001-0x157c56adc810014, quorum=localhost:2181, 
> baseZNode=/hbase Set watcher on existing 
> znode=/hbase/draining/server1,16205,1000
> [3] 2016-10-14 16:02:47,731 WARN  [server1:16001.activeMasterManager] 
> master.ServerManager: Server server1,16205,1000 is not currently online. 
> Ignoring request to add it to draining list.
> [4] 2016-10-14 16:02:47,731 INFO  [server1:16001.activeMasterManager] 
> zookeeper.DrainingServerTracker: Draining RS node created, adding to list 
> [server1,16205,1000]
> 2016-10-14 16:02:47,971 DEBUG [main-EventThread] zookeeper.ZKUtil: 
> master:16001-0x157c56adc810014, quorum=localhost:2181, baseZNode=/hbase Set 
> watcher on existing 
> znode=/hbase/rs/dev6918.prn2.facebook.com,16205,1476486047114
> [5] 2016-10-14 16:02:47,976 DEBUG [main-EventThread] 
> zookeeper.RegionServerTracker: Added tracking of RS 
> /hbase/rs/server1,16205,1000
> [6] 2016-10-14 16:02:52,084 INFO  
> [RpcServer.FifoWFPBQ.default.handler=29,queue=2,port=16001] 
> master.ServerManager: Registering server=server1,16205,1000
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HBASE-16853) Regions are assigned to Region Servers in /hbase/draining after HBase Master failover

2016-10-17 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-16853?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15581703#comment-15581703
 ] 

Hudson commented on HBASE-16853:


SUCCESS: Integrated in Jenkins build HBase-Trunk_matrix #1801 (See 
[https://builds.apache.org/job/HBase-Trunk_matrix/1801/])
HBASE-16853 Regions are assigned to Region Servers in /hbase/draining (tedyu: 
rev 109db38b6ad091b23593ee46b1e919136aed7886)
* (edit) 
hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestAssignmentListener.java
* (edit) 
hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java
* (edit) 
hbase-server/src/main/java/org/apache/hadoop/hbase/zookeeper/DrainingServerTracker.java


> Regions are assigned to Region Servers in /hbase/draining after HBase Master 
> failover
> -
>
> Key: HBASE-16853
> URL: https://issues.apache.org/jira/browse/HBASE-16853
> Project: HBase
>  Issue Type: Bug
>  Components: Balancer, Region Assignment
>Affects Versions: 2.0.0, 1.3.0
>Reporter: David Pope
>Assignee: David Pope
> Fix For: 2.0.0, 1.3.0, 1.4.0
>
> Attachments: 16853.v2.txt, HBASE-16853.branch-1.3-v1.patch, 
> HBASE-16853.branch-1.3-v2.patch
>
>
> h2. Problem
> If there are Region Servers registered as "draining", they will continue to 
> have "draining" znodes after a HMaster failover; however, the balancer will 
> assign regions to them.
> h2. How to reproduce (on hbase master):
> # Add regionserver to /hbase/draining: {{bin/hbase-jruby 
> bin/draining_servers.rb add server1:16205}}
> # Unload the regionserver:  {{bin/hbase-jruby bin/region_mover.rb unload 
> server1:16205}}
> # Kill the Active HMaster and failover to the Backup HMaster
> # Run the balancer: {{hbase shell <<< "balancer"}}
> # Notice regions get assigned on new Active Master to Region Servers in 
> /hbase/draining
> h2. Root Cause
> The Backup HMaster initializes the {{DrainingServerTracker}} before the 
> Region Servers are registered as "online" with the {{ServerManager}}.  As a 
> result, the {{ServerManager.drainingServers}} isn't populated with existing 
> Region Servers in draining when we have an HMaster failover.
> E.g., 
> # We have a region server in draining: {{server1,16205,1000}}
> # The {{RegionServerTracker}} starts up and adds a ZK watcher on the Znode 
> for this RegionServer: {{/hbase/rs/server1,16205,1000}}
> # The {{DrainingServerTracker}} starts and processes each Znode under 
> {{/hbase/draining}}, but the Region Server isn't registered as "online" so it 
> isn't added to the {{ServerManager.drainingServers}} list.
> # The Region Server is added to the {{DrainingServerTracker.drainingServers}} 
> list.
> # The Region Server's Znode watcher is triggered and the ZK watcher is 
> restarted.
> # The Region Server is registered with {{ServerManager}} as "online".
> *END STATE:* The Region Server has a Znode in {{/hbase/draining}}, but it is 
> registered as "online" and the Balancer will start assigning regions to it.
> {code}
> $ bin/hbase-jruby bin/draining_servers.rb list
> [1] server1,16205,1000
> $ grep server1,16205,1000 logs/master-server1.log
> 2016-10-14 16:02:47,713 DEBUG [server1:16001.activeMasterManager] 
> zookeeper.ZKUtil: master:16001-0x157c56adc810014, quorum=localhost:2181, 
> baseZNode=/hbase Set watcher on existing znode=/hbase/rs/server1,16205,1000
> [2] 2016-10-14 16:02:47,722 DEBUG [server1:16001.activeMasterManager] 
> zookeeper.RegionServerTracker: Added tracking of RS 
> /hbase/rs/server1,16205,1000
> 2016-10-14 16:02:47,730 DEBUG [server1:16001.activeMasterManager] 
> zookeeper.ZKUtil: master:16001-0x157c56adc810014, quorum=localhost:2181, 
> baseZNode=/hbase Set watcher on existing 
> znode=/hbase/draining/server1,16205,1000
> [3] 2016-10-14 16:02:47,731 WARN  [server1:16001.activeMasterManager] 
> master.ServerManager: Server server1,16205,1000 is not currently online. 
> Ignoring request to add it to draining list.
> [4] 2016-10-14 16:02:47,731 INFO  [server1:16001.activeMasterManager] 
> zookeeper.DrainingServerTracker: Draining RS node created, adding to list 
> [server1,16205,1000]
> 2016-10-14 16:02:47,971 DEBUG [main-EventThread] zookeeper.ZKUtil: 
> master:16001-0x157c56adc810014, quorum=localhost:2181, baseZNode=/hbase Set 
> watcher on existing 
> znode=/hbase/rs/dev6918.prn2.facebook.com,16205,1476486047114
> [5] 2016-10-14 16:02:47,976 DEBUG [main-EventThread] 
> zookeeper.RegionServerTracker: Added tracking of RS 
> /hbase/rs/server1,16205,1000
> [6] 2016-10-14 16:02:52,084 INFO  
> [RpcServer.FifoWFPBQ.default.handler=29,queue=2,port=16001] 
> master.ServerManager: Registering server=server1,16205,1000
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HBASE-16853) Regions are assigned to Region Servers in /hbase/draining after HBase Master failover

2016-10-17 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-16853?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15581642#comment-15581642
 ] 

Hudson commented on HBASE-16853:


SUCCESS: Integrated in Jenkins build HBase-1.3-JDK7 #44 (See 
[https://builds.apache.org/job/HBase-1.3-JDK7/44/])
HBASE-16853 Regions are assigned to Region Servers in /hbase/draining (tedyu: 
rev d38310aa4d97c0be0e2d1f5113630510a02f7ade)
* (edit) 
hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestAssignmentListener.java
* (edit) 
hbase-server/src/main/java/org/apache/hadoop/hbase/zookeeper/DrainingServerTracker.java
* (edit) 
hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java


> Regions are assigned to Region Servers in /hbase/draining after HBase Master 
> failover
> -
>
> Key: HBASE-16853
> URL: https://issues.apache.org/jira/browse/HBASE-16853
> Project: HBase
>  Issue Type: Bug
>  Components: Balancer, Region Assignment
>Affects Versions: 2.0.0, 1.3.0
>Reporter: David Pope
>Assignee: David Pope
> Fix For: 2.0.0, 1.3.0, 1.4.0
>
> Attachments: 16853.v2.txt, HBASE-16853.branch-1.3-v1.patch, 
> HBASE-16853.branch-1.3-v2.patch
>
>
> h2. Problem
> If there are Region Servers registered as "draining", they will continue to 
> have "draining" znodes after a HMaster failover; however, the balancer will 
> assign regions to them.
> h2. How to reproduce (on hbase master):
> # Add regionserver to /hbase/draining: {{bin/hbase-jruby 
> bin/draining_servers.rb add server1:16205}}
> # Unload the regionserver:  {{bin/hbase-jruby bin/region_mover.rb unload 
> server1:16205}}
> # Kill the Active HMaster and failover to the Backup HMaster
> # Run the balancer: {{hbase shell <<< "balancer"}}
> # Notice regions get assigned on new Active Master to Region Servers in 
> /hbase/draining
> h2. Root Cause
> The Backup HMaster initializes the {{DrainingServerTracker}} before the 
> Region Servers are registered as "online" with the {{ServerManager}}.  As a 
> result, the {{ServerManager.drainingServers}} isn't populated with existing 
> Region Servers in draining when we have an HMaster failover.
> E.g., 
> # We have a region server in draining: {{server1,16205,1000}}
> # The {{RegionServerTracker}} starts up and adds a ZK watcher on the Znode 
> for this RegionServer: {{/hbase/rs/server1,16205,1000}}
> # The {{DrainingServerTracker}} starts and processes each Znode under 
> {{/hbase/draining}}, but the Region Server isn't registered as "online" so it 
> isn't added to the {{ServerManager.drainingServers}} list.
> # The Region Server is added to the {{DrainingServerTracker.drainingServers}} 
> list.
> # The Region Server's Znode watcher is triggered and the ZK watcher is 
> restarted.
> # The Region Server is registered with {{ServerManager}} as "online".
> *END STATE:* The Region Server has a Znode in {{/hbase/draining}}, but it is 
> registered as "online" and the Balancer will start assigning regions to it.
> {code}
> $ bin/hbase-jruby bin/draining_servers.rb list
> [1] server1,16205,1000
> $ grep server1,16205,1000 logs/master-server1.log
> 2016-10-14 16:02:47,713 DEBUG [server1:16001.activeMasterManager] 
> zookeeper.ZKUtil: master:16001-0x157c56adc810014, quorum=localhost:2181, 
> baseZNode=/hbase Set watcher on existing znode=/hbase/rs/server1,16205,1000
> [2] 2016-10-14 16:02:47,722 DEBUG [server1:16001.activeMasterManager] 
> zookeeper.RegionServerTracker: Added tracking of RS 
> /hbase/rs/server1,16205,1000
> 2016-10-14 16:02:47,730 DEBUG [server1:16001.activeMasterManager] 
> zookeeper.ZKUtil: master:16001-0x157c56adc810014, quorum=localhost:2181, 
> baseZNode=/hbase Set watcher on existing 
> znode=/hbase/draining/server1,16205,1000
> [3] 2016-10-14 16:02:47,731 WARN  [server1:16001.activeMasterManager] 
> master.ServerManager: Server server1,16205,1000 is not currently online. 
> Ignoring request to add it to draining list.
> [4] 2016-10-14 16:02:47,731 INFO  [server1:16001.activeMasterManager] 
> zookeeper.DrainingServerTracker: Draining RS node created, adding to list 
> [server1,16205,1000]
> 2016-10-14 16:02:47,971 DEBUG [main-EventThread] zookeeper.ZKUtil: 
> master:16001-0x157c56adc810014, quorum=localhost:2181, baseZNode=/hbase Set 
> watcher on existing 
> znode=/hbase/rs/dev6918.prn2.facebook.com,16205,1476486047114
> [5] 2016-10-14 16:02:47,976 DEBUG [main-EventThread] 
> zookeeper.RegionServerTracker: Added tracking of RS 
> /hbase/rs/server1,16205,1000
> [6] 2016-10-14 16:02:52,084 INFO  
> [RpcServer.FifoWFPBQ.default.handler=29,queue=2,port=16001] 
> master.ServerManager: Registering server=server1,16205,1000
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HBASE-16853) Regions are assigned to Region Servers in /hbase/draining after HBase Master failover

2016-10-16 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-16853?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15580993#comment-15580993
 ] 

Hadoop QA commented on HBASE-16853:
---

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 13s 
{color} | {color:blue} Docker mode activated. {color} |
| {color:blue}0{color} | {color:blue} patch {color} | {color:blue} 0m 3s 
{color} | {color:blue} The patch file was not named according to hbase's naming 
conventions. Please see 
https://yetus.apache.org/documentation/0.3.0/precommit-patchnames for 
instructions. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s 
{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 
0s {color} | {color:green} The patch appears to include 1 new or modified test 
files. {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 2m 
57s {color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 33s 
{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 
44s {color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 
13s {color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 
39s {color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 25s 
{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 
44s {color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 34s 
{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 34s 
{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 
45s {color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 
13s {color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 
0s {color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} hadoopcheck {color} | {color:green} 
26m 34s {color} | {color:green} Patch does not cause any errors with Hadoop 
2.4.0 2.4.1 2.5.0 2.5.1 2.5.2 2.6.1 2.6.2 2.6.3 2.7.1. {color} |
| {color:green}+1{color} | {color:green} hbaseprotoc {color} | {color:green} 0m 
12s {color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 
47s {color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 25s 
{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red} 79m 12s {color} 
| {color:red} hbase-server in the patch failed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 
14s {color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 117m 41s {color} 
| {color:black} {color} |
\\
\\
|| Reason || Tests ||
| Timed out junit tests | 
org.apache.hadoop.hbase.security.visibility.TestVisibilityLabelsWithDeletes |
|   | org.apache.hadoop.hbase.filter.TestMultiRowRangeFilter |
\\
\\
|| Subsystem || Report/Notes ||
| Docker | Client=1.11.2 Server=1.11.2 Image:yetus/hbase:7bda515 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12833663/16853.v2.txt |
| JIRA Issue | HBASE-16853 |
| Optional Tests |  asflicense  javac  javadoc  unit  findbugs  hadoopcheck  
hbaseanti  checkstyle  compile  |
| uname | Linux dbd83b6f41e1 3.13.0-95-generic #142-Ubuntu SMP Fri Aug 12 
17:00:09 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/home/jenkins/jenkins-slave/workspace/PreCommit-HBASE-Build/component/dev-support/hbase-personality.sh
 |
| git revision | master / 0f38415 |
| Default Java | 1.8.0_101 |
| findbugs | v3.0.0 |
| unit | 
https://builds.apache.org/job/PreCommit-HBASE-Build/4042/artifact/patchprocess/patch-unit-hbase-server.txt
 |
| unit test logs |  
https://builds.apache.org/job/PreCommit-HBASE-Build/4042/artifact/patchprocess/patch-unit-hbase-server.txt
 |
|  Test Results | 
https://builds.apache.org/job/PreCommit-HBASE-Build/4042/testReport/ |
| modules | C: hbase-server U: hbase-server |
| Console 

[jira] [Commented] (HBASE-16853) Regions are assigned to Region Servers in /hbase/draining after HBase Master failover

2016-10-16 Thread Heng Chen (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-16853?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15580931#comment-15580931
 ] 

Heng Chen commented on HBASE-16853:
---

+1

> Regions are assigned to Region Servers in /hbase/draining after HBase Master 
> failover
> -
>
> Key: HBASE-16853
> URL: https://issues.apache.org/jira/browse/HBASE-16853
> Project: HBase
>  Issue Type: Bug
>  Components: Balancer, Region Assignment
>Affects Versions: 2.0.0, 1.3.0
>Reporter: David Pope
>Assignee: David Pope
> Fix For: 2.0.0, 1.3.0, 1.4.0
>
> Attachments: 16853.v2.txt, HBASE-16853.branch-1.3-v1.patch, 
> HBASE-16853.branch-1.3-v2.patch
>
>
> h2. Problem
> If there are Region Servers registered as "draining", they will continue to 
> have "draining" znodes after a HMaster failover; however, the balancer will 
> assign regions to them.
> h2. How to reproduce (on hbase master):
> # Add regionserver to /hbase/draining: {{bin/hbase-jruby 
> bin/draining_servers.rb add server1:16205}}
> # Unload the regionserver:  {{bin/hbase-jruby bin/region_mover.rb unload 
> server1:16205}}
> # Kill the Active HMaster and failover to the Backup HMaster
> # Run the balancer: {{hbase shell <<< "balancer"}}
> # Notice regions get assigned on new Active Master to Region Servers in 
> /hbase/draining
> h2. Root Cause
> The Backup HMaster initializes the {{DrainingServerTracker}} before the 
> Region Servers are registered as "online" with the {{ServerManager}}.  As a 
> result, the {{ServerManager.drainingServers}} isn't populated with existing 
> Region Servers in draining when we have an HMaster failover.
> E.g., 
> # We have a region server in draining: {{server1,16205,1000}}
> # The {{RegionServerTracker}} starts up and adds a ZK watcher on the Znode 
> for this RegionServer: {{/hbase/rs/server1,16205,1000}}
> # The {{DrainingServerTracker}} starts and processes each Znode under 
> {{/hbase/draining}}, but the Region Server isn't registered as "online" so it 
> isn't added to the {{ServerManager.drainingServers}} list.
> # The Region Server is added to the {{DrainingServerTracker.drainingServers}} 
> list.
> # The Region Server's Znode watcher is triggered and the ZK watcher is 
> restarted.
> # The Region Server is registered with {{ServerManager}} as "online".
> *END STATE:* The Region Server has a Znode in {{/hbase/draining}}, but it is 
> registered as "online" and the Balancer will start assigning regions to it.
> {code}
> $ bin/hbase-jruby bin/draining_servers.rb list
> [1] server1,16205,1000
> $ grep server1,16205,1000 logs/master-server1.log
> 2016-10-14 16:02:47,713 DEBUG [server1:16001.activeMasterManager] 
> zookeeper.ZKUtil: master:16001-0x157c56adc810014, quorum=localhost:2181, 
> baseZNode=/hbase Set watcher on existing znode=/hbase/rs/server1,16205,1000
> [2] 2016-10-14 16:02:47,722 DEBUG [server1:16001.activeMasterManager] 
> zookeeper.RegionServerTracker: Added tracking of RS 
> /hbase/rs/server1,16205,1000
> 2016-10-14 16:02:47,730 DEBUG [server1:16001.activeMasterManager] 
> zookeeper.ZKUtil: master:16001-0x157c56adc810014, quorum=localhost:2181, 
> baseZNode=/hbase Set watcher on existing 
> znode=/hbase/draining/server1,16205,1000
> [3] 2016-10-14 16:02:47,731 WARN  [server1:16001.activeMasterManager] 
> master.ServerManager: Server server1,16205,1000 is not currently online. 
> Ignoring request to add it to draining list.
> [4] 2016-10-14 16:02:47,731 INFO  [server1:16001.activeMasterManager] 
> zookeeper.DrainingServerTracker: Draining RS node created, adding to list 
> [server1,16205,1000]
> 2016-10-14 16:02:47,971 DEBUG [main-EventThread] zookeeper.ZKUtil: 
> master:16001-0x157c56adc810014, quorum=localhost:2181, baseZNode=/hbase Set 
> watcher on existing 
> znode=/hbase/rs/dev6918.prn2.facebook.com,16205,1476486047114
> [5] 2016-10-14 16:02:47,976 DEBUG [main-EventThread] 
> zookeeper.RegionServerTracker: Added tracking of RS 
> /hbase/rs/server1,16205,1000
> [6] 2016-10-14 16:02:52,084 INFO  
> [RpcServer.FifoWFPBQ.default.handler=29,queue=2,port=16001] 
> master.ServerManager: Registering server=server1,16205,1000
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HBASE-16853) Regions are assigned to Region Servers in /hbase/draining after HBase Master failover

2016-10-16 Thread Ted Yu (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-16853?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15580933#comment-15580933
 ] 

Ted Yu commented on HBASE-16853:


Integrated to branch-1.3 and branch-1

Waiting for QA on master patch.

> Regions are assigned to Region Servers in /hbase/draining after HBase Master 
> failover
> -
>
> Key: HBASE-16853
> URL: https://issues.apache.org/jira/browse/HBASE-16853
> Project: HBase
>  Issue Type: Bug
>  Components: Balancer, Region Assignment
>Affects Versions: 2.0.0, 1.3.0
>Reporter: David Pope
>Assignee: David Pope
> Fix For: 2.0.0, 1.3.0, 1.4.0
>
> Attachments: 16853.v2.txt, HBASE-16853.branch-1.3-v1.patch, 
> HBASE-16853.branch-1.3-v2.patch
>
>
> h2. Problem
> If there are Region Servers registered as "draining", they will continue to 
> have "draining" znodes after a HMaster failover; however, the balancer will 
> assign regions to them.
> h2. How to reproduce (on hbase master):
> # Add regionserver to /hbase/draining: {{bin/hbase-jruby 
> bin/draining_servers.rb add server1:16205}}
> # Unload the regionserver:  {{bin/hbase-jruby bin/region_mover.rb unload 
> server1:16205}}
> # Kill the Active HMaster and failover to the Backup HMaster
> # Run the balancer: {{hbase shell <<< "balancer"}}
> # Notice regions get assigned on new Active Master to Region Servers in 
> /hbase/draining
> h2. Root Cause
> The Backup HMaster initializes the {{DrainingServerTracker}} before the 
> Region Servers are registered as "online" with the {{ServerManager}}.  As a 
> result, the {{ServerManager.drainingServers}} isn't populated with existing 
> Region Servers in draining when we have an HMaster failover.
> E.g., 
> # We have a region server in draining: {{server1,16205,1000}}
> # The {{RegionServerTracker}} starts up and adds a ZK watcher on the Znode 
> for this RegionServer: {{/hbase/rs/server1,16205,1000}}
> # The {{DrainingServerTracker}} starts and processes each Znode under 
> {{/hbase/draining}}, but the Region Server isn't registered as "online" so it 
> isn't added to the {{ServerManager.drainingServers}} list.
> # The Region Server is added to the {{DrainingServerTracker.drainingServers}} 
> list.
> # The Region Server's Znode watcher is triggered and the ZK watcher is 
> restarted.
> # The Region Server is registered with {{ServerManager}} as "online".
> *END STATE:* The Region Server has a Znode in {{/hbase/draining}}, but it is 
> registered as "online" and the Balancer will start assigning regions to it.
> {code}
> $ bin/hbase-jruby bin/draining_servers.rb list
> [1] server1,16205,1000
> $ grep server1,16205,1000 logs/master-server1.log
> 2016-10-14 16:02:47,713 DEBUG [server1:16001.activeMasterManager] 
> zookeeper.ZKUtil: master:16001-0x157c56adc810014, quorum=localhost:2181, 
> baseZNode=/hbase Set watcher on existing znode=/hbase/rs/server1,16205,1000
> [2] 2016-10-14 16:02:47,722 DEBUG [server1:16001.activeMasterManager] 
> zookeeper.RegionServerTracker: Added tracking of RS 
> /hbase/rs/server1,16205,1000
> 2016-10-14 16:02:47,730 DEBUG [server1:16001.activeMasterManager] 
> zookeeper.ZKUtil: master:16001-0x157c56adc810014, quorum=localhost:2181, 
> baseZNode=/hbase Set watcher on existing 
> znode=/hbase/draining/server1,16205,1000
> [3] 2016-10-14 16:02:47,731 WARN  [server1:16001.activeMasterManager] 
> master.ServerManager: Server server1,16205,1000 is not currently online. 
> Ignoring request to add it to draining list.
> [4] 2016-10-14 16:02:47,731 INFO  [server1:16001.activeMasterManager] 
> zookeeper.DrainingServerTracker: Draining RS node created, adding to list 
> [server1,16205,1000]
> 2016-10-14 16:02:47,971 DEBUG [main-EventThread] zookeeper.ZKUtil: 
> master:16001-0x157c56adc810014, quorum=localhost:2181, baseZNode=/hbase Set 
> watcher on existing 
> znode=/hbase/rs/dev6918.prn2.facebook.com,16205,1476486047114
> [5] 2016-10-14 16:02:47,976 DEBUG [main-EventThread] 
> zookeeper.RegionServerTracker: Added tracking of RS 
> /hbase/rs/server1,16205,1000
> [6] 2016-10-14 16:02:52,084 INFO  
> [RpcServer.FifoWFPBQ.default.handler=29,queue=2,port=16001] 
> master.ServerManager: Registering server=server1,16205,1000
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HBASE-16853) Regions are assigned to Region Servers in /hbase/draining after HBase Master failover

2016-10-16 Thread David Pope (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-16853?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15580883#comment-15580883
 ] 

David Pope commented on HBASE-16853:


Thanks, Ted.  That's a patch for master, correct?  Do you want any others, such 
as branch-1?

> Regions are assigned to Region Servers in /hbase/draining after HBase Master 
> failover
> -
>
> Key: HBASE-16853
> URL: https://issues.apache.org/jira/browse/HBASE-16853
> Project: HBase
>  Issue Type: Bug
>  Components: Balancer, Region Assignment
>Affects Versions: 2.0.0, 1.3.0
>Reporter: David Pope
>Assignee: David Pope
> Fix For: 2.0.0, 1.3.0, 1.4.0
>
> Attachments: 16853.v2.txt, HBASE-16853.branch-1.3-v1.patch, 
> HBASE-16853.branch-1.3-v2.patch
>
>
> h2. Problem
> If there are Region Servers registered as "draining", they will continue to 
> have "draining" znodes after a HMaster failover; however, the balancer will 
> assign regions to them.
> h2. How to reproduce (on hbase master):
> # Add regionserver to /hbase/draining: {{bin/hbase-jruby 
> bin/draining_servers.rb add server1:16205}}
> # Unload the regionserver:  {{bin/hbase-jruby bin/region_mover.rb unload 
> server1:16205}}
> # Kill the Active HMaster and failover to the Backup HMaster
> # Run the balancer: {{hbase shell <<< "balancer"}}
> # Notice regions get assigned on new Active Master to Region Servers in 
> /hbase/draining
> h2. Root Cause
> The Backup HMaster initializes the {{DrainingServerTracker}} before the 
> Region Servers are registered as "online" with the {{ServerManager}}.  As a 
> result, the {{ServerManager.drainingServers}} isn't populated with existing 
> Region Servers in draining when we have an HMaster failover.
> E.g., 
> # We have a region server in draining: {{server1,16205,1000}}
> # The {{RegionServerTracker}} starts up and adds a ZK watcher on the Znode 
> for this RegionServer: {{/hbase/rs/server1,16205,1000}}
> # The {{DrainingServerTracker}} starts and processes each Znode under 
> {{/hbase/draining}}, but the Region Server isn't registered as "online" so it 
> isn't added to the {{ServerManager.drainingServers}} list.
> # The Region Server is added to the {{DrainingServerTracker.drainingServers}} 
> list.
> # The Region Server's Znode watcher is triggered and the ZK watcher is 
> restarted.
> # The Region Server is registered with {{ServerManager}} as "online".
> *END STATE:* The Region Server has a Znode in {{/hbase/draining}}, but it is 
> registered as "online" and the Balancer will start assigning regions to it.
> {code}
> $ bin/hbase-jruby bin/draining_servers.rb list
> [1] server1,16205,1000
> $ grep server1,16205,1000 logs/master-server1.log
> 2016-10-14 16:02:47,713 DEBUG [server1:16001.activeMasterManager] 
> zookeeper.ZKUtil: master:16001-0x157c56adc810014, quorum=localhost:2181, 
> baseZNode=/hbase Set watcher on existing znode=/hbase/rs/server1,16205,1000
> [2] 2016-10-14 16:02:47,722 DEBUG [server1:16001.activeMasterManager] 
> zookeeper.RegionServerTracker: Added tracking of RS 
> /hbase/rs/server1,16205,1000
> 2016-10-14 16:02:47,730 DEBUG [server1:16001.activeMasterManager] 
> zookeeper.ZKUtil: master:16001-0x157c56adc810014, quorum=localhost:2181, 
> baseZNode=/hbase Set watcher on existing 
> znode=/hbase/draining/server1,16205,1000
> [3] 2016-10-14 16:02:47,731 WARN  [server1:16001.activeMasterManager] 
> master.ServerManager: Server server1,16205,1000 is not currently online. 
> Ignoring request to add it to draining list.
> [4] 2016-10-14 16:02:47,731 INFO  [server1:16001.activeMasterManager] 
> zookeeper.DrainingServerTracker: Draining RS node created, adding to list 
> [server1,16205,1000]
> 2016-10-14 16:02:47,971 DEBUG [main-EventThread] zookeeper.ZKUtil: 
> master:16001-0x157c56adc810014, quorum=localhost:2181, baseZNode=/hbase Set 
> watcher on existing 
> znode=/hbase/rs/dev6918.prn2.facebook.com,16205,1476486047114
> [5] 2016-10-14 16:02:47,976 DEBUG [main-EventThread] 
> zookeeper.RegionServerTracker: Added tracking of RS 
> /hbase/rs/server1,16205,1000
> [6] 2016-10-14 16:02:52,084 INFO  
> [RpcServer.FifoWFPBQ.default.handler=29,queue=2,port=16001] 
> master.ServerManager: Registering server=server1,16205,1000
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HBASE-16853) Regions are assigned to Region Servers in /hbase/draining after HBase Master failover

2016-10-16 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-16853?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15580709#comment-15580709
 ] 

Hadoop QA commented on HBASE-16853:
---

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 16s 
{color} | {color:blue} Docker mode activated. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s 
{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 
0s {color} | {color:green} The patch appears to include 1 new or modified test 
files. {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 1m 
49s {color} | {color:green} branch-1 passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 30s 
{color} | {color:green} branch-1 passed with JDK v1.8.0_101 {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 35s 
{color} | {color:green} branch-1 passed with JDK v1.7.0_80 {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 
55s {color} | {color:green} branch-1 passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 
16s {color} | {color:green} branch-1 passed {color} |
| {color:red}-1{color} | {color:red} findbugs {color} | {color:red} 1m 53s 
{color} | {color:red} hbase-server in branch-1 has 1 extant Findbugs warnings. 
{color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 25s 
{color} | {color:green} branch-1 passed with JDK v1.8.0_101 {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 33s 
{color} | {color:green} branch-1 passed with JDK v1.7.0_80 {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 
45s {color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 30s 
{color} | {color:green} the patch passed with JDK v1.8.0_101 {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 30s 
{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 33s 
{color} | {color:green} the patch passed with JDK v1.7.0_80 {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 33s 
{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 
56s {color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 
16s {color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 
0s {color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} hadoopcheck {color} | {color:green} 
16m 24s {color} | {color:green} The patch does not cause any errors with Hadoop 
2.4.0 2.4.1 2.5.0 2.5.1 2.5.2 2.6.1 2.6.2 2.6.3 2.7.1. {color} |
| {color:green}+1{color} | {color:green} hbaseprotoc {color} | {color:green} 0m 
15s {color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 2m 5s 
{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 24s 
{color} | {color:green} the patch passed with JDK v1.8.0_101 {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 33s 
{color} | {color:green} the patch passed with JDK v1.7.0_80 {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red} 72m 36s {color} 
| {color:red} hbase-server in the patch failed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 
15s {color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 103m 12s {color} 
| {color:black} {color} |
\\
\\
|| Reason || Tests ||
| Failed junit tests | hadoop.hbase.mapreduce.TestMultiTableSnapshotInputFormat 
|
| Timed out junit tests | org.apache.hadoop.hbase.TestHBaseTestingUtility |
|   | org.apache.hadoop.hbase.regionserver.TestClusterId |
|   | org.apache.hadoop.hbase.quotas.TestQuotaAdmin |
|   | org.apache.hadoop.hbase.filter.TestFuzzyRowFilterEndToEnd |
\\
\\
|| Subsystem || Report/Notes ||
| Docker | Client=1.11.2 Server=1.11.2 Image:yetus/hbase:b2c5d84 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12833648/HBASE-16853.branch-1.3-v2.patch
 |
| JIRA Issue | HBASE-16853 |
| Optional Tests |  asflicense  javac  javadoc  unit  findbugs  hadoopcheck  
hbaseanti  

[jira] [Commented] (HBASE-16853) Regions are assigned to Region Servers in /hbase/draining after HBase Master failover

2016-10-16 Thread Ted Yu (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-16853?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15580034#comment-15580034
 ] 

Ted Yu commented on HBASE-16853:


lgtm

Can you add the new test to TestAssignmentListener (which uses ServerListener) ?

Please attach patch for master.

> Regions are assigned to Region Servers in /hbase/draining after HBase Master 
> failover
> -
>
> Key: HBASE-16853
> URL: https://issues.apache.org/jira/browse/HBASE-16853
> Project: HBase
>  Issue Type: Bug
>  Components: Balancer, Region Assignment
>Affects Versions: 2.0.0, 1.3.0
>Reporter: David Pope
>Assignee: David Pope
> Attachments: HBASE-16853.branch-1.3-v1.patch
>
>
> h2. Problem
> If there are Region Servers registered as "draining", they will continue to 
> have "draining" znodes after a HMaster failover; however, the balancer will 
> assign regions to them.
> h2. How to reproduce (on hbase master):
> # Add regionserver to /hbase/draining: {{bin/hbase-jruby 
> bin/draining_servers.rb add server1:16205}}
> # Unload the regionserver:  {{bin/hbase-jruby bin/region_mover.rb unload 
> server1:16205}}
> # Kill the Active HMaster and failover to the Backup HMaster
> # Run the balancer: {{hbase shell <<< "balancer"}}
> # Notice regions get assigned on new Active Master to Region Servers in 
> /hbase/draining
> h2. Root Cause
> The Backup HMaster initializes the {{DrainingServerTracker}} before the 
> Region Servers are registered as "online" with the {{ServerManager}}.  As a 
> result, the {{ServerManager.drainingServers}} isn't populated with existing 
> Region Servers in draining when we have an HMaster failover.
> E.g., 
> # We have a region server in draining: {{server1,16205,1000}}
> # The {{RegionServerTracker}} starts up and adds a ZK watcher on the Znode 
> for this RegionServer: {{/hbase/rs/server1,16205,1000}}
> # The {{DrainingServerTracker}} starts and processes each Znode under 
> {{/hbase/draining}}, but the Region Server isn't registered as "online" so it 
> isn't added to the {{ServerManager.drainingServers}} list.
> # The Region Server is added to the {{DrainingServerTracker.drainingServers}} 
> list.
> # The Region Server's Znode watcher is triggered and the ZK watcher is 
> restarted.
> # The Region Server is registered with {{ServerManager}} as "online".
> *END STATE:* The Region Server has a Znode in {{/hbase/draining}}, but it is 
> registered as "online" and the Balancer will start assigning regions to it.
> {code}
> $ bin/hbase-jruby bin/draining_servers.rb list
> [1] server1,16205,1000
> $ grep server1,16205,1000 logs/master-server1.log
> 2016-10-14 16:02:47,713 DEBUG [server1:16001.activeMasterManager] 
> zookeeper.ZKUtil: master:16001-0x157c56adc810014, quorum=localhost:2181, 
> baseZNode=/hbase Set watcher on existing znode=/hbase/rs/server1,16205,1000
> [2] 2016-10-14 16:02:47,722 DEBUG [server1:16001.activeMasterManager] 
> zookeeper.RegionServerTracker: Added tracking of RS 
> /hbase/rs/server1,16205,1000
> 2016-10-14 16:02:47,730 DEBUG [server1:16001.activeMasterManager] 
> zookeeper.ZKUtil: master:16001-0x157c56adc810014, quorum=localhost:2181, 
> baseZNode=/hbase Set watcher on existing 
> znode=/hbase/draining/server1,16205,1000
> [3] 2016-10-14 16:02:47,731 WARN  [server1:16001.activeMasterManager] 
> master.ServerManager: Server server1,16205,1000 is not currently online. 
> Ignoring request to add it to draining list.
> [4] 2016-10-14 16:02:47,731 INFO  [server1:16001.activeMasterManager] 
> zookeeper.DrainingServerTracker: Draining RS node created, adding to list 
> [server1,16205,1000]
> 2016-10-14 16:02:47,971 DEBUG [main-EventThread] zookeeper.ZKUtil: 
> master:16001-0x157c56adc810014, quorum=localhost:2181, baseZNode=/hbase Set 
> watcher on existing 
> znode=/hbase/rs/dev6918.prn2.facebook.com,16205,1476486047114
> [5] 2016-10-14 16:02:47,976 DEBUG [main-EventThread] 
> zookeeper.RegionServerTracker: Added tracking of RS 
> /hbase/rs/server1,16205,1000
> [6] 2016-10-14 16:02:52,084 INFO  
> [RpcServer.FifoWFPBQ.default.handler=29,queue=2,port=16001] 
> master.ServerManager: Registering server=server1,16205,1000
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HBASE-16853) Regions are assigned to Region Servers in /hbase/draining after HBase Master failover

2016-10-16 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-16853?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15579860#comment-15579860
 ] 

Hadoop QA commented on HBASE-16853:
---

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 16s 
{color} | {color:blue} Docker mode activated. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s 
{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 
0s {color} | {color:green} The patch appears to include 1 new or modified test 
files. {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 2m 
1s {color} | {color:green} branch-1 passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 30s 
{color} | {color:green} branch-1 passed with JDK v1.8.0_101 {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 34s 
{color} | {color:green} branch-1 passed with JDK v1.7.0_80 {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 
56s {color} | {color:green} branch-1 passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 
16s {color} | {color:green} branch-1 passed {color} |
| {color:red}-1{color} | {color:red} findbugs {color} | {color:red} 1m 54s 
{color} | {color:red} hbase-server in branch-1 has 1 extant Findbugs warnings. 
{color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 24s 
{color} | {color:green} branch-1 passed with JDK v1.8.0_101 {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 32s 
{color} | {color:green} branch-1 passed with JDK v1.7.0_80 {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 
46s {color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 30s 
{color} | {color:green} the patch passed with JDK v1.8.0_101 {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 30s 
{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 33s 
{color} | {color:green} the patch passed with JDK v1.7.0_80 {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 33s 
{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 
56s {color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 
16s {color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 
0s {color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} hadoopcheck {color} | {color:green} 
16m 29s {color} | {color:green} The patch does not cause any errors with Hadoop 
2.4.0 2.4.1 2.5.0 2.5.1 2.5.2 2.6.1 2.6.2 2.6.3 2.7.1. {color} |
| {color:green}+1{color} | {color:green} hbaseprotoc {color} | {color:green} 0m 
14s {color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 2m 7s 
{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 25s 
{color} | {color:green} the patch passed with JDK v1.8.0_101 {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 33s 
{color} | {color:green} the patch passed with JDK v1.7.0_80 {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red} 72m 20s {color} 
| {color:red} hbase-server in the patch failed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 
15s {color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 103m 15s {color} 
| {color:black} {color} |
\\
\\
|| Reason || Tests ||
| Timed out junit tests | org.apache.hadoop.hbase.filter.TestScanRowPrefix |
|   | 
org.apache.hadoop.hbase.security.visibility.TestVisibilityLabelsWithDeletes |
\\
\\
|| Subsystem || Report/Notes ||
| Docker | Client=1.11.2 Server=1.11.2 Image:yetus/hbase:b2c5d84 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12833637/HBASE-16853.branch-1.3-v1.patch
 |
| JIRA Issue | HBASE-16853 |
| Optional Tests |  asflicense  javac  javadoc  unit  findbugs  hadoopcheck  
hbaseanti  checkstyle  compile  |
| uname | Linux abcf67c6de6a 3.13.0-95-generic #142-Ubuntu SMP Fri Aug 12 
17:00:09 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
|