[jira] [Updated] (HBASE-21465) Retry on reportRegionStateTransition can lead to unexpected errors

2018-11-29 Thread Duo Zhang (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-21465?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Duo Zhang updated HBASE-21465:
--
Resolution: Fixed
Status: Resolved  (was: Patch Available)

HBASE-21534 will take care of the flaky TestAssignmentManager related UTs.

> Retry on reportRegionStateTransition can lead to unexpected errors
> --
>
> Key: HBASE-21465
> URL: https://issues.apache.org/jira/browse/HBASE-21465
> Project: HBase
>  Issue Type: Sub-task
>  Components: amv2
>Reporter: Duo Zhang
>Assignee: Duo Zhang
>Priority: Major
> Fix For: 3.0.0, 2.2.0
>
> Attachments: HBASE-21465-UT.patch, HBASE-21465-v1.patch, 
> HBASE-21465-v2.patch, HBASE-21465-v3.patch, HBASE-21465-v4.patch, 
> HBASE-21465.patch
>
>
> It is possible that the reportRegionStateTransition method is succeeded at 
> master side, but before returning the the region server, the rpc connection 
> is broken, or the master restarts. So next when the region server try 
> again,we will find that the state for the region and the TRSP is not correct, 
> and can lead to a RS abort or something even worse.
> We should be able to determine whether a reportRegionStateTransition call is 
> just a retry and has already been succeeded, and just ignore it.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21512) Introduce an AsyncClusterConnection and replace the usage of ClusterConnection

2018-11-29 Thread Duo Zhang (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21512?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16704362#comment-16704362
 ] 

Duo Zhang commented on HBASE-21512:
---

Let me open a HBASE-21512 to land HBASE-21515.

> Introduce an AsyncClusterConnection and replace the usage of ClusterConnection
> --
>
> Key: HBASE-21512
> URL: https://issues.apache.org/jira/browse/HBASE-21512
> Project: HBase
>  Issue Type: Umbrella
>Reporter: Duo Zhang
>Priority: Major
> Fix For: 3.0.0
>
>
> At least for the RSProcedureDispatcher, with CompletableFuture we do not need 
> to set a delay and use a thread pool any more, which could reduce the 
> resource usage and also the latency.
> Once this is done, I think we can remove the ClusterConnection completely, 
> and start to rewrite the old sync client based on the async client, which 
> could reduce the code base a lot for our client.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-21534) TestAssignmentManager is flakey

2018-11-29 Thread Duo Zhang (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-21534?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Duo Zhang updated HBASE-21534:
--
Attachment: HBASE-21534.patch

> TestAssignmentManager is flakey
> ---
>
> Key: HBASE-21534
> URL: https://issues.apache.org/jira/browse/HBASE-21534
> Project: HBase
>  Issue Type: Task
>Reporter: Duo Zhang
>Priority: Major
> Attachments: HBASE-21534.patch
>
>
> See this in the outout and then the test hang
> {noformat}
> 2018-11-29 20:47:50,061 WARN  [MockRSProcedureDispatcher-pool5-t10] 
> assignment.AssignmentManager(894): The region server localhost,102,1 is 
> already dead, skip reportRegionStateTransition call
> {noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21518) TestMasterFailoverWithProcedures is flaky

2018-11-29 Thread Sean Busbey (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21518?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16704343#comment-16704343
 ] 

Sean Busbey commented on HBASE-21518:
-

Yes, any UT that has multiple masters and knocks one over.

> TestMasterFailoverWithProcedures is flaky
> -
>
> Key: HBASE-21518
> URL: https://issues.apache.org/jira/browse/HBASE-21518
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 2.2.0, 2.0.3, 2.1.2
>Reporter: Peter Somogyi
>Assignee: Peter Somogyi
>Priority: Major
> Attachments: HBASE-21518-v1.patch, output.txt
>
>
> TestMasterFailoverWithProcedures test is failing frequently, times out. I 
> faced this failure on 2.0.3RC0 vote and it also appears on multiple flaky 
> dashboards.
> branch-2: 
> [https://builds.apache.org/view/H-L/view/HBase/job/HBase-Flaky-Tests/job/branch-2/2007/]
> branch-2.1: 
> [https://builds.apache.org/view/H-L/view/HBase/job/HBase-Flaky-Tests/job/branch-2.1/2002/]
> branch-2.0: 
> [https://builds.apache.org/view/H-L/view/HBase/job/HBase-Flaky-Tests/job/branch-2.0/1988/]
>   
> {noformat}
> [INFO] Running 
> org.apache.hadoop.hbase.master.procedure.TestMasterFailoverWithProcedures
> [ERROR] Tests run: 4, Failures: 0, Errors: 2, Skipped: 0, Time elapsed: 
> 780.648 s <<< FAILURE! - in 
> org.apache.hadoop.hbase.master.procedure.TestMasterFailoverWithProcedures
> [ERROR] 
> org.apache.hadoop.hbase.master.procedure.TestMasterFailoverWithProcedures  
> Time elapsed: 749.024 s  <<< ERROR!
> org.junit.runners.model.TestTimedOutException: test timed out after 780 
> seconds
>   at 
> org.apache.hadoop.hbase.master.procedure.TestMasterFailoverWithProcedures.tearDown(TestMasterFailoverWithProcedures.java:86)
> [ERROR] 
> org.apache.hadoop.hbase.master.procedure.TestMasterFailoverWithProcedures  
> Time elapsed: 749.051 s  <<< ERROR!
> java.lang.Exception: Appears to be stuck in thread RS-EventLoopGroup-3-2
> {noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21534) TestAssignmentManager is flakey

2018-11-29 Thread Duo Zhang (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21534?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16704339#comment-16704339
 ] 

Duo Zhang commented on HBASE-21534:
---

Look at the log, it seems that we assign meta to the server which is already 
marked as dead. This should be a test issue, let me provide a fix.

{noformat}
2018-11-29 20:47:49,751 INFO  [PEWorker-63] 
procedure.ServerCrashProcedure(131): Start pid=14, 
state=RUNNABLE:SERVER_CRASH_START, hasLock=true; ServerCrashProcedure 
server=localhost,102,1, splitWal=false, meta=true
2018-11-29 20:47:49,751 DEBUG [PEWorker-63] procedure2.RootProcedureState(153): 
Add procedure pid=14, state=RUNNABLE:SERVER_CRASH_SPLIT_META_LOGS, 
hasLock=true; ServerCrashProcedure server=localhost,102,1, splitWal=false, 
meta=true as the 0th rollback step
2018-11-29 20:47:49,751 DEBUG [PEWorker-63] 
procedure.MasterProcedureScheduler(351): Add ServerQueue(localhost,102,1, 
xlock=false sharedLock=0 size=0) to run queue because: pid=14, 
state=RUNNABLE:SERVER_CRASH_SPLIT_META_LOGS, hasLock=false; 
ServerCrashProcedure server=localhost,102,1, splitWal=false, meta=true released 
exclusive lock
2018-11-29 20:47:49,751 DEBUG [PEWorker-63] 
procedure.MasterProcedureScheduler(351): Add ServerQueue(localhost,102,1, 
xlock=false sharedLock=0 size=1) to run queue because: the exclusive lock is 
not held by anyone when adding pid=14, 
state=RUNNABLE:SERVER_CRASH_SPLIT_META_LOGS, hasLock=false; 
ServerCrashProcedure server=localhost,102,1, splitWal=false, meta=true
2018-11-29 20:47:49,751 DEBUG [PEWorker-63] 
procedure.MasterProcedureScheduler(361): Remove ServerQueue(localhost,102,1, 
xlock=false sharedLock=0 size=0) from run queue because: queue is empty after 
polling out pid=14, state=RUNNABLE:SERVER_CRASH_SPLIT_META_LOGS, hasLock=false; 
ServerCrashProcedure server=localhost,102,1, splitWal=false, meta=true
2018-11-29 20:47:49,751 DEBUG [PEWorker-63] 
procedure.MasterProcedureScheduler(361): Remove ServerQueue(localhost,102,1, 
xlock=true (14) sharedLock=0 size=0) from run queue because: pid=14, 
state=RUNNABLE:SERVER_CRASH_SPLIT_META_LOGS, hasLock=false; 
ServerCrashProcedure server=localhost,102,1, splitWal=false, meta=true held 
exclusive lock
2018-11-29 20:47:49,751 DEBUG [PEWorker-63] 
procedure.ServerCrashProcedure(207): Splitting meta WALs pid=14, 
state=RUNNABLE:SERVER_CRASH_SPLIT_META_LOGS, hasLock=true; ServerCrashProcedure 
server=localhost,102,1, splitWal=false, meta=true
2018-11-29 20:47:49,755 INFO  [PEWorker-63] master.MasterWalManager(314): Log 
dir for server localhost,102,1 does not exist
2018-11-29 20:47:49,755 INFO  [PEWorker-63] master.SplitLogManager(465): dead 
splitlog workers [localhost,102,1]
2018-11-29 20:47:49,755 INFO  [PEWorker-63] master.SplitLogManager(297): 
Finished splitting (more than or equal to) 0 bytes in 0 log files in [] in 0ms
2018-11-29 20:47:49,756 DEBUG [PEWorker-63] 
procedure.ServerCrashProcedure(213): Done splitting meta WALs pid=14, 
state=RUNNABLE:SERVER_CRASH_SPLIT_META_LOGS, hasLock=true; ServerCrashProcedure 
server=localhost,102,1, splitWal=false, meta=true
2018-11-29 20:47:49,756 DEBUG [PEWorker-63] procedure2.RootProcedureState(153): 
Add procedure pid=14, state=RUNNABLE:SERVER_CRASH_ASSIGN_META, hasLock=true; 
ServerCrashProcedure server=localhost,102,1, splitWal=false, meta=true as the 
1th rollback step
2018-11-29 20:47:49,756 DEBUG [PEWorker-63] 
procedure.MasterProcedureScheduler(351): Add ServerQueue(localhost,102,1, 
xlock=false sharedLock=0 size=0) to run queue because: pid=14, 
state=RUNNABLE:SERVER_CRASH_ASSIGN_META, hasLock=false; ServerCrashProcedure 
server=localhost,102,1, splitWal=false, meta=true released exclusive lock
2018-11-29 20:47:49,756 DEBUG [PEWorker-63] 
procedure.MasterProcedureScheduler(351): Add ServerQueue(localhost,102,1, 
xlock=false sharedLock=0 size=1) to run queue because: the exclusive lock is 
not held by anyone when adding pid=14, state=RUNNABLE:SERVER_CRASH_ASSIGN_META, 
hasLock=false; ServerCrashProcedure server=localhost,102,1, splitWal=false, 
meta=true
2018-11-29 20:47:49,756 DEBUG [PEWorker-63] 
procedure.MasterProcedureScheduler(361): Remove ServerQueue(localhost,102,1, 
xlock=false sharedLock=0 size=0) from run queue because: queue is empty after 
polling out pid=14, state=RUNNABLE:SERVER_CRASH_ASSIGN_META, hasLock=false; 
ServerCrashProcedure server=localhost,102,1, splitWal=false, meta=true
2018-11-29 20:47:49,756 DEBUG [PEWorker-63] 
procedure.MasterProcedureScheduler(361): Remove ServerQueue(localhost,102,1, 
xlock=true (14) sharedLock=0 size=0) from run queue because: pid=14, 
state=RUNNABLE:SERVER_CRASH_ASSIGN_META, hasLock=false; ServerCrashProcedure 
server=localhost,102,1, splitWal=false, meta=true held exclusive lock
2018-11-29 20:47:49,758 INFO  [PEWorker-63] procedure2.ProcedureExecutor(1680): 
Initialized subprocedures=[{pid=15, ppid=14, 
state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, 

[jira] [Created] (HBASE-21534) TestAssignmentManager is flakey

2018-11-29 Thread Duo Zhang (JIRA)
Duo Zhang created HBASE-21534:
-

 Summary: TestAssignmentManager is flakey
 Key: HBASE-21534
 URL: https://issues.apache.org/jira/browse/HBASE-21534
 Project: HBase
  Issue Type: Task
Reporter: Duo Zhang


See this in the outout and then the test hang
{noformat}
2018-11-29 20:47:50,061 WARN  [MockRSProcedureDispatcher-pool5-t10] 
assignment.AssignmentManager(894): The region server localhost,102,1 is already 
dead, skip reportRegionStateTransition call
{noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21518) TestMasterFailoverWithProcedures is flaky

2018-11-29 Thread Allan Yang (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21518?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16704319#comment-16704319
 ] 

Allan Yang commented on HBASE-21518:


So, other UTs can be flaky because of this too? 

> TestMasterFailoverWithProcedures is flaky
> -
>
> Key: HBASE-21518
> URL: https://issues.apache.org/jira/browse/HBASE-21518
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 2.2.0, 2.0.3, 2.1.2
>Reporter: Peter Somogyi
>Assignee: Peter Somogyi
>Priority: Major
> Attachments: HBASE-21518-v1.patch, output.txt
>
>
> TestMasterFailoverWithProcedures test is failing frequently, times out. I 
> faced this failure on 2.0.3RC0 vote and it also appears on multiple flaky 
> dashboards.
> branch-2: 
> [https://builds.apache.org/view/H-L/view/HBase/job/HBase-Flaky-Tests/job/branch-2/2007/]
> branch-2.1: 
> [https://builds.apache.org/view/H-L/view/HBase/job/HBase-Flaky-Tests/job/branch-2.1/2002/]
> branch-2.0: 
> [https://builds.apache.org/view/H-L/view/HBase/job/HBase-Flaky-Tests/job/branch-2.0/1988/]
>   
> {noformat}
> [INFO] Running 
> org.apache.hadoop.hbase.master.procedure.TestMasterFailoverWithProcedures
> [ERROR] Tests run: 4, Failures: 0, Errors: 2, Skipped: 0, Time elapsed: 
> 780.648 s <<< FAILURE! - in 
> org.apache.hadoop.hbase.master.procedure.TestMasterFailoverWithProcedures
> [ERROR] 
> org.apache.hadoop.hbase.master.procedure.TestMasterFailoverWithProcedures  
> Time elapsed: 749.024 s  <<< ERROR!
> org.junit.runners.model.TestTimedOutException: test timed out after 780 
> seconds
>   at 
> org.apache.hadoop.hbase.master.procedure.TestMasterFailoverWithProcedures.tearDown(TestMasterFailoverWithProcedures.java:86)
> [ERROR] 
> org.apache.hadoop.hbase.master.procedure.TestMasterFailoverWithProcedures  
> Time elapsed: 749.051 s  <<< ERROR!
> java.lang.Exception: Appears to be stuck in thread RS-EventLoopGroup-3-2
> {noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-21515) Also initialize an AsyncClusterConnection in HRegionServer

2018-11-29 Thread Duo Zhang (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-21515?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Duo Zhang updated HBASE-21515:
--
Attachment: HBASE-21515-v4.patch

> Also initialize an AsyncClusterConnection in HRegionServer
> --
>
> Key: HBASE-21515
> URL: https://issues.apache.org/jira/browse/HBASE-21515
> Project: HBase
>  Issue Type: Sub-task
>  Components: regionserver
>Reporter: Duo Zhang
>Assignee: Duo Zhang
>Priority: Major
> Attachments: HBASE-21515-v1.patch, HBASE-21515-v2.patch, 
> HBASE-21515-v3.patch, HBASE-21515-v4.patch, HBASE-21515-v4.patch, 
> HBASE-21515.patch
>
>
> Plan to do this incrementally, so first we will only introduce the new 
> AsyncClusterConnection, without deleting the old ClusterConnection. And we 
> can move the code which uses the ClusterConnection to use 
> AsyncClusterConnection, through different sub tasks here. And finally we can 
> remove the ClusterConnection completely.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21083) Introduce a mechanism to bypass the execution of a stuck procedure

2018-11-29 Thread Allan Yang (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21083?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16704316#comment-16704316
 ] 

Allan Yang commented on HBASE-21083:


And besides, bypassing a procedure in the middle without fix the inconsistency 
is dangerous. We don't want to expose this method to user themself.

> Introduce a mechanism to bypass the execution of a stuck procedure
> --
>
> Key: HBASE-21083
> URL: https://issues.apache.org/jira/browse/HBASE-21083
> Project: HBase
>  Issue Type: Sub-task
>  Components: amv2
>Affects Versions: 2.1.0, 2.0.1
>Reporter: Allan Yang
>Assignee: Allan Yang
>Priority: Major
> Fix For: 3.0.0, 2.1.1, 2.0.2
>
> Attachments: HBASE-21083.branch-2.0.001.patch, 
> HBASE-21083.branch-2.0.002.patch, HBASE-21083.branch-2.0.003.patch, 
> HBASE-21083.branch-2.0.003.patch, HBASE-21083.branch-2.0.003.patch, 
> HBASE-21083.branch-2.1.001.patch
>
>
> Offline discussed with [~stack] and [~Apache9]. We all agreed that we need to 
> introduce a mechanism to 'force complete' a stuck procedure, so the AMv2 can 
> continue running.
> we still have some unrevealed bugs hiding in our AMv2 and procedureV2 system, 
> we need something to interfere with stuck procedures before HBCK2 can work. 
> This is very crucial for a production ready system.
> For now, we have little ways to interfere with running procedures. Aborting 
> them is not a good choice, since some procedures are not abort-able. And some 
> procedure may have overridden the abort() method, which will ignore the abort 
> request.
> So, here, I will introduce a mechanism to bypass the execution of a stuck 
> procedure.
> Basically, I added a field called 'bypass' to Procedure class. If we set this 
> field to true, all the logic in execute/rollback will be skipped, letting 
> this procedure and its ancestors complete normally and releasing the lock 
> resources at last.
> Notice that bypassing a procedure may leave the cluster in a middle state, 
> e.g. the region not assigned, or some hdfs files left behind. 
> The Operators need know the side effect of bypassing and recover the 
> inconsistent state of the cluster themselves, like issuing new procedures to 
> assign the regions.
> A patch will be uploaded and review board will be open. For now, only APIs in 
> ProcedureExecutor are provided. If anything is fine, I will add it to master 
> service and add a shell command to bypass a procedure. Or, maybe we can use 
> dynamically compiled JSPs to execute those APIs as mentioned in HBASE-20679.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21531) race between region report and region move causes master to kill RS

2018-11-29 Thread Duo Zhang (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21531?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16704315#comment-16704315
 ] 

Duo Zhang commented on HBASE-21531:
---

Anyway if you think the region server could have incorrect state then the most 
safe way is to kill the region server...

But since there will be race which leads to a warning in the code but actually 
there is nothing wrong, then I do not think we should try to do any 'fix'...

> race between region report and region move causes master to kill RS
> ---
>
> Key: HBASE-21531
> URL: https://issues.apache.org/jira/browse/HBASE-21531
> Project: HBase
>  Issue Type: Bug
>Reporter: Sergey Shelukhin
>Priority: Major
>
> In this case the delay between the ack from RS and the region report was 
> 1.5s, so I'm not sure what caused the race (network hiccup? unreported retry 
> by protobuf transport?) but in any case I don't see anything that prevents 
> this from happening in a normal case, with a narrowed time window. Any delay 
> (e.g. a GC pause on RS right after the report is built, and ack is sent for 
> the close) or retries expands the window.
> Master starts moving the region and the source RS acks by 21:51,206
> {noformat}
> Master:
> 2018-11-21 21:21:49,024 INFO  [master/6:17000.Chore.1] master.HMaster: 
> balance hri=, source=,17020,1542754626176, destination= 2>,17020,1542863268158
> ...
> Server:
> 2018-11-21 21:21:49,394 INFO  [RS_CLOSE_REGION-regionserver/:17020-1] 
> handler.UnassignRegionHandler: Close 
> ...
> 2018-11-21 21:21:51,095 INFO  [RS_CLOSE_REGION-regionserver/:17020-1] 
> handler.UnassignRegionHandler: Closed 
> {noformat}
> By then the region is removed from onlineRegions, so the master proceeds.
> {noformat}
> 2018-11-21 21:21:51,206 INFO  [PEWorker-4] procedure2.ProcedureExecutor: 
> Finished subprocedure(s) of pid=667, 
> state=RUNNABLE:REGION_STATE_TRANSITION_CONFIRM_CLOSED, hasLock=true; 
> TransitRegionStateProcedure table=, region=, REOPEN/MOVE; 
> resume parent processing.
> 2018-11-21 21:21:51,386 INFO  [PEWorker-13] assignment.RegionStateStore: 
> pid=667 updating hbase:meta row=, regionState=OPENING, 
> regionLocation=,17020,1542863268158
> {noformat}
> There are no obvious errors/delays that I see in RS log, and it doesn't log 
> starting to send the report.
> However, at 21:52.853 the report is processed that still contains this region.
> {noformat}
> 2018-11-21 21:21:52,853 WARN  
> [RpcServer.default.FPBQ.Fifo.handler=48,queue=3,port=17000] 
> assignment.AssignmentManager: Killing ,17020,1542754626176: 
> rit=OPENING, location=,17020,1542863268158, table=, 
> region= reported OPEN on server=,17020,1542754626176 but 
> state has otherwise.
> * ABORTING region server ,17020,1542754626176: 
> org.apache.hadoop.hbase.YouAreDeadException: rit=OPENING, location= 2>,17020,1542863268158, table=, region= reported OPEN on 
> server=,17020,1542754626176 but state has otherwise.
> {noformat}
> RS shuts down in an orderly manner and it can be seen from the log that this 
> region is actually not present (there's no line indicating it's being closed, 
> unlike for other regions).
> I think there needs to be some sort of versioning for region operations 
> and/or in RS reports to allow master to account for concurrent operations and 
> avoid races. Probably per region with either a grace period or an additional 
> global version, so that master could avoid killing RS based on stale reports, 
> but still kill RS if it did retain an old version of the region state due to 
> some bug after acking a new version.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21083) Introduce a mechanism to bypass the execution of a stuck procedure

2018-11-29 Thread Allan Yang (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21083?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16704314#comment-16704314
 ] 

Allan Yang commented on HBASE-21083:


[~xucang], It should not be very hard to back port this patch, but this feature 
is only used by HBCK2 in 2.x, and it does not provide a Admin API(over RPC) or 
something else, it can't be directly called by shell or HBaseAdmin class. 

> Introduce a mechanism to bypass the execution of a stuck procedure
> --
>
> Key: HBASE-21083
> URL: https://issues.apache.org/jira/browse/HBASE-21083
> Project: HBase
>  Issue Type: Sub-task
>  Components: amv2
>Affects Versions: 2.1.0, 2.0.1
>Reporter: Allan Yang
>Assignee: Allan Yang
>Priority: Major
> Fix For: 3.0.0, 2.1.1, 2.0.2
>
> Attachments: HBASE-21083.branch-2.0.001.patch, 
> HBASE-21083.branch-2.0.002.patch, HBASE-21083.branch-2.0.003.patch, 
> HBASE-21083.branch-2.0.003.patch, HBASE-21083.branch-2.0.003.patch, 
> HBASE-21083.branch-2.1.001.patch
>
>
> Offline discussed with [~stack] and [~Apache9]. We all agreed that we need to 
> introduce a mechanism to 'force complete' a stuck procedure, so the AMv2 can 
> continue running.
> we still have some unrevealed bugs hiding in our AMv2 and procedureV2 system, 
> we need something to interfere with stuck procedures before HBCK2 can work. 
> This is very crucial for a production ready system.
> For now, we have little ways to interfere with running procedures. Aborting 
> them is not a good choice, since some procedures are not abort-able. And some 
> procedure may have overridden the abort() method, which will ignore the abort 
> request.
> So, here, I will introduce a mechanism to bypass the execution of a stuck 
> procedure.
> Basically, I added a field called 'bypass' to Procedure class. If we set this 
> field to true, all the logic in execute/rollback will be skipped, letting 
> this procedure and its ancestors complete normally and releasing the lock 
> resources at last.
> Notice that bypassing a procedure may leave the cluster in a middle state, 
> e.g. the region not assigned, or some hdfs files left behind. 
> The Operators need know the side effect of bypassing and recover the 
> inconsistent state of the cluster themselves, like issuing new procedures to 
> assign the regions.
> A patch will be uploaded and review board will be open. For now, only APIs in 
> ProcedureExecutor are provided. If anything is fine, I will add it to master 
> service and add a shell command to bypass a procedure. Or, maybe we can use 
> dynamically compiled JSPs to execute those APIs as mentioned in HBASE-20679.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21524) Unnecessary DEBUG log in ConnectionImplementation#isTableEnabled

2018-11-29 Thread Hudson (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21524?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16704241#comment-16704241
 ] 

Hudson commented on HBASE-21524:


Results for branch branch-2.1
[build #646 on 
builds.a.o|https://builds.apache.org/job/HBase%20Nightly/job/branch-2.1/646/]: 
(x) *{color:red}-1 overall{color}*

details (if available):

(/) {color:green}+1 general checks{color}
-- For more information [see general 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-2.1/646//General_Nightly_Build_Report/]




(x) {color:red}-1 jdk8 hadoop2 checks{color}
-- For more information [see jdk8 (hadoop2) 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-2.1/646//JDK8_Nightly_Build_Report_(Hadoop2)/]


(x) {color:red}-1 jdk8 hadoop3 checks{color}
-- For more information [see jdk8 (hadoop3) 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-2.1/646//JDK8_Nightly_Build_Report_(Hadoop3)/]


(/) {color:green}+1 source release artifact{color}
-- See build output for details.


(/) {color:green}+1 client integration test{color}


> Unnecessary DEBUG log in ConnectionImplementation#isTableEnabled
> 
>
> Key: HBASE-21524
> URL: https://issues.apache.org/jira/browse/HBASE-21524
> Project: HBase
>  Issue Type: Improvement
>  Components: Client
>Reporter: Josh Elser
>Assignee: Josh Elser
>Priority: Major
> Fix For: 3.0.0, 2.2.0, 2.1.2
>
> Attachments: HBASE-21524.001.branch-2.0.patch
>
>
> Too much crap coming out of isTableEnabled for "normal" situations. Don't 
> need a DEBUG message to tell us that the table is available.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-18735) Provide a fast mechanism for shutting down mini cluster

2018-11-29 Thread Hudson (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-18735?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16704242#comment-16704242
 ] 

Hudson commented on HBASE-18735:


Results for branch branch-2.1
[build #646 on 
builds.a.o|https://builds.apache.org/job/HBase%20Nightly/job/branch-2.1/646/]: 
(x) *{color:red}-1 overall{color}*

details (if available):

(/) {color:green}+1 general checks{color}
-- For more information [see general 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-2.1/646//General_Nightly_Build_Report/]




(x) {color:red}-1 jdk8 hadoop2 checks{color}
-- For more information [see jdk8 (hadoop2) 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-2.1/646//JDK8_Nightly_Build_Report_(Hadoop2)/]


(x) {color:red}-1 jdk8 hadoop3 checks{color}
-- For more information [see jdk8 (hadoop3) 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-2.1/646//JDK8_Nightly_Build_Report_(Hadoop3)/]


(/) {color:green}+1 source release artifact{color}
-- See build output for details.


(/) {color:green}+1 client integration test{color}


> Provide a fast mechanism for shutting down mini cluster
> ---
>
> Key: HBASE-18735
> URL: https://issues.apache.org/jira/browse/HBASE-18735
> Project: HBase
>  Issue Type: Wish
>Reporter: Samarth Jain
>Assignee: Artem Ervits
>Priority: Major
> Fix For: 3.0.0, 2.2.0, 2.0.3, 2.1.2
>
> Attachments: HBASE-18735.v01.patch, HBASE-18735.v02.patch, 
> HBASE-18735.v03.patch
>
>
> The current mechanism of shutting down a mini cluster through 
> HBaseTestingUtility.shutDownMiniCluster can take a lot of time when the mini 
> cluster almost has a lot of tables. A lot of this time is spent in closing 
> all the user regions. It would be nice to have a mechanism where this 
> shutdown can happen quickly without having to worry about closing these user 
> regions. At the same time, this mechanism would need to make sure that all 
> the critical system resources like file handles and network ports are still 
> released so that subsequently initialized mini clusters on the same JVM or 
> system won't run into resource issues. This would make testing using HBase 
> mini clusters much faster and immensely help out test frameworks of dependent 
> projects like Phoenix.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-18735) Provide a fast mechanism for shutting down mini cluster

2018-11-29 Thread Hudson (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-18735?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16704240#comment-16704240
 ] 

Hudson commented on HBASE-18735:


Results for branch branch-2.0
[build #1124 on 
builds.a.o|https://builds.apache.org/job/HBase%20Nightly/job/branch-2.0/1124/]: 
(x) *{color:red}-1 overall{color}*

details (if available):

(/) {color:green}+1 general checks{color}
-- For more information [see general 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-2.0/1124//General_Nightly_Build_Report/]




(/) {color:green}+1 jdk8 hadoop2 checks{color}
-- For more information [see jdk8 (hadoop2) 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-2.0/1124//JDK8_Nightly_Build_Report_(Hadoop2)/]


(x) {color:red}-1 jdk8 hadoop3 checks{color}
-- For more information [see jdk8 (hadoop3) 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-2.0/1124//JDK8_Nightly_Build_Report_(Hadoop3)/]


(/) {color:green}+1 source release artifact{color}
-- See build output for details.


> Provide a fast mechanism for shutting down mini cluster
> ---
>
> Key: HBASE-18735
> URL: https://issues.apache.org/jira/browse/HBASE-18735
> Project: HBase
>  Issue Type: Wish
>Reporter: Samarth Jain
>Assignee: Artem Ervits
>Priority: Major
> Fix For: 3.0.0, 2.2.0, 2.0.3, 2.1.2
>
> Attachments: HBASE-18735.v01.patch, HBASE-18735.v02.patch, 
> HBASE-18735.v03.patch
>
>
> The current mechanism of shutting down a mini cluster through 
> HBaseTestingUtility.shutDownMiniCluster can take a lot of time when the mini 
> cluster almost has a lot of tables. A lot of this time is spent in closing 
> all the user regions. It would be nice to have a mechanism where this 
> shutdown can happen quickly without having to worry about closing these user 
> regions. At the same time, this mechanism would need to make sure that all 
> the critical system resources like file handles and network ports are still 
> released so that subsequently initialized mini clusters on the same JVM or 
> system won't run into resource issues. This would make testing using HBase 
> mini clusters much faster and immensely help out test frameworks of dependent 
> projects like Phoenix.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21530) Abort_Procedure should be able to take a list of proc IDs

2018-11-29 Thread Geoffrey Jacoby (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21530?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16704235#comment-16704235
 ] 

Geoffrey Jacoby commented on HBASE-21530:
-

I'm using a variant of HBase 1.3. 

If abort procedure has been deprecated later on, I disagree with that decision. 
I don't think you should ever have something which:

1. Ties up a finite resource on the cluster
2. Doesn't have a timeout and
3. Can't be canceled. 

> Abort_Procedure should be able to take a list of proc IDs
> -
>
> Key: HBASE-21530
> URL: https://issues.apache.org/jira/browse/HBASE-21530
> Project: HBase
>  Issue Type: Improvement
>Reporter: Geoffrey Jacoby
>Priority: Minor
>
> As a convenience, it would be helpful if the HBase shell's abort_procedure 
> call had the option of taking in multiple procedure ids at the same time, 
> rather than relying on operators to use a loop in an external script. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21524) Unnecessary DEBUG log in ConnectionImplementation#isTableEnabled

2018-11-29 Thread Hudson (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21524?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16704239#comment-16704239
 ] 

Hudson commented on HBASE-21524:


Results for branch branch-2.0
[build #1124 on 
builds.a.o|https://builds.apache.org/job/HBase%20Nightly/job/branch-2.0/1124/]: 
(x) *{color:red}-1 overall{color}*

details (if available):

(/) {color:green}+1 general checks{color}
-- For more information [see general 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-2.0/1124//General_Nightly_Build_Report/]




(/) {color:green}+1 jdk8 hadoop2 checks{color}
-- For more information [see jdk8 (hadoop2) 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-2.0/1124//JDK8_Nightly_Build_Report_(Hadoop2)/]


(x) {color:red}-1 jdk8 hadoop3 checks{color}
-- For more information [see jdk8 (hadoop3) 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-2.0/1124//JDK8_Nightly_Build_Report_(Hadoop3)/]


(/) {color:green}+1 source release artifact{color}
-- See build output for details.


> Unnecessary DEBUG log in ConnectionImplementation#isTableEnabled
> 
>
> Key: HBASE-21524
> URL: https://issues.apache.org/jira/browse/HBASE-21524
> Project: HBase
>  Issue Type: Improvement
>  Components: Client
>Reporter: Josh Elser
>Assignee: Josh Elser
>Priority: Major
> Fix For: 3.0.0, 2.2.0, 2.1.2
>
> Attachments: HBASE-21524.001.branch-2.0.patch
>
>
> Too much crap coming out of isTableEnabled for "normal" situations. Don't 
> need a DEBUG message to tell us that the table is available.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21530) Abort_Procedure should be able to take a list of proc IDs

2018-11-29 Thread Geoffrey Jacoby (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21530?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16704237#comment-16704237
 ] 

Geoffrey Jacoby commented on HBASE-21530:
-

In addition, even if aborting a running procedure is dangerous, aborting a 
queued (RUNNABLE) procedure should always be OK. 

> Abort_Procedure should be able to take a list of proc IDs
> -
>
> Key: HBASE-21530
> URL: https://issues.apache.org/jira/browse/HBASE-21530
> Project: HBase
>  Issue Type: Improvement
>Reporter: Geoffrey Jacoby
>Priority: Minor
>
> As a convenience, it would be helpful if the HBase shell's abort_procedure 
> call had the option of taking in multiple procedure ids at the same time, 
> rather than relying on operators to use a loop in an external script. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-18222) Tests fail as lock on table is not released

2018-11-29 Thread Sakthi (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-18222?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16704224#comment-16704224
 ] 

Sakthi commented on HBASE-18222:


[~anuphal] Are you still able to repro this? I tried, but couldn't repro.

> Tests fail as lock on table is not released 
> 
>
> Key: HBASE-18222
> URL: https://issues.apache.org/jira/browse/HBASE-18222
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 3.0.0
> Environment: [INFO] 
> 
> [INFO] Detecting the operating system and CPU architecture
> [INFO] 
> 
> [INFO] os.detected.name: linux
> [INFO] os.detected.arch: ppcle_64
> [INFO] os.detected.version: 3.16
> [INFO] os.detected.version.major: 3
> [INFO] os.detected.version.minor: 16
> [INFO] os.detected.release: ubuntu
> [INFO] os.detected.release.version: 14.04
> [INFO] os.detected.release.like.ubuntu: true
> [INFO] os.detected.release.like.debian: true
> [INFO] os.detected.classifier: linux-ppcle_64
>Reporter: Anup Halarnkar
>Priority: Major
> Fix For: 3.0.0
>
>
> *All Failed Tests*
> 1. 
> org.apache.hadoop.hbase.coprocessor.TestCoprocessorMetrics.testRegionObserverAfterRegionClosed
> _org.apache.hadoop.hbase.exceptions.TimeoutIOException: 
> java.util.concurrent.TimeoutException: The procedure 14 is still running
>   at 
> org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.waitProcedureResult(HBaseAdmin.java:3418)
>   at 
> org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.get(HBaseAdmin.java:3339)
>   at org.apache.hadoop.hbase.client.HBaseAdmin.get(HBaseAdmin.java:1962)
>   at 
> org.apache.hadoop.hbase.client.HBaseAdmin.disableTable(HBaseAdmin.java:764)
>   at 
> org.apache.hadoop.hbase.coprocessor.TestCoprocessorMetrics.testRegionObserverAfterRegionClosed(TestCoprocessorMetrics.java:487)
> _
> 2.org.apache.hadoop.hbase.coprocessor.TestCoprocessorMetrics.testMasterObserver
> 3.org.apache.hadoop.hbase.coprocessor.TestCoprocessorMetrics.testRegionObserverEndpoint
> 4.org.apache.hadoop.hbase.coprocessor.TestCoprocessorMetrics.testRegionObserverMultiRegion
> 5.org.apache.hadoop.hbase.coprocessor.TestCoprocessorMetrics.testRegionObserverSingleRegion
> 6.org.apache.hadoop.hbase.coprocessor.TestCoprocessorMetrics.testWALObserver
> 7.org.apache.hadoop.hbase.coprocessor.TestCoprocessorMetrics.testRegionServerObserver
> 8.org.apache.hadoop.hbase.coprocessor.TestCoprocessorMetrics.testRegionObserverMultiTable
> *-*
> *Stack Traces of all tests from 2 -8 is given below*
> org.apache.hadoop.hbase.TableNotDisabledException: 
> testRegionObserverAfterRegionClosed
>   at 
> org.apache.hadoop.hbase.master.HMaster.checkTableModifiable(HMaster.java:2470)
>   at 
> org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.prepareDelete(DeleteTableProcedure.java:241)
>   at 
> org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:91)
>   at 
> org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure.executeFromState(DeleteTableProcedure.java:58)
>   at 
> org.apache.hadoop.hbase.procedure2.StateMachineProcedure.execute(StateMachineProcedure.java:155)
>   at 
> org.apache.hadoop.hbase.procedure2.Procedure.doExecute(Procedure.java:843)
>   at 
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execProcedure(ProcedureExecutor.java:1373)
>   at 
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor.executeProcedure(ProcedureExecutor.java:1142)
>   at 
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor.access$800(ProcedureExecutor.java:78)
>   at 
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor$WorkerThread.run(ProcedureExecutor.java:1651)



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21515) Also initialize an AsyncClusterConnection in HRegionServer

2018-11-29 Thread Hadoop QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21515?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16704214#comment-16704214
 ] 

Hadoop QA commented on HBASE-21515:
---

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
12s{color} | {color:blue} Docker mode activated. {color} |
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} hbaseanti {color} | {color:green}  0m  
0s{color} | {color:green} Patch does not have any anti-patterns. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green}  0m 
 0s{color} | {color:green} The patch appears to include 20 new or modified test 
files. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
25s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  4m 
11s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  2m 
51s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  2m 
 2s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} shadedjars {color} | {color:green}  3m 
47s{color} | {color:green} branch has no errors when building our shaded 
downstream artifacts. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  3m 
28s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m  
8s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
16s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  4m 
 3s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  2m 
53s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  2m 
53s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
22s{color} | {color:green} The patch passed checkstyle in hbase-common {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
31s{color} | {color:green} hbase-client: The patch generated 0 new + 3 
unchanged - 3 fixed = 3 total (was 6) {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  1m 
10s{color} | {color:green} hbase-server: The patch generated 0 new + 285 
unchanged - 2 fixed = 285 total (was 287) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} shadedjars {color} | {color:green}  3m 
47s{color} | {color:green} patch has no errors when building our shaded 
downstream artifacts. {color} |
| {color:green}+1{color} | {color:green} hadoopcheck {color} | {color:green}  
8m 18s{color} | {color:green} Patch does not cause any errors with Hadoop 2.7.4 
or 3.0.0. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  3m 
39s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m  
1s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} unit {color} | {color:green}  2m 
31s{color} | {color:green} hbase-common in the patch passed. {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green}  3m 
13s{color} | {color:green} hbase-client in the patch passed. {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red}130m 32s{color} 
| {color:red} hbase-server in the patch failed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  1m 
 1s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black}182m 27s{color} | 
{color:black} {color} |
\\
\\
|| Reason || Tests ||
| Failed junit tests | hadoop.hbase.client.TestAsyncClusterAdminApi |
\\
\\
|| Subsystem || Report/Notes ||
| Docker | Client=17.05.0-ce Server=17.05.0-ce Image:yetus/hbase:b002b0b |
| JIRA Issue | HBASE-21515 |
| JIRA 

[jira] [Comment Edited] (HBASE-21083) Introduce a mechanism to bypass the execution of a stuck procedure

2018-11-29 Thread Xu Cang (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21083?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16704195#comment-16704195
 ] 

Xu Cang edited comment on HBASE-21083 at 11/30/18 3:32 AM:
---

[~allan163] thanks for response

The reason I am asking is, seems for hbase branch-1 there is no reliable way to 
unblock or bypass stuck Procedures. And this feature is something potentially 
can be applied to branch-1 to alleviate engineering burden such as manually 
operating on WAL files.

I briefly skimmed your patch and I see most of the code change is related to 
ProcedureV2, not AMv2. So since branch-1 has ProcedureV2, I wanted to ask how 
hard to port this high-level logic to branch-1. 

 


was (Author: xucang):
[~allan163] thanks for response

The reason I am asking is, seems for hbase branch-1 there is no reliable way to 
unblock or bypass stuck Procedures. And this feature is something potentially 
can be applied to branch-1 to alleviate engineering burden such as manually 
operating on WAL files.

I briefly skimmed your patch and I see most of the code change is related to 
ProcedureV2, not AMv2. So since branch-1 has ProcedureV2, so I was asking how 
hard to port this high-level logic to branch-1. 

 

> Introduce a mechanism to bypass the execution of a stuck procedure
> --
>
> Key: HBASE-21083
> URL: https://issues.apache.org/jira/browse/HBASE-21083
> Project: HBase
>  Issue Type: Sub-task
>  Components: amv2
>Affects Versions: 2.1.0, 2.0.1
>Reporter: Allan Yang
>Assignee: Allan Yang
>Priority: Major
> Fix For: 3.0.0, 2.1.1, 2.0.2
>
> Attachments: HBASE-21083.branch-2.0.001.patch, 
> HBASE-21083.branch-2.0.002.patch, HBASE-21083.branch-2.0.003.patch, 
> HBASE-21083.branch-2.0.003.patch, HBASE-21083.branch-2.0.003.patch, 
> HBASE-21083.branch-2.1.001.patch
>
>
> Offline discussed with [~stack] and [~Apache9]. We all agreed that we need to 
> introduce a mechanism to 'force complete' a stuck procedure, so the AMv2 can 
> continue running.
> we still have some unrevealed bugs hiding in our AMv2 and procedureV2 system, 
> we need something to interfere with stuck procedures before HBCK2 can work. 
> This is very crucial for a production ready system.
> For now, we have little ways to interfere with running procedures. Aborting 
> them is not a good choice, since some procedures are not abort-able. And some 
> procedure may have overridden the abort() method, which will ignore the abort 
> request.
> So, here, I will introduce a mechanism to bypass the execution of a stuck 
> procedure.
> Basically, I added a field called 'bypass' to Procedure class. If we set this 
> field to true, all the logic in execute/rollback will be skipped, letting 
> this procedure and its ancestors complete normally and releasing the lock 
> resources at last.
> Notice that bypassing a procedure may leave the cluster in a middle state, 
> e.g. the region not assigned, or some hdfs files left behind. 
> The Operators need know the side effect of bypassing and recover the 
> inconsistent state of the cluster themselves, like issuing new procedures to 
> assign the regions.
> A patch will be uploaded and review board will be open. For now, only APIs in 
> ProcedureExecutor are provided. If anything is fine, I will add it to master 
> service and add a shell command to bypass a procedure. Or, maybe we can use 
> dynamically compiled JSPs to execute those APIs as mentioned in HBASE-20679.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Work started] (HBASE-21453) Convert ReadOnlyZKClient to DEBUG instead of INFO

2018-11-29 Thread Sakthi (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-21453?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on HBASE-21453 started by Sakthi.
--
> Convert ReadOnlyZKClient to DEBUG instead of INFO
> -
>
> Key: HBASE-21453
> URL: https://issues.apache.org/jira/browse/HBASE-21453
> Project: HBase
>  Issue Type: Bug
>Reporter: stack
>Assignee: Sakthi
>Priority: Major
> Attachments: hbase-21453.master.001.patch
>
>
> Running commands in spark-shell, this is what it looks like on each 
> invocation:
> {code}
> scala> val count = rdd.count()
> 2018-11-07 21:01:46,026 INFO  [Executor task launch worker for task 1] 
> zookeeper.ReadOnlyZKClient: Connect 0x18f3d868 to localhost:2181 with session 
> timeout=9ms, retries 30, retry interval 1000ms, keepAlive=6ms
> 2018-11-07 21:01:46,027 INFO  [ReadOnlyZKClient-localhost:2181@0x18f3d868] 
> zookeeper.ZooKeeper: Initiating client connection, 
> connectString=localhost:2181 sessionTimeout=9 
> watcher=org.apache.hadoop.hbase.zookeeper.ReadOnlyZKClient$$Lambda$20/1362339879@743dab9f
> 2018-11-07 21:01:46,030 INFO  
> [ReadOnlyZKClient-localhost:2181@0x18f3d868-SendThread(localhost:2181)] 
> zookeeper.ClientCnxn: Opening socket connection to server 
> localhost/127.0.0.1:2181. Will not attempt to authenticate using SASL 
> (unknown error)
> 2018-11-07 21:01:46,031 INFO  
> [ReadOnlyZKClient-localhost:2181@0x18f3d868-SendThread(localhost:2181)] 
> zookeeper.ClientCnxn: Socket connection established to 
> localhost/127.0.0.1:2181, initiating session
> 2018-11-07 21:01:46,033 INFO  
> [ReadOnlyZKClient-localhost:2181@0x18f3d868-SendThread(localhost:2181)] 
> zookeeper.ClientCnxn: Session establishment complete on server 
> localhost/127.0.0.1:2181, sessionid = 0x166f1b283080005, negotiated timeout = 
> 4
> 2018-11-07 21:01:46,035 INFO  [Executor task launch worker for task 1] 
> mapreduce.TableInputFormatBase: Input split length: 0 bytes.
> [Stage 1:>  (0 + 1) / 
> 1]2018-11-07 21:01:48,074 INFO  [Executor task launch worker for task 1] 
> zookeeper.ReadOnlyZKClient: Close zookeeper connection 0x18f3d868 to 
> localhost:2181
> 2018-11-07 21:01:48,075 INFO  [ReadOnlyZKClient-localhost:2181@0x18f3d868] 
> zookeeper.ZooKeeper: Session: 0x166f1b283080005 closed
> 2018-11-07 21:01:48,076 INFO  [ReadOnlyZKClient 
> -localhost:2181@0x18f3d868-EventThread] zookeeper.ClientCnxn: EventThread 
> shut down for session: 0x166f1b283080005
> count: Long = 10
> {code}
> Let me shut down the ReadOnlyZKClient log level.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-21453) Convert ReadOnlyZKClient to DEBUG instead of INFO

2018-11-29 Thread Sakthi (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-21453?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sakthi updated HBASE-21453:
---
Attachment: hbase-21453.master.001.patch

> Convert ReadOnlyZKClient to DEBUG instead of INFO
> -
>
> Key: HBASE-21453
> URL: https://issues.apache.org/jira/browse/HBASE-21453
> Project: HBase
>  Issue Type: Bug
>Reporter: stack
>Priority: Major
> Attachments: hbase-21453.master.001.patch
>
>
> Running commands in spark-shell, this is what it looks like on each 
> invocation:
> {code}
> scala> val count = rdd.count()
> 2018-11-07 21:01:46,026 INFO  [Executor task launch worker for task 1] 
> zookeeper.ReadOnlyZKClient: Connect 0x18f3d868 to localhost:2181 with session 
> timeout=9ms, retries 30, retry interval 1000ms, keepAlive=6ms
> 2018-11-07 21:01:46,027 INFO  [ReadOnlyZKClient-localhost:2181@0x18f3d868] 
> zookeeper.ZooKeeper: Initiating client connection, 
> connectString=localhost:2181 sessionTimeout=9 
> watcher=org.apache.hadoop.hbase.zookeeper.ReadOnlyZKClient$$Lambda$20/1362339879@743dab9f
> 2018-11-07 21:01:46,030 INFO  
> [ReadOnlyZKClient-localhost:2181@0x18f3d868-SendThread(localhost:2181)] 
> zookeeper.ClientCnxn: Opening socket connection to server 
> localhost/127.0.0.1:2181. Will not attempt to authenticate using SASL 
> (unknown error)
> 2018-11-07 21:01:46,031 INFO  
> [ReadOnlyZKClient-localhost:2181@0x18f3d868-SendThread(localhost:2181)] 
> zookeeper.ClientCnxn: Socket connection established to 
> localhost/127.0.0.1:2181, initiating session
> 2018-11-07 21:01:46,033 INFO  
> [ReadOnlyZKClient-localhost:2181@0x18f3d868-SendThread(localhost:2181)] 
> zookeeper.ClientCnxn: Session establishment complete on server 
> localhost/127.0.0.1:2181, sessionid = 0x166f1b283080005, negotiated timeout = 
> 4
> 2018-11-07 21:01:46,035 INFO  [Executor task launch worker for task 1] 
> mapreduce.TableInputFormatBase: Input split length: 0 bytes.
> [Stage 1:>  (0 + 1) / 
> 1]2018-11-07 21:01:48,074 INFO  [Executor task launch worker for task 1] 
> zookeeper.ReadOnlyZKClient: Close zookeeper connection 0x18f3d868 to 
> localhost:2181
> 2018-11-07 21:01:48,075 INFO  [ReadOnlyZKClient-localhost:2181@0x18f3d868] 
> zookeeper.ZooKeeper: Session: 0x166f1b283080005 closed
> 2018-11-07 21:01:48,076 INFO  [ReadOnlyZKClient 
> -localhost:2181@0x18f3d868-EventThread] zookeeper.ClientCnxn: EventThread 
> shut down for session: 0x166f1b283080005
> count: Long = 10
> {code}
> Let me shut down the ReadOnlyZKClient log level.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21453) Convert ReadOnlyZKClient to DEBUG instead of INFO

2018-11-29 Thread Sakthi (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21453?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16704208#comment-16704208
 ] 

Sakthi commented on HBASE-21453:


Converted the info level logs in ReadOnlyZKClient to debug level.

> Convert ReadOnlyZKClient to DEBUG instead of INFO
> -
>
> Key: HBASE-21453
> URL: https://issues.apache.org/jira/browse/HBASE-21453
> Project: HBase
>  Issue Type: Bug
>Reporter: stack
>Priority: Major
> Attachments: hbase-21453.master.001.patch
>
>
> Running commands in spark-shell, this is what it looks like on each 
> invocation:
> {code}
> scala> val count = rdd.count()
> 2018-11-07 21:01:46,026 INFO  [Executor task launch worker for task 1] 
> zookeeper.ReadOnlyZKClient: Connect 0x18f3d868 to localhost:2181 with session 
> timeout=9ms, retries 30, retry interval 1000ms, keepAlive=6ms
> 2018-11-07 21:01:46,027 INFO  [ReadOnlyZKClient-localhost:2181@0x18f3d868] 
> zookeeper.ZooKeeper: Initiating client connection, 
> connectString=localhost:2181 sessionTimeout=9 
> watcher=org.apache.hadoop.hbase.zookeeper.ReadOnlyZKClient$$Lambda$20/1362339879@743dab9f
> 2018-11-07 21:01:46,030 INFO  
> [ReadOnlyZKClient-localhost:2181@0x18f3d868-SendThread(localhost:2181)] 
> zookeeper.ClientCnxn: Opening socket connection to server 
> localhost/127.0.0.1:2181. Will not attempt to authenticate using SASL 
> (unknown error)
> 2018-11-07 21:01:46,031 INFO  
> [ReadOnlyZKClient-localhost:2181@0x18f3d868-SendThread(localhost:2181)] 
> zookeeper.ClientCnxn: Socket connection established to 
> localhost/127.0.0.1:2181, initiating session
> 2018-11-07 21:01:46,033 INFO  
> [ReadOnlyZKClient-localhost:2181@0x18f3d868-SendThread(localhost:2181)] 
> zookeeper.ClientCnxn: Session establishment complete on server 
> localhost/127.0.0.1:2181, sessionid = 0x166f1b283080005, negotiated timeout = 
> 4
> 2018-11-07 21:01:46,035 INFO  [Executor task launch worker for task 1] 
> mapreduce.TableInputFormatBase: Input split length: 0 bytes.
> [Stage 1:>  (0 + 1) / 
> 1]2018-11-07 21:01:48,074 INFO  [Executor task launch worker for task 1] 
> zookeeper.ReadOnlyZKClient: Close zookeeper connection 0x18f3d868 to 
> localhost:2181
> 2018-11-07 21:01:48,075 INFO  [ReadOnlyZKClient-localhost:2181@0x18f3d868] 
> zookeeper.ZooKeeper: Session: 0x166f1b283080005 closed
> 2018-11-07 21:01:48,076 INFO  [ReadOnlyZKClient 
> -localhost:2181@0x18f3d868-EventThread] zookeeper.ClientCnxn: EventThread 
> shut down for session: 0x166f1b283080005
> count: Long = 10
> {code}
> Let me shut down the ReadOnlyZKClient log level.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (HBASE-21453) Convert ReadOnlyZKClient to DEBUG instead of INFO

2018-11-29 Thread Sakthi (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-21453?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sakthi reassigned HBASE-21453:
--

Assignee: Sakthi

> Convert ReadOnlyZKClient to DEBUG instead of INFO
> -
>
> Key: HBASE-21453
> URL: https://issues.apache.org/jira/browse/HBASE-21453
> Project: HBase
>  Issue Type: Bug
>Reporter: stack
>Assignee: Sakthi
>Priority: Major
> Attachments: hbase-21453.master.001.patch
>
>
> Running commands in spark-shell, this is what it looks like on each 
> invocation:
> {code}
> scala> val count = rdd.count()
> 2018-11-07 21:01:46,026 INFO  [Executor task launch worker for task 1] 
> zookeeper.ReadOnlyZKClient: Connect 0x18f3d868 to localhost:2181 with session 
> timeout=9ms, retries 30, retry interval 1000ms, keepAlive=6ms
> 2018-11-07 21:01:46,027 INFO  [ReadOnlyZKClient-localhost:2181@0x18f3d868] 
> zookeeper.ZooKeeper: Initiating client connection, 
> connectString=localhost:2181 sessionTimeout=9 
> watcher=org.apache.hadoop.hbase.zookeeper.ReadOnlyZKClient$$Lambda$20/1362339879@743dab9f
> 2018-11-07 21:01:46,030 INFO  
> [ReadOnlyZKClient-localhost:2181@0x18f3d868-SendThread(localhost:2181)] 
> zookeeper.ClientCnxn: Opening socket connection to server 
> localhost/127.0.0.1:2181. Will not attempt to authenticate using SASL 
> (unknown error)
> 2018-11-07 21:01:46,031 INFO  
> [ReadOnlyZKClient-localhost:2181@0x18f3d868-SendThread(localhost:2181)] 
> zookeeper.ClientCnxn: Socket connection established to 
> localhost/127.0.0.1:2181, initiating session
> 2018-11-07 21:01:46,033 INFO  
> [ReadOnlyZKClient-localhost:2181@0x18f3d868-SendThread(localhost:2181)] 
> zookeeper.ClientCnxn: Session establishment complete on server 
> localhost/127.0.0.1:2181, sessionid = 0x166f1b283080005, negotiated timeout = 
> 4
> 2018-11-07 21:01:46,035 INFO  [Executor task launch worker for task 1] 
> mapreduce.TableInputFormatBase: Input split length: 0 bytes.
> [Stage 1:>  (0 + 1) / 
> 1]2018-11-07 21:01:48,074 INFO  [Executor task launch worker for task 1] 
> zookeeper.ReadOnlyZKClient: Close zookeeper connection 0x18f3d868 to 
> localhost:2181
> 2018-11-07 21:01:48,075 INFO  [ReadOnlyZKClient-localhost:2181@0x18f3d868] 
> zookeeper.ZooKeeper: Session: 0x166f1b283080005 closed
> 2018-11-07 21:01:48,076 INFO  [ReadOnlyZKClient 
> -localhost:2181@0x18f3d868-EventThread] zookeeper.ClientCnxn: EventThread 
> shut down for session: 0x166f1b283080005
> count: Long = 10
> {code}
> Let me shut down the ReadOnlyZKClient log level.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21083) Introduce a mechanism to bypass the execution of a stuck procedure

2018-11-29 Thread Xu Cang (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21083?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16704195#comment-16704195
 ] 

Xu Cang commented on HBASE-21083:
-

[~allan163] thanks for response

The reason I am asking is, seems for hbase branch-1 there is no reliable way to 
unblock or bypass stuck Procedures. And this feature is something potentially 
can be applied to branch-1 to alleviate engineering burden such as manually 
operating on WAL files.

I briefly skimmed your patch and I see most of the code change is related to 
ProcedureV2, not AMv2. So since branch-1 has ProcedureV2, so I was asking how 
hard to port this high-level logic to branch-1. 

 

> Introduce a mechanism to bypass the execution of a stuck procedure
> --
>
> Key: HBASE-21083
> URL: https://issues.apache.org/jira/browse/HBASE-21083
> Project: HBase
>  Issue Type: Sub-task
>  Components: amv2
>Affects Versions: 2.1.0, 2.0.1
>Reporter: Allan Yang
>Assignee: Allan Yang
>Priority: Major
> Fix For: 3.0.0, 2.1.1, 2.0.2
>
> Attachments: HBASE-21083.branch-2.0.001.patch, 
> HBASE-21083.branch-2.0.002.patch, HBASE-21083.branch-2.0.003.patch, 
> HBASE-21083.branch-2.0.003.patch, HBASE-21083.branch-2.0.003.patch, 
> HBASE-21083.branch-2.1.001.patch
>
>
> Offline discussed with [~stack] and [~Apache9]. We all agreed that we need to 
> introduce a mechanism to 'force complete' a stuck procedure, so the AMv2 can 
> continue running.
> we still have some unrevealed bugs hiding in our AMv2 and procedureV2 system, 
> we need something to interfere with stuck procedures before HBCK2 can work. 
> This is very crucial for a production ready system.
> For now, we have little ways to interfere with running procedures. Aborting 
> them is not a good choice, since some procedures are not abort-able. And some 
> procedure may have overridden the abort() method, which will ignore the abort 
> request.
> So, here, I will introduce a mechanism to bypass the execution of a stuck 
> procedure.
> Basically, I added a field called 'bypass' to Procedure class. If we set this 
> field to true, all the logic in execute/rollback will be skipped, letting 
> this procedure and its ancestors complete normally and releasing the lock 
> resources at last.
> Notice that bypassing a procedure may leave the cluster in a middle state, 
> e.g. the region not assigned, or some hdfs files left behind. 
> The Operators need know the side effect of bypassing and recover the 
> inconsistent state of the cluster themselves, like issuing new procedures to 
> assign the regions.
> A patch will be uploaded and review board will be open. For now, only APIs in 
> ProcedureExecutor are provided. If anything is fine, I will add it to master 
> service and add a shell command to bypass a procedure. Or, maybe we can use 
> dynamically compiled JSPs to execute those APIs as mentioned in HBASE-20679.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21481) [acl] Superuser's permissions should not be granted or revoked by any non-su global admin

2018-11-29 Thread Reid Chan (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21481?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16704187#comment-16704187
 ] 

Reid Chan commented on HBASE-21481:
---

ping [~dbist13], [~psomogyi], WDYT.

> [acl] Superuser's permissions should not be granted or revoked by any non-su 
> global admin
> -
>
> Key: HBASE-21481
> URL: https://issues.apache.org/jira/browse/HBASE-21481
> Project: HBase
>  Issue Type: Improvement
>Reporter: Reid Chan
>Assignee: Reid Chan
>Priority: Major
>  Labels: ACL, security-issue
> Fix For: 3.0.0, 2.2.0
>
> Attachments: HBASE-21481.master.001.patch, 
> HBASE-21481.master.002.patch, HBASE-21481.master.003.patch, 
> HBASE-21481.master.004.patch, HBASE-21481.master.005.patch, 
> HBASE-21481.master.006.patch, HBASE-21481.master.007.patch, 
> HBASE-21481.master.008.patch
>
>
> Superusers are {{hbase.superuser}} listed in configuration and plus the one 
> who start master process, these two may be overlap.
> A superuser must be a global admin, but a global admin may not be a 
> superuser, possibly granted afterwards.
> For now, an non-su global admin with a Global.ADMIN permission can grant or 
> revoke any superuser's permission, accidentally or deliberately.
> The purpose of this issue is to ban this action.
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21481) [acl] Superuser's permissions should not be granted or revoked by any non-su global admin

2018-11-29 Thread Reid Chan (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21481?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16704185#comment-16704185
 ] 

Reid Chan commented on HBASE-21481:
---

Failed test passed locally, not related i think.
{code}
[INFO] ---
[INFO]  T E S T S
[INFO] ---
[INFO] Running org.apache.hadoop.hbase.client.TestRestoreSnapshotFromClientClone
[INFO] Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 81.502 s 
- in org.apache.hadoop.hbase.client.TestRestoreSnapshotFromClientClone
[INFO]
[INFO] Results:
[INFO]
[INFO] Tests run: 4, Failures: 0, Errors: 0, Skipped: 0
{code}

> [acl] Superuser's permissions should not be granted or revoked by any non-su 
> global admin
> -
>
> Key: HBASE-21481
> URL: https://issues.apache.org/jira/browse/HBASE-21481
> Project: HBase
>  Issue Type: Improvement
>Reporter: Reid Chan
>Assignee: Reid Chan
>Priority: Major
>  Labels: ACL, security-issue
> Fix For: 3.0.0, 2.2.0
>
> Attachments: HBASE-21481.master.001.patch, 
> HBASE-21481.master.002.patch, HBASE-21481.master.003.patch, 
> HBASE-21481.master.004.patch, HBASE-21481.master.005.patch, 
> HBASE-21481.master.006.patch, HBASE-21481.master.007.patch, 
> HBASE-21481.master.008.patch
>
>
> Superusers are {{hbase.superuser}} listed in configuration and plus the one 
> who start master process, these two may be overlap.
> A superuser must be a global admin, but a global admin may not be a 
> superuser, possibly granted afterwards.
> For now, an non-su global admin with a Global.ADMIN permission can grant or 
> revoke any superuser's permission, accidentally or deliberately.
> The purpose of this issue is to ban this action.
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21083) Introduce a mechanism to bypass the execution of a stuck procedure

2018-11-29 Thread Allan Yang (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21083?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16704171#comment-16704171
 ] 

Allan Yang commented on HBASE-21083:


[~xucang] this feature is mostly used by HBCK2? how are preparing to use it in 
1.x?

> Introduce a mechanism to bypass the execution of a stuck procedure
> --
>
> Key: HBASE-21083
> URL: https://issues.apache.org/jira/browse/HBASE-21083
> Project: HBase
>  Issue Type: Sub-task
>  Components: amv2
>Affects Versions: 2.1.0, 2.0.1
>Reporter: Allan Yang
>Assignee: Allan Yang
>Priority: Major
> Fix For: 3.0.0, 2.1.1, 2.0.2
>
> Attachments: HBASE-21083.branch-2.0.001.patch, 
> HBASE-21083.branch-2.0.002.patch, HBASE-21083.branch-2.0.003.patch, 
> HBASE-21083.branch-2.0.003.patch, HBASE-21083.branch-2.0.003.patch, 
> HBASE-21083.branch-2.1.001.patch
>
>
> Offline discussed with [~stack] and [~Apache9]. We all agreed that we need to 
> introduce a mechanism to 'force complete' a stuck procedure, so the AMv2 can 
> continue running.
> we still have some unrevealed bugs hiding in our AMv2 and procedureV2 system, 
> we need something to interfere with stuck procedures before HBCK2 can work. 
> This is very crucial for a production ready system.
> For now, we have little ways to interfere with running procedures. Aborting 
> them is not a good choice, since some procedures are not abort-able. And some 
> procedure may have overridden the abort() method, which will ignore the abort 
> request.
> So, here, I will introduce a mechanism to bypass the execution of a stuck 
> procedure.
> Basically, I added a field called 'bypass' to Procedure class. If we set this 
> field to true, all the logic in execute/rollback will be skipped, letting 
> this procedure and its ancestors complete normally and releasing the lock 
> resources at last.
> Notice that bypassing a procedure may leave the cluster in a middle state, 
> e.g. the region not assigned, or some hdfs files left behind. 
> The Operators need know the side effect of bypassing and recover the 
> inconsistent state of the cluster themselves, like issuing new procedures to 
> assign the regions.
> A patch will be uploaded and review board will be open. For now, only APIs in 
> ProcedureExecutor are provided. If anything is fine, I will add it to master 
> service and add a shell command to bypass a procedure. Or, maybe we can use 
> dynamically compiled JSPs to execute those APIs as mentioned in HBASE-20679.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21504) If enable FIFOCompactionPolicy, a compaction may write a "empty" hfile whose maxTimeStamp is long max. This kind of hfile will never be archived.

2018-11-29 Thread Hudson (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21504?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16704179#comment-16704179
 ] 

Hudson commented on HBASE-21504:


SUCCESS: Integrated in Jenkins build HBase-1.2-IT #1187 (See 
[https://builds.apache.org/job/HBase-1.2-IT/1187/])
HBASE-21504 If enable FIFOCompactionPolicy, a compaction may write a (openinx: 
rev f49dbae937379888653e9417a314b0b9ead28d82)
* (edit) 
hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestFIFOCompactionPolicy.java
* (edit) 
hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/compactions/FIFOCompactionPolicy.java


> If enable FIFOCompactionPolicy, a compaction may write a "empty" hfile whose 
> maxTimeStamp is long max. This kind of hfile will never be archived.
> -
>
> Key: HBASE-21504
> URL: https://issues.apache.org/jira/browse/HBASE-21504
> Project: HBase
>  Issue Type: Bug
>  Components: Compaction
>Affects Versions: 2.1.0
>Reporter: xuming
>Assignee: Zheng Hu
>Priority: Critical
> Fix For: 3.0.0, 1.5.0, 1.3.3, 2.2.0, 1.4.9, 2.1.2, 2.0.4
>
> Attachments: 1.patch, HBASE-21504.v1.patch
>
>
> When i use FIFOCompactionPolicy, and if all hfiles(>1) are TTL expired in a 
> region, once we do a compaction on the region, the compaction policy will 
> select the latest hfile to do compaction.But beacuse the latest hfile is 
> already TTL expired, compactor only write a "empty" hfile(whose entry counter 
> is 0 and maxTimeStamp is Long.MAX_VALUE) finally. Because maxTimeStamp is 
> long max, so the "empty" hfile will never be TTL expired, more seriously we 
> can not archive it by FIFOCompactionPolicy forever.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Comment Edited] (HBASE-21083) Introduce a mechanism to bypass the execution of a stuck procedure

2018-11-29 Thread Allan Yang (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21083?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16704171#comment-16704171
 ] 

Allan Yang edited comment on HBASE-21083 at 11/30/18 2:40 AM:
--

[~xucang] this feature is mostly used by HBCK2. how are preparing to use it in 
1.x?


was (Author: allan163):
[~xucang] this feature is mostly used by HBCK2? how are preparing to use it in 
1.x?

> Introduce a mechanism to bypass the execution of a stuck procedure
> --
>
> Key: HBASE-21083
> URL: https://issues.apache.org/jira/browse/HBASE-21083
> Project: HBase
>  Issue Type: Sub-task
>  Components: amv2
>Affects Versions: 2.1.0, 2.0.1
>Reporter: Allan Yang
>Assignee: Allan Yang
>Priority: Major
> Fix For: 3.0.0, 2.1.1, 2.0.2
>
> Attachments: HBASE-21083.branch-2.0.001.patch, 
> HBASE-21083.branch-2.0.002.patch, HBASE-21083.branch-2.0.003.patch, 
> HBASE-21083.branch-2.0.003.patch, HBASE-21083.branch-2.0.003.patch, 
> HBASE-21083.branch-2.1.001.patch
>
>
> Offline discussed with [~stack] and [~Apache9]. We all agreed that we need to 
> introduce a mechanism to 'force complete' a stuck procedure, so the AMv2 can 
> continue running.
> we still have some unrevealed bugs hiding in our AMv2 and procedureV2 system, 
> we need something to interfere with stuck procedures before HBCK2 can work. 
> This is very crucial for a production ready system.
> For now, we have little ways to interfere with running procedures. Aborting 
> them is not a good choice, since some procedures are not abort-able. And some 
> procedure may have overridden the abort() method, which will ignore the abort 
> request.
> So, here, I will introduce a mechanism to bypass the execution of a stuck 
> procedure.
> Basically, I added a field called 'bypass' to Procedure class. If we set this 
> field to true, all the logic in execute/rollback will be skipped, letting 
> this procedure and its ancestors complete normally and releasing the lock 
> resources at last.
> Notice that bypassing a procedure may leave the cluster in a middle state, 
> e.g. the region not assigned, or some hdfs files left behind. 
> The Operators need know the side effect of bypassing and recover the 
> inconsistent state of the cluster themselves, like issuing new procedures to 
> assign the regions.
> A patch will be uploaded and review board will be open. For now, only APIs in 
> ProcedureExecutor are provided. If anything is fine, I will add it to master 
> service and add a shell command to bypass a procedure. Or, maybe we can use 
> dynamically compiled JSPs to execute those APIs as mentioned in HBASE-20679.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Comment Edited] (HBASE-21083) Introduce a mechanism to bypass the execution of a stuck procedure

2018-11-29 Thread Allan Yang (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21083?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16704171#comment-16704171
 ] 

Allan Yang edited comment on HBASE-21083 at 11/30/18 2:40 AM:
--

[~xucang] this feature is mostly used by HBCK2. how are you preparing to use it 
in 1.x?


was (Author: allan163):
[~xucang] this feature is mostly used by HBCK2. how are preparing to use it in 
1.x?

> Introduce a mechanism to bypass the execution of a stuck procedure
> --
>
> Key: HBASE-21083
> URL: https://issues.apache.org/jira/browse/HBASE-21083
> Project: HBase
>  Issue Type: Sub-task
>  Components: amv2
>Affects Versions: 2.1.0, 2.0.1
>Reporter: Allan Yang
>Assignee: Allan Yang
>Priority: Major
> Fix For: 3.0.0, 2.1.1, 2.0.2
>
> Attachments: HBASE-21083.branch-2.0.001.patch, 
> HBASE-21083.branch-2.0.002.patch, HBASE-21083.branch-2.0.003.patch, 
> HBASE-21083.branch-2.0.003.patch, HBASE-21083.branch-2.0.003.patch, 
> HBASE-21083.branch-2.1.001.patch
>
>
> Offline discussed with [~stack] and [~Apache9]. We all agreed that we need to 
> introduce a mechanism to 'force complete' a stuck procedure, so the AMv2 can 
> continue running.
> we still have some unrevealed bugs hiding in our AMv2 and procedureV2 system, 
> we need something to interfere with stuck procedures before HBCK2 can work. 
> This is very crucial for a production ready system.
> For now, we have little ways to interfere with running procedures. Aborting 
> them is not a good choice, since some procedures are not abort-able. And some 
> procedure may have overridden the abort() method, which will ignore the abort 
> request.
> So, here, I will introduce a mechanism to bypass the execution of a stuck 
> procedure.
> Basically, I added a field called 'bypass' to Procedure class. If we set this 
> field to true, all the logic in execute/rollback will be skipped, letting 
> this procedure and its ancestors complete normally and releasing the lock 
> resources at last.
> Notice that bypassing a procedure may leave the cluster in a middle state, 
> e.g. the region not assigned, or some hdfs files left behind. 
> The Operators need know the side effect of bypassing and recover the 
> inconsistent state of the cluster themselves, like issuing new procedures to 
> assign the regions.
> A patch will be uploaded and review board will be open. For now, only APIs in 
> ProcedureExecutor are provided. If anything is fine, I will add it to master 
> service and add a shell command to bypass a procedure. Or, maybe we can use 
> dynamically compiled JSPs to execute those APIs as mentioned in HBASE-20679.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-21504) If enable FIFOCompactionPolicy, a compaction may write a "empty" hfile whose maxTimeStamp is long max. This kind of hfile will never be archived.

2018-11-29 Thread Zheng Hu (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-21504?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zheng Hu updated HBASE-21504:
-
  Resolution: Fixed
Hadoop Flags: Reviewed
  Status: Resolved  (was: Patch Available)

> If enable FIFOCompactionPolicy, a compaction may write a "empty" hfile whose 
> maxTimeStamp is long max. This kind of hfile will never be archived.
> -
>
> Key: HBASE-21504
> URL: https://issues.apache.org/jira/browse/HBASE-21504
> Project: HBase
>  Issue Type: Bug
>  Components: Compaction
>Affects Versions: 2.1.0
>Reporter: xuming
>Assignee: Zheng Hu
>Priority: Critical
> Fix For: 3.0.0, 1.5.0, 1.3.3, 2.2.0, 1.4.9, 2.1.2, 2.0.4
>
> Attachments: 1.patch, HBASE-21504.v1.patch
>
>
> When i use FIFOCompactionPolicy, and if all hfiles(>1) are TTL expired in a 
> region, once we do a compaction on the region, the compaction policy will 
> select the latest hfile to do compaction.But beacuse the latest hfile is 
> already TTL expired, compactor only write a "empty" hfile(whose entry counter 
> is 0 and maxTimeStamp is Long.MAX_VALUE) finally. Because maxTimeStamp is 
> long max, so the "empty" hfile will never be TTL expired, more seriously we 
> can not archive it by FIFOCompactionPolicy forever.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21504) If enable FIFOCompactionPolicy, a compaction may write a "empty" hfile whose maxTimeStamp is long max. This kind of hfile will never be archived.

2018-11-29 Thread Hudson (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21504?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16704167#comment-16704167
 ] 

Hudson commented on HBASE-21504:


SUCCESS: Integrated in Jenkins build HBase-1.3-IT #506 (See 
[https://builds.apache.org/job/HBase-1.3-IT/506/])
HBASE-21504 If enable FIFOCompactionPolicy, a compaction may write a (openinx: 
rev 57222ff0aa66047fc405c488f86a27b622beea64)
* (edit) 
hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/compactions/FIFOCompactionPolicy.java
* (edit) 
hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/compactions/TestFIFOCompactionPolicy.java


> If enable FIFOCompactionPolicy, a compaction may write a "empty" hfile whose 
> maxTimeStamp is long max. This kind of hfile will never be archived.
> -
>
> Key: HBASE-21504
> URL: https://issues.apache.org/jira/browse/HBASE-21504
> Project: HBase
>  Issue Type: Bug
>  Components: Compaction
>Affects Versions: 2.1.0
>Reporter: xuming
>Assignee: Zheng Hu
>Priority: Critical
> Fix For: 3.0.0, 1.5.0, 1.3.3, 2.2.0, 1.4.9, 2.1.2, 2.0.4
>
> Attachments: 1.patch, HBASE-21504.v1.patch
>
>
> When i use FIFOCompactionPolicy, and if all hfiles(>1) are TTL expired in a 
> region, once we do a compaction on the region, the compaction policy will 
> select the latest hfile to do compaction.But beacuse the latest hfile is 
> already TTL expired, compactor only write a "empty" hfile(whose entry counter 
> is 0 and maxTimeStamp is Long.MAX_VALUE) finally. Because maxTimeStamp is 
> long max, so the "empty" hfile will never be TTL expired, more seriously we 
> can not archive it by FIFOCompactionPolicy forever.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21530) Abort_Procedure should be able to take a list of proc IDs

2018-11-29 Thread Allan Yang (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21530?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16704164#comment-16704164
 ] 

Allan Yang commented on HBASE-21530:


What version of HBase are you using?abort procedure should deprecated in 
HBase2.0+, this feature is problematic, most of the procedure is not abortable.

> Abort_Procedure should be able to take a list of proc IDs
> -
>
> Key: HBASE-21530
> URL: https://issues.apache.org/jira/browse/HBASE-21530
> Project: HBase
>  Issue Type: Improvement
>Reporter: Geoffrey Jacoby
>Priority: Minor
>
> As a convenience, it would be helpful if the HBase shell's abort_procedure 
> call had the option of taking in multiple procedure ids at the same time, 
> rather than relying on operators to use a loop in an external script. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21504) If enable FIFOCompactionPolicy, a compaction may write a "empty" hfile whose maxTimeStamp is long max. This kind of hfile will never be archived.

2018-11-29 Thread Zheng Hu (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21504?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16704160#comment-16704160
 ] 

Zheng Hu commented on HBASE-21504:
--

Pushed to branch-2.x & branch-1.4 & branch-1.3 branch-1.2 & branch-1.  Thanks 
[~allan163] for reviewing. 

> If enable FIFOCompactionPolicy, a compaction may write a "empty" hfile whose 
> maxTimeStamp is long max. This kind of hfile will never be archived.
> -
>
> Key: HBASE-21504
> URL: https://issues.apache.org/jira/browse/HBASE-21504
> Project: HBase
>  Issue Type: Bug
>  Components: Compaction
>Affects Versions: 2.1.0
>Reporter: xuming
>Assignee: Zheng Hu
>Priority: Critical
> Fix For: 3.0.0, 1.5.0, 1.3.3, 2.2.0, 1.4.9, 2.1.2, 2.0.4
>
> Attachments: 1.patch, HBASE-21504.v1.patch
>
>
> When i use FIFOCompactionPolicy, and if all hfiles(>1) are TTL expired in a 
> region, once we do a compaction on the region, the compaction policy will 
> select the latest hfile to do compaction.But beacuse the latest hfile is 
> already TTL expired, compactor only write a "empty" hfile(whose entry counter 
> is 0 and maxTimeStamp is Long.MAX_VALUE) finally. Because maxTimeStamp is 
> long max, so the "empty" hfile will never be TTL expired, more seriously we 
> can not archive it by FIFOCompactionPolicy forever.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (HBASE-21533) Add a section in our ref guide for the folding(removal) of hbase:namespace table

2018-11-29 Thread Duo Zhang (JIRA)
Duo Zhang created HBASE-21533:
-

 Summary: Add a section in our ref guide for the folding(removal) 
of hbase:namespace table
 Key: HBASE-21533
 URL: https://issues.apache.org/jira/browse/HBASE-21533
 Project: HBase
  Issue Type: Task
  Components: documentation
Reporter: Duo Zhang
 Fix For: 3.0.0






--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (HBASE-21532) HFileCorruptionChecker#checkHFile does not enumerate the whole file

2018-11-29 Thread Andrew Purtell (JIRA)
Andrew Purtell created HBASE-21532:
--

 Summary: HFileCorruptionChecker#checkHFile does not enumerate the 
whole file
 Key: HBASE-21532
 URL: https://issues.apache.org/jira/browse/HBASE-21532
 Project: HBase
  Issue Type: Bug
  Components: hbck, hbck2
Reporter: Andrew Purtell


HFileCorruptionChecker#checkHFile can potentially miss some cases of file 
corruption because it does not enumerate all the data blocks of the file. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21464) Splitting blocked with meta NSRE during split transaction

2018-11-29 Thread Andrew Purtell (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21464?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16704102#comment-16704102
 ] 

Andrew Purtell commented on HBASE-21464:


[~lhofhansl] Per [~allan163]'s suggestion no changes were made in that regard 
after all. The code isn't further away from the place where it was before. Same 
number of assumptions. 

> Splitting blocked with meta NSRE during split transaction
> -
>
> Key: HBASE-21464
> URL: https://issues.apache.org/jira/browse/HBASE-21464
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.5.0, 1.4.3, 1.4.4, 1.4.5, 1.4.6, 1.4.8, 1.4.7
>Reporter: Andrew Purtell
>Assignee: Andrew Purtell
>Priority: Blocker
> Fix For: 1.5.0, 1.4.9
>
> Attachments: HBASE-21464-branch-1.patch, HBASE-21464-branch-1.patch
>
>
> Splitting is blocked during split transaction. The split worker is trying to 
> update meta but isn't able to relocate it after NSRE:
> {noformat}
> 2018-11-09 17:50:45,277 INFO  
> [regionserver/ip-172-31-5-92.us-west-2.compute.internal/172.31.5.92:8120-splits-1541785709434]
>  client.RpcRetryingCaller: Call exception, tries=13, retries=350, 
> started=88590 ms ago, cancelled=false, 
> msg=org.apache.hadoop.hbase.NotServingRegionException: Region hbase:meta,,1 
> is not online on ip-172-31-13-83.us-west-2.compute.internal,8120,1541785618832
>      at 
> org.apache.hadoop.hbase.regionserver.HRegionServer.getRegionByEncodedName(HRegionServer.java:3088)
>         at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.getRegion(RSRpcServices.java:1271)
>         at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:2198)
>         at 
> org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:36617)
>         at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2396)
>         at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124)
>         at 
> org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:297)
>         at 
> org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:277)row 
> 'test,,1541785709452.5ba6596f0050c2dab969d152829227c6.44' on table 
> 'hbase:meta' at region=hbase:meta,1.1588230740, 
> hostname=ip-172-31-15-225.us-west-2.compute.internal,8120,1541785640586, 
> seqNum=0{noformat}
> Clients, in this case YCSB, are hung with part of the keyspace missing:
> {noformat}
> 2018-11-09 17:51:06,033 DEBUG [hconnection-0x5739e567-shared--pool1-t165] 
> client.ConnectionManager$HConnectionImplementation: locateRegionInMeta 
> parentTable=hbase:meta, metaLocation=, attempt=14 of 35 failed; retrying 
> after sleep of 20158 because: No server address listed in hbase:meta for 
> region 
> test,user307326104267982763,1541785754600.ef90030b05cb02305b75e9bfbc3ee081. 
> containing row user3301635648728421323{noformat}
> Balancing cannot run indefinitely because the split transaction is stuck
> {noformat}
> 2018-11-09 17:49:55,478 DEBUG 
> [RpcServer.default.FPBQ.Fifo.handler=29,queue=2,port=8100] master.HMaster: 
> Not running balancer because 3 region(s) in transition: 
> [{ef90030b05cb02305b75e9bfbc3ee081 state=SPLITTING_NEW, ts=1541785754606, 
> server=ip-172-31-5-92.us-west-2.compute.internal,8120,1541785626417}, 
> {5ba6596f0050c2dab969d152829227c6 state=SPLITTING, ts=1541785754606, 
> server=ip-172-31-5-92.us-west-2.compute{noformat}
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-21464) Splitting blocked with meta NSRE during split transaction

2018-11-29 Thread Andrew Purtell (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-21464?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Purtell updated HBASE-21464:
---
  Resolution: Fixed
Hadoop Flags: Reviewed
  Status: Resolved  (was: Patch Available)

Pushed to branch-1.4 and branch-1

> Splitting blocked with meta NSRE during split transaction
> -
>
> Key: HBASE-21464
> URL: https://issues.apache.org/jira/browse/HBASE-21464
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.5.0, 1.4.3, 1.4.4, 1.4.5, 1.4.6, 1.4.8, 1.4.7
>Reporter: Andrew Purtell
>Assignee: Andrew Purtell
>Priority: Blocker
> Fix For: 1.5.0, 1.4.9
>
> Attachments: HBASE-21464-branch-1.patch, HBASE-21464-branch-1.patch
>
>
> Splitting is blocked during split transaction. The split worker is trying to 
> update meta but isn't able to relocate it after NSRE:
> {noformat}
> 2018-11-09 17:50:45,277 INFO  
> [regionserver/ip-172-31-5-92.us-west-2.compute.internal/172.31.5.92:8120-splits-1541785709434]
>  client.RpcRetryingCaller: Call exception, tries=13, retries=350, 
> started=88590 ms ago, cancelled=false, 
> msg=org.apache.hadoop.hbase.NotServingRegionException: Region hbase:meta,,1 
> is not online on ip-172-31-13-83.us-west-2.compute.internal,8120,1541785618832
>      at 
> org.apache.hadoop.hbase.regionserver.HRegionServer.getRegionByEncodedName(HRegionServer.java:3088)
>         at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.getRegion(RSRpcServices.java:1271)
>         at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:2198)
>         at 
> org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:36617)
>         at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2396)
>         at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124)
>         at 
> org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:297)
>         at 
> org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:277)row 
> 'test,,1541785709452.5ba6596f0050c2dab969d152829227c6.44' on table 
> 'hbase:meta' at region=hbase:meta,1.1588230740, 
> hostname=ip-172-31-15-225.us-west-2.compute.internal,8120,1541785640586, 
> seqNum=0{noformat}
> Clients, in this case YCSB, are hung with part of the keyspace missing:
> {noformat}
> 2018-11-09 17:51:06,033 DEBUG [hconnection-0x5739e567-shared--pool1-t165] 
> client.ConnectionManager$HConnectionImplementation: locateRegionInMeta 
> parentTable=hbase:meta, metaLocation=, attempt=14 of 35 failed; retrying 
> after sleep of 20158 because: No server address listed in hbase:meta for 
> region 
> test,user307326104267982763,1541785754600.ef90030b05cb02305b75e9bfbc3ee081. 
> containing row user3301635648728421323{noformat}
> Balancing cannot run indefinitely because the split transaction is stuck
> {noformat}
> 2018-11-09 17:49:55,478 DEBUG 
> [RpcServer.default.FPBQ.Fifo.handler=29,queue=2,port=8100] master.HMaster: 
> Not running balancer because 3 region(s) in transition: 
> [{ef90030b05cb02305b75e9bfbc3ee081 state=SPLITTING_NEW, ts=1541785754606, 
> server=ip-172-31-5-92.us-west-2.compute.internal,8120,1541785626417}, 
> {5ba6596f0050c2dab969d152829227c6 state=SPLITTING, ts=1541785754606, 
> server=ip-172-31-5-92.us-west-2.compute{noformat}
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21531) race between region report and region move causes master to kill RS

2018-11-29 Thread Sergey Shelukhin (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21531?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16704078#comment-16704078
 ] 

Sergey Shelukhin commented on HBASE-21531:
--

It's the snapshot of 3.0 before HBASE-21421. That fix looks a little bit 
suspect to me... what if the state of RS is actually invalid and it's serving 
some region it shouldn't be serving? Master should at least try to close the 
region again.

> race between region report and region move causes master to kill RS
> ---
>
> Key: HBASE-21531
> URL: https://issues.apache.org/jira/browse/HBASE-21531
> Project: HBase
>  Issue Type: Bug
>Reporter: Sergey Shelukhin
>Priority: Major
>
> In this case the delay between the ack from RS and the region report was 
> 1.5s, so I'm not sure what caused the race (network hiccup? unreported retry 
> by protobuf transport?) but in any case I don't see anything that prevents 
> this from happening in a normal case, with a narrowed time window. Any delay 
> (e.g. a GC pause on RS right after the report is built, and ack is sent for 
> the close) or retries expands the window.
> Master starts moving the region and the source RS acks by 21:51,206
> {noformat}
> Master:
> 2018-11-21 21:21:49,024 INFO  [master/6:17000.Chore.1] master.HMaster: 
> balance hri=, source=,17020,1542754626176, destination= 2>,17020,1542863268158
> ...
> Server:
> 2018-11-21 21:21:49,394 INFO  [RS_CLOSE_REGION-regionserver/:17020-1] 
> handler.UnassignRegionHandler: Close 
> ...
> 2018-11-21 21:21:51,095 INFO  [RS_CLOSE_REGION-regionserver/:17020-1] 
> handler.UnassignRegionHandler: Closed 
> {noformat}
> By then the region is removed from onlineRegions, so the master proceeds.
> {noformat}
> 2018-11-21 21:21:51,206 INFO  [PEWorker-4] procedure2.ProcedureExecutor: 
> Finished subprocedure(s) of pid=667, 
> state=RUNNABLE:REGION_STATE_TRANSITION_CONFIRM_CLOSED, hasLock=true; 
> TransitRegionStateProcedure table=, region=, REOPEN/MOVE; 
> resume parent processing.
> 2018-11-21 21:21:51,386 INFO  [PEWorker-13] assignment.RegionStateStore: 
> pid=667 updating hbase:meta row=, regionState=OPENING, 
> regionLocation=,17020,1542863268158
> {noformat}
> There are no obvious errors/delays that I see in RS log, and it doesn't log 
> starting to send the report.
> However, at 21:52.853 the report is processed that still contains this region.
> {noformat}
> 2018-11-21 21:21:52,853 WARN  
> [RpcServer.default.FPBQ.Fifo.handler=48,queue=3,port=17000] 
> assignment.AssignmentManager: Killing ,17020,1542754626176: 
> rit=OPENING, location=,17020,1542863268158, table=, 
> region= reported OPEN on server=,17020,1542754626176 but 
> state has otherwise.
> * ABORTING region server ,17020,1542754626176: 
> org.apache.hadoop.hbase.YouAreDeadException: rit=OPENING, location= 2>,17020,1542863268158, table=, region= reported OPEN on 
> server=,17020,1542754626176 but state has otherwise.
> {noformat}
> RS shuts down in an orderly manner and it can be seen from the log that this 
> region is actually not present (there's no line indicating it's being closed, 
> unlike for other regions).
> I think there needs to be some sort of versioning for region operations 
> and/or in RS reports to allow master to account for concurrent operations and 
> avoid races. Probably per region with either a grace period or an additional 
> global version, so that master could avoid killing RS based on stale reports, 
> but still kill RS if it did retain an old version of the region state due to 
> some bug after acking a new version.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21464) Splitting blocked with meta NSRE during split transaction

2018-11-29 Thread Lars Hofhansl (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21464?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16704076#comment-16704076
 ] 

Lars Hofhansl commented on HBASE-21464:
---

+1

(as a small comment: Will this remove us even further from a splittable meta by 
baking in more assumptions about a single region meta?)

> Splitting blocked with meta NSRE during split transaction
> -
>
> Key: HBASE-21464
> URL: https://issues.apache.org/jira/browse/HBASE-21464
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.5.0, 1.4.3, 1.4.4, 1.4.5, 1.4.6, 1.4.8, 1.4.7
>Reporter: Andrew Purtell
>Assignee: Andrew Purtell
>Priority: Blocker
> Fix For: 1.5.0, 1.4.9
>
> Attachments: HBASE-21464-branch-1.patch, HBASE-21464-branch-1.patch
>
>
> Splitting is blocked during split transaction. The split worker is trying to 
> update meta but isn't able to relocate it after NSRE:
> {noformat}
> 2018-11-09 17:50:45,277 INFO  
> [regionserver/ip-172-31-5-92.us-west-2.compute.internal/172.31.5.92:8120-splits-1541785709434]
>  client.RpcRetryingCaller: Call exception, tries=13, retries=350, 
> started=88590 ms ago, cancelled=false, 
> msg=org.apache.hadoop.hbase.NotServingRegionException: Region hbase:meta,,1 
> is not online on ip-172-31-13-83.us-west-2.compute.internal,8120,1541785618832
>      at 
> org.apache.hadoop.hbase.regionserver.HRegionServer.getRegionByEncodedName(HRegionServer.java:3088)
>         at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.getRegion(RSRpcServices.java:1271)
>         at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:2198)
>         at 
> org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:36617)
>         at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2396)
>         at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124)
>         at 
> org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:297)
>         at 
> org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:277)row 
> 'test,,1541785709452.5ba6596f0050c2dab969d152829227c6.44' on table 
> 'hbase:meta' at region=hbase:meta,1.1588230740, 
> hostname=ip-172-31-15-225.us-west-2.compute.internal,8120,1541785640586, 
> seqNum=0{noformat}
> Clients, in this case YCSB, are hung with part of the keyspace missing:
> {noformat}
> 2018-11-09 17:51:06,033 DEBUG [hconnection-0x5739e567-shared--pool1-t165] 
> client.ConnectionManager$HConnectionImplementation: locateRegionInMeta 
> parentTable=hbase:meta, metaLocation=, attempt=14 of 35 failed; retrying 
> after sleep of 20158 because: No server address listed in hbase:meta for 
> region 
> test,user307326104267982763,1541785754600.ef90030b05cb02305b75e9bfbc3ee081. 
> containing row user3301635648728421323{noformat}
> Balancing cannot run indefinitely because the split transaction is stuck
> {noformat}
> 2018-11-09 17:49:55,478 DEBUG 
> [RpcServer.default.FPBQ.Fifo.handler=29,queue=2,port=8100] master.HMaster: 
> Not running balancer because 3 region(s) in transition: 
> [{ef90030b05cb02305b75e9bfbc3ee081 state=SPLITTING_NEW, ts=1541785754606, 
> server=ip-172-31-5-92.us-west-2.compute.internal,8120,1541785626417}, 
> {5ba6596f0050c2dab969d152829227c6 state=SPLITTING, ts=1541785754606, 
> server=ip-172-31-5-92.us-west-2.compute{noformat}
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21531) race between region report and region move causes master to kill RS

2018-11-29 Thread Duo Zhang (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21531?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16704071#comment-16704071
 ] 

Duo Zhang commented on HBASE-21531:
---

What is the HBase version?

I think HBASE-21421 is the solution, where we just remove the killing of RS in 
reportOnlineRegions...

> race between region report and region move causes master to kill RS
> ---
>
> Key: HBASE-21531
> URL: https://issues.apache.org/jira/browse/HBASE-21531
> Project: HBase
>  Issue Type: Bug
>Reporter: Sergey Shelukhin
>Priority: Major
>
> In this case the delay between the ack from RS and the region report was 
> 1.5s, so I'm not sure what caused the race (network hiccup? unreported retry 
> by protobuf transport?) but in any case I don't see anything that prevents 
> this from happening in a normal case, with a narrowed time window. Any delay 
> (e.g. a GC pause on RS right after the report is built, and ack is sent for 
> the close) or retries expands the window.
> Master starts moving the region and the source RS acks by 21:51,206
> {noformat}
> Master:
> 2018-11-21 21:21:49,024 INFO  [master/6:17000.Chore.1] master.HMaster: 
> balance hri=, source=,17020,1542754626176, destination= 2>,17020,1542863268158
> ...
> Server:
> 2018-11-21 21:21:49,394 INFO  [RS_CLOSE_REGION-regionserver/:17020-1] 
> handler.UnassignRegionHandler: Close 
> ...
> 2018-11-21 21:21:51,095 INFO  [RS_CLOSE_REGION-regionserver/:17020-1] 
> handler.UnassignRegionHandler: Closed 
> {noformat}
> By then the region is removed from onlineRegions, so the master proceeds.
> {noformat}
> 2018-11-21 21:21:51,206 INFO  [PEWorker-4] procedure2.ProcedureExecutor: 
> Finished subprocedure(s) of pid=667, 
> state=RUNNABLE:REGION_STATE_TRANSITION_CONFIRM_CLOSED, hasLock=true; 
> TransitRegionStateProcedure table=, region=, REOPEN/MOVE; 
> resume parent processing.
> 2018-11-21 21:21:51,386 INFO  [PEWorker-13] assignment.RegionStateStore: 
> pid=667 updating hbase:meta row=, regionState=OPENING, 
> regionLocation=,17020,1542863268158
> {noformat}
> There are no obvious errors/delays that I see in RS log, and it doesn't log 
> starting to send the report.
> However, at 21:52.853 the report is processed that still contains this region.
> {noformat}
> 2018-11-21 21:21:52,853 WARN  
> [RpcServer.default.FPBQ.Fifo.handler=48,queue=3,port=17000] 
> assignment.AssignmentManager: Killing ,17020,1542754626176: 
> rit=OPENING, location=,17020,1542863268158, table=, 
> region= reported OPEN on server=,17020,1542754626176 but 
> state has otherwise.
> * ABORTING region server ,17020,1542754626176: 
> org.apache.hadoop.hbase.YouAreDeadException: rit=OPENING, location= 2>,17020,1542863268158, table=, region= reported OPEN on 
> server=,17020,1542754626176 but state has otherwise.
> {noformat}
> RS shuts down in an orderly manner and it can be seen from the log that this 
> region is actually not present (there's no line indicating it's being closed, 
> unlike for other regions).
> I think there needs to be some sort of versioning for region operations 
> and/or in RS reports to allow master to account for concurrent operations and 
> avoid races. Probably per region with either a grace period or an additional 
> global version, so that master could avoid killing RS based on stale reports, 
> but still kill RS if it did retain an old version of the region state due to 
> some bug after acking a new version.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-17914) Create a new reader instead of cloning a new StoreFile when compaction

2018-11-29 Thread Duo Zhang (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-17914?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16704064#comment-16704064
 ] 

Duo Zhang commented on HBASE-17914:
---

The preStoreFileReaderOpen method has already been marked as Deprecated, which 
means this method has already exposed too many internal states and we want to 
stop doing this in the future.

For now imaybe it is fine to add these parameters as this method is only 
suppose to be used by Phoenix , but could you please find another way to 
implement the feature? The StoreFileReader is IA.Private so the methods and 
behavior will not be stable...

Thanks.

> Create a new reader instead of cloning a new StoreFile when compaction
> --
>
> Key: HBASE-17914
> URL: https://issues.apache.org/jira/browse/HBASE-17914
> Project: HBase
>  Issue Type: Sub-task
>  Components: Compaction, regionserver
>Affects Versions: 2.0.0
>Reporter: Duo Zhang
>Assignee: Duo Zhang
>Priority: Major
> Fix For: 2.0.0
>
> Attachments: HBASE-17914-v1.patch, HBASE-17914-v1.patch, 
> HBASE-17914-v2.patch, HBASE-17914-v3.patch, HBASE-17914.patch
>
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-21515) Also initialize an AsyncClusterConnection in HRegionServer

2018-11-29 Thread Duo Zhang (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-21515?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Duo Zhang updated HBASE-21515:
--
Attachment: HBASE-21515-v4.patch

> Also initialize an AsyncClusterConnection in HRegionServer
> --
>
> Key: HBASE-21515
> URL: https://issues.apache.org/jira/browse/HBASE-21515
> Project: HBase
>  Issue Type: Sub-task
>  Components: regionserver
>Reporter: Duo Zhang
>Assignee: Duo Zhang
>Priority: Major
> Attachments: HBASE-21515-v1.patch, HBASE-21515-v2.patch, 
> HBASE-21515-v3.patch, HBASE-21515-v4.patch, HBASE-21515.patch
>
>
> Plan to do this incrementally, so first we will only introduce the new 
> AsyncClusterConnection, without deleting the old ClusterConnection. And we 
> can move the code which uses the ClusterConnection to use 
> AsyncClusterConnection, through different sub tasks here. And finally we can 
> remove the ClusterConnection completely.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-21531) race between region report and region move causes master to kill RS

2018-11-29 Thread Sergey Shelukhin (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-21531?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Shelukhin updated HBASE-21531:
-
Description: 
In this case the delay between the ack from RS and the region report was 1.5s, 
so I'm not sure what caused the race (network hiccup? unreported retry by 
protobuf transport?) but in any case I don't see anything that prevents this 
from happening in a normal case, with a narrowed time window. Any delay (e.g. a 
GC pause on RS right after the report is built, and ack is sent for the close) 
or retries expands the window.


Master starts moving the region and the source RS acks by 21:51,206
{noformat}
Master:
2018-11-21 21:21:49,024 INFO  [master/6:17000.Chore.1] master.HMaster: balance 
hri=, source=,17020,1542754626176, destination=,17020,1542863268158
...
Server:
2018-11-21 21:21:49,394 INFO  [RS_CLOSE_REGION-regionserver/:17020-1] 
handler.UnassignRegionHandler: Close 
...
2018-11-21 21:21:51,095 INFO  [RS_CLOSE_REGION-regionserver/:17020-1] 
handler.UnassignRegionHandler: Closed 
{noformat}
By then the region is removed from onlineRegions, so the master proceeds.
{noformat}
2018-11-21 21:21:51,206 INFO  [PEWorker-4] procedure2.ProcedureExecutor: 
Finished subprocedure(s) of pid=667, 
state=RUNNABLE:REGION_STATE_TRANSITION_CONFIRM_CLOSED, hasLock=true; 
TransitRegionStateProcedure table=, region=, REOPEN/MOVE; 
resume parent processing.
2018-11-21 21:21:51,386 INFO  [PEWorker-13] assignment.RegionStateStore: 
pid=667 updating hbase:meta row=, regionState=OPENING, 
regionLocation=,17020,1542863268158
{noformat}

There are no obvious errors/delays that I see in RS log, and it doesn't log 
starting to send the report.
However, at 21:52.853 the report is processed that still contains this region.
{noformat}
2018-11-21 21:21:52,853 WARN  
[RpcServer.default.FPBQ.Fifo.handler=48,queue=3,port=17000] 
assignment.AssignmentManager: Killing ,17020,1542754626176: rit=OPENING, 
location=,17020,1542863268158, table=, region= 
reported OPEN on server=,17020,1542754626176 but state has otherwise.
* ABORTING region server ,17020,1542754626176: 
org.apache.hadoop.hbase.YouAreDeadException: rit=OPENING, location=,17020,1542863268158, table=, region= reported OPEN on 
server=,17020,1542754626176 but state has otherwise.
{noformat}

RS shuts down in an orderly manner and it can be seen from the log that this 
region is actually not present (there's no line indicating it's being closed, 
unlike for other regions).

I think there needs to be some sort of versioning for region operations and/or 
in RS reports to allow master to account for concurrent operations and avoid 
races. Probably per region with either a grace period or an additional global 
version, so that master could avoid killing RS based on stale reports, but 
still kill RS if it did retain an old version of the region state due to some 
bug after acking a new version.

  was:
In this case the delay between the ack from RS and the region report was 1.5s, 
so I'm not sure what caused the race (network hiccup? unreported retry by 
protobuf transport?) but in any case I don't see anything that prevents this 
from happening in a normal case, with a narrowed time window. Any delay (e.g. a 
GC pause on RS right after the report is built, and ack is sent for the close) 
or retries expands the window.


Master starts moving the region and the source RS acks by 21:51,206
{noformat}
Master:
2018-11-21 21:21:49,024 INFO  [master/6:17000.Chore.1] master.HMaster: balance 
hri=, source=,17020,1542754626176, destination=,17020,1542863268158
...
Server:
2018-11-21 21:21:49,394 INFO  [RS_CLOSE_REGION-regionserver/:17020-1] 
handler.UnassignRegionHandler: Close 
...
2018-11-21 21:21:51,095 INFO  [RS_CLOSE_REGION-regionserver/:17020-1] 
handler.UnassignRegionHandler: Closed 
{noformat}
By then the region is removed from onlineRegions, so the master proceeds.
{noformat}
2018-11-21 21:21:51,206 INFO  [PEWorker-4] procedure2.ProcedureExecutor: 
Finished subprocedure(s) of pid=667, 
state=RUNNABLE:REGION_STATE_TRANSITION_CONFIRM_CLOSED, hasLock=true; 
TransitRegionStateProcedure table=dummy_table, region=, 
REOPEN/MOVE; resume parent processing.
2018-11-21 21:21:51,386 INFO  [PEWorker-13] assignment.RegionStateStore: 
pid=667 updating hbase:meta row=, regionState=OPENING, 
regionLocation=,17020,1542863268158
{noformat}

There are no obvious errors/delays that I see in RS log, and it doesn't log 
starting to send the report.
However, at 21:52.853 the report is processed that still contains this region.
{noformat}
2018-11-21 21:21:52,853 WARN  
[RpcServer.default.FPBQ.Fifo.handler=48,queue=3,port=17000] 
assignment.AssignmentManager: Killing ,17020,1542754626176: rit=OPENING, 
location=,17020,1542863268158, table=dummy_table, region= 
reported OPEN on server=,17020,1542754626176 but state has otherwise.
* ABORTING region server ,17020,1542754626176: 

[jira] [Created] (HBASE-21531) race between region report and region move causes master to kill RS

2018-11-29 Thread Sergey Shelukhin (JIRA)
Sergey Shelukhin created HBASE-21531:


 Summary: race between region report and region move causes master 
to kill RS
 Key: HBASE-21531
 URL: https://issues.apache.org/jira/browse/HBASE-21531
 Project: HBase
  Issue Type: Bug
Reporter: Sergey Shelukhin


In this case the delay between the ack from RS and the region report was 1.5s, 
so I'm not sure what caused the race (network hiccup? unreported retry by 
protobuf transport?) but in any case I don't see anything that prevents this 
from happening in a normal case, with a narrowed time window. Any delay (e.g. a 
GC pause on RS right after the report is built, and ack is sent for the close) 
or retries expands the window.


Master starts moving the region and the source RS acks by 21:51,206
{noformat}
Master:
2018-11-21 21:21:49,024 INFO  [master/6:17000.Chore.1] master.HMaster: balance 
hri=, source=,17020,1542754626176, destination=,17020,1542863268158
...
Server:
2018-11-21 21:21:49,394 INFO  [RS_CLOSE_REGION-regionserver/:17020-1] 
handler.UnassignRegionHandler: Close 
...
2018-11-21 21:21:51,095 INFO  [RS_CLOSE_REGION-regionserver/:17020-1] 
handler.UnassignRegionHandler: Closed 
{noformat}
By then the region is removed from onlineRegions, so the master proceeds.
{noformat}
2018-11-21 21:21:51,206 INFO  [PEWorker-4] procedure2.ProcedureExecutor: 
Finished subprocedure(s) of pid=667, 
state=RUNNABLE:REGION_STATE_TRANSITION_CONFIRM_CLOSED, hasLock=true; 
TransitRegionStateProcedure table=dummy_table, region=, 
REOPEN/MOVE; resume parent processing.
2018-11-21 21:21:51,386 INFO  [PEWorker-13] assignment.RegionStateStore: 
pid=667 updating hbase:meta row=, regionState=OPENING, 
regionLocation=,17020,1542863268158
{noformat}

There are no obvious errors/delays that I see in RS log, and it doesn't log 
starting to send the report.
However, at 21:52.853 the report is processed that still contains this region.
{noformat}
2018-11-21 21:21:52,853 WARN  
[RpcServer.default.FPBQ.Fifo.handler=48,queue=3,port=17000] 
assignment.AssignmentManager: Killing ,17020,1542754626176: rit=OPENING, 
location=,17020,1542863268158, table=dummy_table, region= 
reported OPEN on server=,17020,1542754626176 but state has otherwise.
* ABORTING region server ,17020,1542754626176: 
org.apache.hadoop.hbase.YouAreDeadException: rit=OPENING, location=,17020,1542863268158, table=dummy_table, region= reported OPEN 
on server=,17020,1542754626176 but state has otherwise.
{noformat}

RS shuts down in an orderly manner and it can be seen from the log that this 
region is actually not present (there's no line indicating it's being closed, 
unlike for other regions).

I think there needs to be some sort of versioning in RS reports to allow master 
to account for concurrent operations and avoid races.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-21531) race between region report and region move causes master to kill RS

2018-11-29 Thread Sergey Shelukhin (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-21531?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Shelukhin updated HBASE-21531:
-
Description: 
In this case the delay between the ack from RS and the region report was 1.5s, 
so I'm not sure what caused the race (network hiccup? unreported retry by 
protobuf transport?) but in any case I don't see anything that prevents this 
from happening in a normal case, with a narrowed time window. Any delay (e.g. a 
GC pause on RS right after the report is built, and ack is sent for the close) 
or retries expands the window.


Master starts moving the region and the source RS acks by 21:51,206
{noformat}
Master:
2018-11-21 21:21:49,024 INFO  [master/6:17000.Chore.1] master.HMaster: balance 
hri=, source=,17020,1542754626176, destination=,17020,1542863268158
...
Server:
2018-11-21 21:21:49,394 INFO  [RS_CLOSE_REGION-regionserver/:17020-1] 
handler.UnassignRegionHandler: Close 
...
2018-11-21 21:21:51,095 INFO  [RS_CLOSE_REGION-regionserver/:17020-1] 
handler.UnassignRegionHandler: Closed 
{noformat}
By then the region is removed from onlineRegions, so the master proceeds.
{noformat}
2018-11-21 21:21:51,206 INFO  [PEWorker-4] procedure2.ProcedureExecutor: 
Finished subprocedure(s) of pid=667, 
state=RUNNABLE:REGION_STATE_TRANSITION_CONFIRM_CLOSED, hasLock=true; 
TransitRegionStateProcedure table=dummy_table, region=, 
REOPEN/MOVE; resume parent processing.
2018-11-21 21:21:51,386 INFO  [PEWorker-13] assignment.RegionStateStore: 
pid=667 updating hbase:meta row=, regionState=OPENING, 
regionLocation=,17020,1542863268158
{noformat}

There are no obvious errors/delays that I see in RS log, and it doesn't log 
starting to send the report.
However, at 21:52.853 the report is processed that still contains this region.
{noformat}
2018-11-21 21:21:52,853 WARN  
[RpcServer.default.FPBQ.Fifo.handler=48,queue=3,port=17000] 
assignment.AssignmentManager: Killing ,17020,1542754626176: rit=OPENING, 
location=,17020,1542863268158, table=dummy_table, region= 
reported OPEN on server=,17020,1542754626176 but state has otherwise.
* ABORTING region server ,17020,1542754626176: 
org.apache.hadoop.hbase.YouAreDeadException: rit=OPENING, location=,17020,1542863268158, table=dummy_table, region= reported OPEN 
on server=,17020,1542754626176 but state has otherwise.
{noformat}

RS shuts down in an orderly manner and it can be seen from the log that this 
region is actually not present (there's no line indicating it's being closed, 
unlike for other regions).

I think there needs to be some sort of versioning for region operations and/or 
in RS reports to allow master to account for concurrent operations and avoid 
races. Probably per region with either a grace period or an additional global 
version, so that master could avoid killing RS based on stale reports, but 
still kill RS if it did retain an old version of the region state due to some 
bug after acking a new version.

  was:
In this case the delay between the ack from RS and the region report was 1.5s, 
so I'm not sure what caused the race (network hiccup? unreported retry by 
protobuf transport?) but in any case I don't see anything that prevents this 
from happening in a normal case, with a narrowed time window. Any delay (e.g. a 
GC pause on RS right after the report is built, and ack is sent for the close) 
or retries expands the window.


Master starts moving the region and the source RS acks by 21:51,206
{noformat}
Master:
2018-11-21 21:21:49,024 INFO  [master/6:17000.Chore.1] master.HMaster: balance 
hri=, source=,17020,1542754626176, destination=,17020,1542863268158
...
Server:
2018-11-21 21:21:49,394 INFO  [RS_CLOSE_REGION-regionserver/:17020-1] 
handler.UnassignRegionHandler: Close 
...
2018-11-21 21:21:51,095 INFO  [RS_CLOSE_REGION-regionserver/:17020-1] 
handler.UnassignRegionHandler: Closed 
{noformat}
By then the region is removed from onlineRegions, so the master proceeds.
{noformat}
2018-11-21 21:21:51,206 INFO  [PEWorker-4] procedure2.ProcedureExecutor: 
Finished subprocedure(s) of pid=667, 
state=RUNNABLE:REGION_STATE_TRANSITION_CONFIRM_CLOSED, hasLock=true; 
TransitRegionStateProcedure table=dummy_table, region=, 
REOPEN/MOVE; resume parent processing.
2018-11-21 21:21:51,386 INFO  [PEWorker-13] assignment.RegionStateStore: 
pid=667 updating hbase:meta row=, regionState=OPENING, 
regionLocation=,17020,1542863268158
{noformat}

There are no obvious errors/delays that I see in RS log, and it doesn't log 
starting to send the report.
However, at 21:52.853 the report is processed that still contains this region.
{noformat}
2018-11-21 21:21:52,853 WARN  
[RpcServer.default.FPBQ.Fifo.handler=48,queue=3,port=17000] 
assignment.AssignmentManager: Killing ,17020,1542754626176: rit=OPENING, 
location=,17020,1542863268158, table=dummy_table, region= 
reported OPEN on server=,17020,1542754626176 but state has otherwise.
* ABORTING region 

[jira] [Updated] (HBASE-21531) race between region report and region move causes master to kill RS

2018-11-29 Thread Sergey Shelukhin (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-21531?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Shelukhin updated HBASE-21531:
-
Description: 
In this case the delay between the ack from RS and the region report was 1.5s, 
so I'm not sure what caused the race (network hiccup? unreported retry by 
protobuf transport?) but in any case I don't see anything that prevents this 
from happening in a normal case, with a narrowed time window. Any delay (e.g. a 
GC pause on RS right after the report is built, and ack is sent for the close) 
or retries expands the window.


Master starts moving the region and the source RS acks by 21:51,206
{noformat}
Master:
2018-11-21 21:21:49,024 INFO  [master/6:17000.Chore.1] master.HMaster: balance 
hri=, source=,17020,1542754626176, destination=,17020,1542863268158
...
Server:
2018-11-21 21:21:49,394 INFO  [RS_CLOSE_REGION-regionserver/:17020-1] 
handler.UnassignRegionHandler: Close 
...
2018-11-21 21:21:51,095 INFO  [RS_CLOSE_REGION-regionserver/:17020-1] 
handler.UnassignRegionHandler: Closed 
{noformat}
By then the region is removed from onlineRegions, so the master proceeds.
{noformat}
2018-11-21 21:21:51,206 INFO  [PEWorker-4] procedure2.ProcedureExecutor: 
Finished subprocedure(s) of pid=667, 
state=RUNNABLE:REGION_STATE_TRANSITION_CONFIRM_CLOSED, hasLock=true; 
TransitRegionStateProcedure table=dummy_table, region=, 
REOPEN/MOVE; resume parent processing.
2018-11-21 21:21:51,386 INFO  [PEWorker-13] assignment.RegionStateStore: 
pid=667 updating hbase:meta row=, regionState=OPENING, 
regionLocation=,17020,1542863268158
{noformat}

There are no obvious errors/delays that I see in RS log, and it doesn't log 
starting to send the report.
However, at 21:52.853 the report is processed that still contains this region.
{noformat}
2018-11-21 21:21:52,853 WARN  
[RpcServer.default.FPBQ.Fifo.handler=48,queue=3,port=17000] 
assignment.AssignmentManager: Killing ,17020,1542754626176: rit=OPENING, 
location=,17020,1542863268158, table=dummy_table, region= 
reported OPEN on server=,17020,1542754626176 but state has otherwise.
* ABORTING region server ,17020,1542754626176: 
org.apache.hadoop.hbase.YouAreDeadException: rit=OPENING, location=,17020,1542863268158, table=dummy_table, region= reported OPEN 
on server=,17020,1542754626176 but state has otherwise.
{noformat}

RS shuts down in an orderly manner and it can be seen from the log that this 
region is actually not present (there's no line indicating it's being closed, 
unlike for other regions).

I think there needs to be some sort of versioning for region operations and/or 
in RS reports to allow master to account for concurrent operations and avoid 
races. Probably per region with either a grace period or an additional global 
version, so that master could avoid killing RS based on stale reports, but 
still kill RS if it did retain an old version of the region state due to some 
bug after acking a new version state.

  was:
In this case the delay between the ack from RS and the region report was 1.5s, 
so I'm not sure what caused the race (network hiccup? unreported retry by 
protobuf transport?) but in any case I don't see anything that prevents this 
from happening in a normal case, with a narrowed time window. Any delay (e.g. a 
GC pause on RS right after the report is built, and ack is sent for the close) 
or retries expands the window.


Master starts moving the region and the source RS acks by 21:51,206
{noformat}
Master:
2018-11-21 21:21:49,024 INFO  [master/6:17000.Chore.1] master.HMaster: balance 
hri=, source=,17020,1542754626176, destination=,17020,1542863268158
...
Server:
2018-11-21 21:21:49,394 INFO  [RS_CLOSE_REGION-regionserver/:17020-1] 
handler.UnassignRegionHandler: Close 
...
2018-11-21 21:21:51,095 INFO  [RS_CLOSE_REGION-regionserver/:17020-1] 
handler.UnassignRegionHandler: Closed 
{noformat}
By then the region is removed from onlineRegions, so the master proceeds.
{noformat}
2018-11-21 21:21:51,206 INFO  [PEWorker-4] procedure2.ProcedureExecutor: 
Finished subprocedure(s) of pid=667, 
state=RUNNABLE:REGION_STATE_TRANSITION_CONFIRM_CLOSED, hasLock=true; 
TransitRegionStateProcedure table=dummy_table, region=, 
REOPEN/MOVE; resume parent processing.
2018-11-21 21:21:51,386 INFO  [PEWorker-13] assignment.RegionStateStore: 
pid=667 updating hbase:meta row=, regionState=OPENING, 
regionLocation=,17020,1542863268158
{noformat}

There are no obvious errors/delays that I see in RS log, and it doesn't log 
starting to send the report.
However, at 21:52.853 the report is processed that still contains this region.
{noformat}
2018-11-21 21:21:52,853 WARN  
[RpcServer.default.FPBQ.Fifo.handler=48,queue=3,port=17000] 
assignment.AssignmentManager: Killing ,17020,1542754626176: rit=OPENING, 
location=,17020,1542863268158, table=dummy_table, region= 
reported OPEN on server=,17020,1542754626176 but state has otherwise.
* ABORTING 

[jira] [Commented] (HBASE-21524) Unnecessary DEBUG log in ConnectionImplementation#isTableEnabled

2018-11-29 Thread Hudson (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21524?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16704018#comment-16704018
 ] 

Hudson commented on HBASE-21524:


Results for branch branch-2
[build #1531 on 
builds.a.o|https://builds.apache.org/job/HBase%20Nightly/job/branch-2/1531/]: 
(x) *{color:red}-1 overall{color}*

details (if available):

(/) {color:green}+1 general checks{color}
-- For more information [see general 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-2/1531//General_Nightly_Build_Report/]




(x) {color:red}-1 jdk8 hadoop2 checks{color}
-- For more information [see jdk8 (hadoop2) 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-2/1531//JDK8_Nightly_Build_Report_(Hadoop2)/]


(/) {color:green}+1 jdk8 hadoop3 checks{color}
-- For more information [see jdk8 (hadoop3) 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-2/1531//JDK8_Nightly_Build_Report_(Hadoop3)/]


(/) {color:green}+1 source release artifact{color}
-- See build output for details.


(/) {color:green}+1 client integration test{color}


> Unnecessary DEBUG log in ConnectionImplementation#isTableEnabled
> 
>
> Key: HBASE-21524
> URL: https://issues.apache.org/jira/browse/HBASE-21524
> Project: HBase
>  Issue Type: Improvement
>  Components: Client
>Reporter: Josh Elser
>Assignee: Josh Elser
>Priority: Major
> Fix For: 3.0.0, 2.2.0, 2.1.2
>
> Attachments: HBASE-21524.001.branch-2.0.patch
>
>
> Too much crap coming out of isTableEnabled for "normal" situations. Don't 
> need a DEBUG message to tell us that the table is available.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (HBASE-21530) Abort_Procedure should be able to take a list of proc IDs

2018-11-29 Thread Geoffrey Jacoby (JIRA)
Geoffrey Jacoby created HBASE-21530:
---

 Summary: Abort_Procedure should be able to take a list of proc IDs
 Key: HBASE-21530
 URL: https://issues.apache.org/jira/browse/HBASE-21530
 Project: HBase
  Issue Type: Improvement
Reporter: Geoffrey Jacoby


As a convenience, it would be helpful if the HBase shell's abort_procedure call 
had the option of taking in multiple procedure ids at the same time, rather 
than relying on operators to use a loop in an external script. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (HBASE-21529) Abort_Procedure in HBase shell fails when all handlers are taken up by Procedures

2018-11-29 Thread Geoffrey Jacoby (JIRA)
Geoffrey Jacoby created HBASE-21529:
---

 Summary: Abort_Procedure in HBase shell fails when all handlers 
are taken up by Procedures
 Key: HBASE-21529
 URL: https://issues.apache.org/jira/browse/HBASE-21529
 Project: HBase
  Issue Type: Bug
  Components: proc-v2
Affects Versions: 1.4.8, 1.3.2
Reporter: Geoffrey Jacoby


If all RPC handlers are taken up by stuck procedures, the operator will not be 
able to abort them, because the abort procedure call will try to use the same 
(exhausted) pool of RPC handlers. 

The abort procedure call should use a dedicated pool of RPC handlers. This pool 
can probably be small, but it needs to be separate. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-18735) Provide a fast mechanism for shutting down mini cluster

2018-11-29 Thread Josh Elser (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-18735?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Josh Elser updated HBASE-18735:
---
Fix Version/s: 2.1.2
   2.0.3
   2.2.0
   3.0.0

> Provide a fast mechanism for shutting down mini cluster
> ---
>
> Key: HBASE-18735
> URL: https://issues.apache.org/jira/browse/HBASE-18735
> Project: HBase
>  Issue Type: Wish
>Reporter: Samarth Jain
>Assignee: Artem Ervits
>Priority: Major
> Fix For: 3.0.0, 2.2.0, 2.0.3, 2.1.2
>
> Attachments: HBASE-18735.v01.patch, HBASE-18735.v02.patch, 
> HBASE-18735.v03.patch
>
>
> The current mechanism of shutting down a mini cluster through 
> HBaseTestingUtility.shutDownMiniCluster can take a lot of time when the mini 
> cluster almost has a lot of tables. A lot of this time is spent in closing 
> all the user regions. It would be nice to have a mechanism where this 
> shutdown can happen quickly without having to worry about closing these user 
> regions. At the same time, this mechanism would need to make sure that all 
> the critical system resources like file handles and network ports are still 
> released so that subsequently initialized mini clusters on the same JVM or 
> system won't run into resource issues. This would make testing using HBase 
> mini clusters much faster and immensely help out test frameworks of dependent 
> projects like Phoenix.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-18735) Provide a fast mechanism for shutting down mini cluster

2018-11-29 Thread Josh Elser (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-18735?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Josh Elser updated HBASE-18735:
---
  Resolution: Fixed
Hadoop Flags: Reviewed
  Status: Resolved  (was: Patch Available)

Pushing this one to keep Artem chugging along. Seems low risk to the rest of 
the population.

Thanks for the patch, Artem!

> Provide a fast mechanism for shutting down mini cluster
> ---
>
> Key: HBASE-18735
> URL: https://issues.apache.org/jira/browse/HBASE-18735
> Project: HBase
>  Issue Type: Wish
>Reporter: Samarth Jain
>Assignee: Artem Ervits
>Priority: Major
> Fix For: 3.0.0, 2.2.0, 2.0.3, 2.1.2
>
> Attachments: HBASE-18735.v01.patch, HBASE-18735.v02.patch, 
> HBASE-18735.v03.patch
>
>
> The current mechanism of shutting down a mini cluster through 
> HBaseTestingUtility.shutDownMiniCluster can take a lot of time when the mini 
> cluster almost has a lot of tables. A lot of this time is spent in closing 
> all the user regions. It would be nice to have a mechanism where this 
> shutdown can happen quickly without having to worry about closing these user 
> regions. At the same time, this mechanism would need to make sure that all 
> the critical system resources like file handles and network ports are still 
> released so that subsequently initialized mini clusters on the same JVM or 
> system won't run into resource issues. This would make testing using HBase 
> mini clusters much faster and immensely help out test frameworks of dependent 
> projects like Phoenix.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-18735) Provide a fast mechanism for shutting down mini cluster

2018-11-29 Thread Josh Elser (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-18735?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16703850#comment-16703850
 ] 

Josh Elser commented on HBASE-18735:


Sounds good enough to me. I don't see an obvious reason why your change would 
have any direct impact on TestMultiColumnScanner.

Let's try this on master and 2.x? Would say we can put it on Master only, but I 
don't think we have a corresponding Phoenix build for HBase 3 yet.

> Provide a fast mechanism for shutting down mini cluster
> ---
>
> Key: HBASE-18735
> URL: https://issues.apache.org/jira/browse/HBASE-18735
> Project: HBase
>  Issue Type: Wish
>Reporter: Samarth Jain
>Assignee: Artem Ervits
>Priority: Major
> Attachments: HBASE-18735.v01.patch, HBASE-18735.v02.patch, 
> HBASE-18735.v03.patch
>
>
> The current mechanism of shutting down a mini cluster through 
> HBaseTestingUtility.shutDownMiniCluster can take a lot of time when the mini 
> cluster almost has a lot of tables. A lot of this time is spent in closing 
> all the user regions. It would be nice to have a mechanism where this 
> shutdown can happen quickly without having to worry about closing these user 
> regions. At the same time, this mechanism would need to make sure that all 
> the critical system resources like file handles and network ports are still 
> released so that subsequently initialized mini clusters on the same JVM or 
> system won't run into resource issues. This would make testing using HBase 
> mini clusters much faster and immensely help out test frameworks of dependent 
> projects like Phoenix.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Comment Edited] (HBASE-21083) Introduce a mechanism to bypass the execution of a stuck procedure

2018-11-29 Thread Xu Cang (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21083?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16702299#comment-16702299
 ] 

Xu Cang edited comment on HBASE-21083 at 11/29/18 9:14 PM:
---

[~allan163]

Do you see a value backporting this to branch-1? 

We have seen MasterProcedure stuck in production cluster and have very few ways 
to safely resolve it. 


was (Author: xucang):
[~allan163]

Do you see a value backporting this to branch-1? 

We have seen MasterProcedure stuck in production cluster and have very few ways 
to safely resolve it. 

(I am not very familiar with AMv2 but seems AMv2 is available in branch-1 too? 
I can see procedure2.Procedure class and so on) Thanks.

> Introduce a mechanism to bypass the execution of a stuck procedure
> --
>
> Key: HBASE-21083
> URL: https://issues.apache.org/jira/browse/HBASE-21083
> Project: HBase
>  Issue Type: Sub-task
>  Components: amv2
>Affects Versions: 2.1.0, 2.0.1
>Reporter: Allan Yang
>Assignee: Allan Yang
>Priority: Major
> Fix For: 3.0.0, 2.1.1, 2.0.2
>
> Attachments: HBASE-21083.branch-2.0.001.patch, 
> HBASE-21083.branch-2.0.002.patch, HBASE-21083.branch-2.0.003.patch, 
> HBASE-21083.branch-2.0.003.patch, HBASE-21083.branch-2.0.003.patch, 
> HBASE-21083.branch-2.1.001.patch
>
>
> Offline discussed with [~stack] and [~Apache9]. We all agreed that we need to 
> introduce a mechanism to 'force complete' a stuck procedure, so the AMv2 can 
> continue running.
> we still have some unrevealed bugs hiding in our AMv2 and procedureV2 system, 
> we need something to interfere with stuck procedures before HBCK2 can work. 
> This is very crucial for a production ready system.
> For now, we have little ways to interfere with running procedures. Aborting 
> them is not a good choice, since some procedures are not abort-able. And some 
> procedure may have overridden the abort() method, which will ignore the abort 
> request.
> So, here, I will introduce a mechanism to bypass the execution of a stuck 
> procedure.
> Basically, I added a field called 'bypass' to Procedure class. If we set this 
> field to true, all the logic in execute/rollback will be skipped, letting 
> this procedure and its ancestors complete normally and releasing the lock 
> resources at last.
> Notice that bypassing a procedure may leave the cluster in a middle state, 
> e.g. the region not assigned, or some hdfs files left behind. 
> The Operators need know the side effect of bypassing and recover the 
> inconsistent state of the cluster themselves, like issuing new procedures to 
> assign the regions.
> A patch will be uploaded and review board will be open. For now, only APIs in 
> ProcedureExecutor are provided. If anything is fine, I will add it to master 
> service and add a shell command to bypass a procedure. Or, maybe we can use 
> dynamically compiled JSPs to execute those APIs as mentioned in HBASE-20679.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21528) hbase-connectors need scala dependency convergence

2018-11-29 Thread Peter Somogyi (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21528?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16703747#comment-16703747
 ] 

Peter Somogyi commented on HBASE-21528:
---

Currently there is no precommit check for hbase-connectors and 
hbase-operator-tools repos and that's why both Hadoop QA tests failed. The plan 
is to add GitHub Pull Request verification because Yetus does not support 
multiple repositories connected to a single Jira project. 

> hbase-connectors need scala dependency convergence
> --
>
> Key: HBASE-21528
> URL: https://issues.apache.org/jira/browse/HBASE-21528
> Project: HBase
>  Issue Type: Task
>  Components: hbase-connectors
>Reporter: Artem Ervits
>Assignee: Artem Ervits
>Priority: Minor
> Attachments: HBASE-21528.v01.patch, HBASE-21528.v02.patch
>
>
> this issue reappeared after migration of hbase-spark to hbase-connectors 
> repo. Originally addressed in 
> https://issues.apache.org/jira/browse/HBASE-20175 
> introducing scala-maven-plugin and adding a flag to honor compatible major 
> scala version makes warning disappear.
>  
> {code:java}
> [WARNING] Expected all dependencies to require Scala version: 2.11.12
> [WARNING] org.apache.hbase.connectors.spark:hbase-spark:1.0.0-SNAPSHOT 
> requires scala version: 2.11.12
> [WARNING] org.apache.spark:spark-streaming_2.11:2.4.0 requires scala version: 
> 2.11.12
> [WARNING] org.apache.spark:spark-streaming_2.11:2.4.0 requires scala version: 
> 2.11.12
> [WARNING] org.scala-lang:scala-reflect:2.11.12 requires scala version: 2.11.12
> [WARNING] org.apache.spark:spark-tags_2.11:2.4.0 requires scala version: 
> 2.11.12
> [WARNING] com.twitter:chill_2.11:0.9.3 requires scala version: 2.11.12
> [WARNING] org.scala-lang.modules:scala-parser-combinators_2.11:1.1.0 requires 
> scala version: 2.11.12
> [WARNING] org.json4s:json4s-core_2.11:3.2.11 requires scala version: 2.11.0
> [WARNING] Multiple versions of scala libraries detected!
> [WARNING] warning: there were 17 deprecation warnings; re-run with 
> -deprecation for details
> [WARNING] one warning found
> [WARNING] Expected all dependencies to require Scala version: 2.11.12
> [WARNING] org.apache.hbase.connectors.spark:hbase-spark:1.0.0-SNAPSHOT 
> requires scala version: 2.11.12
> [WARNING] org.apache.spark:spark-streaming_2.11:2.4.0 requires scala version: 
> 2.11.12
> [WARNING] org.apache.spark:spark-streaming_2.11:2.4.0 requires scala version: 
> 2.11.12
> [WARNING] org.scala-lang:scala-reflect:2.11.12 requires scala version: 2.11.12
> [WARNING] org.apache.spark:spark-tags_2.11:2.4.0 requires scala version: 
> 2.11.12
> [WARNING] com.twitter:chill_2.11:0.9.3 requires scala version: 2.11.12
> [WARNING] org.scala-lang.modules:scala-parser-combinators_2.11:1.1.0 requires 
> scala version: 2.11.12
> [WARNING] org.json4s:json4s-core_2.11:3.2.11 requires scala version: 2.11.0
> [WARNING] Multiple versions of scala libraries detected!{code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Comment Edited] (HBASE-21528) hbase-connectors need scala dependency convergence

2018-11-29 Thread Artem Ervits (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21528?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16703731#comment-16703731
 ] 

Artem Ervits edited comment on HBASE-21528 at 11/29/18 8:08 PM:


[~stack] btw, had to generate the patch with git diff --no-prefix as there's no 
dev-tools/make_patch.sh script. Uploaded v.02 patch in case I fat-fingered it 
first time.


was (Author: dbist13):
[~stack] btw, had to generate the patch with git diff --no-prefix as there's no 
yetus integration yet. Uploaded v.02 patch in case I fat-fingered it first time.

> hbase-connectors need scala dependency convergence
> --
>
> Key: HBASE-21528
> URL: https://issues.apache.org/jira/browse/HBASE-21528
> Project: HBase
>  Issue Type: Task
>  Components: hbase-connectors
>Reporter: Artem Ervits
>Assignee: Artem Ervits
>Priority: Minor
> Attachments: HBASE-21528.v01.patch, HBASE-21528.v02.patch
>
>
> this issue reappeared after migration of hbase-spark to hbase-connectors 
> repo. Originally addressed in 
> https://issues.apache.org/jira/browse/HBASE-20175 
> introducing scala-maven-plugin and adding a flag to honor compatible major 
> scala version makes warning disappear.
>  
> {code:java}
> [WARNING] Expected all dependencies to require Scala version: 2.11.12
> [WARNING] org.apache.hbase.connectors.spark:hbase-spark:1.0.0-SNAPSHOT 
> requires scala version: 2.11.12
> [WARNING] org.apache.spark:spark-streaming_2.11:2.4.0 requires scala version: 
> 2.11.12
> [WARNING] org.apache.spark:spark-streaming_2.11:2.4.0 requires scala version: 
> 2.11.12
> [WARNING] org.scala-lang:scala-reflect:2.11.12 requires scala version: 2.11.12
> [WARNING] org.apache.spark:spark-tags_2.11:2.4.0 requires scala version: 
> 2.11.12
> [WARNING] com.twitter:chill_2.11:0.9.3 requires scala version: 2.11.12
> [WARNING] org.scala-lang.modules:scala-parser-combinators_2.11:1.1.0 requires 
> scala version: 2.11.12
> [WARNING] org.json4s:json4s-core_2.11:3.2.11 requires scala version: 2.11.0
> [WARNING] Multiple versions of scala libraries detected!
> [WARNING] warning: there were 17 deprecation warnings; re-run with 
> -deprecation for details
> [WARNING] one warning found
> [WARNING] Expected all dependencies to require Scala version: 2.11.12
> [WARNING] org.apache.hbase.connectors.spark:hbase-spark:1.0.0-SNAPSHOT 
> requires scala version: 2.11.12
> [WARNING] org.apache.spark:spark-streaming_2.11:2.4.0 requires scala version: 
> 2.11.12
> [WARNING] org.apache.spark:spark-streaming_2.11:2.4.0 requires scala version: 
> 2.11.12
> [WARNING] org.scala-lang:scala-reflect:2.11.12 requires scala version: 2.11.12
> [WARNING] org.apache.spark:spark-tags_2.11:2.4.0 requires scala version: 
> 2.11.12
> [WARNING] com.twitter:chill_2.11:0.9.3 requires scala version: 2.11.12
> [WARNING] org.scala-lang.modules:scala-parser-combinators_2.11:1.1.0 requires 
> scala version: 2.11.12
> [WARNING] org.json4s:json4s-core_2.11:3.2.11 requires scala version: 2.11.0
> [WARNING] Multiple versions of scala libraries detected!{code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21528) hbase-connectors need scala dependency convergence

2018-11-29 Thread Artem Ervits (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21528?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16703731#comment-16703731
 ] 

Artem Ervits commented on HBASE-21528:
--

[~stack] btw, had to generate the patch with git diff --no-prefix as there's no 
yetus integration yet. Uploaded v.02 patch in case I fat-fingered it first time.

> hbase-connectors need scala dependency convergence
> --
>
> Key: HBASE-21528
> URL: https://issues.apache.org/jira/browse/HBASE-21528
> Project: HBase
>  Issue Type: Task
>  Components: hbase-connectors
>Reporter: Artem Ervits
>Assignee: Artem Ervits
>Priority: Minor
> Attachments: HBASE-21528.v01.patch, HBASE-21528.v02.patch
>
>
> this issue reappeared after migration of hbase-spark to hbase-connectors 
> repo. Originally addressed in 
> https://issues.apache.org/jira/browse/HBASE-20175 
> introducing scala-maven-plugin and adding a flag to honor compatible major 
> scala version makes warning disappear.
>  
> {code:java}
> [WARNING] Expected all dependencies to require Scala version: 2.11.12
> [WARNING] org.apache.hbase.connectors.spark:hbase-spark:1.0.0-SNAPSHOT 
> requires scala version: 2.11.12
> [WARNING] org.apache.spark:spark-streaming_2.11:2.4.0 requires scala version: 
> 2.11.12
> [WARNING] org.apache.spark:spark-streaming_2.11:2.4.0 requires scala version: 
> 2.11.12
> [WARNING] org.scala-lang:scala-reflect:2.11.12 requires scala version: 2.11.12
> [WARNING] org.apache.spark:spark-tags_2.11:2.4.0 requires scala version: 
> 2.11.12
> [WARNING] com.twitter:chill_2.11:0.9.3 requires scala version: 2.11.12
> [WARNING] org.scala-lang.modules:scala-parser-combinators_2.11:1.1.0 requires 
> scala version: 2.11.12
> [WARNING] org.json4s:json4s-core_2.11:3.2.11 requires scala version: 2.11.0
> [WARNING] Multiple versions of scala libraries detected!
> [WARNING] warning: there were 17 deprecation warnings; re-run with 
> -deprecation for details
> [WARNING] one warning found
> [WARNING] Expected all dependencies to require Scala version: 2.11.12
> [WARNING] org.apache.hbase.connectors.spark:hbase-spark:1.0.0-SNAPSHOT 
> requires scala version: 2.11.12
> [WARNING] org.apache.spark:spark-streaming_2.11:2.4.0 requires scala version: 
> 2.11.12
> [WARNING] org.apache.spark:spark-streaming_2.11:2.4.0 requires scala version: 
> 2.11.12
> [WARNING] org.scala-lang:scala-reflect:2.11.12 requires scala version: 2.11.12
> [WARNING] org.apache.spark:spark-tags_2.11:2.4.0 requires scala version: 
> 2.11.12
> [WARNING] com.twitter:chill_2.11:0.9.3 requires scala version: 2.11.12
> [WARNING] org.scala-lang.modules:scala-parser-combinators_2.11:1.1.0 requires 
> scala version: 2.11.12
> [WARNING] org.json4s:json4s-core_2.11:3.2.11 requires scala version: 2.11.0
> [WARNING] Multiple versions of scala libraries detected!{code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21518) TestMasterFailoverWithProcedures is flaky

2018-11-29 Thread Peter Somogyi (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21518?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16703730#comment-16703730
 ] 

Peter Somogyi commented on HBASE-21518:
---

[~stack]: Shall I commit it back until branch-2.0?

> TestMasterFailoverWithProcedures is flaky
> -
>
> Key: HBASE-21518
> URL: https://issues.apache.org/jira/browse/HBASE-21518
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 2.2.0, 2.0.3, 2.1.2
>Reporter: Peter Somogyi
>Assignee: Peter Somogyi
>Priority: Major
> Attachments: HBASE-21518-v1.patch, output.txt
>
>
> TestMasterFailoverWithProcedures test is failing frequently, times out. I 
> faced this failure on 2.0.3RC0 vote and it also appears on multiple flaky 
> dashboards.
> branch-2: 
> [https://builds.apache.org/view/H-L/view/HBase/job/HBase-Flaky-Tests/job/branch-2/2007/]
> branch-2.1: 
> [https://builds.apache.org/view/H-L/view/HBase/job/HBase-Flaky-Tests/job/branch-2.1/2002/]
> branch-2.0: 
> [https://builds.apache.org/view/H-L/view/HBase/job/HBase-Flaky-Tests/job/branch-2.0/1988/]
>   
> {noformat}
> [INFO] Running 
> org.apache.hadoop.hbase.master.procedure.TestMasterFailoverWithProcedures
> [ERROR] Tests run: 4, Failures: 0, Errors: 2, Skipped: 0, Time elapsed: 
> 780.648 s <<< FAILURE! - in 
> org.apache.hadoop.hbase.master.procedure.TestMasterFailoverWithProcedures
> [ERROR] 
> org.apache.hadoop.hbase.master.procedure.TestMasterFailoverWithProcedures  
> Time elapsed: 749.024 s  <<< ERROR!
> org.junit.runners.model.TestTimedOutException: test timed out after 780 
> seconds
>   at 
> org.apache.hadoop.hbase.master.procedure.TestMasterFailoverWithProcedures.tearDown(TestMasterFailoverWithProcedures.java:86)
> [ERROR] 
> org.apache.hadoop.hbase.master.procedure.TestMasterFailoverWithProcedures  
> Time elapsed: 749.051 s  <<< ERROR!
> java.lang.Exception: Appears to be stuck in thread RS-EventLoopGroup-3-2
> {noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21518) TestMasterFailoverWithProcedures is flaky

2018-11-29 Thread Hadoop QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21518?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16703728#comment-16703728
 ] 

Hadoop QA commented on HBASE-21518:
---

| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
15s{color} | {color:blue} Docker mode activated. {color} |
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} hbaseanti {color} | {color:green}  0m  
0s{color} | {color:green} Patch does not have any anti-patterns. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:orange}-0{color} | {color:orange} test4tests {color} | {color:orange}  
0m  0s{color} | {color:orange} The patch doesn't appear to include any new or 
modified tests. Please justify why no new tests are needed for this patch. Also 
please list what manual steps were performed to verify this patch. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  5m 
11s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  2m  
4s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  1m 
14s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} shadedjars {color} | {color:green}  4m 
15s{color} | {color:green} branch has no errors when building our shaded 
downstream artifacts. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  2m  
5s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
29s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  4m 
26s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m 
58s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  1m 
58s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  1m 
13s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} shadedjars {color} | {color:green}  4m 
11s{color} | {color:green} patch has no errors when building our shaded 
downstream artifacts. {color} |
| {color:green}+1{color} | {color:green} hadoopcheck {color} | {color:green}  
9m 15s{color} | {color:green} Patch does not cause any errors with Hadoop 2.7.4 
or 3.0.0. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  2m  
8s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
30s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} unit {color} | {color:green}134m 
14s{color} | {color:green} hbase-server in the patch passed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
19s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black}174m 12s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Docker | Client=17.05.0-ce Server=17.05.0-ce Image:yetus/hbase:b002b0b |
| JIRA Issue | HBASE-21518 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12950045/HBASE-21518-v1.patch |
| Optional Tests |  dupname  asflicense  javac  javadoc  unit  findbugs  
shadedjars  hadoopcheck  hbaseanti  checkstyle  compile  |
| uname | Linux 345b56ce27bf 4.4.0-139-generic #165~14.04.1-Ubuntu SMP Wed Oct 
31 10:55:11 UTC 2018 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/home/jenkins/jenkins-slave/workspace/PreCommit-HBASE-Build@2/component/dev-support/hbase-personality.sh
 |
| git revision | master / fdddc47e77 |
| maven | version: Apache Maven 3.5.4 
(1edded0938998edf8bf061f1ceb3cfdeccf443fe; 2018-06-17T18:33:14Z) |
| Default Java | 1.8.0_181 |
| findbugs | v3.1.0-RC3 |
|  Test Results | 
https://builds.apache.org/job/PreCommit-HBASE-Build/15152/testReport/ |
| Max. process+thread count | 4421 (vs. ulimit of 1) |
| modules | C: hbase-server U: hbase-server |
| Console output | 

[jira] [Commented] (HBASE-21528) hbase-connectors need scala dependency convergence

2018-11-29 Thread Hadoop QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21528?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16703716#comment-16703716
 ] 

Hadoop QA commented on HBASE-21528:
---

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
19s{color} | {color:blue} Docker mode activated. {color} |
| {color:red}-1{color} | {color:red} patch {color} | {color:red}  0m  4s{color} 
| {color:red} HBASE-21528 does not apply to master. Rebase required? Wrong 
Branch? See https://yetus.apache.org/documentation/0.8.0/precommit-patchnames 
for help. {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Docker | Client=17.05.0-ce Server=17.05.0-ce Image:yetus/hbase:b002b0b |
| JIRA Issue | HBASE-21528 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12950063/HBASE-21528.v02.patch 
|
| Console output | 
https://builds.apache.org/job/PreCommit-HBASE-Build/15154/console |
| Powered by | Apache Yetus 0.8.0   http://yetus.apache.org |


This message was automatically generated.



> hbase-connectors need scala dependency convergence
> --
>
> Key: HBASE-21528
> URL: https://issues.apache.org/jira/browse/HBASE-21528
> Project: HBase
>  Issue Type: Task
>  Components: hbase-connectors
>Reporter: Artem Ervits
>Assignee: Artem Ervits
>Priority: Minor
> Attachments: HBASE-21528.v01.patch, HBASE-21528.v02.patch
>
>
> this issue reappeared after migration of hbase-spark to hbase-connectors 
> repo. Originally addressed in 
> https://issues.apache.org/jira/browse/HBASE-20175 
> introducing scala-maven-plugin and adding a flag to honor compatible major 
> scala version makes warning disappear.
>  
> {code:java}
> [WARNING] Expected all dependencies to require Scala version: 2.11.12
> [WARNING] org.apache.hbase.connectors.spark:hbase-spark:1.0.0-SNAPSHOT 
> requires scala version: 2.11.12
> [WARNING] org.apache.spark:spark-streaming_2.11:2.4.0 requires scala version: 
> 2.11.12
> [WARNING] org.apache.spark:spark-streaming_2.11:2.4.0 requires scala version: 
> 2.11.12
> [WARNING] org.scala-lang:scala-reflect:2.11.12 requires scala version: 2.11.12
> [WARNING] org.apache.spark:spark-tags_2.11:2.4.0 requires scala version: 
> 2.11.12
> [WARNING] com.twitter:chill_2.11:0.9.3 requires scala version: 2.11.12
> [WARNING] org.scala-lang.modules:scala-parser-combinators_2.11:1.1.0 requires 
> scala version: 2.11.12
> [WARNING] org.json4s:json4s-core_2.11:3.2.11 requires scala version: 2.11.0
> [WARNING] Multiple versions of scala libraries detected!
> [WARNING] warning: there were 17 deprecation warnings; re-run with 
> -deprecation for details
> [WARNING] one warning found
> [WARNING] Expected all dependencies to require Scala version: 2.11.12
> [WARNING] org.apache.hbase.connectors.spark:hbase-spark:1.0.0-SNAPSHOT 
> requires scala version: 2.11.12
> [WARNING] org.apache.spark:spark-streaming_2.11:2.4.0 requires scala version: 
> 2.11.12
> [WARNING] org.apache.spark:spark-streaming_2.11:2.4.0 requires scala version: 
> 2.11.12
> [WARNING] org.scala-lang:scala-reflect:2.11.12 requires scala version: 2.11.12
> [WARNING] org.apache.spark:spark-tags_2.11:2.4.0 requires scala version: 
> 2.11.12
> [WARNING] com.twitter:chill_2.11:0.9.3 requires scala version: 2.11.12
> [WARNING] org.scala-lang.modules:scala-parser-combinators_2.11:1.1.0 requires 
> scala version: 2.11.12
> [WARNING] org.json4s:json4s-core_2.11:3.2.11 requires scala version: 2.11.0
> [WARNING] Multiple versions of scala libraries detected!{code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-21528) hbase-connectors need scala dependency convergence

2018-11-29 Thread Artem Ervits (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-21528?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Artem Ervits updated HBASE-21528:
-
Attachment: HBASE-21528.v02.patch

> hbase-connectors need scala dependency convergence
> --
>
> Key: HBASE-21528
> URL: https://issues.apache.org/jira/browse/HBASE-21528
> Project: HBase
>  Issue Type: Task
>  Components: hbase-connectors
>Reporter: Artem Ervits
>Assignee: Artem Ervits
>Priority: Minor
> Attachments: HBASE-21528.v01.patch, HBASE-21528.v02.patch
>
>
> this issue reappeared after migration of hbase-spark to hbase-connectors 
> repo. Originally addressed in 
> https://issues.apache.org/jira/browse/HBASE-20175 
> introducing scala-maven-plugin and adding a flag to honor compatible major 
> scala version makes warning disappear.
>  
> {code:java}
> [WARNING] Expected all dependencies to require Scala version: 2.11.12
> [WARNING] org.apache.hbase.connectors.spark:hbase-spark:1.0.0-SNAPSHOT 
> requires scala version: 2.11.12
> [WARNING] org.apache.spark:spark-streaming_2.11:2.4.0 requires scala version: 
> 2.11.12
> [WARNING] org.apache.spark:spark-streaming_2.11:2.4.0 requires scala version: 
> 2.11.12
> [WARNING] org.scala-lang:scala-reflect:2.11.12 requires scala version: 2.11.12
> [WARNING] org.apache.spark:spark-tags_2.11:2.4.0 requires scala version: 
> 2.11.12
> [WARNING] com.twitter:chill_2.11:0.9.3 requires scala version: 2.11.12
> [WARNING] org.scala-lang.modules:scala-parser-combinators_2.11:1.1.0 requires 
> scala version: 2.11.12
> [WARNING] org.json4s:json4s-core_2.11:3.2.11 requires scala version: 2.11.0
> [WARNING] Multiple versions of scala libraries detected!
> [WARNING] warning: there were 17 deprecation warnings; re-run with 
> -deprecation for details
> [WARNING] one warning found
> [WARNING] Expected all dependencies to require Scala version: 2.11.12
> [WARNING] org.apache.hbase.connectors.spark:hbase-spark:1.0.0-SNAPSHOT 
> requires scala version: 2.11.12
> [WARNING] org.apache.spark:spark-streaming_2.11:2.4.0 requires scala version: 
> 2.11.12
> [WARNING] org.apache.spark:spark-streaming_2.11:2.4.0 requires scala version: 
> 2.11.12
> [WARNING] org.scala-lang:scala-reflect:2.11.12 requires scala version: 2.11.12
> [WARNING] org.apache.spark:spark-tags_2.11:2.4.0 requires scala version: 
> 2.11.12
> [WARNING] com.twitter:chill_2.11:0.9.3 requires scala version: 2.11.12
> [WARNING] org.scala-lang.modules:scala-parser-combinators_2.11:1.1.0 requires 
> scala version: 2.11.12
> [WARNING] org.json4s:json4s-core_2.11:3.2.11 requires scala version: 2.11.0
> [WARNING] Multiple versions of scala libraries detected!{code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21528) hbase-connectors need scala dependency convergence

2018-11-29 Thread Hadoop QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21528?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16703688#comment-16703688
 ] 

Hadoop QA commented on HBASE-21528:
---

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
11s{color} | {color:blue} Docker mode activated. {color} |
| {color:red}-1{color} | {color:red} patch {color} | {color:red}  0m  4s{color} 
| {color:red} HBASE-21528 does not apply to master. Rebase required? Wrong 
Branch? See https://yetus.apache.org/documentation/0.8.0/precommit-patchnames 
for help. {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Docker | Client=17.05.0-ce Server=17.05.0-ce Image:yetus/hbase:b002b0b |
| JIRA Issue | HBASE-21528 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12950062/HBASE-21528.v01.patch 
|
| Console output | 
https://builds.apache.org/job/PreCommit-HBASE-Build/15153/console |
| Powered by | Apache Yetus 0.8.0   http://yetus.apache.org |


This message was automatically generated.



> hbase-connectors need scala dependency convergence
> --
>
> Key: HBASE-21528
> URL: https://issues.apache.org/jira/browse/HBASE-21528
> Project: HBase
>  Issue Type: Task
>  Components: hbase-connectors
>Reporter: Artem Ervits
>Assignee: Artem Ervits
>Priority: Minor
> Attachments: HBASE-21528.v01.patch
>
>
> this issue reappeared after migration of hbase-spark to hbase-connectors 
> repo. Originally addressed in 
> https://issues.apache.org/jira/browse/HBASE-20175 
> introducing scala-maven-plugin and adding a flag to honor compatible major 
> scala version makes warning disappear.
>  
> {code:java}
> [WARNING] Expected all dependencies to require Scala version: 2.11.12
> [WARNING] org.apache.hbase.connectors.spark:hbase-spark:1.0.0-SNAPSHOT 
> requires scala version: 2.11.12
> [WARNING] org.apache.spark:spark-streaming_2.11:2.4.0 requires scala version: 
> 2.11.12
> [WARNING] org.apache.spark:spark-streaming_2.11:2.4.0 requires scala version: 
> 2.11.12
> [WARNING] org.scala-lang:scala-reflect:2.11.12 requires scala version: 2.11.12
> [WARNING] org.apache.spark:spark-tags_2.11:2.4.0 requires scala version: 
> 2.11.12
> [WARNING] com.twitter:chill_2.11:0.9.3 requires scala version: 2.11.12
> [WARNING] org.scala-lang.modules:scala-parser-combinators_2.11:1.1.0 requires 
> scala version: 2.11.12
> [WARNING] org.json4s:json4s-core_2.11:3.2.11 requires scala version: 2.11.0
> [WARNING] Multiple versions of scala libraries detected!
> [WARNING] warning: there were 17 deprecation warnings; re-run with 
> -deprecation for details
> [WARNING] one warning found
> [WARNING] Expected all dependencies to require Scala version: 2.11.12
> [WARNING] org.apache.hbase.connectors.spark:hbase-spark:1.0.0-SNAPSHOT 
> requires scala version: 2.11.12
> [WARNING] org.apache.spark:spark-streaming_2.11:2.4.0 requires scala version: 
> 2.11.12
> [WARNING] org.apache.spark:spark-streaming_2.11:2.4.0 requires scala version: 
> 2.11.12
> [WARNING] org.scala-lang:scala-reflect:2.11.12 requires scala version: 2.11.12
> [WARNING] org.apache.spark:spark-tags_2.11:2.4.0 requires scala version: 
> 2.11.12
> [WARNING] com.twitter:chill_2.11:0.9.3 requires scala version: 2.11.12
> [WARNING] org.scala-lang.modules:scala-parser-combinators_2.11:1.1.0 requires 
> scala version: 2.11.12
> [WARNING] org.json4s:json4s-core_2.11:3.2.11 requires scala version: 2.11.0
> [WARNING] Multiple versions of scala libraries detected!{code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-21528) hbase-connectors need scala dependency convergence

2018-11-29 Thread Artem Ervits (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-21528?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Artem Ervits updated HBASE-21528:
-
Description: 
this issue reappeared after migration of hbase-spark to hbase-connectors repo. 
Originally addressed in https://issues.apache.org/jira/browse/HBASE-20175 

introducing scala-maven-plugin and adding a flag to honor compatible major 
scala version makes warning disappear.

 
{code:java}
[WARNING] Expected all dependencies to require Scala version: 2.11.12
[WARNING] org.apache.hbase.connectors.spark:hbase-spark:1.0.0-SNAPSHOT requires 
scala version: 2.11.12
[WARNING] org.apache.spark:spark-streaming_2.11:2.4.0 requires scala version: 
2.11.12
[WARNING] org.apache.spark:spark-streaming_2.11:2.4.0 requires scala version: 
2.11.12
[WARNING] org.scala-lang:scala-reflect:2.11.12 requires scala version: 2.11.12
[WARNING] org.apache.spark:spark-tags_2.11:2.4.0 requires scala version: 2.11.12
[WARNING] com.twitter:chill_2.11:0.9.3 requires scala version: 2.11.12
[WARNING] org.scala-lang.modules:scala-parser-combinators_2.11:1.1.0 requires 
scala version: 2.11.12
[WARNING] org.json4s:json4s-core_2.11:3.2.11 requires scala version: 2.11.0
[WARNING] Multiple versions of scala libraries detected!
[WARNING] warning: there were 17 deprecation warnings; re-run with -deprecation 
for details
[WARNING] one warning found
[WARNING] Expected all dependencies to require Scala version: 2.11.12
[WARNING] org.apache.hbase.connectors.spark:hbase-spark:1.0.0-SNAPSHOT requires 
scala version: 2.11.12
[WARNING] org.apache.spark:spark-streaming_2.11:2.4.0 requires scala version: 
2.11.12
[WARNING] org.apache.spark:spark-streaming_2.11:2.4.0 requires scala version: 
2.11.12
[WARNING] org.scala-lang:scala-reflect:2.11.12 requires scala version: 2.11.12
[WARNING] org.apache.spark:spark-tags_2.11:2.4.0 requires scala version: 2.11.12
[WARNING] com.twitter:chill_2.11:0.9.3 requires scala version: 2.11.12
[WARNING] org.scala-lang.modules:scala-parser-combinators_2.11:1.1.0 requires 
scala version: 2.11.12
[WARNING] org.json4s:json4s-core_2.11:3.2.11 requires scala version: 2.11.0
[WARNING] Multiple versions of scala libraries detected!{code}

  was:
this issue reappeared after migration of hbase-spark to hbase-connectors repo. 
Originally addressed in https://issues.apache.org/jira/browse/HBASE-20175 

introducing scala-maven-plugin and adding a flag to honor compatible major 
scala version makes warning disappear.


> hbase-connectors need scala dependency convergence
> --
>
> Key: HBASE-21528
> URL: https://issues.apache.org/jira/browse/HBASE-21528
> Project: HBase
>  Issue Type: Task
>  Components: hbase-connectors
>Reporter: Artem Ervits
>Assignee: Artem Ervits
>Priority: Minor
> Attachments: HBASE-21528.v01.patch
>
>
> this issue reappeared after migration of hbase-spark to hbase-connectors 
> repo. Originally addressed in 
> https://issues.apache.org/jira/browse/HBASE-20175 
> introducing scala-maven-plugin and adding a flag to honor compatible major 
> scala version makes warning disappear.
>  
> {code:java}
> [WARNING] Expected all dependencies to require Scala version: 2.11.12
> [WARNING] org.apache.hbase.connectors.spark:hbase-spark:1.0.0-SNAPSHOT 
> requires scala version: 2.11.12
> [WARNING] org.apache.spark:spark-streaming_2.11:2.4.0 requires scala version: 
> 2.11.12
> [WARNING] org.apache.spark:spark-streaming_2.11:2.4.0 requires scala version: 
> 2.11.12
> [WARNING] org.scala-lang:scala-reflect:2.11.12 requires scala version: 2.11.12
> [WARNING] org.apache.spark:spark-tags_2.11:2.4.0 requires scala version: 
> 2.11.12
> [WARNING] com.twitter:chill_2.11:0.9.3 requires scala version: 2.11.12
> [WARNING] org.scala-lang.modules:scala-parser-combinators_2.11:1.1.0 requires 
> scala version: 2.11.12
> [WARNING] org.json4s:json4s-core_2.11:3.2.11 requires scala version: 2.11.0
> [WARNING] Multiple versions of scala libraries detected!
> [WARNING] warning: there were 17 deprecation warnings; re-run with 
> -deprecation for details
> [WARNING] one warning found
> [WARNING] Expected all dependencies to require Scala version: 2.11.12
> [WARNING] org.apache.hbase.connectors.spark:hbase-spark:1.0.0-SNAPSHOT 
> requires scala version: 2.11.12
> [WARNING] org.apache.spark:spark-streaming_2.11:2.4.0 requires scala version: 
> 2.11.12
> [WARNING] org.apache.spark:spark-streaming_2.11:2.4.0 requires scala version: 
> 2.11.12
> [WARNING] org.scala-lang:scala-reflect:2.11.12 requires scala version: 2.11.12
> [WARNING] org.apache.spark:spark-tags_2.11:2.4.0 requires scala version: 
> 2.11.12
> [WARNING] com.twitter:chill_2.11:0.9.3 requires scala version: 2.11.12
> [WARNING] org.scala-lang.modules:scala-parser-combinators_2.11:1.1.0 requires 
> scala version: 2.11.12
> [WARNING] 

[jira] [Updated] (HBASE-21528) hbase-connectors need scala dependency convergence

2018-11-29 Thread Artem Ervits (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-21528?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Artem Ervits updated HBASE-21528:
-
Attachment: HBASE-21528.v01.patch

> hbase-connectors need scala dependency convergence
> --
>
> Key: HBASE-21528
> URL: https://issues.apache.org/jira/browse/HBASE-21528
> Project: HBase
>  Issue Type: Task
>  Components: hbase-connectors
>Reporter: Artem Ervits
>Assignee: Artem Ervits
>Priority: Minor
> Attachments: HBASE-21528.v01.patch
>
>
> this issue reappeared after migration of hbase-spark to hbase-connectors 
> repo. Originally addressed in 
> https://issues.apache.org/jira/browse/HBASE-20175 
> introducing scala-maven-plugin and adding a flag to honor compatible major 
> scala version makes warning disappear.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-21528) hbase-connectors need scala dependency convergence

2018-11-29 Thread Artem Ervits (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-21528?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Artem Ervits updated HBASE-21528:
-
Status: Patch Available  (was: Open)

> hbase-connectors need scala dependency convergence
> --
>
> Key: HBASE-21528
> URL: https://issues.apache.org/jira/browse/HBASE-21528
> Project: HBase
>  Issue Type: Task
>  Components: hbase-connectors
>Reporter: Artem Ervits
>Assignee: Artem Ervits
>Priority: Minor
> Attachments: HBASE-21528.v01.patch
>
>
> this issue reappeared after migration of hbase-spark to hbase-connectors 
> repo. Originally addressed in 
> https://issues.apache.org/jira/browse/HBASE-20175 
> introducing scala-maven-plugin and adding a flag to honor compatible major 
> scala version makes warning disappear.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-21528) hbase-connectors need scala dependency convergence

2018-11-29 Thread Artem Ervits (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-21528?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Artem Ervits updated HBASE-21528:
-
Summary: hbase-connectors need scala dependency convergence  (was: 
hbase-connectors need scala library convergence)

> hbase-connectors need scala dependency convergence
> --
>
> Key: HBASE-21528
> URL: https://issues.apache.org/jira/browse/HBASE-21528
> Project: HBase
>  Issue Type: Task
>  Components: hbase-connectors
>Reporter: Artem Ervits
>Assignee: Artem Ervits
>Priority: Minor
>
> this issue reappeared after migration of hbase-spark to hbase-connectors 
> repo. Originally addressed in 
> https://issues.apache.org/jira/browse/HBASE-20175 
> introducing scala-maven-plugin and adding a flag to honor compatible major 
> scala version makes warning disappear.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (HBASE-21528) hbase-connectors need scala library convergence

2018-11-29 Thread Artem Ervits (JIRA)
Artem Ervits created HBASE-21528:


 Summary: hbase-connectors need scala library convergence
 Key: HBASE-21528
 URL: https://issues.apache.org/jira/browse/HBASE-21528
 Project: HBase
  Issue Type: Task
  Components: hbase-connectors
Reporter: Artem Ervits
Assignee: Artem Ervits


this issue reappeared after migration of hbase-spark to hbase-connectors repo. 
Originally addressed in https://issues.apache.org/jira/browse/HBASE-20175 

introducing scala-maven-plugin and adding a flag to honor compatible major 
scala version makes warning disappear.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21522) meta replicas appear to cause master restart to kill regionservers

2018-11-29 Thread Sergey Shelukhin (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21522?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16703665#comment-16703665
 ] 

Sergey Shelukhin commented on HBASE-21522:
--

Meta loaded event is set at the end of loadMeta, in joinCluster, that is in 
turn called in finishActiveMasterInitialization before assigning meta replicas.
So, there's a (significant) window where a region server report will go into 
else path (i.e. not "else if !isMetaLoaded()") and trigger this condition.

> meta replicas appear to cause master restart to kill regionservers
> --
>
> Key: HBASE-21522
> URL: https://issues.apache.org/jira/browse/HBASE-21522
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 3.0.0
>Reporter: Sergey Shelukhin
>Priority: Major
>
> On master restart, AM.start adds FIRST_META_REGIONINFO to regionStates; that 
> has replica ID of 0. Before the meta is loaded, 
> AssignmentManager.checkOnlineRegionsReportForMeta is called for RS reports, 
> and that also only checks for 0th replica of meta and loads it once 
> discovered.
> Once the meta is loaded, RS reports are processed normally; however nobody 
> appears to add meta replicas to regionStates.
> So, when an RS hosting one reports in, it gets killed: 
> {noformat}
> * ABORTING region server : 
> org.apache.hadoop.hbase.YouAreDeadException: Not online: hbase:meta,,1_0001
> * ABORTING region server : 
> org.apache.hadoop.hbase.YouAreDeadException: Not online: hbase:meta,,1_0002
> {noformat}
> This exception is thrown when regionStates has no record for the region.
> RS in question shut down in an orderly manner and they do have the 
> corresponding regions, that master then assigns to someone else in a few 
> minutes.
> Still, this seems less than ideal.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (HBASE-21527) [2.0] Unnecessary DEBUG log in ConnectionImplementation#isTableEnabled

2018-11-29 Thread Josh Elser (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-21527?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Josh Elser resolved HBASE-21527.

  Resolution: Fixed
Hadoop Flags: Reviewed

Pushed to branch-2.0. Thanks for the input, Stack!

> [2.0] Unnecessary DEBUG log in ConnectionImplementation#isTableEnabled
> --
>
> Key: HBASE-21527
> URL: https://issues.apache.org/jira/browse/HBASE-21527
> Project: HBase
>  Issue Type: Improvement
>  Components: Client
>Reporter: Josh Elser
>Assignee: Josh Elser
>Priority: Major
> Fix For: 2.0.3
>
> Attachments: HBASE-21524.001.branch-2.0.patch
>
>
> Too much crap coming out of isTableEnabled for "normal" situations. Don't 
> need a DEBUG message to tell us that the table is available.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Comment Edited] (HBASE-18735) Provide a fast mechanism for shutting down mini cluster

2018-11-29 Thread Artem Ervits (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-18735?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16703634#comment-16703634
 ] 

Artem Ervits edited comment on HBASE-18735 at 11/29/18 6:39 PM:


The
{code:java}
org.apache.hadoop.hbase.regionserver.TestMultiColumnScanner{code}
test passes locally, albeit runs for 575 seconds on my Mac and unrelated to my 
patch.


was (Author: dbist13):
The {org.apache.hadoop.hbase.regionserver.TestMultiColumnScanner} test passes 
locally, albeit runs for 575 seconds on my Mac and unrelated to my patch.

> Provide a fast mechanism for shutting down mini cluster
> ---
>
> Key: HBASE-18735
> URL: https://issues.apache.org/jira/browse/HBASE-18735
> Project: HBase
>  Issue Type: Wish
>Reporter: Samarth Jain
>Assignee: Artem Ervits
>Priority: Major
> Attachments: HBASE-18735.v01.patch, HBASE-18735.v02.patch, 
> HBASE-18735.v03.patch
>
>
> The current mechanism of shutting down a mini cluster through 
> HBaseTestingUtility.shutDownMiniCluster can take a lot of time when the mini 
> cluster almost has a lot of tables. A lot of this time is spent in closing 
> all the user regions. It would be nice to have a mechanism where this 
> shutdown can happen quickly without having to worry about closing these user 
> regions. At the same time, this mechanism would need to make sure that all 
> the critical system resources like file handles and network ports are still 
> released so that subsequently initialized mini clusters on the same JVM or 
> system won't run into resource issues. This would make testing using HBase 
> mini clusters much faster and immensely help out test frameworks of dependent 
> projects like Phoenix.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Comment Edited] (HBASE-18735) Provide a fast mechanism for shutting down mini cluster

2018-11-29 Thread Artem Ervits (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-18735?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16703634#comment-16703634
 ] 

Artem Ervits edited comment on HBASE-18735 at 11/29/18 6:38 PM:


The {org.apache.hadoop.hbase.regionserver.TestMultiColumnScanner} test passes 
locally, albeit runs for 575 seconds on my Mac and unrelated to my patch.


was (Author: dbist13):
The [org.apache.hadoop.hbase.regionserver.TestMultiColumnScanner] test passes 
locally, albeit runs for 575 seconds on my Mac and unrelated to my patch.

> Provide a fast mechanism for shutting down mini cluster
> ---
>
> Key: HBASE-18735
> URL: https://issues.apache.org/jira/browse/HBASE-18735
> Project: HBase
>  Issue Type: Wish
>Reporter: Samarth Jain
>Assignee: Artem Ervits
>Priority: Major
> Attachments: HBASE-18735.v01.patch, HBASE-18735.v02.patch, 
> HBASE-18735.v03.patch
>
>
> The current mechanism of shutting down a mini cluster through 
> HBaseTestingUtility.shutDownMiniCluster can take a lot of time when the mini 
> cluster almost has a lot of tables. A lot of this time is spent in closing 
> all the user regions. It would be nice to have a mechanism where this 
> shutdown can happen quickly without having to worry about closing these user 
> regions. At the same time, this mechanism would need to make sure that all 
> the critical system resources like file handles and network ports are still 
> released so that subsequently initialized mini clusters on the same JVM or 
> system won't run into resource issues. This would make testing using HBase 
> mini clusters much faster and immensely help out test frameworks of dependent 
> projects like Phoenix.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-18735) Provide a fast mechanism for shutting down mini cluster

2018-11-29 Thread Artem Ervits (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-18735?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16703634#comment-16703634
 ] 

Artem Ervits commented on HBASE-18735:
--

The [org.apache.hadoop.hbase.regionserver.TestMultiColumnScanner] test passes 
locally, albeit runs for 575 seconds on my Mac and unrelated to my patch.

> Provide a fast mechanism for shutting down mini cluster
> ---
>
> Key: HBASE-18735
> URL: https://issues.apache.org/jira/browse/HBASE-18735
> Project: HBase
>  Issue Type: Wish
>Reporter: Samarth Jain
>Assignee: Artem Ervits
>Priority: Major
> Attachments: HBASE-18735.v01.patch, HBASE-18735.v02.patch, 
> HBASE-18735.v03.patch
>
>
> The current mechanism of shutting down a mini cluster through 
> HBaseTestingUtility.shutDownMiniCluster can take a lot of time when the mini 
> cluster almost has a lot of tables. A lot of this time is spent in closing 
> all the user regions. It would be nice to have a mechanism where this 
> shutdown can happen quickly without having to worry about closing these user 
> regions. At the same time, this mechanism would need to make sure that all 
> the critical system resources like file handles and network ports are still 
> released so that subsequently initialized mini clusters on the same JVM or 
> system won't run into resource issues. This would make testing using HBase 
> mini clusters much faster and immensely help out test frameworks of dependent 
> projects like Phoenix.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21464) Splitting blocked with meta NSRE during split transaction

2018-11-29 Thread Andrew Purtell (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21464?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16703602#comment-16703602
 ] 

Andrew Purtell commented on HBASE-21464:


Thanks for the review [~allan163]. Will check the latest with ITBLL and commit 
if good.

> Splitting blocked with meta NSRE during split transaction
> -
>
> Key: HBASE-21464
> URL: https://issues.apache.org/jira/browse/HBASE-21464
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.5.0, 1.4.3, 1.4.4, 1.4.5, 1.4.6, 1.4.8, 1.4.7
>Reporter: Andrew Purtell
>Assignee: Andrew Purtell
>Priority: Blocker
> Fix For: 1.5.0, 1.4.9
>
> Attachments: HBASE-21464-branch-1.patch, HBASE-21464-branch-1.patch
>
>
> Splitting is blocked during split transaction. The split worker is trying to 
> update meta but isn't able to relocate it after NSRE:
> {noformat}
> 2018-11-09 17:50:45,277 INFO  
> [regionserver/ip-172-31-5-92.us-west-2.compute.internal/172.31.5.92:8120-splits-1541785709434]
>  client.RpcRetryingCaller: Call exception, tries=13, retries=350, 
> started=88590 ms ago, cancelled=false, 
> msg=org.apache.hadoop.hbase.NotServingRegionException: Region hbase:meta,,1 
> is not online on ip-172-31-13-83.us-west-2.compute.internal,8120,1541785618832
>      at 
> org.apache.hadoop.hbase.regionserver.HRegionServer.getRegionByEncodedName(HRegionServer.java:3088)
>         at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.getRegion(RSRpcServices.java:1271)
>         at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:2198)
>         at 
> org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:36617)
>         at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2396)
>         at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124)
>         at 
> org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:297)
>         at 
> org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:277)row 
> 'test,,1541785709452.5ba6596f0050c2dab969d152829227c6.44' on table 
> 'hbase:meta' at region=hbase:meta,1.1588230740, 
> hostname=ip-172-31-15-225.us-west-2.compute.internal,8120,1541785640586, 
> seqNum=0{noformat}
> Clients, in this case YCSB, are hung with part of the keyspace missing:
> {noformat}
> 2018-11-09 17:51:06,033 DEBUG [hconnection-0x5739e567-shared--pool1-t165] 
> client.ConnectionManager$HConnectionImplementation: locateRegionInMeta 
> parentTable=hbase:meta, metaLocation=, attempt=14 of 35 failed; retrying 
> after sleep of 20158 because: No server address listed in hbase:meta for 
> region 
> test,user307326104267982763,1541785754600.ef90030b05cb02305b75e9bfbc3ee081. 
> containing row user3301635648728421323{noformat}
> Balancing cannot run indefinitely because the split transaction is stuck
> {noformat}
> 2018-11-09 17:49:55,478 DEBUG 
> [RpcServer.default.FPBQ.Fifo.handler=29,queue=2,port=8100] master.HMaster: 
> Not running balancer because 3 region(s) in transition: 
> [{ef90030b05cb02305b75e9bfbc3ee081 state=SPLITTING_NEW, ts=1541785754606, 
> server=ip-172-31-5-92.us-west-2.compute.internal,8120,1541785626417}, 
> {5ba6596f0050c2dab969d152829227c6 state=SPLITTING, ts=1541785754606, 
> server=ip-172-31-5-92.us-west-2.compute{noformat}
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21518) TestMasterFailoverWithProcedures is flaky

2018-11-29 Thread Sean Busbey (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21518?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16703571#comment-16703571
 ] 

Sean Busbey commented on HBASE-21518:
-

+1

> TestMasterFailoverWithProcedures is flaky
> -
>
> Key: HBASE-21518
> URL: https://issues.apache.org/jira/browse/HBASE-21518
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 2.2.0, 2.0.3, 2.1.2
>Reporter: Peter Somogyi
>Assignee: Peter Somogyi
>Priority: Major
> Attachments: HBASE-21518-v1.patch, output.txt
>
>
> TestMasterFailoverWithProcedures test is failing frequently, times out. I 
> faced this failure on 2.0.3RC0 vote and it also appears on multiple flaky 
> dashboards.
> branch-2: 
> [https://builds.apache.org/view/H-L/view/HBase/job/HBase-Flaky-Tests/job/branch-2/2007/]
> branch-2.1: 
> [https://builds.apache.org/view/H-L/view/HBase/job/HBase-Flaky-Tests/job/branch-2.1/2002/]
> branch-2.0: 
> [https://builds.apache.org/view/H-L/view/HBase/job/HBase-Flaky-Tests/job/branch-2.0/1988/]
>   
> {noformat}
> [INFO] Running 
> org.apache.hadoop.hbase.master.procedure.TestMasterFailoverWithProcedures
> [ERROR] Tests run: 4, Failures: 0, Errors: 2, Skipped: 0, Time elapsed: 
> 780.648 s <<< FAILURE! - in 
> org.apache.hadoop.hbase.master.procedure.TestMasterFailoverWithProcedures
> [ERROR] 
> org.apache.hadoop.hbase.master.procedure.TestMasterFailoverWithProcedures  
> Time elapsed: 749.024 s  <<< ERROR!
> org.junit.runners.model.TestTimedOutException: test timed out after 780 
> seconds
>   at 
> org.apache.hadoop.hbase.master.procedure.TestMasterFailoverWithProcedures.tearDown(TestMasterFailoverWithProcedures.java:86)
> [ERROR] 
> org.apache.hadoop.hbase.master.procedure.TestMasterFailoverWithProcedures  
> Time elapsed: 749.051 s  <<< ERROR!
> java.lang.Exception: Appears to be stuck in thread RS-EventLoopGroup-3-2
> {noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-21521) Expose master startup status via JMX and web UI

2018-11-29 Thread Andrew Purtell (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-21521?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Purtell updated HBASE-21521:
---
Description: 
Add an internal API to the master for tracking startup progress. Expose this 
information via JMX.

Modify the master to bring the web UI up sooner. Will require tweaks to various 
views to prevent attempts to retrieve state before the master fully up (or else 
expect NPEs). Currently, before the master has fully initialized an attempt to 
use the web UI will return a 500 error code and display an error page.

Finally, update the web UI to display startup progress, like HDFS-4249. 

Filing this for branch-1. Need to check what if anything is available or 
improved in branch-2 and master.

  was:
Add an internal API to the master for tracking startup progress. Expose this 
information via JMX.

Modify the master to bring the web UI up sooner. Will require tweaks to various 
views to prevent attempts to retrieve state before the master fully up (or else 
expect NPEs). Currently, before the master has fully initialized an attempt to 
use the web UI will return a 500 error code and display an error page.

Finally, update the web UI to display startup progress, like HDFS-4249. 
(Example: [^HDFS-4249-2.png)]

Filing this for branch-1. Need to check what if anything is available or 
improved in branch-2 and master.


> Expose master startup status via JMX and web UI
> ---
>
> Key: HBASE-21521
> URL: https://issues.apache.org/jira/browse/HBASE-21521
> Project: HBase
>  Issue Type: Improvement
>  Components: master, UI
>Reporter: Andrew Purtell
>Priority: Major
>
> Add an internal API to the master for tracking startup progress. Expose this 
> information via JMX.
> Modify the master to bring the web UI up sooner. Will require tweaks to 
> various views to prevent attempts to retrieve state before the master fully 
> up (or else expect NPEs). Currently, before the master has fully initialized 
> an attempt to use the web UI will return a 500 error code and display an 
> error page.
> Finally, update the web UI to display startup progress, like HDFS-4249. 
> Filing this for branch-1. Need to check what if anything is available or 
> improved in branch-2 and master.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-18735) Provide a fast mechanism for shutting down mini cluster

2018-11-29 Thread Hadoop QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-18735?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16703531#comment-16703531
 ] 

Hadoop QA commented on HBASE-18735:
---

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
20s{color} | {color:blue} Docker mode activated. {color} |
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} hbaseanti {color} | {color:green}  0m  
0s{color} | {color:green} Patch does not have any anti-patterns. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green}  0m 
 0s{color} | {color:green} The patch appears to include 2 new or modified test 
files. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  4m 
39s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m 
58s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  1m 
17s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} shadedjars {color} | {color:green}  4m 
23s{color} | {color:green} branch has no errors when building our shaded 
downstream artifacts. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  2m 
22s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
34s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  4m 
30s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m 
58s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  1m 
58s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  1m 
15s{color} | {color:green} hbase-server: The patch generated 0 new + 228 
unchanged - 4 fixed = 228 total (was 232) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} shadedjars {color} | {color:green}  4m 
18s{color} | {color:green} patch has no errors when building our shaded 
downstream artifacts. {color} |
| {color:green}+1{color} | {color:green} hadoopcheck {color} | {color:green}  
9m 49s{color} | {color:green} Patch does not cause any errors with Hadoop 2.7.4 
or 3.0.0. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  2m 
28s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
36s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:red}-1{color} | {color:red} unit {color} | {color:red}137m 51s{color} 
| {color:red} hbase-server in the patch failed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
23s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black}179m 14s{color} | 
{color:black} {color} |
\\
\\
|| Reason || Tests ||
| Failed junit tests | hadoop.hbase.regionserver.TestMultiColumnScanner |
\\
\\
|| Subsystem || Report/Notes ||
| Docker | Client=17.05.0-ce Server=17.05.0-ce Image:yetus/hbase:b002b0b |
| JIRA Issue | HBASE-18735 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12950018/HBASE-18735.v03.patch 
|
| Optional Tests |  dupname  asflicense  javac  javadoc  unit  findbugs  
shadedjars  hadoopcheck  hbaseanti  checkstyle  compile  |
| uname | Linux db5673e13012 4.4.0-139-generic #165~14.04.1-Ubuntu SMP Wed Oct 
31 10:55:11 UTC 2018 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/home/jenkins/jenkins-slave/workspace/PreCommit-HBASE-Build/component/dev-support/hbase-personality.sh
 |
| git revision | master / f1f2b5a038 |
| maven | version: Apache Maven 3.5.4 
(1edded0938998edf8bf061f1ceb3cfdeccf443fe; 2018-06-17T18:33:14Z) |
| Default Java | 1.8.0_181 |
| findbugs | v3.1.0-RC3 |
| unit | 
https://builds.apache.org/job/PreCommit-HBASE-Build/15150/artifact/patchprocess/patch-unit-hbase-server.txt
 |
|  Test Results | 
https://builds.apache.org/job/PreCommit-HBASE-Build/15150/testReport/ |
| Max. process+thread count | 4668 (vs. 

[jira] [Commented] (HBASE-21527) [2.0] Unnecessary DEBUG log in ConnectionImplementation#isTableEnabled

2018-11-29 Thread stack (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21527?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16703526#comment-16703526
 ] 

stack commented on HBASE-21527:
---

Thanks [~elserj] +1 for branch-2.0. Will include in next RC if one... else 
2.0.4.

> [2.0] Unnecessary DEBUG log in ConnectionImplementation#isTableEnabled
> --
>
> Key: HBASE-21527
> URL: https://issues.apache.org/jira/browse/HBASE-21527
> Project: HBase
>  Issue Type: Improvement
>  Components: Client
>Reporter: Josh Elser
>Assignee: Josh Elser
>Priority: Major
> Fix For: 2.0.3
>
> Attachments: HBASE-21524.001.branch-2.0.patch
>
>
> Too much crap coming out of isTableEnabled for "normal" situations. Don't 
> need a DEBUG message to tell us that the table is available.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21518) TestMasterFailoverWithProcedures is flaky

2018-11-29 Thread Peter Somogyi (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21518?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16703523#comment-16703523
 ] 

Peter Somogyi commented on HBASE-21518:
---

With the help of [~busbey] we figured out that both masters were considered as 
active masters. This patch adds an extra check whether master is already 
stopped so a previously killed master cannot claim it is still active.

> TestMasterFailoverWithProcedures is flaky
> -
>
> Key: HBASE-21518
> URL: https://issues.apache.org/jira/browse/HBASE-21518
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 2.2.0, 2.0.3, 2.1.2
>Reporter: Peter Somogyi
>Assignee: Peter Somogyi
>Priority: Major
> Attachments: HBASE-21518-v1.patch, output.txt
>
>
> TestMasterFailoverWithProcedures test is failing frequently, times out. I 
> faced this failure on 2.0.3RC0 vote and it also appears on multiple flaky 
> dashboards.
> branch-2: 
> [https://builds.apache.org/view/H-L/view/HBase/job/HBase-Flaky-Tests/job/branch-2/2007/]
> branch-2.1: 
> [https://builds.apache.org/view/H-L/view/HBase/job/HBase-Flaky-Tests/job/branch-2.1/2002/]
> branch-2.0: 
> [https://builds.apache.org/view/H-L/view/HBase/job/HBase-Flaky-Tests/job/branch-2.0/1988/]
>   
> {noformat}
> [INFO] Running 
> org.apache.hadoop.hbase.master.procedure.TestMasterFailoverWithProcedures
> [ERROR] Tests run: 4, Failures: 0, Errors: 2, Skipped: 0, Time elapsed: 
> 780.648 s <<< FAILURE! - in 
> org.apache.hadoop.hbase.master.procedure.TestMasterFailoverWithProcedures
> [ERROR] 
> org.apache.hadoop.hbase.master.procedure.TestMasterFailoverWithProcedures  
> Time elapsed: 749.024 s  <<< ERROR!
> org.junit.runners.model.TestTimedOutException: test timed out after 780 
> seconds
>   at 
> org.apache.hadoop.hbase.master.procedure.TestMasterFailoverWithProcedures.tearDown(TestMasterFailoverWithProcedures.java:86)
> [ERROR] 
> org.apache.hadoop.hbase.master.procedure.TestMasterFailoverWithProcedures  
> Time elapsed: 749.051 s  <<< ERROR!
> java.lang.Exception: Appears to be stuck in thread RS-EventLoopGroup-3-2
> {noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-21518) TestMasterFailoverWithProcedures is flaky

2018-11-29 Thread Peter Somogyi (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-21518?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Peter Somogyi updated HBASE-21518:
--
Status: Patch Available  (was: In Progress)

> TestMasterFailoverWithProcedures is flaky
> -
>
> Key: HBASE-21518
> URL: https://issues.apache.org/jira/browse/HBASE-21518
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 2.2.0, 2.0.3, 2.1.2
>Reporter: Peter Somogyi
>Assignee: Peter Somogyi
>Priority: Major
> Attachments: HBASE-21518-v1.patch, output.txt
>
>
> TestMasterFailoverWithProcedures test is failing frequently, times out. I 
> faced this failure on 2.0.3RC0 vote and it also appears on multiple flaky 
> dashboards.
> branch-2: 
> [https://builds.apache.org/view/H-L/view/HBase/job/HBase-Flaky-Tests/job/branch-2/2007/]
> branch-2.1: 
> [https://builds.apache.org/view/H-L/view/HBase/job/HBase-Flaky-Tests/job/branch-2.1/2002/]
> branch-2.0: 
> [https://builds.apache.org/view/H-L/view/HBase/job/HBase-Flaky-Tests/job/branch-2.0/1988/]
>   
> {noformat}
> [INFO] Running 
> org.apache.hadoop.hbase.master.procedure.TestMasterFailoverWithProcedures
> [ERROR] Tests run: 4, Failures: 0, Errors: 2, Skipped: 0, Time elapsed: 
> 780.648 s <<< FAILURE! - in 
> org.apache.hadoop.hbase.master.procedure.TestMasterFailoverWithProcedures
> [ERROR] 
> org.apache.hadoop.hbase.master.procedure.TestMasterFailoverWithProcedures  
> Time elapsed: 749.024 s  <<< ERROR!
> org.junit.runners.model.TestTimedOutException: test timed out after 780 
> seconds
>   at 
> org.apache.hadoop.hbase.master.procedure.TestMasterFailoverWithProcedures.tearDown(TestMasterFailoverWithProcedures.java:86)
> [ERROR] 
> org.apache.hadoop.hbase.master.procedure.TestMasterFailoverWithProcedures  
> Time elapsed: 749.051 s  <<< ERROR!
> java.lang.Exception: Appears to be stuck in thread RS-EventLoopGroup-3-2
> {noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-21518) TestMasterFailoverWithProcedures is flaky

2018-11-29 Thread Peter Somogyi (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-21518?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Peter Somogyi updated HBASE-21518:
--
Attachment: HBASE-21518-v1.patch

> TestMasterFailoverWithProcedures is flaky
> -
>
> Key: HBASE-21518
> URL: https://issues.apache.org/jira/browse/HBASE-21518
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 2.2.0, 2.0.3, 2.1.2
>Reporter: Peter Somogyi
>Assignee: Peter Somogyi
>Priority: Major
> Attachments: HBASE-21518-v1.patch, output.txt
>
>
> TestMasterFailoverWithProcedures test is failing frequently, times out. I 
> faced this failure on 2.0.3RC0 vote and it also appears on multiple flaky 
> dashboards.
> branch-2: 
> [https://builds.apache.org/view/H-L/view/HBase/job/HBase-Flaky-Tests/job/branch-2/2007/]
> branch-2.1: 
> [https://builds.apache.org/view/H-L/view/HBase/job/HBase-Flaky-Tests/job/branch-2.1/2002/]
> branch-2.0: 
> [https://builds.apache.org/view/H-L/view/HBase/job/HBase-Flaky-Tests/job/branch-2.0/1988/]
>   
> {noformat}
> [INFO] Running 
> org.apache.hadoop.hbase.master.procedure.TestMasterFailoverWithProcedures
> [ERROR] Tests run: 4, Failures: 0, Errors: 2, Skipped: 0, Time elapsed: 
> 780.648 s <<< FAILURE! - in 
> org.apache.hadoop.hbase.master.procedure.TestMasterFailoverWithProcedures
> [ERROR] 
> org.apache.hadoop.hbase.master.procedure.TestMasterFailoverWithProcedures  
> Time elapsed: 749.024 s  <<< ERROR!
> org.junit.runners.model.TestTimedOutException: test timed out after 780 
> seconds
>   at 
> org.apache.hadoop.hbase.master.procedure.TestMasterFailoverWithProcedures.tearDown(TestMasterFailoverWithProcedures.java:86)
> [ERROR] 
> org.apache.hadoop.hbase.master.procedure.TestMasterFailoverWithProcedures  
> Time elapsed: 749.051 s  <<< ERROR!
> java.lang.Exception: Appears to be stuck in thread RS-EventLoopGroup-3-2
> {noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-18735) Provide a fast mechanism for shutting down mini cluster

2018-11-29 Thread Artem Ervits (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-18735?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16703508#comment-16703508
 ] 

Artem Ervits commented on HBASE-18735:
--

[~jamestaylor] it's hard to say as I imagine the proportion of improvement is 
tied to number of tables and rows ingested. I originally tested with 1, 10 and 
100 tables. Improvement was 1 second faster for 1 table, 5 seconds or so for 10 
tables and 50 seconds for 100 tables. On 100 tables, improvement is 12%. If you 
want to identify a few tests in Phoenix to test this on, I will be happy to 
take this upon myself, let's do this in a Phoenix Jira, feel free to assign to 
me?

> Provide a fast mechanism for shutting down mini cluster
> ---
>
> Key: HBASE-18735
> URL: https://issues.apache.org/jira/browse/HBASE-18735
> Project: HBase
>  Issue Type: Wish
>Reporter: Samarth Jain
>Assignee: Artem Ervits
>Priority: Major
> Attachments: HBASE-18735.v01.patch, HBASE-18735.v02.patch, 
> HBASE-18735.v03.patch
>
>
> The current mechanism of shutting down a mini cluster through 
> HBaseTestingUtility.shutDownMiniCluster can take a lot of time when the mini 
> cluster almost has a lot of tables. A lot of this time is spent in closing 
> all the user regions. It would be nice to have a mechanism where this 
> shutdown can happen quickly without having to worry about closing these user 
> regions. At the same time, this mechanism would need to make sure that all 
> the critical system resources like file handles and network ports are still 
> released so that subsequently initialized mini clusters on the same JVM or 
> system won't run into resource issues. This would make testing using HBase 
> mini clusters much faster and immensely help out test frameworks of dependent 
> projects like Phoenix.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-21524) Unnecessary DEBUG log in ConnectionImplementation#isTableEnabled

2018-11-29 Thread Josh Elser (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-21524?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Josh Elser updated HBASE-21524:
---
  Resolution: Fixed
Hadoop Flags: Reviewed
  Status: Resolved  (was: Patch Available)

Thanks for the review, Stack. Created HBASE-21527 for 2.0 backport.

> Unnecessary DEBUG log in ConnectionImplementation#isTableEnabled
> 
>
> Key: HBASE-21524
> URL: https://issues.apache.org/jira/browse/HBASE-21524
> Project: HBase
>  Issue Type: Improvement
>  Components: Client
>Reporter: Josh Elser
>Assignee: Josh Elser
>Priority: Major
> Fix For: 3.0.0, 2.2.0, 2.1.2
>
> Attachments: HBASE-21524.001.branch-2.0.patch
>
>
> Too much crap coming out of isTableEnabled for "normal" situations. Don't 
> need a DEBUG message to tell us that the table is available.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-21524) Unnecessary DEBUG log in ConnectionImplementation#isTableEnabled

2018-11-29 Thread Josh Elser (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-21524?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Josh Elser updated HBASE-21524:
---
Fix Version/s: (was: 2.0.3)
   3.0.0

> Unnecessary DEBUG log in ConnectionImplementation#isTableEnabled
> 
>
> Key: HBASE-21524
> URL: https://issues.apache.org/jira/browse/HBASE-21524
> Project: HBase
>  Issue Type: Improvement
>  Components: Client
>Reporter: Josh Elser
>Assignee: Josh Elser
>Priority: Major
> Fix For: 3.0.0, 2.2.0, 2.1.2
>
> Attachments: HBASE-21524.001.branch-2.0.patch
>
>
> Too much crap coming out of isTableEnabled for "normal" situations. Don't 
> need a DEBUG message to tell us that the table is available.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21527) [2.0] Unnecessary DEBUG log in ConnectionImplementation#isTableEnabled

2018-11-29 Thread Josh Elser (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21527?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16703495#comment-16703495
 ] 

Josh Elser commented on HBASE-21527:


FYI [~stack], spun this out for 2.0.

> [2.0] Unnecessary DEBUG log in ConnectionImplementation#isTableEnabled
> --
>
> Key: HBASE-21527
> URL: https://issues.apache.org/jira/browse/HBASE-21527
> Project: HBase
>  Issue Type: Improvement
>  Components: Client
>Reporter: Josh Elser
>Assignee: Josh Elser
>Priority: Major
> Fix For: 2.0.3
>
> Attachments: HBASE-21524.001.branch-2.0.patch
>
>
> Too much crap coming out of isTableEnabled for "normal" situations. Don't 
> need a DEBUG message to tell us that the table is available.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-21527) [2.0] Unnecessary DEBUG log in ConnectionImplementation#isTableEnabled

2018-11-29 Thread Josh Elser (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-21527?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Josh Elser updated HBASE-21527:
---
Fix Version/s: (was: 2.1.2)
   (was: 2.2.0)

> [2.0] Unnecessary DEBUG log in ConnectionImplementation#isTableEnabled
> --
>
> Key: HBASE-21527
> URL: https://issues.apache.org/jira/browse/HBASE-21527
> Project: HBase
>  Issue Type: Improvement
>  Components: Client
>Reporter: Josh Elser
>Assignee: Josh Elser
>Priority: Major
> Fix For: 2.0.3
>
> Attachments: HBASE-21524.001.branch-2.0.patch
>
>
> Too much crap coming out of isTableEnabled for "normal" situations. Don't 
> need a DEBUG message to tell us that the table is available.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (HBASE-21527) [2.0] Unnecessary DEBUG log in ConnectionImplementation#isTableEnabled

2018-11-29 Thread Josh Elser (JIRA)
Josh Elser created HBASE-21527:
--

 Summary: [2.0] Unnecessary DEBUG log in 
ConnectionImplementation#isTableEnabled
 Key: HBASE-21527
 URL: https://issues.apache.org/jira/browse/HBASE-21527
 Project: HBase
  Issue Type: Improvement
  Components: Client
Reporter: Josh Elser
Assignee: Josh Elser
 Fix For: 2.2.0, 2.0.3, 2.1.2
 Attachments: HBASE-21524.001.branch-2.0.patch

Too much crap coming out of isTableEnabled for "normal" situations. Don't need 
a DEBUG message to tell us that the table is available.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-21523) Chatty DEBUG logging in Master log

2018-11-29 Thread Josh Elser (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-21523?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Josh Elser updated HBASE-21523:
---
  Resolution: Fixed
Hadoop Flags: Reviewed
  Status: Resolved  (was: Patch Available)

Pushed to master. Thanks, Peter!

> Chatty DEBUG logging in Master log
> --
>
> Key: HBASE-21523
> URL: https://issues.apache.org/jira/browse/HBASE-21523
> Project: HBase
>  Issue Type: Improvement
>  Components: backuprestore
>Reporter: Josh Elser
>Assignee: Josh Elser
>Priority: Major
> Fix For: 3.0.0
>
> Attachments: HBASE-21523.001.patch, HBASE-21523.002.patch
>
>
> Noticed this in testing, a lot of spam in the Master log at DEBUG.
> {noformat}
> 2018-11-27 01:12:28,322 DEBUG [ForkJoinPool-1-worker-6] impl.BackupMetaTable: 
> Backup table backup:system exists and available
> 2018-11-27 01:12:28,326 DEBUG [ForkJoinPool-1-worker-6] impl.BackupMetaTable: 
> Backup table backup:system_bulk exists and available
> 2018-11-27 01:12:28,341 DEBUG [ForkJoinPool-1-worker-6] impl.BackupMetaTable: 
> Backup table backup:system exists and available
> 2018-11-27 01:12:28,345 DEBUG [ForkJoinPool-1-worker-6] impl.BackupMetaTable: 
> Backup table backup:system_bulk exists and available
> {noformat}
> FYI [~vrodionov]



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21523) Chatty DEBUG logging in Master log

2018-11-29 Thread Hadoop QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21523?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16703470#comment-16703470
 ] 

Hadoop QA commented on HBASE-21523:
---

| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
14s{color} | {color:blue} Docker mode activated. {color} |
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} hbaseanti {color} | {color:green}  0m  
0s{color} | {color:green} Patch does not have any anti-patterns. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:orange}-0{color} | {color:orange} test4tests {color} | {color:orange}  
0m  0s{color} | {color:orange} The patch doesn't appear to include any new or 
modified tests. Please justify why no new tests are needed for this patch. Also 
please list what manual steps were performed to verify this patch. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  5m 
29s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
37s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
16s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} shadedjars {color} | {color:green}  4m 
38s{color} | {color:green} branch has no errors when building our shaded 
downstream artifacts. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  0m 
46s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
18s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  4m 
56s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
36s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
36s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
17s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} shadedjars {color} | {color:green}  4m 
55s{color} | {color:green} patch has no errors when building our shaded 
downstream artifacts. {color} |
| {color:green}+1{color} | {color:green} hadoopcheck {color} | {color:green} 
10m 31s{color} | {color:green} Patch does not cause any errors with Hadoop 
2.7.4 or 3.0.0. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  0m 
53s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
18s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} unit {color} | {color:green} 12m 
45s{color} | {color:green} hbase-backup in the patch passed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
18s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 48m 18s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Docker | Client=17.05.0-ce Server=17.05.0-ce Image:yetus/hbase:b002b0b |
| JIRA Issue | HBASE-21523 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12950026/HBASE-21523.002.patch 
|
| Optional Tests |  dupname  asflicense  javac  javadoc  unit  findbugs  
shadedjars  hadoopcheck  hbaseanti  checkstyle  compile  |
| uname | Linux 048dc562518d 4.4.0-139-generic #165~14.04.1-Ubuntu SMP Wed Oct 
31 10:55:11 UTC 2018 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/home/jenkins/jenkins-slave/workspace/PreCommit-HBASE-Build@2/component/dev-support/hbase-personality.sh
 |
| git revision | master / f1f2b5a038 |
| maven | version: Apache Maven 3.5.4 
(1edded0938998edf8bf061f1ceb3cfdeccf443fe; 2018-06-17T18:33:14Z) |
| Default Java | 1.8.0_181 |
| findbugs | v3.1.0-RC3 |
|  Test Results | 
https://builds.apache.org/job/PreCommit-HBASE-Build/15151/testReport/ |
| Max. process+thread count | 4125 (vs. ulimit of 1) |
| modules | C: hbase-backup U: hbase-backup |
| Console output | 

[jira] [Commented] (HBASE-21523) Chatty DEBUG logging in Master log

2018-11-29 Thread Josh Elser (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21523?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16703472#comment-16703472
 ] 

Josh Elser commented on HBASE-21523:


Thanks much, Peter!

> Chatty DEBUG logging in Master log
> --
>
> Key: HBASE-21523
> URL: https://issues.apache.org/jira/browse/HBASE-21523
> Project: HBase
>  Issue Type: Improvement
>  Components: backuprestore
>Reporter: Josh Elser
>Assignee: Josh Elser
>Priority: Major
> Fix For: 3.0.0
>
> Attachments: HBASE-21523.001.patch, HBASE-21523.002.patch
>
>
> Noticed this in testing, a lot of spam in the Master log at DEBUG.
> {noformat}
> 2018-11-27 01:12:28,322 DEBUG [ForkJoinPool-1-worker-6] impl.BackupMetaTable: 
> Backup table backup:system exists and available
> 2018-11-27 01:12:28,326 DEBUG [ForkJoinPool-1-worker-6] impl.BackupMetaTable: 
> Backup table backup:system_bulk exists and available
> 2018-11-27 01:12:28,341 DEBUG [ForkJoinPool-1-worker-6] impl.BackupMetaTable: 
> Backup table backup:system exists and available
> 2018-11-27 01:12:28,345 DEBUG [ForkJoinPool-1-worker-6] impl.BackupMetaTable: 
> Backup table backup:system_bulk exists and available
> {noformat}
> FYI [~vrodionov]



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-18735) Provide a fast mechanism for shutting down mini cluster

2018-11-29 Thread James Taylor (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-18735?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16703468#comment-16703468
 ] 

James Taylor commented on HBASE-18735:
--

Thanks, [~dbist]. This is definitely useful. You mentioned that it’s 40-50 
seconds faster. What’s the percentage improvement? If you change Phoenix to use 
this new mechanism, are the test runs faster? In Phoenix, to prevent the long 
shutdown time, we currently keep the mini cluster up across test suites and 
asynchronously drop tables to prevent OOM errors (since we create so many of 
them). If startup/shutdown time is fast, we could instead do that.

> Provide a fast mechanism for shutting down mini cluster
> ---
>
> Key: HBASE-18735
> URL: https://issues.apache.org/jira/browse/HBASE-18735
> Project: HBase
>  Issue Type: Wish
>Reporter: Samarth Jain
>Assignee: Artem Ervits
>Priority: Major
> Attachments: HBASE-18735.v01.patch, HBASE-18735.v02.patch, 
> HBASE-18735.v03.patch
>
>
> The current mechanism of shutting down a mini cluster through 
> HBaseTestingUtility.shutDownMiniCluster can take a lot of time when the mini 
> cluster almost has a lot of tables. A lot of this time is spent in closing 
> all the user regions. It would be nice to have a mechanism where this 
> shutdown can happen quickly without having to worry about closing these user 
> regions. At the same time, this mechanism would need to make sure that all 
> the critical system resources like file handles and network ports are still 
> released so that subsequently initialized mini clusters on the same JVM or 
> system won't run into resource issues. This would make testing using HBase 
> mini clusters much faster and immensely help out test frameworks of dependent 
> projects like Phoenix.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21523) Chatty DEBUG logging in Master log

2018-11-29 Thread Peter Somogyi (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21523?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16703469#comment-16703469
 ] 

Peter Somogyi commented on HBASE-21523:
---

+1 on v2.

> Chatty DEBUG logging in Master log
> --
>
> Key: HBASE-21523
> URL: https://issues.apache.org/jira/browse/HBASE-21523
> Project: HBase
>  Issue Type: Improvement
>  Components: backuprestore
>Reporter: Josh Elser
>Assignee: Josh Elser
>Priority: Major
> Fix For: 3.0.0
>
> Attachments: HBASE-21523.001.patch, HBASE-21523.002.patch
>
>
> Noticed this in testing, a lot of spam in the Master log at DEBUG.
> {noformat}
> 2018-11-27 01:12:28,322 DEBUG [ForkJoinPool-1-worker-6] impl.BackupMetaTable: 
> Backup table backup:system exists and available
> 2018-11-27 01:12:28,326 DEBUG [ForkJoinPool-1-worker-6] impl.BackupMetaTable: 
> Backup table backup:system_bulk exists and available
> 2018-11-27 01:12:28,341 DEBUG [ForkJoinPool-1-worker-6] impl.BackupMetaTable: 
> Backup table backup:system exists and available
> 2018-11-27 01:12:28,345 DEBUG [ForkJoinPool-1-worker-6] impl.BackupMetaTable: 
> Backup table backup:system_bulk exists and available
> {noformat}
> FYI [~vrodionov]



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21300) Fix the wrong reference file path when restoring snapshots for tables with MOB columns

2018-11-29 Thread Hudson (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21300?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16703446#comment-16703446
 ] 

Hudson commented on HBASE-21300:


Results for branch branch-2.1
[build #645 on 
builds.a.o|https://builds.apache.org/job/HBase%20Nightly/job/branch-2.1/645/]: 
(x) *{color:red}-1 overall{color}*

details (if available):

(/) {color:green}+1 general checks{color}
-- For more information [see general 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-2.1/645//General_Nightly_Build_Report/]




(/) {color:green}+1 jdk8 hadoop2 checks{color}
-- For more information [see jdk8 (hadoop2) 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-2.1/645//JDK8_Nightly_Build_Report_(Hadoop2)/]


(x) {color:red}-1 jdk8 hadoop3 checks{color}
-- For more information [see jdk8 (hadoop3) 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-2.1/645//JDK8_Nightly_Build_Report_(Hadoop3)/]


(/) {color:green}+1 source release artifact{color}
-- See build output for details.


(/) {color:green}+1 client integration test{color}


> Fix the wrong reference file path when restoring snapshots for tables with 
> MOB columns
> --
>
> Key: HBASE-21300
> URL: https://issues.apache.org/jira/browse/HBASE-21300
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 3.0.0, 2.2.0
>Reporter: Yi Mei
>Assignee: Yi Mei
>Priority: Major
> Fix For: 3.0.0, 2.2.0, 2.0.3, 2.1.2
>
> Attachments: HBASE-21300.branch-2.0.001.patch, 
> HBASE-21300.branch-2.1.001.patch, HBASE-21300.master.001.patch, 
> HBASE-21300.master.002.patch, HBASE-21300.master.003.patch, 
> HBASE-21300.master.004.patch, HBASE-21300.master.005.patch, 
> HBASE-21300.master.006.patch, HBASE-21300.master.007.patch, 
> HBASE-21300.master.008.patch
>
>
> When restoring snapshots for tables with MOB columns, the reference files for 
> mob region are created under hbase root dir, rather than restore dir.
> Some of the mob reference file paths are as follows:
> {quote}hdfs:/7ae0d109-3ca4-d0e7-7250-62ed234ab247/mobdir/data/ns_testMob/testMob
> hdfs:/7ae0d109-3ca4-d0e7-7250-62ed234ab247/mobdir/data/ns_testMob/testMob/057a856eb65753c6e6bdb168ba58a0b2
> hdfs:/7ae0d109-3ca4-d0e7-7250-62ed234ab247/mobdir/data/ns_testMob/testMob/057a856eb65753c6e6bdb168ba58a0b2/A
> hdfs:/7ae0d109-3ca4-d0e7-7250-62ed234ab247/mobdir/data/ns_testMob/testMob/057a856eb65753c6e6bdb168ba58a0b2/A/d41d8cd98f00b204e9800998ecf8427e201810120fc8e2446f174598a7280a81b1134cee
> hdfs:/7ae0d109-3ca4-d0e7-7250-62ed234ab247/mobdir/data/ns_testMob/testMob/057a856eb65753c6e6bdb168ba58a0b2/A/ns_testMob=testMob=057a856eb65753c6e6bdb168ba58a0b2-d41d8cd98f00b204e9800998ecf8427e201810120fc8e2446f174598a7280a81b1134cee
> {quote}
> The restore dir files are as follows:
> {quote}hdfs://hbase/.tmpdir-to-restore-snapshot/856e06fa-e018-4e95-9647-2cfbd5161e7e
> hdfs://hbase/.tmpdir-to-restore-snapshot/856e06fa-e018-4e95-9647-2cfbd5161e7e/data
> hdfs://hbase/.tmpdir-to-restore-snapshot/856e06fa-e018-4e95-9647-2cfbd5161e7e/data/ns_testMob
> hdfs://hbase/.tmpdir-to-restore-snapshot/856e06fa-e018-4e95-9647-2cfbd5161e7e/data/ns_testMob/testMob
> hdfs://hbase/.tmpdir-to-restore-snapshot/856e06fa-e018-4e95-9647-2cfbd5161e7e/data/ns_testMob/testMob/ecdf66f0d8c09a816faf37336ad262e1
> hdfs://hbase/.tmpdir-to-restore-snapshot/856e06fa-e018-4e95-9647-2cfbd5161e7e/data/ns_testMob/testMob/ecdf66f0d8c09a816faf37336ad262e1/.regioninfo
> hdfs://hbase/.tmpdir-to-restore-snapshot/856e06fa-e018-4e95-9647-2cfbd5161e7e/data/ns_testMob/testMob/ecdf66f0d8c09a816faf37336ad262e1/A
> hdfs://hbase/.tmpdir-to-restore-snapshot/856e06fa-e018-4e95-9647-2cfbd5161e7e/data/ns_testMob/testMob/ecdf66f0d8c09a816faf37336ad262e1/A/ns_testMob=testMob=ecdf66f0d8c09a816faf37336ad262e1-7208172df03b46518370643aa28ffd05
> {quote}
>  
>  
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21300) Fix the wrong reference file path when restoring snapshots for tables with MOB columns

2018-11-29 Thread Hudson (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21300?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16703400#comment-16703400
 ] 

Hudson commented on HBASE-21300:


Results for branch branch-2.0
[build #1123 on 
builds.a.o|https://builds.apache.org/job/HBase%20Nightly/job/branch-2.0/1123/]: 
(/) *{color:green}+1 overall{color}*

details (if available):

(/) {color:green}+1 general checks{color}
-- For more information [see general 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-2.0/1123//General_Nightly_Build_Report/]




(/) {color:green}+1 jdk8 hadoop2 checks{color}
-- For more information [see jdk8 (hadoop2) 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-2.0/1123//JDK8_Nightly_Build_Report_(Hadoop2)/]


(/) {color:green}+1 jdk8 hadoop3 checks{color}
-- For more information [see jdk8 (hadoop3) 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-2.0/1123//JDK8_Nightly_Build_Report_(Hadoop3)/]


(/) {color:green}+1 source release artifact{color}
-- See build output for details.


> Fix the wrong reference file path when restoring snapshots for tables with 
> MOB columns
> --
>
> Key: HBASE-21300
> URL: https://issues.apache.org/jira/browse/HBASE-21300
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 3.0.0, 2.2.0
>Reporter: Yi Mei
>Assignee: Yi Mei
>Priority: Major
> Fix For: 3.0.0, 2.2.0, 2.0.3, 2.1.2
>
> Attachments: HBASE-21300.branch-2.0.001.patch, 
> HBASE-21300.branch-2.1.001.patch, HBASE-21300.master.001.patch, 
> HBASE-21300.master.002.patch, HBASE-21300.master.003.patch, 
> HBASE-21300.master.004.patch, HBASE-21300.master.005.patch, 
> HBASE-21300.master.006.patch, HBASE-21300.master.007.patch, 
> HBASE-21300.master.008.patch
>
>
> When restoring snapshots for tables with MOB columns, the reference files for 
> mob region are created under hbase root dir, rather than restore dir.
> Some of the mob reference file paths are as follows:
> {quote}hdfs:/7ae0d109-3ca4-d0e7-7250-62ed234ab247/mobdir/data/ns_testMob/testMob
> hdfs:/7ae0d109-3ca4-d0e7-7250-62ed234ab247/mobdir/data/ns_testMob/testMob/057a856eb65753c6e6bdb168ba58a0b2
> hdfs:/7ae0d109-3ca4-d0e7-7250-62ed234ab247/mobdir/data/ns_testMob/testMob/057a856eb65753c6e6bdb168ba58a0b2/A
> hdfs:/7ae0d109-3ca4-d0e7-7250-62ed234ab247/mobdir/data/ns_testMob/testMob/057a856eb65753c6e6bdb168ba58a0b2/A/d41d8cd98f00b204e9800998ecf8427e201810120fc8e2446f174598a7280a81b1134cee
> hdfs:/7ae0d109-3ca4-d0e7-7250-62ed234ab247/mobdir/data/ns_testMob/testMob/057a856eb65753c6e6bdb168ba58a0b2/A/ns_testMob=testMob=057a856eb65753c6e6bdb168ba58a0b2-d41d8cd98f00b204e9800998ecf8427e201810120fc8e2446f174598a7280a81b1134cee
> {quote}
> The restore dir files are as follows:
> {quote}hdfs://hbase/.tmpdir-to-restore-snapshot/856e06fa-e018-4e95-9647-2cfbd5161e7e
> hdfs://hbase/.tmpdir-to-restore-snapshot/856e06fa-e018-4e95-9647-2cfbd5161e7e/data
> hdfs://hbase/.tmpdir-to-restore-snapshot/856e06fa-e018-4e95-9647-2cfbd5161e7e/data/ns_testMob
> hdfs://hbase/.tmpdir-to-restore-snapshot/856e06fa-e018-4e95-9647-2cfbd5161e7e/data/ns_testMob/testMob
> hdfs://hbase/.tmpdir-to-restore-snapshot/856e06fa-e018-4e95-9647-2cfbd5161e7e/data/ns_testMob/testMob/ecdf66f0d8c09a816faf37336ad262e1
> hdfs://hbase/.tmpdir-to-restore-snapshot/856e06fa-e018-4e95-9647-2cfbd5161e7e/data/ns_testMob/testMob/ecdf66f0d8c09a816faf37336ad262e1/.regioninfo
> hdfs://hbase/.tmpdir-to-restore-snapshot/856e06fa-e018-4e95-9647-2cfbd5161e7e/data/ns_testMob/testMob/ecdf66f0d8c09a816faf37336ad262e1/A
> hdfs://hbase/.tmpdir-to-restore-snapshot/856e06fa-e018-4e95-9647-2cfbd5161e7e/data/ns_testMob/testMob/ecdf66f0d8c09a816faf37336ad262e1/A/ns_testMob=testMob=ecdf66f0d8c09a816faf37336ad262e1-7208172df03b46518370643aa28ffd05
> {quote}
>  
>  
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21524) Unnecessary DEBUG log in ConnectionImplementation#isTableEnabled

2018-11-29 Thread Josh Elser (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21524?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16703401#comment-16703401
 ] 

Josh Elser commented on HBASE-21524:


Thanks [~stack]. What would you like me to do for branch-2.0 with your 2.0.3 
VOTE thread out?

> Unnecessary DEBUG log in ConnectionImplementation#isTableEnabled
> 
>
> Key: HBASE-21524
> URL: https://issues.apache.org/jira/browse/HBASE-21524
> Project: HBase
>  Issue Type: Improvement
>  Components: Client
>Reporter: Josh Elser
>Assignee: Josh Elser
>Priority: Major
> Fix For: 2.2.0, 2.0.3, 2.1.2
>
> Attachments: HBASE-21524.001.branch-2.0.patch
>
>
> Too much crap coming out of isTableEnabled for "normal" situations. Don't 
> need a DEBUG message to tell us that the table is available.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-21523) Chatty DEBUG logging in Master log

2018-11-29 Thread Josh Elser (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-21523?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Josh Elser updated HBASE-21523:
---
Attachment: HBASE-21523.002.patch

> Chatty DEBUG logging in Master log
> --
>
> Key: HBASE-21523
> URL: https://issues.apache.org/jira/browse/HBASE-21523
> Project: HBase
>  Issue Type: Improvement
>  Components: backuprestore
>Reporter: Josh Elser
>Assignee: Josh Elser
>Priority: Major
> Fix For: 3.0.0
>
> Attachments: HBASE-21523.001.patch, HBASE-21523.002.patch
>
>
> Noticed this in testing, a lot of spam in the Master log at DEBUG.
> {noformat}
> 2018-11-27 01:12:28,322 DEBUG [ForkJoinPool-1-worker-6] impl.BackupMetaTable: 
> Backup table backup:system exists and available
> 2018-11-27 01:12:28,326 DEBUG [ForkJoinPool-1-worker-6] impl.BackupMetaTable: 
> Backup table backup:system_bulk exists and available
> 2018-11-27 01:12:28,341 DEBUG [ForkJoinPool-1-worker-6] impl.BackupMetaTable: 
> Backup table backup:system exists and available
> 2018-11-27 01:12:28,345 DEBUG [ForkJoinPool-1-worker-6] impl.BackupMetaTable: 
> Backup table backup:system_bulk exists and available
> {noformat}
> FYI [~vrodionov]



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21523) Chatty DEBUG logging in Master log

2018-11-29 Thread Josh Elser (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21523?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16703351#comment-16703351
 ] 

Josh Elser commented on HBASE-21523:


bq. Please use parameterized logging and remove LOG.isDebugEnabled().

Sure. Didnt' do it the first time because it was still using the old logger (I 
thought?). Let me dbl check and fix the checkstyle too.

> Chatty DEBUG logging in Master log
> --
>
> Key: HBASE-21523
> URL: https://issues.apache.org/jira/browse/HBASE-21523
> Project: HBase
>  Issue Type: Improvement
>  Components: backuprestore
>Reporter: Josh Elser
>Assignee: Josh Elser
>Priority: Major
> Fix For: 3.0.0
>
> Attachments: HBASE-21523.001.patch
>
>
> Noticed this in testing, a lot of spam in the Master log at DEBUG.
> {noformat}
> 2018-11-27 01:12:28,322 DEBUG [ForkJoinPool-1-worker-6] impl.BackupMetaTable: 
> Backup table backup:system exists and available
> 2018-11-27 01:12:28,326 DEBUG [ForkJoinPool-1-worker-6] impl.BackupMetaTable: 
> Backup table backup:system_bulk exists and available
> 2018-11-27 01:12:28,341 DEBUG [ForkJoinPool-1-worker-6] impl.BackupMetaTable: 
> Backup table backup:system exists and available
> 2018-11-27 01:12:28,345 DEBUG [ForkJoinPool-1-worker-6] impl.BackupMetaTable: 
> Backup table backup:system_bulk exists and available
> {noformat}
> FYI [~vrodionov]



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21518) TestMasterFailoverWithProcedures is flaky

2018-11-29 Thread Peter Somogyi (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21518?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16703332#comment-16703332
 ] 

Peter Somogyi commented on HBASE-21518:
---

ServerManager#expireServer checks if cluster is shutting down and in this case 
does not create ServerCrashProcedure for dead servers. This AtomicBoolean 
variable is set to true when ServerManager#shutdownCluster method is called, 
however, there are 2 ServerManager instances and on expireServer a different 
one is checked.

I added some debug logs with hashcodes where you can see that clusterShutdown 
was set to true (hash=1980707837) but later on during shutdown the variable 
contains false (hash=416244779) that's why ServerCrashProcedure is created 
which hangs since there are no Master.
{noformat}
2018-11-29 15:43:21,948 INFO [Thread-81] master.ServerManager(160): 
ServerManager initialized. clusterShutdown false, thread 210, hash 1980707837
2018-11-29 15:43:23,929 INFO [Thread-81] master.ServerManager(913): Called 
isClusterShutdown. clusterShutdown=false, thread 210, hash=1980707837
2018-11-29 15:43:23,929 INFO [Thread-81] master.ServerManager(913): Called 
isClusterShutdown. clusterShutdown=false, thread 210, hash=1980707837
2018-11-29 15:43:29,732 INFO [Thread-80] master.ServerManager(160): 
ServerManager initialized. clusterShutdown false, thread 209, hash 416244779
2018-11-29 15:43:29,820 INFO [Thread-80] master.ServerManager(913): Called 
isClusterShutdown. clusterShutdown=false, thread 209, hash=416244779
2018-11-29 15:43:29,820 INFO [Thread-80] master.ServerManager(913): Called 
isClusterShutdown. clusterShutdown=false, thread 209, hash=416244779
2018-11-29 15:43:29,937 INFO [Time-limited test] master.ServerManager(904): Set 
clusterShutdown to true, thread 14, hash 1980707837
2018-11-29 15:43:30,985 INFO [RegionServerTracker-0] master.ServerManager(913): 
Called isClusterShutdown. clusterShutdown=false, thread 461, hash=416244779
2018-11-29 15:43:30,986 INFO [RegionServerTracker-0] master.ServerManager(913): 
Called isClusterShutdown. clusterShutdown=false, thread 461, hash=416244779
2018-11-29 15:48:29,851 INFO [master/172.30.65.195:0.Chore.1] 
master.ServerManager(913): Called isClusterShutdown. clusterShutdown=false, 
thread 417, hash=416244779
2018-11-29 15:48:29,852 INFO [master/172.30.65.195:0.Chore.1] 
master.ServerManager(913): Called isClusterShutdown. clusterShutdown=false, 
thread 417, hash=416244779
2018-11-29 15:48:29,852 INFO [master/172.30.65.195:0.Chore.1] 
master.ServerManager(913): Called isClusterShutdown. clusterShutdown=false, 
thread 417, hash=416244779
2018-11-29 15:53:32,277 INFO [master/172.30.65.195:0.Chore.1] 
master.ServerManager(913): Called isClusterShutdown. clusterShutdown=false, 
thread 417, hash=416244779
2018-11-29 15:53:32,277 INFO [master/172.30.65.195:0.Chore.1] 
master.ServerManager(913): Called isClusterShutdown. clusterShutdown=false, 
thread 417, hash=416244779
2018-11-29 15:53:32,277 INFO [master/172.30.65.195:0.Chore.1] 
master.ServerManager(913): Called isClusterShutdown. clusterShutdown=false, 
thread 417, hash=416244779{noformat}

> TestMasterFailoverWithProcedures is flaky
> -
>
> Key: HBASE-21518
> URL: https://issues.apache.org/jira/browse/HBASE-21518
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 2.2.0, 2.0.3, 2.1.2
>Reporter: Peter Somogyi
>Assignee: Peter Somogyi
>Priority: Major
> Attachments: output.txt
>
>
> TestMasterFailoverWithProcedures test is failing frequently, times out. I 
> faced this failure on 2.0.3RC0 vote and it also appears on multiple flaky 
> dashboards.
> branch-2: 
> [https://builds.apache.org/view/H-L/view/HBase/job/HBase-Flaky-Tests/job/branch-2/2007/]
> branch-2.1: 
> [https://builds.apache.org/view/H-L/view/HBase/job/HBase-Flaky-Tests/job/branch-2.1/2002/]
> branch-2.0: 
> [https://builds.apache.org/view/H-L/view/HBase/job/HBase-Flaky-Tests/job/branch-2.0/1988/]
>   
> {noformat}
> [INFO] Running 
> org.apache.hadoop.hbase.master.procedure.TestMasterFailoverWithProcedures
> [ERROR] Tests run: 4, Failures: 0, Errors: 2, Skipped: 0, Time elapsed: 
> 780.648 s <<< FAILURE! - in 
> org.apache.hadoop.hbase.master.procedure.TestMasterFailoverWithProcedures
> [ERROR] 
> org.apache.hadoop.hbase.master.procedure.TestMasterFailoverWithProcedures  
> Time elapsed: 749.024 s  <<< ERROR!
> org.junit.runners.model.TestTimedOutException: test timed out after 780 
> seconds
>   at 
> org.apache.hadoop.hbase.master.procedure.TestMasterFailoverWithProcedures.tearDown(TestMasterFailoverWithProcedures.java:86)
> [ERROR] 
> org.apache.hadoop.hbase.master.procedure.TestMasterFailoverWithProcedures  
> Time elapsed: 749.051 s  <<< ERROR!
> java.lang.Exception: Appears to be stuck in thread RS-EventLoopGroup-3-2
> {noformat}



--
This message was 

[jira] [Commented] (HBASE-21401) Sanity check when constructing the KeyValue

2018-11-29 Thread Hadoop QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21401?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16703289#comment-16703289
 ] 

Hadoop QA commented on HBASE-21401:
---

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
24s{color} | {color:blue} Docker mode activated. {color} |
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} hbaseanti {color} | {color:green}  0m  
0s{color} | {color:green} Patch does not have any anti-patterns. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green}  0m 
 0s{color} | {color:green} The patch appears to include 1 new or modified test 
files. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
28s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  4m 
32s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  2m 
22s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  1m 
31s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} shadedjars {color} | {color:green}  3m 
51s{color} | {color:green} branch has no errors when building our shaded 
downstream artifacts. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  2m 
36s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
43s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
14s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  3m 
57s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  2m 
16s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  2m 
16s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
24s{color} | {color:red} hbase-common: The patch generated 21 new + 148 
unchanged - 1 fixed = 169 total (was 149) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} shadedjars {color} | {color:green}  3m 
49s{color} | {color:green} patch has no errors when building our shaded 
downstream artifacts. {color} |
| {color:green}+1{color} | {color:green} hadoopcheck {color} | {color:green}  
8m 41s{color} | {color:green} Patch does not cause any errors with Hadoop 2.7.4 
or 3.0.0. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  2m 
51s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
47s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:red}-1{color} | {color:red} unit {color} | {color:red}  2m 34s{color} 
| {color:red} hbase-common in the patch failed. {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red}248m 19s{color} 
| {color:red} hbase-server in the patch failed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
41s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black}292m 52s{color} | 
{color:black} {color} |
\\
\\
|| Reason || Tests ||
| Failed junit tests | hadoop.hbase.TestKeyValue |
|   | hadoop.hbase.master.TestAssignmentManagerMetrics |
|   | hadoop.hbase.io.hfile.TestCacheOnWrite |
|   | hadoop.hbase.io.encoding.TestDataBlockEncoders |
|   | hadoop.hbase.io.hfile.TestHFileSeek |
|   | hadoop.hbase.client.TestAdmin1 |
\\
\\
|| Subsystem || Report/Notes ||
| Docker | Client=17.05.0-ce Server=17.05.0-ce Image:yetus/hbase:b002b0b |
| JIRA Issue | HBASE-21401 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12949984/HBASE-21401.v6.patch |
| Optional Tests |  dupname  asflicense  javac  javadoc  unit  findbugs  
shadedjars  hadoopcheck  hbaseanti  checkstyle  compile  |
| uname | Linux ab114c9e23ba 4.4.0-138-generic #164-Ubuntu SMP Tue Oct 2 

[jira] [Commented] (HBASE-21515) Also initialize an AsyncClusterConnection in HRegionServer

2018-11-29 Thread Hadoop QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21515?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16703269#comment-16703269
 ] 

Hadoop QA commented on HBASE-21515:
---

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
12s{color} | {color:blue} Docker mode activated. {color} |
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} hbaseanti {color} | {color:green}  0m  
0s{color} | {color:green} Patch does not have any anti-patterns. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green}  0m 
 0s{color} | {color:green} The patch appears to include 20 new or modified test 
files. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
27s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  4m 
 7s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  2m 
53s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  2m 
 4s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} shadedjars {color} | {color:green}  3m 
51s{color} | {color:green} branch has no errors when building our shaded 
downstream artifacts. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  3m 
23s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m  
9s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
16s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:red}-1{color} | {color:red} mvninstall {color} | {color:red}  2m 
25s{color} | {color:red} root in the patch failed. {color} |
| {color:red}-1{color} | {color:red} compile {color} | {color:red}  0m 
32s{color} | {color:red} hbase-server in the patch failed. {color} |
| {color:red}-1{color} | {color:red} javac {color} | {color:red}  0m 32s{color} 
| {color:red} hbase-server in the patch failed. {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
22s{color} | {color:green} The patch passed checkstyle in hbase-common {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
31s{color} | {color:red} hbase-client: The patch generated 3 new + 3 unchanged 
- 3 fixed = 6 total (was 6) {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  1m 
10s{color} | {color:green} hbase-server: The patch generated 0 new + 285 
unchanged - 2 fixed = 285 total (was 287) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:red}-1{color} | {color:red} shadedjars {color} | {color:red}  2m 
50s{color} | {color:red} patch has 16 errors when building our shaded 
downstream artifacts. {color} |
| {color:red}-1{color} | {color:red} hadoopcheck {color} | {color:red}  1m 
34s{color} | {color:red} The patch causes 16 errors with Hadoop v2.7.4. {color} 
|
| {color:red}-1{color} | {color:red} hadoopcheck {color} | {color:red}  3m 
10s{color} | {color:red} The patch causes 16 errors with Hadoop v3.0.0. {color} 
|
| {color:red}-1{color} | {color:red} findbugs {color} | {color:red}  0m 
28s{color} | {color:red} hbase-server in the patch failed. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m  
7s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} unit {color} | {color:green}  2m 
45s{color} | {color:green} hbase-common in the patch passed. {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green}  3m 
14s{color} | {color:green} hbase-client in the patch passed. {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red}  0m 29s{color} 
| {color:red} hbase-server in the patch failed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
28s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 41m 44s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Docker | Client=17.05.0-ce Server=17.05.0-ce 

  1   2   >