[jira] [Resolved] (HBASE-16641) QA tests for hbase-client skip the second part.

2016-10-12 Thread Heng Chen (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-16641?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Heng Chen resolved HBASE-16641.
---
Resolution: Duplicate

> QA tests for hbase-client skip the second part.
> ---
>
> Key: HBASE-16641
> URL: https://issues.apache.org/jira/browse/HBASE-16641
> Project: HBase
>  Issue Type: Bug
>Reporter: Heng Chen
>
> See 
> https://builds.apache.org/job/PreCommit-HBASE-Build/3547/artifact/patchprocess/patch-unit-hbase-client.txt
> {code}
> [INFO] --- maven-surefire-plugin:2.18.1:test (secondPartTestsExecution) @ 
> hbase-client ---
> [INFO] Tests are skipped.
> {code}
> The first part passed fine,  but second parts is skipped. 
> Notice hbase-client/pom.xml 
> {code}
>  
>   
> secondPartTestsExecution
> test
> 
>   test
> 
> 
>   true
> 
>   
> 
> {code}
> If i change the 'skip' to be false,  the second part could be triggered.  But 
> this configuration existed for a long time,  is the cmd line on build box 
> updated recently? 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-16702) TestBlockEvictionFromClient is broken

2016-09-23 Thread Heng Chen (JIRA)
Heng Chen created HBASE-16702:
-

 Summary: TestBlockEvictionFromClient is broken
 Key: HBASE-16702
 URL: https://issues.apache.org/jira/browse/HBASE-16702
 Project: HBase
  Issue Type: Bug
Reporter: Heng Chen






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-16665) Check whether KeyValueUtil.createXXX could be replaced by CellUtil without copy

2016-09-21 Thread Heng Chen (JIRA)
Heng Chen created HBASE-16665:
-

 Summary: Check whether KeyValueUtil.createXXX could be replaced by 
CellUtil without copy
 Key: HBASE-16665
 URL: https://issues.apache.org/jira/browse/HBASE-16665
 Project: HBase
  Issue Type: Bug
Reporter: Heng Chen






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-16652) Figure out performance difference between increment and append

2016-09-18 Thread Heng Chen (JIRA)
Heng Chen created HBASE-16652:
-

 Summary: Figure out performance difference between increment and 
append
 Key: HBASE-16652
 URL: https://issues.apache.org/jira/browse/HBASE-16652
 Project: HBase
  Issue Type: Sub-task
Reporter: Heng Chen


When do performance test in HBASE-16625,  i found it has the very big 
difference between Append and Increment (append is about 37% faster than 
increment).

As [~stack] mentioned in 
https://issues.apache.org/jira/browse/HBASE-16610?focusedCommentId=15493166=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-15493166,
   append and increment has been unified in server-side,  and they looks the 
same in client-side. 

This issue is to figure out why the performance looks different between them.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-16641) QA tests for hbase-client skip the second part.

2016-09-15 Thread Heng Chen (JIRA)
Heng Chen created HBASE-16641:
-

 Summary: QA tests for hbase-client skip the second part.
 Key: HBASE-16641
 URL: https://issues.apache.org/jira/browse/HBASE-16641
 Project: HBase
  Issue Type: Bug
Reporter: Heng Chen






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-16631) Extract AsyncRequestFuture relates code from AsyncProcess

2016-09-14 Thread Heng Chen (JIRA)
Heng Chen created HBASE-16631:
-

 Summary: Extract AsyncRequestFuture relates code from AsyncProcess
 Key: HBASE-16631
 URL: https://issues.apache.org/jira/browse/HBASE-16631
 Project: HBase
  Issue Type: Bug
Reporter: Heng Chen
Assignee: Heng Chen


Now, AsyncProcess class is too large (over 2000+ lines),  and there are so many 
sub classes in it.  

AsyncRequestFutureImpl is the most biggest subclass in AP,  we could extract it 
out from AP to reduce the AP size.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-16625) Performance test for interface unified with AP

2016-09-13 Thread Heng Chen (JIRA)
Heng Chen created HBASE-16625:
-

 Summary: Performance test for interface unified with AP
 Key: HBASE-16625
 URL: https://issues.apache.org/jira/browse/HBASE-16625
 Project: HBase
  Issue Type: Sub-task
Reporter: Heng Chen






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-16623) Unify Get with AP

2016-09-12 Thread Heng Chen (JIRA)
Heng Chen created HBASE-16623:
-

 Summary: Unify Get with AP
 Key: HBASE-16623
 URL: https://issues.apache.org/jira/browse/HBASE-16623
 Project: HBase
  Issue Type: Sub-task
Reporter: Heng Chen
Assignee: Heng Chen






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-16611) Flakey org.apache.hadoop.hbase.client.TestReplicasClient.testCancelOfMultiGet

2016-09-11 Thread Heng Chen (JIRA)
Heng Chen created HBASE-16611:
-

 Summary: Flakey 
org.apache.hadoop.hbase.client.TestReplicasClient.testCancelOfMultiGet
 Key: HBASE-16611
 URL: https://issues.apache.org/jira/browse/HBASE-16611
 Project: HBase
  Issue Type: Bug
Reporter: Heng Chen


see 
https://builds.apache.org/job/PreCommit-HBASE-Build/3494/artifact/patchprocess/patch-unit-hbase-server.txt

{code}
testCancelOfMultiGet(org.apache.hadoop.hbase.client.TestReplicasClient)  Time 
elapsed: 4.026 sec  <<< FAILURE!
java.lang.AssertionError: null
at org.junit.Assert.fail(Assert.java:86)
at org.junit.Assert.assertTrue(Assert.java:41)
at org.junit.Assert.assertTrue(Assert.java:52)
at 
org.apache.hadoop.hbase.client.TestReplicasClient.testCancelOfMultiGet(TestReplicasClient.java:579)

Tests run: 26, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 94.401 sec - 
in org.apache.hadoop.hbase.client.TestAdmin2
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 11.861 sec - in 
org.apache.hadoop.hbase.client.TestClientScannerRPCTimeout
Running 
org.apache.hadoop.hbase.client.TestRestoreSnapshotFromClientWithRegionReplicas
Tests run: 1, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 261.925 sec <<< 
FAILURE! - in org.apache.hadoop.hbase.client.TestReplicasClient
testCancelOfMultiGet(org.apache.hadoop.hbase.client.TestReplicasClient)  Time 
elapsed: 4.522 sec  <<< FAILURE!
java.lang.AssertionError: null
at org.junit.Assert.fail(Assert.java:86)
at org.junit.Assert.assertTrue(Assert.java:41)
at org.junit.Assert.assertTrue(Assert.java:52)
at 
org.apache.hadoop.hbase.client.TestReplicasClient.testCancelOfMultiGet(TestReplicasClient.java:581)

Running org.apache.hadoop.hbase.client.TestFastFail
Tests run: 2, Failures: 0, Errors: 0, Skipped: 1, Time elapsed: 3.648 sec - in 
org.apache.hadoop.hbase.client.TestFastFail
Tests run: 1, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 277.894 sec <<< 
FAILURE! - in org.apache.hadoop.hbase.client.TestReplicasClient
testCancelOfMultiGet(org.apache.hadoop.hbase.client.TestReplicasClient)  Time 
elapsed: 5.359 sec  <<< FAILURE!
java.lang.AssertionError: null
at org.junit.Assert.fail(Assert.java:86)
at org.junit.Assert.assertTrue(Assert.java:41)
at org.junit.Assert.assertTrue(Assert.java:52)
at 
org.apache.hadoop.hbase.client.TestReplicasClient.testCancelOfMultiGet(TestReplicasClient.java:579)
{code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-16610) Unify append, increment with AP

2016-09-11 Thread Heng Chen (JIRA)
Heng Chen created HBASE-16610:
-

 Summary: Unify append, increment with AP
 Key: HBASE-16610
 URL: https://issues.apache.org/jira/browse/HBASE-16610
 Project: HBase
  Issue Type: Sub-task
Reporter: Heng Chen
Assignee: Heng Chen






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-16607) Make NoncedRegionServerCallable extends CancellableRegionServerCallable

2016-09-10 Thread Heng Chen (JIRA)
Heng Chen created HBASE-16607:
-

 Summary: Make NoncedRegionServerCallable extends 
CancellableRegionServerCallable
 Key: HBASE-16607
 URL: https://issues.apache.org/jira/browse/HBASE-16607
 Project: HBase
  Issue Type: Sub-task
Reporter: Heng Chen


This is the first step to unify append, increment with AP.

And after extends CancellableRegionServerCallable,  we could remove lots of 
duplicate code in NoncedRegionServerCallable



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-16606) Remove some duplicate code in HTable

2016-09-10 Thread Heng Chen (JIRA)
Heng Chen created HBASE-16606:
-

 Summary: Remove some duplicate code in HTable
 Key: HBASE-16606
 URL: https://issues.apache.org/jira/browse/HBASE-16606
 Project: HBase
  Issue Type: Sub-task
Reporter: Heng Chen
Assignee: Heng Chen






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HBASE-16597) Revisit the threadPool is really needed in submitAll and submit interface in AsyncProcess

2016-09-10 Thread Heng Chen (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-16597?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Heng Chen resolved HBASE-16597.
---
Resolution: Invalid

> Revisit the threadPool is really needed in submitAll and submit interface in 
> AsyncProcess
> -
>
> Key: HBASE-16597
> URL: https://issues.apache.org/jira/browse/HBASE-16597
> Project: HBase
>  Issue Type: Sub-task
>Reporter: Heng Chen
>Assignee: Heng Chen
>
> Currently,  threadPool could be passed into AP by constructor and submitAll, 
> submit interfaces,  but i notice in HTable the pool passed to AP though 
> submitAll looks same with the one in AP as default,  Let me revisit it to 
> ensure whether the pool is really needed in submitAll and submit interface.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-16597) Revisit the threadPool is really needed in submitAll and submit interface in AsyncProcess

2016-09-09 Thread Heng Chen (JIRA)
Heng Chen created HBASE-16597:
-

 Summary: Revisit the threadPool is really needed in submitAll and 
submit interface in AsyncProcess
 Key: HBASE-16597
 URL: https://issues.apache.org/jira/browse/HBASE-16597
 Project: HBase
  Issue Type: Sub-task
Reporter: Heng Chen


Currently,  threadPool could be passed into AP by constructor and submitAll, 
submit interfaces,  but i notice in HTable the pool passed to AP though 
submitAll looks same with the one in AP as default,  Let me revisit it to 
ensure whether the pool is really needed in submitAll and submit interface.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-16596) Reduce redundant interfaces in AP

2016-09-09 Thread Heng Chen (JIRA)
Heng Chen created HBASE-16596:
-

 Summary: Reduce redundant interfaces in AP
 Key: HBASE-16596
 URL: https://issues.apache.org/jira/browse/HBASE-16596
 Project: HBase
  Issue Type: Sub-task
Reporter: Heng Chen
Assignee: Heng Chen


Currently, there are lots of interfaces in AP,  some of them only be used only 
in test case, or be used only once,  let's remove the redundant ones to keep 
the AP more clear



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-16593) Unify HTable with AP

2016-09-09 Thread Heng Chen (JIRA)
Heng Chen created HBASE-16593:
-

 Summary: Unify HTable with AP
 Key: HBASE-16593
 URL: https://issues.apache.org/jira/browse/HBASE-16593
 Project: HBase
  Issue Type: Umbrella
Reporter: Heng Chen
Assignee: Heng Chen


Currently, HTable has two ways to deal with request,  one is call RPC directly, 
it is used to processed single action request such as Get, Delete, Append, 
Increment.  Another one is through AP to deal with multi action requests, such 
as batch, mutation etc.

This issue is to unify them with AP only. It has some benefits, for example we 
could implements async interface easily with AP,  and we could make the client 
logic more clear just use AP to communicate with Server.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-16592) Unify Delete request with AP

2016-09-09 Thread Heng Chen (JIRA)
Heng Chen created HBASE-16592:
-

 Summary: Unify Delete request with AP
 Key: HBASE-16592
 URL: https://issues.apache.org/jira/browse/HBASE-16592
 Project: HBase
  Issue Type: Task
Reporter: Heng Chen
Assignee: Heng Chen


This is the first step try to unify the HTable with AP only,  to extend AP 
could process single action, i introduced AbstractResponse,  multiResponse and 
singleResponse (introduced to deal with single result) will extend this class. 





--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HBASE-16562) ITBLL should fail to start if misconfigured

2016-09-07 Thread Heng Chen (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-16562?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Heng Chen resolved HBASE-16562.
---
  Resolution: Fixed
Hadoop Flags: Reviewed

> ITBLL should fail to start if misconfigured
> ---
>
> Key: HBASE-16562
> URL: https://issues.apache.org/jira/browse/HBASE-16562
> Project: HBase
>  Issue Type: Improvement
>  Components: integration tests
>Reporter: Andrew Purtell
>Assignee: Heng Chen
> Fix For: 2.0.0, 1.0.4, 1.4.0, 1.3.1, 1.1.7, 0.98.23, 1.2.4
>
> Attachments: HBASE-16562-branch-1.2.patch, 
> HBASE-16562-branch-1.2.v1.patch, HBASE-16562.patch, HBASE-16562.v1.patch, 
> HBASE-16562.v1.patch-addendum
>
>
> The number of nodes in ITBLL must a multiple of width*wrap (defaults to 25M, 
> but can be configured by adding two more args to the test invocation) or else 
> verification will fail. This can be very expensive in terms of time or hourly 
> billing for on demand test resources. Check the sanity of test parameters 
> before launching any MR jobs and fail fast if invariants aren't met with an 
> indication what parameter(s) need fixing. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HBASE-16564) ITBLL run failed with hadoop 2.7.2 on branch 0.98

2016-09-06 Thread Heng Chen (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-16564?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Heng Chen resolved HBASE-16564.
---
Resolution: Invalid

As [~Apache9] said, the best solution is upgrade hadoop client,  so close this 
issue as invalid.

> ITBLL run failed with hadoop 2.7.2 on branch 0.98
> -
>
> Key: HBASE-16564
> URL: https://issues.apache.org/jira/browse/HBASE-16564
> Project: HBase
>  Issue Type: Bug
>Reporter: Heng Chen
>Priority: Minor
>
> 0.98 compiled with hadoop 2.2.0,   so it has some compatibility issues with 
> hadoop 2.7.2 (it seems 2.5.0+ has the same issue),  some counter has been 
> removed.  
> IMO we should catch the exception so our ITBLL could go on.
> {code}
> 16/09/06 15:39:33 INFO hbase.HBaseCluster: Added new HBaseAdmin
> 16/09/06 15:39:33 INFO hbase.HBaseCluster: Restoring cluster - done
> 16/09/06 15:39:33 INFO hbase.HBaseCommonTestingUtility: Stopping mini 
> mapreduce cluster...
> 16/09/06 15:39:33 INFO Configuration.deprecation: mapred.job.tracker is 
> deprecated. Instead, use mapreduce.jobtracker.address
> 16/09/06 15:39:33 INFO hbase.HBaseCommonTestingUtility: Mini mapreduce 
> cluster stopped
> 16/09/06 15:39:33 ERROR util.AbstractHBaseTool: Error running command-line 
> tool
> java.lang.IllegalArgumentException: No enum constant 
> org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_MAPS
>   at java.lang.Enum.valueOf(Enum.java:238)
>   at 
> org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup.valueOf(FrameworkCounterGroup.java:148)
>   at 
> org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup.findCounter(FrameworkCounterGroup.java:182)
>   at 
> org.apache.hadoop.mapreduce.counters.AbstractCounters.findCounter(AbstractCounters.java:154)
>   at 
> org.apache.hadoop.mapreduce.TypeConverter.fromYarn(TypeConverter.java:240)
>   at 
> org.apache.hadoop.mapred.ClientServiceDelegate.getJobCounters(ClientServiceDelegate.java:370)
>   at 
> org.apache.hadoop.mapred.YARNRunner.getJobCounters(YARNRunner.java:511)
>   at org.apache.hadoop.mapreduce.Job$7.run(Job.java:756)
>   at org.apache.hadoop.mapreduce.Job$7.run(Job.java:753)
>   at java.security.AccessController.doPrivileged(Native Method)
>   at javax.security.auth.Subject.doAs(Subject.java:422)
>   at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1491)
>   at org.apache.hadoop.mapreduce.Job.getCounters(Job.java:753)
>   at 
> org.apache.hadoop.mapreduce.Job.monitorAndPrintJob(Job.java:1361)
>   at 
> org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1289)
>   at 
> org.apache.hadoop.hbase.test.IntegrationTestBigLinkedList$Generator.jobCompletion(IntegrationTestBigLinkedList.java:543)
>   at 
> org.apache.hadoop.hbase.test.IntegrationTestBigLinkedList$Generator.runRandomInputGenerator(IntegrationTestBigLinkedList.java:505)
>   at 
> org.apache.hadoop.hbase.test.IntegrationTestBigLinkedList$Generator.run(IntegrationTestBigLinkedList.java:553)
>   at 
> org.apache.hadoop.hbase.test.IntegrationTestBigLinkedList$Loop.runGenerator(IntegrationTestBigLinkedList.java:842)
>   at 
> org.apache.hadoop.hbase.test.IntegrationTestBigLinkedList$Loop.run(IntegrationTestBigLinkedList.java:892)
>   at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
>   at 
> org.apache.hadoop.hbase.test.IntegrationTestBigLinkedList.runTestFromCommandLine(IntegrationTestBigLinkedList.java:1237)
>   at 
> org.apache.hadoop.hbase.IntegrationTestBase.doWork(IntegrationTestBase.java:115)
>   at 
> org.apache.hadoop.hbase.util.AbstractHBaseTool.run(AbstractHBaseTool.java:112)
>   at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
>   at 
> org.apache.hadoop.hbase.test.IntegrationTestBigLinkedList.main(IntegrationTestBigLinkedList.java:1272)
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-16564) ITBLL run failed with hdfs 2.7.2 on branch 0.98

2016-09-06 Thread Heng Chen (JIRA)
Heng Chen created HBASE-16564:
-

 Summary: ITBLL run failed with hdfs 2.7.2 on branch 0.98
 Key: HBASE-16564
 URL: https://issues.apache.org/jira/browse/HBASE-16564
 Project: HBase
  Issue Type: Bug
Reporter: Heng Chen
Priority: Minor


0.98 compiled with hdfs 2.2.0,   so it has some compatibility issues with hdfs 
2.7.2 (it seems 2.5.0+ has the same issue),  some counter has been removed.  

IMO we should catch the exception so our ITBLL could go on.

{code}
16/09/06 15:39:33 INFO hbase.HBaseCluster: Added new HBaseAdmin
16/09/06 15:39:33 INFO hbase.HBaseCluster: Restoring cluster - done
16/09/06 15:39:33 INFO hbase.HBaseCommonTestingUtility: Stopping mini mapreduce 
cluster...
16/09/06 15:39:33 INFO Configuration.deprecation: mapred.job.tracker is 
deprecated. Instead, use mapreduce.jobtracker.address
16/09/06 15:39:33 INFO hbase.HBaseCommonTestingUtility: Mini mapreduce cluster 
stopped
16/09/06 15:39:33 ERROR util.AbstractHBaseTool: Error running command-line tool
java.lang.IllegalArgumentException: No enum constant 
org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_MAPS
at java.lang.Enum.valueOf(Enum.java:238)
at 
org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup.valueOf(FrameworkCounterGroup.java:148)
at 
org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup.findCounter(FrameworkCounterGroup.java:182)
at 
org.apache.hadoop.mapreduce.counters.AbstractCounters.findCounter(AbstractCounters.java:154)
at 
org.apache.hadoop.mapreduce.TypeConverter.fromYarn(TypeConverter.java:240)
at 
org.apache.hadoop.mapred.ClientServiceDelegate.getJobCounters(ClientServiceDelegate.java:370)
at 
org.apache.hadoop.mapred.YARNRunner.getJobCounters(YARNRunner.java:511)
at org.apache.hadoop.mapreduce.Job$7.run(Job.java:756)
at org.apache.hadoop.mapreduce.Job$7.run(Job.java:753)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:422)
at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1491)
at org.apache.hadoop.mapreduce.Job.getCounters(Job.java:753)
at org.apache.hadoop.mapreduce.Job.monitorAndPrintJob(Job.java:1361)
at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1289)
at 
org.apache.hadoop.hbase.test.IntegrationTestBigLinkedList$Generator.jobCompletion(IntegrationTestBigLinkedList.java:543)
at 
org.apache.hadoop.hbase.test.IntegrationTestBigLinkedList$Generator.runRandomInputGenerator(IntegrationTestBigLinkedList.java:505)
at 
org.apache.hadoop.hbase.test.IntegrationTestBigLinkedList$Generator.run(IntegrationTestBigLinkedList.java:553)
at 
org.apache.hadoop.hbase.test.IntegrationTestBigLinkedList$Loop.runGenerator(IntegrationTestBigLinkedList.java:842)
at 
org.apache.hadoop.hbase.test.IntegrationTestBigLinkedList$Loop.run(IntegrationTestBigLinkedList.java:892)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
at 
org.apache.hadoop.hbase.test.IntegrationTestBigLinkedList.runTestFromCommandLine(IntegrationTestBigLinkedList.java:1237)
at 
org.apache.hadoop.hbase.IntegrationTestBase.doWork(IntegrationTestBase.java:115)
at 
org.apache.hadoop.hbase.util.AbstractHBaseTool.run(AbstractHBaseTool.java:112)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
at 
org.apache.hadoop.hbase.test.IntegrationTestBigLinkedList.main(IntegrationTestBigLinkedList.java:1272)
{code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-16512) when locate meta region, we should respect the param "useCache" passed in on 0.98

2016-08-29 Thread Heng Chen (JIRA)
Heng Chen created HBASE-16512:
-

 Summary: when locate meta region, we should respect the param 
"useCache" passed in on 0.98
 Key: HBASE-16512
 URL: https://issues.apache.org/jira/browse/HBASE-16512
 Project: HBase
  Issue Type: Bug
Affects Versions: 0.98.21
Reporter: Heng Chen


we found that when RS with meta crash,  client will retry the same request,  
but it still use the original meta location in cache, so all request retried 
will failed. 

Notice the code in HConnectionMananger#locateRegionInMeta,  the "useCache" 
passed in is not used when try to found the meta region. 

{code}
private HRegionLocation locateRegionInMeta(final TableName parentTable,
  final TableName tableName, final byte [] row, boolean useCache,
  Object regionLockObject, boolean retry)
throws IOException {
  ..
  for (int tries = 0; true; tries++) {
   .
HRegionLocation metaLocation = null;
try {
  // locate the meta region
  metaLocation = locateRegion(parentTable, metaKey, true, false); 
//NOTICE: we should honor the "useCache" passed in when locate the meta region.
  
  }
}
{code}







--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-16490) Fix race condition between SnapshotManager and SnapshotCleaner

2016-08-24 Thread Heng Chen (JIRA)
Heng Chen created HBASE-16490:
-

 Summary: Fix race condition between SnapshotManager and 
SnapshotCleaner
 Key: HBASE-16490
 URL: https://issues.apache.org/jira/browse/HBASE-16490
 Project: HBase
  Issue Type: Bug
Reporter: Heng Chen
 Fix For: 2.0.0


As [~mbertozzi] comments on HBASE-16464,  there maybe race condition between 
SnapshotManager and SnapshotCleaner.  We should use one lock when create 
snapshot,  and cleanup should acquire the lock before take action.

One method is pass HMaster as param into Cleaner through 
{{FileCleanerDelegate.getDeletableFiles}},  suggestions are welcomed.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-16464) archive folder grows bigger and bigger due to corrupt snapshot under tmp dir

2016-08-22 Thread Heng Chen (JIRA)
Heng Chen created HBASE-16464:
-

 Summary: archive folder grows bigger and bigger due to corrupt 
snapshot under tmp dir
 Key: HBASE-16464
 URL: https://issues.apache.org/jira/browse/HBASE-16464
 Project: HBase
  Issue Type: Bug
Reporter: Heng Chen


We met the problem on our real production cluster,  we need to cleanup some 
data on hbase,  we notice the archive folder is much larger than others,  so we 
delete all snapshots of all tables,  but the archive folder still grows bigger 
and bigger. 

After check the hmaster log, we notice the exception below:
{code}
2016-08-22 15:34:33,089 ERROR [f04,16000,1471240833208_ChoreService_1] 
snapshot.SnapshotHFileCleaner: Exception while checking if files were valid, 
keeping them just in case.
org.apache.hadoop.hbase.snapshot.CorruptedSnapshotException: Couldn't read 
snapshot info 
from:hdfs://f04/hbase/.hbase-snapshot/.tmp/frog_stastic_2016-08-17/.snapshotinfo
at 
org.apache.hadoop.hbase.snapshot.SnapshotDescriptionUtils.readSnapshotInfo(SnapshotDescriptionUtils.java:295)
at 
org.apache.hadoop.hbase.snapshot.SnapshotReferenceUtil.getHFileNames(SnapshotReferenceUtil.java:328)
at 
org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner$1.filesUnderSnapshot(SnapshotHFileCleaner.java:85)
at 
org.apache.hadoop.hbase.master.snapshot.SnapshotFileCache.getSnapshotsInProgress(SnapshotFileCache.java:303)
at 
org.apache.hadoop.hbase.master.snapshot.SnapshotFileCache.getUnreferencedFiles(SnapshotFileCache.java:194)
at 
org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner.getDeletableFiles(SnapshotHFileCleaner.java:62)
at 
org.apache.hadoop.hbase.master.cleaner.CleanerChore.checkAndDeleteFiles(CleanerChore.java:233)
at 
org.apache.hadoop.hbase.master.cleaner.CleanerChore.checkAndDeleteEntries(CleanerChore.java:157)
at 
org.apache.hadoop.hbase.master.cleaner.CleanerChore.checkAndDeleteDirectory(CleanerChore.java:180)
at 
org.apache.hadoop.hbase.master.cleaner.CleanerChore.checkAndDeleteEntries(CleanerChore.java:149)
at 
org.apache.hadoop.hbase.master.cleaner.CleanerChore.checkAndDeleteDirectory(CleanerChore.java:180)
at 
org.apache.hadoop.hbase.master.cleaner.CleanerChore.checkAndDeleteEntries(CleanerChore.java:149)
at 
org.apache.hadoop.hbase.master.cleaner.CleanerChore.checkAndDeleteDirectory(CleanerChore.java:180)
at 
org.apache.hadoop.hbase.master.cleaner.CleanerChore.checkAndDeleteEntries(CleanerChore.java:149)
at 
org.apache.hadoop.hbase.master.cleaner.CleanerChore.checkAndDeleteDirectory(CleanerChore.java:180)
at 
org.apache.hadoop.hbase.master.cleaner.CleanerChore.checkAndDeleteEntries(CleanerChore.java:149)
at 
org.apache.hadoop.hbase.master.cleaner.CleanerChore.checkAndDeleteDirectory(CleanerChore.java:180)
at 
org.apache.hadoop.hbase.master.cleaner.CleanerChore.checkAndDeleteEntries(CleanerChore.java:149)
at 
org.apache.hadoop.hbase.master.cleaner.CleanerChore.chore(CleanerChore.java:124)
at org.apache.hadoop.hbase.ScheduledChore.run(ScheduledChore.java:185)
at 
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
at 
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
at 
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
Caused by: java.io.FileNotFoundException: File does not exist: 
/hbase/.hbase-snapshot/.tmp/frog_stastic_2016-08-17/.snapshotinfo
at 
org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:71)
at 
org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:61)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocationsInt(FSNamesystem.java:1828)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocations(FSNamesystem.java:1799)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocations(FSNamesystem.java:1712)
at 
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getBlockLocations(NameNodeRpcServer.java:587)
at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getBlockLocations(ClientNamenodeProtocolServerSideTranslatorPB.java:365)
at 
org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
at 

[jira] [Created] (HBASE-16427) After HBASE-13701, hbase standalone mode start failed due to mkdir failed

2016-08-16 Thread Heng Chen (JIRA)
Heng Chen created HBASE-16427:
-

 Summary: After HBASE-13701,  hbase standalone mode start failed 
due to mkdir failed
 Key: HBASE-16427
 URL: https://issues.apache.org/jira/browse/HBASE-16427
 Project: HBase
  Issue Type: Bug
Reporter: Heng Chen


{code}
2016-08-17 10:26:43,305 ERROR [main] regionserver.SecureBulkLoadManager: Failed 
to create or set permission on staging directory /user/chenheng/hbase-staging
ExitCodeException exitCode=1: chmod: /user/chenheng/hbase-staging: No such file 
or directory

at org.apache.hadoop.util.Shell.runCommand(Shell.java:545)
at org.apache.hadoop.util.Shell.run(Shell.java:456)
at 
org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:722)
at org.apache.hadoop.util.Shell.execCommand(Shell.java:815)
at org.apache.hadoop.util.Shell.execCommand(Shell.java:798)
at 
org.apache.hadoop.fs.RawLocalFileSystem.setPermission(RawLocalFileSystem.java:728)
at 
org.apache.hadoop.fs.FilterFileSystem.setPermission(FilterFileSystem.java:502)
at 
org.apache.hadoop.hbase.regionserver.SecureBulkLoadManager.start(SecureBulkLoadManager.java:124)
at 
org.apache.hadoop.hbase.regionserver.HRegionServer.(HRegionServer.java:626)
at org.apache.hadoop.hbase.master.HMaster.(HMaster.java:406)
at 
org.apache.hadoop.hbase.master.HMasterCommandLine$LocalHMaster.(HMasterCommandLine.java:307)
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
at 
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
at 
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
at java.lang.reflect.Constructor.newInstance(Constructor.java:422)
at 
org.apache.hadoop.hbase.util.JVMClusterUtil.createMasterThread(JVMClusterUtil.java:140)
at 
org.apache.hadoop.hbase.LocalHBaseCluster.addMaster(LocalHBaseCluster.java:221)
at 
org.apache.hadoop.hbase.LocalHBaseCluster.(LocalHBaseCluster.java:156)
at 
org.apache.hadoop.hbase.master.HMasterCommandLine.startMaster(HMasterCommandLine.java:226)
at 
org.apache.hadoop.hbase.master.HMasterCommandLine.run(HMasterCommandLine.java:139)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
at 
org.apache.hadoop.hbase.util.ServerCommandLine.doMain(ServerCommandLine.java:127)
at org.apache.hadoop.hbase.master.HMaster.main(HMaster.java:2421)
2016-08-17 10:26:43,306 ERROR [main] master.HMasterCommandLine: Master exiting
{code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-16258) Some issues about metrices on webUI

2016-07-20 Thread Heng Chen (JIRA)
Heng Chen created HBASE-16258:
-

 Summary: Some issues about metrices on webUI
 Key: HBASE-16258
 URL: https://issues.apache.org/jira/browse/HBASE-16258
 Project: HBase
  Issue Type: Bug
Reporter: Heng Chen






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HBASE-16076) Cannot configure split policy in HBase shell

2016-07-17 Thread Heng Chen (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-16076?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Heng Chen resolved HBASE-16076.
---
Resolution: Fixed
  Assignee: Heng Chen

Commit to master

> Cannot configure split policy in HBase shell
> 
>
> Key: HBASE-16076
> URL: https://issues.apache.org/jira/browse/HBASE-16076
> Project: HBase
>  Issue Type: Bug
>  Components: documentation
>Affects Versions: 2.0.0
>Reporter: Youngjoon Kim
>Assignee: Heng Chen
>Priority: Minor
> Fix For: 2.0.0
>
> Attachments: HBASE-16076.patch, HBASE-16076_v1.patch
>
>
> The reference guide explains how to configure split policy in HBase 
> shell([link|http://hbase.apache.org/book.html#_custom_split_policies]).
> {noformat}
> Configuring the Split Policy On a Table Using HBase Shell
> hbase> create 'test', {METHOD => 'table_att', CONFIG => {'SPLIT_POLICY' => 
> 'org.apache.hadoop.hbase.regionserver.ConstantSizeRegionSplitPolicy'}},
> {NAME => 'cf1'}
> {noformat}
> But if run that command, shell complains 'An argument ignored (unknown or 
> overridden): CONFIG', and the table description has no split policy.
> {noformat}
> hbase(main):067:0* create 'test', {METHOD => 'table_att', CONFIG => 
> {'SPLIT_POLICY' => 
> 'org.apache.hadoop.hbase.regionserver.ConstantSizeRegionSplitPolicy'}}, {NAME 
> => 'cf1'}
> An argument ignored (unknown or overridden): CONFIG
> Created table test
> Took 1.2180 seconds
> hbase(main):068:0> describe 'test'
> Table test is ENABLED
> test
> COLUMN FAMILIES DESCRIPTION
> {NAME => 'cf1', DATA_BLOCK_ENCODING => 'NONE', BLOOMFILTER => 'ROW', 
> REPLICATION_SCOPE => '0', COMPRESSION => 'NONE', VERSIONS => '1', TTL => 
> 'FOREVER', MIN_VERSIONS => '0', IN_MEMORY_COMPACTION => 'false', 
> KEEP_DELETED_CELLS => 'FALSE', BLOCKSIZE => '65536', IN_MEMORY => '
> false', BLOCKCACHE => 'true'}
> 1 row(s)
> Took 0.0200 seconds
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HBASE-16111) Truncate preserve shell command is broken

2016-06-27 Thread Heng Chen (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-16111?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Heng Chen resolved HBASE-16111.
---
  Resolution: Fixed
Assignee: Heng Chen
Hadoop Flags: Reviewed

> Truncate preserve shell command is broken
> -
>
> Key: HBASE-16111
> URL: https://issues.apache.org/jira/browse/HBASE-16111
> Project: HBase
>  Issue Type: Bug
>  Components: shell
>Reporter: Elliott Clark
>Assignee: Heng Chen
>  Labels: shell
> Fix For: 2.0.0
>
> Attachments: HBASE-16111.patch
>
>
> On a recent version of master I get this:
> {code}
> hbase(main):001:0> truncate_preserve 'TestTable'
> ERROR: undefined local variable or method `table' for 
> #
> Here is some help for this command:
>   Disables, drops and recreates the specified table while still maintaing the 
> previous region boundaries.
> Took 0.0290 seconds
> hbase(main):002:0> truncate 'TestTable'
> Truncating 'TestTable' table (it may take a while):
> Disabling table...
> Truncating table...
> Took 10.0040 seconds
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HBASE-15900) RS stuck in get lock of HStore

2016-06-23 Thread Heng Chen (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-15900?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Heng Chen resolved HBASE-15900.
---
Resolution: Duplicate

> RS stuck in get lock of HStore
> --
>
> Key: HBASE-15900
> URL: https://issues.apache.org/jira/browse/HBASE-15900
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.1.1, 1.3.0
>Reporter: Heng Chen
> Attachments: 0d32a6bab354e6cc170cd59a2d485797.jstack.txt, 
> 0d32a6bab354e6cc170cd59a2d485797.rs.log, 9fe15a52_9fe15a52_save, 
> c91324eb_81194e359707acadee2906ffe36ab130.log, dump.txt
>
>
> It happens on my production cluster when i run MR job.  I save the dump.txt 
> from this RS webUI.
> Many threads stuck here:
> {code}
> Thread 133 (B.defaultRpcServer.handler=94,queue=4,port=16020):
>32   State: WAITING
>31   Blocked count: 477816
>30   Waited count: 535255
>29   Waiting on 
> java.util.concurrent.locks.ReentrantReadWriteLock$NonfairSync@6447ba67
>28   Stack:
>27 sun.misc.Unsafe.park(Native Method)
>26 java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
>25 
> java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:836)
>24 
> java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireShared(AbstractQueuedSynchronizer.java:967)
>23 
> java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireShared(AbstractQueuedSynchronizer.java:1283)
>22 
> java.util.concurrent.locks.ReentrantReadWriteLock$ReadLock.lock(ReentrantReadWriteLock.java:727)
>21 org.apache.hadoop.hbase.regionserver.HStore.add(HStore.java:666)
>20 
> org.apache.hadoop.hbase.regionserver.HRegion.applyFamilyMapToMemstore(HRegion.java:3621)
>19 
> org.apache.hadoop.hbase.regionserver.HRegion.doMiniBatchMutation(HRegion.java:3038)
>18 
> org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:2793)
>17 
> org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:2735)
>16 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.doBatchOp(RSRpcServices.java:692)
>15 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.doNonAtomicRegionMutation(RSRpcServices.java:654)
>14 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.multi(RSRpcServices.java:2029)
>13 
> org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:32213)
>12 org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2112)
>11 org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
>10 
> org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
> 9 org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
> 8 java.lang.Thread.run(Thread.java:745)
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-16040) Remove configuration "hbase.replication"

2016-06-16 Thread Heng Chen (JIRA)
Heng Chen created HBASE-16040:
-

 Summary: Remove configuration "hbase.replication"
 Key: HBASE-16040
 URL: https://issues.apache.org/jira/browse/HBASE-16040
 Project: HBase
  Issue Type: Bug
Reporter: Heng Chen
 Fix For: 2.0.0


This configuration was introduced to reduce overhead of replication. 
Now the overhead of replication is negligible. Besides that, this config is not 
in hbase-default.xml,  user has to read the code to know about it and its' 
default value, this is unfriendly. 

So let's remove it.  suggestions?



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HBASE-16031) Documents about "hbase.replication" default value seems wrong

2016-06-16 Thread Heng Chen (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-16031?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Heng Chen resolved HBASE-16031.
---
   Resolution: Fixed
 Assignee: Heng Chen
Fix Version/s: 2.0.0

> Documents about "hbase.replication" default value seems wrong
> -
>
> Key: HBASE-16031
> URL: https://issues.apache.org/jira/browse/HBASE-16031
> Project: HBase
>  Issue Type: Bug
>Reporter: Heng Chen
>Assignee: Heng Chen
> Fix For: 2.0.0
>
> Attachments: HBASE-16031.patch
>
>
> {code}
>   public static final String
>   REPLICATION_ENABLE_KEY = "hbase.replication";
>   public static final boolean
>   REPLICATION_ENABLE_DEFAULT = true;
> {code}
> The code shows that default value is true, but documents shows the default 
> value is false.
> {code}
> | hbase.replication
> | Whether replication is enabled or disabled on a given
> cluster
> | false
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-16031) Documents about "hbase.replication" default value seems wrong

2016-06-14 Thread Heng Chen (JIRA)
Heng Chen created HBASE-16031:
-

 Summary: Documents about "hbase.replication" default value seems 
wrong
 Key: HBASE-16031
 URL: https://issues.apache.org/jira/browse/HBASE-16031
 Project: HBase
  Issue Type: Bug
Reporter: Heng Chen


{code}
  public static final String
  REPLICATION_ENABLE_KEY = "hbase.replication";
  public static final boolean
  REPLICATION_ENABLE_DEFAULT = true;
{code}

The code shows that default value is true, but documents shows the default 
value is false.

{code}

| hbase.replication
| Whether replication is enabled or disabled on a given
cluster
| false
{code}





--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-15900) RS stuck in get lock of HStore

2016-05-27 Thread Heng Chen (JIRA)
Heng Chen created HBASE-15900:
-

 Summary: RS stuck in get lock of HStore
 Key: HBASE-15900
 URL: https://issues.apache.org/jira/browse/HBASE-15900
 Project: HBase
  Issue Type: Bug
Reporter: Heng Chen
 Attachments: dump.txt

It happens on my production cluster when i run MR job.  I save the dump.txt 
from this RS webUI.

Many threads stuck here:
{code}
Thread 133 (B.defaultRpcServer.handler=94,queue=4,port=16020):
   32   State: WAITING
   31   Blocked count: 477816
   30   Waited count: 535255
   29   Waiting on 
java.util.concurrent.locks.ReentrantReadWriteLock$NonfairSync@6447ba67
   28   Stack:
   27 sun.misc.Unsafe.park(Native Method)
   26 java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
   25 
java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:836)
   24 
java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireShared(AbstractQueuedSynchronizer.java:967)
   23 
java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireShared(AbstractQueuedSynchronizer.java:1283)
   22 
java.util.concurrent.locks.ReentrantReadWriteLock$ReadLock.lock(ReentrantReadWriteLock.java:727)
   21 org.apache.hadoop.hbase.regionserver.HStore.add(HStore.java:666)
   20 
org.apache.hadoop.hbase.regionserver.HRegion.applyFamilyMapToMemstore(HRegion.java:3621)
   19 
org.apache.hadoop.hbase.regionserver.HRegion.doMiniBatchMutation(HRegion.java:3038)
   18 
org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:2793)
   17 
org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:2735)
   16 
org.apache.hadoop.hbase.regionserver.RSRpcServices.doBatchOp(RSRpcServices.java:692)
   15 
org.apache.hadoop.hbase.regionserver.RSRpcServices.doNonAtomicRegionMutation(RSRpcServices.java:654)
   14 
org.apache.hadoop.hbase.regionserver.RSRpcServices.multi(RSRpcServices.java:2029)
   13 
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:32213)
   12 org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2112)
   11 org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
   10 
org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
9 org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
8 java.lang.Thread.run(Thread.java:745)
{code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-15844) We should respect hfile.block.index.cacheonwrite when write intermediate index Block

2016-05-17 Thread Heng Chen (JIRA)
Heng Chen created HBASE-15844:
-

 Summary: We should respect hfile.block.index.cacheonwrite when 
write intermediate index Block
 Key: HBASE-15844
 URL: https://issues.apache.org/jira/browse/HBASE-15844
 Project: HBase
  Issue Type: Bug
Reporter: Heng Chen


{code: title=BlockIndexWriter#writeIntermediateBlock}
  if (cacheConf != null) {
HFileBlock blockForCaching = blockWriter.getBlockForCaching(cacheConf);
cacheConf.getBlockCache().cacheBlock(new BlockCacheKey(nameForCaching,
  beginOffset, true, blockForCaching.getBlockType()), blockForCaching);
  }
{code}

The if condition should be ?
{code}
if (cacheConf != null && cacheConf.shouldCacheIndexesOnWrite()) 
{code} 





--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-15814) Miss important information in Document of HBase Security

2016-05-11 Thread Heng Chen (JIRA)
Heng Chen created HBASE-15814:
-

 Summary: Miss important information in Document of HBase Security
 Key: HBASE-15814
 URL: https://issues.apache.org/jira/browse/HBASE-15814
 Project: HBase
  Issue Type: Bug
  Components: documentation
Reporter: Heng Chen


I have deployed secure cluster recently, and found we miss important 
information in http://hbase.apache.org/book.html#security

Some configurations like 
{code}

  hbase.regionserver.kerberos.principal 
  hbase/_h...@your-realm.com 
 

 
  hbase.regionserver.keytab.file 
  /etc/hbase/conf/hbase.keytab 


 
  hbase.master.kerberos.principal 
  hbase/_h...@your-realm.com 
 

 
hbase.master.keytab.file 
/etc/hbase/conf/hbase.keytab 

{code}

And i found more detailed document in 
http://www.cloudera.com/documentation/enterprise/5-5-x/topics/cdh_sg_hbase_authentication.html



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-15804) Return 404 in some document link

2016-05-09 Thread Heng Chen (JIRA)
Heng Chen created HBASE-15804:
-

 Summary: Return 404 in some document link
 Key: HBASE-15804
 URL: https://issues.apache.org/jira/browse/HBASE-15804
 Project: HBase
  Issue Type: Bug
Reporter: Heng Chen


http://hbase.apache.org/book.html#security

The link to {{Understanding User Authentication and Authorization in Apache 
HBase} } return 404



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HBASE-15720) Print row locks at the debug dump page

2016-05-02 Thread Heng Chen (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-15720?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Heng Chen resolved HBASE-15720.
---
Resolution: Fixed

> Print row locks at the debug dump page
> --
>
> Key: HBASE-15720
> URL: https://issues.apache.org/jira/browse/HBASE-15720
> Project: HBase
>  Issue Type: Improvement
>Affects Versions: 1.2.1
>Reporter: Enis Soztutar
>Assignee: Heng Chen
> Fix For: 2.0.0, 1.3.0, 1.4.0, 1.1.5, 1.2.2, 0.98.20, 1.0.5
>
> Attachments: 4742C21D-B9CE-4921-9B32-CC319488EC64.png, 
> HBASE-15720-branch-1.0-addendum.patch, HBASE-15720-branch-1.2-addendum.patch, 
> HBASE-15720.patch
>
>
> We had to debug cases where some handlers are holding row locks for an 
> extended time (and maybe leak) and other handlers are getting timeouts for 
> obtaining row locks. 
> We should add row lock information at the debug page at the RS UI to be able 
> to live-debug such cases.  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Reopened] (HBASE-15278) AsyncRPCClient hangs if Connection closes before RPC call response

2016-04-29 Thread Heng Chen (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-15278?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Heng Chen reopened HBASE-15278:
---

revert it.  See testcase failed 
https://builds.apache.org/job/PreCommit-HBASE-Build/1691/testReport/

> AsyncRPCClient hangs if Connection closes before RPC call response 
> ---
>
> Key: HBASE-15278
> URL: https://issues.apache.org/jira/browse/HBASE-15278
> Project: HBase
>  Issue Type: Bug
>  Components: rpc, test
>Affects Versions: 2.0.0
>Reporter: Enis Soztutar
>Assignee: Heng Chen
>Priority: Blocker
> Fix For: 2.0.0
>
> Attachments: HBASE-15278.patch, HBASE-15278_v1.patch, 
> HBASE-15278_v2.patch, hbase-15278_v00.patch
>
>
> The test for HBASE-15212 discovered an issue with Async RPC Client. 
> In that test, we are closing the connection if an RPC call writes a call 
> larger than max allowed size, the server closes the connection. However the 
> async client does not seem to handle connection closes with outstanding RPC 
> calls. The client just hangs. 
> Marking this blocker against 2.0 since it is default there. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-15692) We should remove snapshot dir under .tmp when write .snapshotinfo failed.

2016-04-21 Thread Heng Chen (JIRA)
Heng Chen created HBASE-15692:
-

 Summary: We should remove snapshot dir under .tmp when write 
.snapshotinfo failed.
 Key: HBASE-15692
 URL: https://issues.apache.org/jira/browse/HBASE-15692
 Project: HBase
  Issue Type: Bug
Reporter: Heng Chen


we encounter this problem on our production cluster.   
This is exception in HMaster.log
{code}
2016-04-22 11:05:06,390 ERROR [f04,16000,1459941011479_ChoreService_3] 
snapshot.SnapshotHFileCleaner: Exception while checking if files were valid, 
keeping them just in case.
org.apache.hadoop.hbase.snapshot.CorruptedSnapshotException: Couldn't read 
snapshot info 
from:hdfs://f04/hbase/.hbase-snapshot/.tmp/frog_stastic_2016-04-07/.snapshotinfo
at 
org.apache.hadoop.hbase.snapshot.SnapshotDescriptionUtils.readSnapshotInfo(SnapshotDescriptionUtils.java:295)
at 
org.apache.hadoop.hbase.snapshot.SnapshotReferenceUtil.getHFileNames(SnapshotReferenceUtil.java:328)
at 
org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner$1.filesUnderSnapshot(SnapshotHFileCleaner.java:85)
at 
org.apache.hadoop.hbase.master.snapshot.SnapshotFileCache.getSnapshotsInProgress(SnapshotFileCache.java:303)
at 
org.apache.hadoop.hbase.master.snapshot.SnapshotFileCache.getUnreferencedFiles(SnapshotFileCache.java:194)
at 
org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner.getDeletableFiles(SnapshotHFileCleaner.java:62)
at 
org.apache.hadoop.hbase.master.cleaner.CleanerChore.checkAndDeleteFiles(CleanerChore.java:233)
at 
org.apache.hadoop.hbase.master.cleaner.CleanerChore.checkAndDeleteEntries(CleanerChore.java:157)
at 
org.apache.hadoop.hbase.master.cleaner.CleanerChore.checkAndDeleteDirectory(CleanerChore.java:180)
at 
org.apache.hadoop.hbase.master.cleaner.CleanerChore.checkAndDeleteEntries(CleanerChore.java:149)
at 
org.apache.hadoop.hbase.master.cleaner.CleanerChore.checkAndDeleteDirectory(CleanerChore.java:180)
at 
org.apache.hadoop.hbase.master.cleaner.CleanerChore.checkAndDeleteEntries(CleanerChore.java:149)
at 
org.apache.hadoop.hbase.master.cleaner.CleanerChore.checkAndDeleteDirectory(CleanerChore.java:180)
at 
org.apache.hadoop.hbase.master.cleaner.CleanerChore.checkAndDeleteEntries(CleanerChore.java:149)
at 
org.apache.hadoop.hbase.master.cleaner.CleanerChore.checkAndDeleteDirectory(CleanerChore.java:180)
at 
org.apache.hadoop.hbase.master.cleaner.CleanerChore.checkAndDeleteEntries(CleanerChore.java:149)
at 
org.apache.hadoop.hbase.master.cleaner.CleanerChore.checkAndDeleteDirectory(CleanerChore.java:180)
at 
org.apache.hadoop.hbase.master.cleaner.CleanerChore.checkAndDeleteEntries(CleanerChore.java:149)
at 
org.apache.hadoop.hbase.master.cleaner.CleanerChore.chore(CleanerChore.java:124)
at org.apache.hadoop.hbase.ScheduledChore.run(ScheduledChore.java:185)
at 
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
at 
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
at 
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
Caused by: java.io.FileNotFoundException: File does not exist: 
/hbase/.hbase-snapshot/.tmp/frog_stastic_2016-04-07/.snapshotinfo
at 
org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:64)
at 
org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:54)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocationsUpdateTimes(FSNamesystem.java:1795)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocationsInt(FSNamesystem.java:1738)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocations(FSNamesystem.java:1718)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocations(FSNamesystem.java:1690)
at 
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getBlockLocations(NameNodeRpcServer.java:519)
at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getBlockLocations(ClientNamenodeProtocolServerSideTranslatorPB.java:337)
at 
org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:585)
at 

[jira] [Resolved] (HBASE-15642) split region number for one table on webUI never be reduced

2016-04-13 Thread Heng Chen (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-15642?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Heng Chen resolved HBASE-15642.
---
Resolution: Invalid

> split region number for one table on webUI never be reduced 
> 
>
> Key: HBASE-15642
> URL: https://issues.apache.org/jira/browse/HBASE-15642
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 0.98.17
>Reporter: Heng Chen
> Attachments: 268286F8-8BA4-4144-9B66-1A6F70FEC2EB.png
>
>
> This happened after we upgrade our cluster from 0.98.6 to 0.98.17. 
> The number should be reduced,   but it always increases from original 20+ to 
> 49 now. (yesterday it was 48)
> Need to dig.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-15643) Need metrics of cache hit ratio for one table

2016-04-12 Thread Heng Chen (JIRA)
Heng Chen created HBASE-15643:
-

 Summary: Need metrics of cache hit ratio for one table
 Key: HBASE-15643
 URL: https://issues.apache.org/jira/browse/HBASE-15643
 Project: HBase
  Issue Type: Bug
Reporter: Heng Chen


There are many tables on our cluster,  we  only some of them need to be read 
online.  

We could improve the performance of read by cache,  but we need some metrics 
for it at table level 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-15642) split region number for one table on webUI never be reduced

2016-04-12 Thread Heng Chen (JIRA)
Heng Chen created HBASE-15642:
-

 Summary: split region number for one table on webUI never be 
reduced 
 Key: HBASE-15642
 URL: https://issues.apache.org/jira/browse/HBASE-15642
 Project: HBase
  Issue Type: Bug
Affects Versions: 0.98.17
Reporter: Heng Chen


This happened after we upgrade our cluster from 0.98.6 to 0.98.17. 




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-15635) Mean age of Blocks in cache (seconds) on webUI should be greater than zero

2016-04-11 Thread Heng Chen (JIRA)
Heng Chen created HBASE-15635:
-

 Summary: Mean age of Blocks in cache (seconds) on webUI should be 
greater than zero
 Key: HBASE-15635
 URL: https://issues.apache.org/jira/browse/HBASE-15635
 Project: HBase
  Issue Type: Bug
Affects Versions: 0.98.17
Reporter: Heng Chen






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-15629) Backport HBASE-14703 to 0.98+

2016-04-11 Thread Heng Chen (JIRA)
Heng Chen created HBASE-15629:
-

 Summary: Backport HBASE-14703 to 0.98+
 Key: HBASE-15629
 URL: https://issues.apache.org/jira/browse/HBASE-15629
 Project: HBase
  Issue Type: Task
Reporter: Heng Chen






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HBASE-15377) Per-RS Get metric is time based, per-region metric is size-based

2016-03-07 Thread Heng Chen (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-15377?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Heng Chen resolved HBASE-15377.
---
Resolution: Duplicate

> Per-RS Get metric is time based, per-region metric is size-based
> 
>
> Key: HBASE-15377
> URL: https://issues.apache.org/jira/browse/HBASE-15377
> Project: HBase
>  Issue Type: Sub-task
>Reporter: Enis Soztutar
>
> We have metrics for Get operations at the region server level and region 
> level. 
> {code}
>"Get_num_ops" : 4837505,
> "Get_min" : 0,
> "Get_max" : 296,
> "Get_mean" : 0.2934618155433431,
> "Get_median" : 0.0,
> "Get_75th_percentile" : 0.0,
> "Get_95th_percentile" : 1.0,
> "Get_99th_percentile" : 1.0,
> {code}
> and 
> {code}
>"Namespace_hbase_table_meta_region_1588230740_metric_get_num_ops" : 103,
> "Namespace_hbase_table_meta_region_1588230740_metric_get_min" : 450,
> "Namespace_hbase_table_meta_region_1588230740_metric_get_max" : 470,
> "Namespace_hbase_table_meta_region_1588230740_metric_get_mean" : 
> 450.19417475728153,
> "Namespace_hbase_table_meta_region_1588230740_metric_get_median" : 460.0,
> "Namespace_hbase_table_meta_region_1588230740_metric_get_75th_percentile" 
> : 470.0,
> "Namespace_hbase_table_meta_region_1588230740_metric_get_95th_percentile" 
> : 470.0,
> "Namespace_hbase_table_meta_region_1588230740_metric_get_99th_percentile" 
> : 470.0,
> {code}
> The problem is that the report values for the region server shows the 
> latency, versus the reported values for the region shows the response sizes. 
> There is no way of telling this without reading the source code. 
> I think we should deprecate response size histograms in favor of latency 
> histograms. 
> See also HBASE-15376. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HBASE-15329) Cross-Site Scripting: Reflected in table.jsp

2016-03-02 Thread Heng Chen (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-15329?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Heng Chen resolved HBASE-15329.
---
  Resolution: Fixed
Hadoop Flags: Reviewed

> Cross-Site Scripting: Reflected in table.jsp
> 
>
> Key: HBASE-15329
> URL: https://issues.apache.org/jira/browse/HBASE-15329
> Project: HBase
>  Issue Type: Bug
>  Components: security
>Reporter: stack
>Assignee: Samir Ahmic
>Priority: Minor
> Fix For: 2.0.0
>
> Attachments: HBASE-15329_v0.patch
>
>
> Minor issue where we write back table name in a few places. Should clean it 
> up:
> {code}
>  } else { 
>   out.write("\nTable: ");
>   out.print( fqtn );
>   out.write("\n");
>  } 
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-15384) Avoid using '/tmp' directory in TestBulkLoad

2016-03-02 Thread Heng Chen (JIRA)
Heng Chen created HBASE-15384:
-

 Summary: Avoid using '/tmp' directory in TestBulkLoad
 Key: HBASE-15384
 URL: https://issues.apache.org/jira/browse/HBASE-15384
 Project: HBase
  Issue Type: Sub-task
Reporter: Heng Chen






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-15326) NPE in TestHRegion.testBatchPut_whileNoRowLocksHeld

2016-02-25 Thread Heng Chen (JIRA)
Heng Chen created HBASE-15326:
-

 Summary: NPE in TestHRegion.testBatchPut_whileNoRowLocksHeld
 Key: HBASE-15326
 URL: https://issues.apache.org/jira/browse/HBASE-15326
 Project: HBase
  Issue Type: Bug
Reporter: Heng Chen


{code}
java.lang.NullPointerException
at 
org.apache.hadoop.metrics2.lib.MutableHistogram.updateSnapshotMetrics(MutableHistogram.java:72)
at 
org.apache.hadoop.metrics2.lib.MutableRangeHistogram.snapshot(MutableRangeHistogram.java:59)
at 
org.apache.hadoop.metrics2.lib.DynamicMetricsRegistry.snapshot(DynamicMetricsRegistry.java:391)
at 
org.apache.hadoop.hbase.metrics.BaseSourceImpl.getMetrics(BaseSourceImpl.java:146)
at 
org.apache.hadoop.hbase.test.MetricsAssertHelperImpl.getMetrics(MetricsAssertHelperImpl.java:243)
at 
org.apache.hadoop.hbase.test.MetricsAssertHelperImpl.getCounter(MetricsAssertHelperImpl.java:201)
at 
org.apache.hadoop.hbase.regionserver.TestHRegion.testBatchPut_whileNoRowLocksHeld(TestHRegion.java:1498)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:497)
at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
at 
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at 
org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at 
org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:55)
at org.junit.rules.RunRules.evaluate(RunRules.java:20)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
at 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
at 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
at org.junit.runner.JUnitCore.run(JUnitCore.java:137)
at 
com.intellij.junit4.JUnit4IdeaTestRunner.startRunnerWithArgs(JUnit4IdeaTestRunner.java:117)
at 
com.intellij.rt.execution.junit.JUnitStarter.prepareStreamsAndStart(JUnitStarter.java:234)
at 
com.intellij.rt.execution.junit.JUnitStarter.main(JUnitStarter.java:74)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:497)
at com.intellij.rt.execution.application.AppMain.main(AppMain.java:144)
{code}

It seems to be introduced after HBASE-15222,  [~eclark]



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-15308) Flakey TestSplitWalDataLoss on branch-1.1

2016-02-22 Thread Heng Chen (JIRA)
Heng Chen created HBASE-15308:
-

 Summary: Flakey TestSplitWalDataLoss on branch-1.1
 Key: HBASE-15308
 URL: https://issues.apache.org/jira/browse/HBASE-15308
 Project: HBase
  Issue Type: Sub-task
Reporter: Heng Chen


It happens during HBASE-15169 QA test,  see 

https://builds.apache.org/job/PreCommit-HBASE-Build/628/artifact/patchprocess/patch-unit-hbase-server-jdk1.8.0_72.txt

https://builds.apache.org/job/PreCommit-HBASE-Build/547/artifact/patchprocess/patch-unit-hbase-server-jdk1.8.0_72.txt



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-15288) Flakey TestMasterMetrics.testClusterRequests on branch-1.1

2016-02-18 Thread Heng Chen (JIRA)
Heng Chen created HBASE-15288:
-

 Summary: Flakey TestMasterMetrics.testClusterRequests on branch-1.1
 Key: HBASE-15288
 URL: https://issues.apache.org/jira/browse/HBASE-15288
 Project: HBase
  Issue Type: Sub-task
Reporter: Heng Chen


I found it during HBASE-15169, Let me see what's happening.

https://builds.apache.org/job/PreCommit-HBASE-Build/586/testReport/org.apache.hadoop.hbase.master/TestMasterMetrics/testClusterRequests/



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HBASE-15182) unify normalizer switch

2016-02-17 Thread Heng Chen (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-15182?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Heng Chen resolved HBASE-15182.
---
Resolution: Invalid

> unify normalizer switch
> ---
>
> Key: HBASE-15182
> URL: https://issues.apache.org/jira/browse/HBASE-15182
> Project: HBase
>  Issue Type: Sub-task
>Reporter: Heng Chen
> Fix For: 2.0.0, 1.3.0
>
>
> After HBASE-15128,  we will have an uniform way to do switch. Let's unify 
> normalizer into it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-15280) Scan metrics should not ignore empty rows.

2016-02-17 Thread Heng Chen (JIRA)
Heng Chen created HBASE-15280:
-

 Summary: Scan metrics should not ignore empty rows.
 Key: HBASE-15280
 URL: https://issues.apache.org/jira/browse/HBASE-15280
 Project: HBase
  Issue Type: Bug
Reporter: Heng Chen


It is come from HBASE-15267.

we found that as for empty result,  GET increment the readRequestCount, but 
Scan does not do it.   
Detail information, please see  HBASE-15267 comments.

This jira is to fix it.  We should record each request we do.  At least, we 
could make Scan metrics behavior be consistent with Get. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HBASE-15178) metrics on web UI sometimes flakey

2016-01-27 Thread Heng Chen (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-15178?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Heng Chen resolved HBASE-15178.
---
Resolution: Duplicate

> metrics on web UI sometimes flakey
> --
>
> Key: HBASE-15178
> URL: https://issues.apache.org/jira/browse/HBASE-15178
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 0.98.6
>Reporter: Heng Chen
> Attachments: 70DD704D-7E05-49BA-AC84-0646671F67AA.png
>
>
> see attachement



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-15182) unify normalizer switch

2016-01-27 Thread Heng Chen (JIRA)
Heng Chen created HBASE-15182:
-

 Summary: unify normalizer switch
 Key: HBASE-15182
 URL: https://issues.apache.org/jira/browse/HBASE-15182
 Project: HBase
  Issue Type: Sub-task
Reporter: Heng Chen


After HBASE-15128,  we will have an uniform way to do switch. Let's unify 
normalizer into it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-15178) metrics on web UI sometimes flakey

2016-01-26 Thread Heng Chen (JIRA)
Heng Chen created HBASE-15178:
-

 Summary: metrics on web UI sometimes flakey
 Key: HBASE-15178
 URL: https://issues.apache.org/jira/browse/HBASE-15178
 Project: HBase
  Issue Type: Bug
Reporter: Heng Chen






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-15108) TestReplicationAdmin failed on branch-1.0

2016-01-13 Thread Heng Chen (JIRA)
Heng Chen created HBASE-15108:
-

 Summary: TestReplicationAdmin failed on branch-1.0 
 Key: HBASE-15108
 URL: https://issues.apache.org/jira/browse/HBASE-15108
 Project: HBase
  Issue Type: Bug
Reporter: Heng Chen


I notice it on HBASE-15095.

See
https://builds.apache.org/job/PreCommit-HBASE-Build/95/artifact/patchprocess/patch-unit-hbase-server-jdk1.8.0.txt



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-15049) AuthTypes.NONE cause exception after HS2 start

2015-12-29 Thread Heng Chen (JIRA)
Heng Chen created HBASE-15049:
-

 Summary: AuthTypes.NONE cause exception after HS2 start
 Key: HBASE-15049
 URL: https://issues.apache.org/jira/browse/HBASE-15049
 Project: HBase
  Issue Type: Bug
Reporter: Heng Chen


I set {{hive.server2.authentication}} to be {{NONE}}

After HS2 start, i see exception is log below:
{code}
2015-12-29 16:58:42,339 ERROR [HiveServer2-Handler-Pool: Thread-31]: 
server.TThreadPoolServer (TThreadPoolServer.java:run(296)) - Error occurred 
during processing of message.
java.lang.RuntimeException: 
org.apache.thrift.transport.TSaslTransportException: No data or no sasl data in 
the stream
at 
org.apache.thrift.transport.TSaslServerTransport$Factory.getTransport(TSaslServerTransport.java:219)
at 
org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:268)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.thrift.transport.TSaslTransportException: No data or no 
sasl data in the stream
at 
org.apache.thrift.transport.TSaslTransport.open(TSaslTransport.java:328)
at 
org.apache.thrift.transport.TSaslServerTransport.open(TSaslServerTransport.java:41)
at 
org.apache.thrift.transport.TSaslServerTransport$Factory.getTransport(TSaslServerTransport.java:216)
... 4 more
{code}

IMO the problem is we use Sasl transport when authType is NONE, 
{code:title=HiveAuthFactory.java}
  public TTransportFactory getAuthTransFactory() throws LoginException {
TTransportFactory transportFactory;
if (authTypeStr.equalsIgnoreCase(AuthTypes.KERBEROS.getAuthName())) {
  try {
transportFactory = 
saslServer.createTransportFactory(getSaslProperties());
  } catch (TTransportException e) {
throw new LoginException(e.getMessage());
  }
} else if (authTypeStr.equalsIgnoreCase(AuthTypes.NONE.getAuthName())) {
  transportFactory = PlainSaslHelper.getPlainTransportFactory(authTypeStr);
} else if (authTypeStr.equalsIgnoreCase(AuthTypes.LDAP.getAuthName())) {
  transportFactory = PlainSaslHelper.getPlainTransportFactory(authTypeStr);
} else if (authTypeStr.equalsIgnoreCase(AuthTypes.PAM.getAuthName())) {
  transportFactory = PlainSaslHelper.getPlainTransportFactory(authTypeStr);
} else if (authTypeStr.equalsIgnoreCase(AuthTypes.NOSASL.getAuthName())) {
  transportFactory = new TTransportFactory();
} else if (authTypeStr.equalsIgnoreCase(AuthTypes.CUSTOM.getAuthName())) {
  transportFactory = PlainSaslHelper.getPlainTransportFactory(authTypeStr);
} else {
  throw new LoginException("Unsupported authentication type " + 
authTypeStr);
}
return transportFactory;
  }
{code}




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HBASE-15049) AuthTypes.NONE cause exception after HS2 start

2015-12-29 Thread Heng Chen (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-15049?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Heng Chen resolved HBASE-15049.
---
Resolution: Invalid

> AuthTypes.NONE cause exception after HS2 start
> --
>
> Key: HBASE-15049
> URL: https://issues.apache.org/jira/browse/HBASE-15049
> Project: HBase
>  Issue Type: Bug
>Reporter: Heng Chen
>
> I set {{hive.server2.authentication}} to be {{NONE}}
> After HS2 start, i see exception in log below:
> {code}
> 2015-12-29 16:58:42,339 ERROR [HiveServer2-Handler-Pool: Thread-31]: 
> server.TThreadPoolServer (TThreadPoolServer.java:run(296)) - Error occurred 
> during processing of message.
> java.lang.RuntimeException: 
> org.apache.thrift.transport.TSaslTransportException: No data or no sasl data 
> in the stream
> at 
> org.apache.thrift.transport.TSaslServerTransport$Factory.getTransport(TSaslServerTransport.java:219)
> at 
> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:268)
> at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> at java.lang.Thread.run(Thread.java:745)
> Caused by: org.apache.thrift.transport.TSaslTransportException: No data or no 
> sasl data in the stream
> at 
> org.apache.thrift.transport.TSaslTransport.open(TSaslTransport.java:328)
> at 
> org.apache.thrift.transport.TSaslServerTransport.open(TSaslServerTransport.java:41)
> at 
> org.apache.thrift.transport.TSaslServerTransport$Factory.getTransport(TSaslServerTransport.java:216)
> ... 4 more
> {code}
> IMO the problem is we use Sasl transport when authType is NONE, 
> {code:title=HiveAuthFactory.java}
>   public TTransportFactory getAuthTransFactory() throws LoginException {
> TTransportFactory transportFactory;
> if (authTypeStr.equalsIgnoreCase(AuthTypes.KERBEROS.getAuthName())) {
>   try {
> transportFactory = 
> saslServer.createTransportFactory(getSaslProperties());
>   } catch (TTransportException e) {
> throw new LoginException(e.getMessage());
>   }
> } else if (authTypeStr.equalsIgnoreCase(AuthTypes.NONE.getAuthName())) {
>   transportFactory = 
> PlainSaslHelper.getPlainTransportFactory(authTypeStr);
> } else if (authTypeStr.equalsIgnoreCase(AuthTypes.LDAP.getAuthName())) {
>   transportFactory = 
> PlainSaslHelper.getPlainTransportFactory(authTypeStr);
> } else if (authTypeStr.equalsIgnoreCase(AuthTypes.PAM.getAuthName())) {
>   transportFactory = 
> PlainSaslHelper.getPlainTransportFactory(authTypeStr);
> } else if (authTypeStr.equalsIgnoreCase(AuthTypes.NOSASL.getAuthName())) {
>   transportFactory = new TTransportFactory();
> } else if (authTypeStr.equalsIgnoreCase(AuthTypes.CUSTOM.getAuthName())) {
>   transportFactory = 
> PlainSaslHelper.getPlainTransportFactory(authTypeStr);
> } else {
>   throw new LoginException("Unsupported authentication type " + 
> authTypeStr);
> }
> return transportFactory;
>   }
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Reopened] (HBASE-14684) Try to remove all MiniMapReduceCluster in unit tests

2015-12-22 Thread Heng Chen (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-14684?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Heng Chen reopened HBASE-14684:
---

revert branch-1. It seems some problems about some tests

> Try to remove all MiniMapReduceCluster in unit tests
> 
>
> Key: HBASE-14684
> URL: https://issues.apache.org/jira/browse/HBASE-14684
> Project: HBase
>  Issue Type: Improvement
>  Components: test
>Reporter: Heng Chen
>Assignee: Heng Chen
>Priority: Critical
> Fix For: 2.0.0, 1.2.0, 1.3.0
>
> Attachments: 14684.branch-1.txt, 14684.branch-1.txt, 
> 14684.branch-1.txt, HBASE-14684-branch-1.2.patch, HBASE-14684-branch-1.patch, 
> HBASE-14684-branch-1.patch, HBASE-14684-branch-1.patch, 
> HBASE-14684-branch-1_v1.patch, HBASE-14684.patch, HBASE-14684_v1.patch
>
>
> As discussion in dev list,  we will try to do MR job without 
> MiniMapReduceCluster.
> Testcases will run faster and more reliable.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HBASE-14949) Skip duplicate entries when replay WAL.

2015-12-09 Thread Heng Chen (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-14949?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Heng Chen resolved HBASE-14949.
---
Resolution: Invalid

> Skip duplicate entries when replay WAL.
> ---
>
> Key: HBASE-14949
> URL: https://issues.apache.org/jira/browse/HBASE-14949
> Project: HBase
>  Issue Type: Sub-task
>Reporter: Heng Chen
> Attachments: HBASE-14949.patch
>
>
> As HBASE-14004 design,  there will be duplicate entries in different WAL.  It 
> happens when one hflush failed, we will close old WAL with 'acked hflushed' 
> length,  then open a new WAL and write the unacked hlushed entries into it.
> So there maybe some overlap between old WAL and new WAL.
> We should skip the duplicate entries when replay.  I think it has no harm to 
> current logic, maybe we do it first. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Reopened] (HBASE-14949) Skip duplicate entries when replay WAL.

2015-12-09 Thread Heng Chen (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-14949?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Heng Chen reopened HBASE-14949:
---

> Skip duplicate entries when replay WAL.
> ---
>
> Key: HBASE-14949
> URL: https://issues.apache.org/jira/browse/HBASE-14949
> Project: HBase
>  Issue Type: Sub-task
>Reporter: Heng Chen
> Attachments: HBASE-14949.patch
>
>
> As HBASE-14004 design,  there will be duplicate entries in different WAL.  It 
> happens when one hflush failed, we will close old WAL with 'acked hflushed' 
> length,  then open a new WAL and write the unacked hlushed entries into it.
> So there maybe some overlap between old WAL and new WAL.
> We should skip the duplicate entries when replay.  I think it has no harm to 
> current logic, maybe we do it first. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-14949) Skip duplicate entries when replay WAL.

2015-12-08 Thread Heng Chen (JIRA)
Heng Chen created HBASE-14949:
-

 Summary: Skip duplicate entries when replay WAL.
 Key: HBASE-14949
 URL: https://issues.apache.org/jira/browse/HBASE-14949
 Project: HBase
  Issue Type: Improvement
Reporter: Heng Chen


As HBASE-14004 design,  there will be duplicate entries in different WAL.  It 
happens when one hflush failed, we will close old WAL with 'acked hflushed' 
length,  then open a new WAL and write the unacked hlushed entries into it.
So there maybe some overlap between old WAL and new WAL.

We should skip the duplicate entries when replay.  I think it has no harm to 
current logic, maybe we do it first. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-14897) TestTableLockManager.testReapAllTableLocks is flakey

2015-11-30 Thread Heng Chen (JIRA)
Heng Chen created HBASE-14897:
-

 Summary: TestTableLockManager.testReapAllTableLocks is flakey
 Key: HBASE-14897
 URL: https://issues.apache.org/jira/browse/HBASE-14897
 Project: HBase
  Issue Type: Bug
Reporter: Heng Chen


It comes from email list which [~stack] post.

This is the some relates QA information.

https://builds.apache.org/view/H-L/view/HBase/job/HBase-Trunk_matrix/512/jdk=latest1.8,label=Hadoop/testReport/org.apache.hadoop.hbase.master/TestTableLockManager/testReapAllTableLocks/

The reason is here.
{code}
writeLocksObtained.await();
writeLocksAttempted.await();
{code}

writeLocksAttempted maybe count down to 0 before created node on ZK,  and main 
thread will go on to run lockManager.reapWriteLocks(),  And after that node was 
created on ZK,  so relates lock acquire will timeout.

I upload a patch which can reproduce this issue.






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-14843) TestWALProcedureStore.testLoad is flakey

2015-11-19 Thread Heng Chen (JIRA)
Heng Chen created HBASE-14843:
-

 Summary: TestWALProcedureStore.testLoad is flakey
 Key: HBASE-14843
 URL: https://issues.apache.org/jira/browse/HBASE-14843
 Project: HBase
  Issue Type: Bug
Affects Versions: 2.0.0
Reporter: Heng Chen


I see it twice recently, 
see.
https://builds.apache.org/job/PreCommit-HBASE-Build/16589//testReport/org.apache.hadoop.hbase.procedure2.store.wal/TestWALProcedureStore/testLoad/

https://builds.apache.org/job/PreCommit-HBASE-Build/16532/testReport/org.apache.hadoop.hbase.procedure2.store.wal/TestWALProcedureStore/testLoad/

Let's see what's happening.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-14815) TestMobExportSnapshot.testExportFailure timeout occasionally

2015-11-14 Thread Heng Chen (JIRA)
Heng Chen created HBASE-14815:
-

 Summary: TestMobExportSnapshot.testExportFailure timeout 
occasionally
 Key: HBASE-14815
 URL: https://issues.apache.org/jira/browse/HBASE-14815
 Project: HBase
  Issue Type: Bug
Reporter: Heng Chen


On master,  TestMobExportSnapshot.testExportFailure timeout occasionally.

See
https://builds.apache.org/job/PreCommit-HBASE-Build/16514//testReport/org.apache.hadoop.hbase.snapshot/TestMobExportSnapshot/testExportFailure/



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HBASE-14659) Fix flakey TestHFileOutputFormat

2015-11-05 Thread Heng Chen (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-14659?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Heng Chen resolved HBASE-14659.
---
Resolution: Won't Fix

> Fix flakey TestHFileOutputFormat
> 
>
> Key: HBASE-14659
> URL: https://issues.apache.org/jira/browse/HBASE-14659
> Project: HBase
>  Issue Type: Bug
>Reporter: Heng Chen
>Assignee: Heng Chen
>
> As i said in HBASE-14654
> This testcase was hanged twice recently.   
> the two QA information:
> https://builds.apache.org/job/PreCommit-HBASE-Build/16118/console
> https://builds.apache.org/job/PreCommit-HBASE-Build/16117/console
> I notice that the two QA running simultaneously on same machine H0.
> I doubt If there are some common resources the two QA conflicts.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HBASE-14265) we should forbid creating table using 'hbase' namespace except by superuser

2015-10-27 Thread Heng Chen (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-14265?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Heng Chen resolved HBASE-14265.
---
Resolution: Invalid

> we should forbid creating table using 'hbase' namespace except by superuser
> ---
>
> Key: HBASE-14265
> URL: https://issues.apache.org/jira/browse/HBASE-14265
> Project: HBase
>  Issue Type: Bug
>Reporter: Heng Chen
>Assignee: Heng Chen
> Attachments: HBASE-14265.patch, HBASE-14265_v2.patch, 
> HBASE-14265_v3.patch, HBASE-14265_v4.patch
>
>
> Now, there is no limit for users who can create table under 'hbase' 
> NameSpace. I think it has some risk.
> Because we use {{TableName.systemTable}} to decide whether this table is 
> System or not.
> But as code,  {{TableName.systemTable}} will be true, if NS equals "hbase'
> {code}
>  if (Bytes.equals(NamespaceDescriptor.SYSTEM_NAMESPACE_NAME, namespace)) {
> this.namespace = NamespaceDescriptor.SYSTEM_NAMESPACE_NAME;
> this.namespaceAsString = 
> NamespaceDescriptor.SYSTEM_NAMESPACE_NAME_STR;
> this.systemTable = true;
>   } 
> {code}
>  
> And we treat system table and normal table differently. 
> For example,  https://issues.apache.org/jira/browse/HBASE-14257 will flush 
> fast if table belong to system table.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-14703) update the per-region stats twice for the call on return

2015-10-26 Thread Heng Chen (JIRA)
Heng Chen created HBASE-14703:
-

 Summary: update the per-region stats twice for the call on return
 Key: HBASE-14703
 URL: https://issues.apache.org/jira/browse/HBASE-14703
 Project: HBase
  Issue Type: Bug
Reporter: Heng Chen


In {{AsyncProcess.SingleServerRequestRunnable}}, it seems we update 
serverStatistics twice.

The first one is that we wrapper {{RetryingCallable}}  by 
{{StatsTrackingRpcRetryingCaller}}, and do serverStatistics update when we call 
{{callWithRetries}} and {{callWithoutRetries}}. Relates code like below:
{code}
  @Override
  public T callWithRetries(RetryingCallable callable, int callTimeout)
  throws IOException, RuntimeException {
T result = delegate.callWithRetries(callable, callTimeout);
return updateStatsAndUnwrap(result, callable);
  }

  @Override
  public T callWithoutRetries(RetryingCallable callable, int callTimeout)
  throws IOException, RuntimeException {
T result = delegate.callWithRetries(callable, callTimeout);
return updateStatsAndUnwrap(result, callable);
  }
{code}

The secondary one is after we get response, in {{receiveMultiAction}}, we do 
update again. 
{code}
// update the stats about the region, if its a user table. We don't want to 
slow down
// updates to meta tables, especially from internal updates (master, etc).
if (AsyncProcess.this.connection.getStatisticsTracker() != null) {
  result = ResultStatsUtil.updateStats(result,
  AsyncProcess.this.connection.getStatisticsTracker(), server, regionName);
}
{code}

It seems that {{StatsTrackingRpcRetryingCaller}} is NOT necessary,  remove it?




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-14684) Try to remove all MiniMapReduceCluster in unit tests

2015-10-23 Thread Heng Chen (JIRA)
Heng Chen created HBASE-14684:
-

 Summary: Try to remove all MiniMapReduceCluster in unit tests
 Key: HBASE-14684
 URL: https://issues.apache.org/jira/browse/HBASE-14684
 Project: HBase
  Issue Type: Improvement
Reporter: Heng Chen


As discussion in dev list,  we will try to do MR job without 
MiniMapReduceCluster.

Testcases will run faster and more reliable.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-14662) Fix NPE in HFileOutputFormat2

2015-10-21 Thread Heng Chen (JIRA)
Heng Chen created HBASE-14662:
-

 Summary: Fix NPE in HFileOutputFormat2
 Key: HBASE-14662
 URL: https://issues.apache.org/jira/browse/HBASE-14662
 Project: HBase
  Issue Type: Bug
Reporter: Heng Chen


When i dig in HBASE-14659,  i run testWritingPEData. There are a lot of NPE 
thrown and testcase run a long time. 

The reason is that, in {{HFileOutputFormat2}}
{code}
HRegionLocation loc = null;
String tableName = conf.get(OUTPUT_TABLE_NAME_CONF_KEY);

try (Connection connection = 
ConnectionFactory.createConnection(conf);
   RegionLocator locator =
 connection.getRegionLocator(TableName.valueOf(tableName))) 
{
  loc = locator.getRegionLocation(rowKey);
} catch (Throwable e) {
  LOG.warn("there's something wrong when locating rowkey: " +
Bytes.toString(rowKey), e);
  loc = null;
}
{code}

Because we did not set {{OUTPUT_TABLE_NAME_CONF_KEY}}, So tableName is null,  
So NPE thrown.

And RegionLocator will create connection which RegionLocator use to find region 
location. Because zk is not start in this testcase, So it will retry many 
times. 

But all this actions is nesseray,  we can skip create connection by check 
whether tableName is null





--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-14654) Reenable TestMultiParallel#testActiveThreadsCount

2015-10-20 Thread Heng Chen (JIRA)
Heng Chen created HBASE-14654:
-

 Summary: Reenable TestMultiParallel#testActiveThreadsCount
 Key: HBASE-14654
 URL: https://issues.apache.org/jira/browse/HBASE-14654
 Project: HBase
  Issue Type: Bug
Reporter: Heng Chen


It was disabled in HBASE-14642,  this issue should reenable it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-14659) Fix flakey TestHFileOutputFormat

2015-10-20 Thread Heng Chen (JIRA)
Heng Chen created HBASE-14659:
-

 Summary: Fix flakey TestHFileOutputFormat
 Key: HBASE-14659
 URL: https://issues.apache.org/jira/browse/HBASE-14659
 Project: HBase
  Issue Type: Bug
Reporter: Heng Chen







--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-14455) Try to get rid of unused HConnection instance

2015-09-21 Thread Heng Chen (JIRA)
Heng Chen created HBASE-14455:
-

 Summary: Try to get rid of unused HConnection instance 
 Key: HBASE-14455
 URL: https://issues.apache.org/jira/browse/HBASE-14455
 Project: HBase
  Issue Type: Improvement
Reporter: Heng Chen
Priority: Minor


After HBASE-14361,  we get rid of HConnection Instance in ReplicationSink in 
standalone mode. But there are still three HConnection Instance, i will try to 
remove unused ones.

The three instance is created 
{code}
6 2015-09-21 16:05:36,401 INFO  [10.0.3.80:60429.activeMasterManager] 
client.ConnectionImplementation:.
7 java.lang.Throwable
8 >---at 
org.apache.hadoop.hbase.client.ConnectionImplementation.(ConnectionImplementation.java:217)
9 >---at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native 
Method)
   10 >---at 
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
   11 >---at 
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
   12 >---at java.lang.reflect.Constructor.newInstance(Constructor.java:422)
   13 >---at 
org.apache.hadoop.hbase.client.ConnectionFactory.createConnection(ConnectionFactory.java:231)
   14 >---at 
org.apache.hadoop.hbase.client.ConnectionFactory.createConnection(ConnectionFactory.java:118)
   15 >---at 
org.apache.hadoop.hbase.master.ServerManager.(ServerManager.java:222)
   16 >---at 
org.apache.hadoop.hbase.master.ServerManager.(ServerManager.java:212)
   17 >---at 
org.apache.hadoop.hbase.master.HMaster.createServerManager(HMaster.java:853)
   18 >---at 
org.apache.hadoop.hbase.master.HMaster.finishActiveMasterInitialization(HMaster.java:661)
   19 >---at org.apache.hadoop.hbase.master.HMaster.access$500(HMaster.java:180)
   20 >---at org.apache.hadoop.hbase.master.HMaster$1.run(HMaster.java:1670)
   21 >---at java.lang.Thread.run(Thread.java:745)
   22 2015-09-21 16:05:36,401 INFO  [M:0;10.0.3.80:60429] 
client.ConnectionImplementation:.
   23 java.lang.Throwable
   24 >---at 
org.apache.hadoop.hbase.client.ConnectionImplementation.(ConnectionImplementation.java:217)
   25 >---at 
org.apache.hadoop.hbase.client.ConnectionUtils$1.(ConnectionUtils.java:128)
   26 >---at 
org.apache.hadoop.hbase.client.ConnectionUtils.createShortCircuitConnection(ConnectionUtils.java:128)
   27 >---at 
org.apache.hadoop.hbase.regionserver.HRegionServer.createClusterConnection(HRegionServer.java:686)
   28 >---at 
org.apache.hadoop.hbase.regionserver.HRegionServer.setupClusterConnection(HRegionServer.java:717)
   29 >---at 
org.apache.hadoop.hbase.regionserver.HRegionServer.preRegistrationInitialization(HRegionServer.java:730)
   30 >---at 
org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:885)
   31 >---at 
org.apache.hadoop.hbase.master.HMasterCommandLine$LocalHMaster.run(HMasterCommandLine.java:311)
   32 >---at java.lang.Thread.run(Thread.java:745)
   33 2015-09-21 16:05:36,401 INFO  [RS:0;10.0.3.80:60431] 
client.ConnectionImplementation:.
   34 java.lang.Throwable
   35 >---at 
org.apache.hadoop.hbase.client.ConnectionImplementation.(ConnectionImplementation.java:217)
   36 >---at 
org.apache.hadoop.hbase.client.ConnectionUtils$1.(ConnectionUtils.java:128)
   37 >---at 
org.apache.hadoop.hbase.client.ConnectionUtils.createShortCircuitConnection(ConnectionUtils.java:128)
   38 >---at 
org.apache.hadoop.hbase.regionserver.HRegionServer.createClusterConnection(HRegionServer.java:686)
   39 >---at 
org.apache.hadoop.hbase.regionserver.HRegionServer.setupClusterConnection(HRegionServer.java:717)
   40 >---at 
org.apache.hadoop.hbase.regionserver.HRegionServer.preRegistrationInitialization(HRegionServer.java:730)
   41 >---at 
org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:885)
   42 >---at java.lang.Thread.run(Thread.java:745)
{code}




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-14265) we should forbidden create table using 'hbase' namespace

2015-08-19 Thread Heng Chen (JIRA)
Heng Chen created HBASE-14265:
-

 Summary: we should forbidden create table using 'hbase' namespace
 Key: HBASE-14265
 URL: https://issues.apache.org/jira/browse/HBASE-14265
 Project: HBase
  Issue Type: Bug
Reporter: Heng Chen


Now, there is no limit for users who can create table under 'hbase' NameSpace. 
I think it has some risk.

Because we use {{TableName.systemTable}} to decide whether this table is System 
or not.

But as code,  {{TableName.systemTable}} will be true, if NS equals hbase'
{code}
 if (Bytes.equals(NamespaceDescriptor.SYSTEM_NAMESPACE_NAME, namespace)) {
this.namespace = NamespaceDescriptor.SYSTEM_NAMESPACE_NAME;
this.namespaceAsString = NamespaceDescriptor.SYSTEM_NAMESPACE_NAME_STR;
this.systemTable = true;
  } 
{code}
 
And we treat system table and normal table differently. 
For example,  https://issues.apache.org/jira/browse/HBASE-14257 will flush fast 
if table belong to system table.







--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-14230) replace reflection in FSHlog with HdfsDataOutputStream#getCurrentBlockReplication()

2015-08-17 Thread Heng Chen (JIRA)
Heng Chen created HBASE-14230:
-

 Summary: replace reflection in FSHlog with 
HdfsDataOutputStream#getCurrentBlockReplication()
 Key: HBASE-14230
 URL: https://issues.apache.org/jira/browse/HBASE-14230
 Project: HBase
  Issue Type: Improvement
  Components: wal
Reporter: Heng Chen
Priority: Minor


As comment TODO said, we use 
{{HdfsDataOutputStream#getCurrentBlockReplication}} and 
{{DFSOutputStream.getPipeLine}} to replace reflection in FSHlog



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-14203) remove duplicate code getTableDescriptor in HTable

2015-08-10 Thread Heng Chen (JIRA)
Heng Chen created HBASE-14203:
-

 Summary: remove duplicate code getTableDescriptor in HTable
 Key: HBASE-14203
 URL: https://issues.apache.org/jira/browse/HBASE-14203
 Project: HBase
  Issue Type: Improvement
Reporter: Heng Chen
Priority: Trivial


As TODO in comment said, 
{{HTable.getTableDescriptor}} is same as {{HAdmin.getTableDescriptor}}. remove 
the duplicate code.




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-14189) BlockCache options should consider CF Level BlockCacheEnabled setting

2015-08-06 Thread Heng Chen (JIRA)
Heng Chen created HBASE-14189:
-

 Summary: BlockCache options should consider CF Level 
BlockCacheEnabled setting
 Key: HBASE-14189
 URL: https://issues.apache.org/jira/browse/HBASE-14189
 Project: HBase
  Issue Type: Improvement
  Components: BlockCache
Reporter: Heng Chen


While using BlockCache,  we use {{cacheDataOnRead}}({{cacheDataOnWrite}}) 
represents for whether we should cache block after read(write) block from(to) 
hdfs.  We should honour BC setting and CF Level cache setting while using 
BlockCache.




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-14178) regionserver blocks because of waiting for offsetLock

2015-08-02 Thread Heng Chen (JIRA)
Heng Chen created HBASE-14178:
-

 Summary: regionserver blocks because of waiting for offsetLock
 Key: HBASE-14178
 URL: https://issues.apache.org/jira/browse/HBASE-14178
 Project: HBase
  Issue Type: Bug
Reporter: Heng Chen
Priority: Critical


My regionserver blocks, and all client rpc timeout.

I print the regionserver's jstack,  it seems a lot of thread was blocked for 
waiting offsetLock, detail infomation belows:

{code}
B.DefaultRpcServer.handler=2,queue=2,port=60020 #82 daemon prio=5 os_prio=0 
tid=0x01827000 nid=0x2cdc in Object.wait() [0x7f3831b72000]
   java.lang.Thread.State: WAITING (on object monitor)
at java.lang.Object.wait(Native Method)
at java.lang.Object.wait(Object.java:502)
at org.apache.hadoop.hbase.util.IdLock.getLockEntry(IdLock.java:79)
- locked 0x000773af7c18 (a 
org.apache.hadoop.hbase.util.IdLock$Entry)
at 
org.apache.hadoop.hbase.io.hfile.HFileReaderV2.readBlock(HFileReaderV2.java:352)
at 
org.apache.hadoop.hbase.io.hfile.HFileBlockIndex$BlockIndexReader.loadDataBlockWithScanInfo(HFileBlockIndex.java:253)
at 
org.apache.hadoop.hbase.io.hfile.HFileReaderV2$AbstractScannerV2.seekTo(HFileReaderV2.java:524)
at 
org.apache.hadoop.hbase.io.hfile.HFileReaderV2$AbstractScannerV2.reseekTo(HFileReaderV2.java:572)
at 
org.apache.hadoop.hbase.regionserver.StoreFileScanner.reseekAtOrAfter(StoreFileScanner.java:257)
at 
org.apache.hadoop.hbase.regionserver.StoreFileScanner.reseek(StoreFileScanner.java:173)
at 
org.apache.hadoop.hbase.regionserver.NonLazyKeyValueScanner.doRealSeek(NonLazyKeyValueScanner.java:55)
at 
org.apache.hadoop.hbase.regionserver.KeyValueHeap.generalizedSeek(KeyValueHeap.java:313)
at 
org.apache.hadoop.hbase.regionserver.KeyValueHeap.requestSeek(KeyValueHeap.java:269)
at 
org.apache.hadoop.hbase.regionserver.StoreScanner.reseek(StoreScanner.java:695)
at 
org.apache.hadoop.hbase.regionserver.StoreScanner.seekAsDirection(StoreScanner.java:683)
at 
org.apache.hadoop.hbase.regionserver.StoreScanner.next(StoreScanner.java:533)
at 
org.apache.hadoop.hbase.regionserver.KeyValueHeap.next(KeyValueHeap.java:140)
at 
org.apache.hadoop.hbase.regionserver.HRegion$RegionScannerImpl.populateResult(HRegion.java:3889)
at 
org.apache.hadoop.hbase.regionserver.HRegion$RegionScannerImpl.nextInternal(HRegion.java:3969)
at 
org.apache.hadoop.hbase.regionserver.HRegion$RegionScannerImpl.nextRaw(HRegion.java:3847)
at 
org.apache.hadoop.hbase.regionserver.HRegion$RegionScannerImpl.next(HRegion.java:3820)
- locked 0x0005e5c55ad0 (a 
org.apache.hadoop.hbase.regionserver.HRegion$RegionScannerImpl)
at 
org.apache.hadoop.hbase.regionserver.HRegion$RegionScannerImpl.next(HRegion.java:3807)
at org.apache.hadoop.hbase.regionserver.HRegion.get(HRegion.java:4779)
at org.apache.hadoop.hbase.regionserver.HRegion.get(HRegion.java:4753)
at 
org.apache.hadoop.hbase.regionserver.HRegionServer.get(HRegionServer.java:2916)
at 
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:29583)
at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2027)
at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:108)
at 
org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:114)
at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:94)
at java.lang.Thread.run(Thread.java:745)

   Locked ownable synchronizers:
- 0x0005e5c55c08 (a 
java.util.concurrent.locks.ReentrantLock$NonfairSync)
{code}





--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HBASE-14113) Compile failed on master

2015-07-16 Thread Heng Chen (JIRA)
Heng Chen created HBASE-14113:
-

 Summary: Compile failed on master
 Key: HBASE-14113
 URL: https://issues.apache.org/jira/browse/HBASE-14113
 Project: HBase
  Issue Type: Bug
  Components: build
Reporter: Heng Chen


mvn package -DskipTests

Compile failed!  

{code:bash}
[ERROR] Failed to execute goal 
org.apache.maven.plugins:maven-compiler-plugin:3.2:compile (default-compile) on 
project hbase-client: Compilation failure: Compilation failure:
[ERROR] 
/Users/chenheng/apache/hbase/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BinaryComparator.java:[54,3]
 method does not override or implement a method from a supertype
[ERROR] 
/Users/chenheng/apache/hbase/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/NullComparator.java:[64,3]
 method does not override or implement a method from a supertype
[ERROR] 
/Users/chenheng/apache/hbase/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/LongComparator.java:[51,5]
 method does not override or implement a method from a supertype
[ERROR] 
/Users/chenheng/apache/hbase/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BitComparator.java:[137,3]
 method does not override or implement a method from a supertype
[ERROR] 
/Users/chenheng/apache/hbase/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/BinaryPrefixComparator.java:[56,3]
 method does not override or implement a method from a supertype
{code}

My java version:
{code}
java version 1.8.0_40
Java(TM) SE Runtime Environment (build 1.8.0_40-b25)
Java HotSpot(TM) 64-Bit Server VM (build 25.40-b25, mixed mode)
{code}






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)