[jira] [Commented] (HBASE-8519) Backup master will never come up if primary master dies during initialization

2013-05-10 Thread Jerry He (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8519?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655179#comment-13655179
 ] 

Jerry He commented on HBASE-8519:
-

The missing part seems to be this.
The backup master only check the clusterStateZNode once at the end when it 
wakes up and tries to be master again.
It is not notified if the clusterStateZNode has ever been up or not before that.

The real 'stop' case is: clusterStateZNode up --> clusterStateZNode down

The backup master needs to watch for clusterStateZNode. If it is not aware 
clusterStateZNode has ever been up 
or was up in its whole life, it should treat it as not a real 'stop' case at 
the end.

The if condition would become:

{code}
  ..
  if (clusterHasBeenUp && !clusterStatusTracker.isClusterUp()) {
  this.master.stop(
"Cluster went down before this master became active");
}
  ..
{code}

Comments, please.

> Backup master will never come up if primary master dies during initialization
> -
>
> Key: HBASE-8519
> URL: https://issues.apache.org/jira/browse/HBASE-8519
> Project: HBase
>  Issue Type: Bug
>  Components: master
>Affects Versions: 0.94.7, 0.95.0
>Reporter: Jerry He
>Assignee: Jerry He
>Priority: Minor
> Fix For: 0.98.0
>
>
> The problem happens if primary master dies after becoming master but before 
> it completes initialization and calls clusterStatusTracker.setClusterUp(),
> The backup master will try to become the master, but will shutdown itself 
> promptly because it sees 'the cluster is not up'.
> This is the backup master log:
> 2013-05-09 15:08:05,568 INFO 
> org.apache.hadoop.hbase.master.metrics.MasterMetrics: Initialized
> 2013-05-09 15:08:05,573 DEBUG org.apache.hadoop.hbase.master.HMaster: HMaster 
> started in backup mode.  Stalling until master znode is written.
> 2013-05-09 15:08:05,589 INFO 
> org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper: Node /hbase/master 
> already exists and this is not a retry
> 2013-05-09 15:08:05,590 INFO 
> org.apache.hadoop.hbase.master.ActiveMasterManager: Adding ZNode for 
> /hbase/backup-masters/xxx.com,6,1368137285373 in backup master directory
> 2013-05-09 15:08:05,595 INFO 
> org.apache.hadoop.hbase.master.ActiveMasterManager: Another master is the 
> active master, xxx.com,6,1368137283107; waiting to become the next active 
> master
> 2013-05-09 15:09:45,006 DEBUG 
> org.apache.hadoop.hbase.master.ActiveMasterManager: No master available. 
> Notifying waiting threads
> 2013-05-09 15:09:45,006 INFO org.apache.hadoop.hbase.master.HMaster: Cluster 
> went down before this master became active
> 2013-05-09 15:09:45,006 DEBUG org.apache.hadoop.hbase.master.HMaster: 
> Stopping service threads
> 2013-05-09 15:09:45,006 INFO org.apache.hadoop.ipc.HBaseServer: Stopping 
> server on 6
>  
> In ActiveMasterManager::blockUntilBecomingActiveMaster()
> {code}
>   ..
>   if (!clusterStatusTracker.isClusterUp()) {
>   this.master.stop(
> "Cluster went down before this master became active");
> }
>   ..
> {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-8453) TestImportExport failing again due to configuration issues

2013-05-10 Thread Andrew Purtell (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8453?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655178#comment-13655178
 ] 

Andrew Purtell commented on HBASE-8453:
---

I have found that getting a JobConf from the MiniMRCluster and then using 
HBaseConfiguration.merge to merge in HBase minicluster details will fix all 
remaining Hadoop 2 test failures involving mapreduce. It's just a matter of 
finding all of the places Job is instantaned and doing so. Unfortunately the 
result then fails for Hadoop 1 but only in EC2 environments. Might get back to 
that this week. 

> TestImportExport failing again due to configuration issues
> --
>
> Key: HBASE-8453
> URL: https://issues.apache.org/jira/browse/HBASE-8453
> Project: HBase
>  Issue Type: Sub-task
>  Components: mapreduce, test
>Affects Versions: 0.98.0, 0.94.8, 0.95.1
>Reporter: Andrew Purtell
> Fix For: 0.98.0, 0.94.8, 0.95.1
>
> Attachments: 8453.patch, 8453-v1-0.94.patch
>
>
> TestImportExport fails for me with a connection refused exception:
> {noformat}
> java.lang.reflect.UndeclaredThrowableException
>   at 
> org.apache.hadoop.yarn.exceptions.impl.pb.YarnRemoteExceptionPBImpl.unwrapAndThrowException(YarnRemoteExceptionPBImpl.java:135)
>   at 
> org.apache.hadoop.yarn.api.impl.pb.client.ClientRMProtocolPBClientImpl.getNewApplication(ClientRMProtocolPBClientImpl.java:162)
>   at 
> org.apache.hadoop.yarn.client.YarnClientImpl.getNewApplication(YarnClientImpl.java:121)
>   at 
> org.apache.hadoop.mapred.ResourceMgrDelegate.getNewJobID(ResourceMgrDelegate.java:107)
>   at org.apache.hadoop.mapred.YARNRunner.getNewJobID(YARNRunner.java:231)
>   at 
> org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:352)
> [...]
> Caused by: com.google.protobuf.ServiceException: java.net.ConnectException: 
> Call From ip-10-174-75-236/10.174.75.236 to 0.0.0.0:8032 failed on connection 
> exception: java.net.ConnectException: Connection refused; For more details 
> see:  http://wiki.apache.org/hadoop/ConnectionRefused
>   at 
> org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:212)
>   at com.sun.proxy.$Proxy89.getNewApplication(Unknown Source)
>   at 
> org.apache.hadoop.yarn.api.impl.pb.client.ClientRMProtocolPBClientImpl.getNewApplication(ClientRMProtocolPBClientImpl.java:159)
>   ... 42 more
> {noformat}
> Settings in the MiniMRCluster configuration are not properly propagated in 
> this test.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-8497) Protobuf WAL also needs a trailer

2013-05-10 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8497?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655176#comment-13655176
 ] 

Hadoop QA commented on HBASE-8497:
--

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12582770/HBASE-8497-v2.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 9 new 
or modified tests.

{color:green}+1 hadoop1.0{color}.  The patch compiles against the hadoop 
1.0 profile.

{color:green}+1 hadoop2.0{color}.  The patch compiles against the hadoop 
2.0 profile.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 lineLengths{color}.  The patch introduces lines longer than 
100

  {color:green}+1 site{color}.  The mvn site goal succeeds with this patch.

{color:green}+1 core tests{color}.  The patch passed unit tests in .

Test results: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5639//testReport/
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5639//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-protocol.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5639//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-client.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5639//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-examples.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5639//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-hadoop1-compat.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5639//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-prefix-tree.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5639//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-common.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5639//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-server.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5639//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-hadoop-compat.html
Console output: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5639//console

This message is automatically generated.

> Protobuf WAL also needs a trailer 
> --
>
> Key: HBASE-8497
> URL: https://issues.apache.org/jira/browse/HBASE-8497
> Project: HBase
>  Issue Type: Sub-task
>  Components: Protobufs, wal
>Affects Versions: 0.95.1
>Reporter: Enis Soztutar
>Assignee: Himanshu Vashishtha
> Fix For: 0.98.0, 0.95.1
>
> Attachments: HBASE-8497-v0.patch, HBASE-8497-v2.patch
>
>
> New Protobuf WAL has a header, but we will probably need a trailer as well, 
> reserved for later usage. 
> Right now, we can we just serialize an empty trailer, but putting more 
> metadata there, like range of sequence_id's, region names, table names etc 
> might be needed in the future. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HBASE-8531) TestZooKeeper fails in trunk/0.95 builds

2013-05-10 Thread stack (JIRA)
stack created HBASE-8531:


 Summary: TestZooKeeper fails in trunk/0.95 builds
 Key: HBASE-8531
 URL: https://issues.apache.org/jira/browse/HBASE-8531
 Project: HBase
  Issue Type: Bug
  Components: Zookeeper
Reporter: stack
Assignee: stack
 Fix For: 0.95.1


TestZooKeeper fails on occasion.  I caught a good example recently.  See below 
failure stack trace.

It took me a while.  I thought the issue had to do w/ our recent ipc 
refactorings but it looks like a problem we have always had.  In short, 
MetaScanner is not handling DoNotRetryIOEs -- it is letting them out.  DNRIOEs 
when scanning are supposed to force a reset of the scan.  HTable#next catches 
these and does the necessary scanner reset up.  MetaScanner is running some 
subset of what HTable does when it is scanning except the part where it catches 
a DNRIOE and redoes the scan.  Odd.

TestZooKeeper failed in this instance because the test kills a regionserver at 
same time as we are trying to create a table.  In create table we do a meta 
scan using MetaScanner passing a Visitor.  The scan starts and gets a 
RegionServerStoppedException (This is NOT a DNRIOE -- it should be -- but later 
we convert it into one up in ScannerCallable).

DNRIOEs are thrown to the upper layers to handle

Let me look into having MetaScanner just use HTable scanning.  It makes an 
instance just to find where to start the scan... let me try using this instance 
for actually scanning.

TODO: Do this convertion everywhere a DNRIOE could come out.

Here is the stack trace

{code}

org.apache.hadoop.hbase.exceptions.DoNotRetryIOException: Reset scanner
at 
org.apache.hadoop.hbase.client.ScannerCallable.call(ScannerCallable.java:209)
at 
org.apache.hadoop.hbase.client.ScannerCallable.call(ScannerCallable.java:52)
at 
org.apache.hadoop.hbase.client.ServerCallable.withRetries(ServerCallable.java:170)
at 
org.apache.hadoop.hbase.client.MetaScanner.metaScan(MetaScanner.java:212)
at 
org.apache.hadoop.hbase.client.MetaScanner.access$000(MetaScanner.java:52)
at 
org.apache.hadoop.hbase.client.MetaScanner$1.connect(MetaScanner.java:131)
at 
org.apache.hadoop.hbase.client.MetaScanner$1.connect(MetaScanner.java:128)
at 
org.apache.hadoop.hbase.client.HConnectionManager.execute(HConnectionManager.java:398)
at 
org.apache.hadoop.hbase.client.MetaScanner.metaScan(MetaScanner.java:128)
at 
org.apache.hadoop.hbase.client.MetaScanner.metaScan(MetaScanner.java:103)
at 
org.apache.hadoop.hbase.client.MetaScanner.metaScan(MetaScanner.java:81)
at 
org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:448)
at 
org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:348)
at 
org.apache.hadoop.hbase.TestZooKeeper.testSanity(TestZooKeeper.java:242)
at 
org.apache.hadoop.hbase.TestZooKeeper.testRegionServerSessionExpired(TestZooKeeper.java:203)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
at 
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at 
org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:271)
at 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:70)
at 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
at 
org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at 
org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
at org.junit.runners.Suite.runChild(Suite.java:127)
at org.junit.runners.Suite.runChild(Suite.java:26)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
at 
java.util.concurrent.Executor

[jira] [Updated] (HBASE-8497) Protobuf WAL also needs a trailer

2013-05-10 Thread Himanshu Vashishtha (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-8497?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Himanshu Vashishtha updated HBASE-8497:
---

Attachment: HBASE-8497-v2.patch

Incorporated feedback on the initial version. 
Apart from the nits that are fixed:
a) Modified exception handling as suggested.
b) Added sanity checks on the trailer size. Currently, this patch makes 
trailer_max_size to 1MB, and didn't make it configurable; This is to avoid 
cases where a user lowers the max size and a reader tries to read old logs. 
Moreover, I see 1MB size a good enough size for the use case (6774). If there 
is any requirement to increase it, it could be done in some other jira. Please 
suggest if you think differently.
c) Trailer is instantiated in the init() method. While writing the trailer, a 
check is made in case trailer is set to null or its size exceeds max_size. In 
this case, a default trailer is written rather with a WARN message.
d) Added test case for  corrupt trailer.

Test: I tested it on my local by killing the RS and doing the split, etc. Ran 
unit tests for classes HLog* many times.


> Protobuf WAL also needs a trailer 
> --
>
> Key: HBASE-8497
> URL: https://issues.apache.org/jira/browse/HBASE-8497
> Project: HBase
>  Issue Type: Sub-task
>  Components: Protobufs, wal
>Affects Versions: 0.95.1
>Reporter: Enis Soztutar
>Assignee: Himanshu Vashishtha
> Fix For: 0.98.0, 0.95.1
>
> Attachments: HBASE-8497-v0.patch, HBASE-8497-v2.patch
>
>
> New Protobuf WAL has a header, but we will probably need a trailer as well, 
> reserved for later usage. 
> Right now, we can we just serialize an empty trailer, but putting more 
> metadata there, like range of sequence_id's, region names, table names etc 
> might be needed in the future. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-8530) Refine error message from ExportSnapshot when there is leftover snapshot in target cluster

2013-05-10 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8530?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655147#comment-13655147
 ] 

Hadoop QA commented on HBASE-8530:
--

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12582763/8530-v1.txt
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:red}-1 tests included{color}.  The patch doesn't appear to include 
any new or modified tests.
Please justify why no new tests are needed for this 
patch.
Also please list what manual steps were performed to 
verify this patch.

{color:green}+1 hadoop1.0{color}.  The patch compiles against the hadoop 
1.0 profile.

{color:green}+1 hadoop2.0{color}.  The patch compiles against the hadoop 
2.0 profile.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 lineLengths{color}.  The patch does not introduce lines 
longer than 100

  {color:green}+1 site{color}.  The mvn site goal succeeds with this patch.

 {color:red}-1 core tests{color}.  The patch failed these unit tests:
   org.apache.hadoop.hbase.client.TestHCM

 {color:red}-1 core zombie tests{color}.  There are 1 zombie test(s): 

Test results: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5638//testReport/
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5638//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-protocol.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5638//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-client.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5638//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-examples.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5638//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-hadoop1-compat.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5638//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-prefix-tree.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5638//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-common.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5638//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-server.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5638//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-hadoop-compat.html
Console output: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5638//console

This message is automatically generated.

> Refine error message from ExportSnapshot when there is leftover snapshot in 
> target cluster
> --
>
> Key: HBASE-8530
> URL: https://issues.apache.org/jira/browse/HBASE-8530
> Project: HBase
>  Issue Type: Bug
>Reporter: Ted Yu
>Assignee: Ted Yu
>Priority: Minor
> Fix For: 0.98.0
>
> Attachments: 8530-v1.txt
>
>
> ExportSnapshot performs the following validation check before starting export:
> {code}
> // Check if the snapshot already in-progress
> if (outputFs.exists(snapshotTmpDir)) {
>   System.err.println("A snapshot with the same name '" + snapshotName + 
> "' is in-progress");
>   return 1;
> }
> {code}
> The intention was to prevent concurrent snapshot export when snapshot with 
> the same name may be outstanding.
> Here is sample output:
> {code}
> hbase@ip-10-165-36-226 root$ hbase 
> org.apache.hadoop.hbase.snapshot.ExportSnapshot -snapshot snapone -copy-to 
> hdfs://10.12.113.22:8020/apps/hbase/data -mappers 1
> A snapshot with the same name 'snapone' is in-progress
> {code}
> We should note the actual path in the error message so that user can check 
> whether the snapshot with same name is still pending.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-8450) Update hbase-default.xml and general recommendations to better suit current hw, h2, experience, etc.

2013-05-10 Thread binlijin (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8450?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655148#comment-13655148
 ] 

binlijin commented on HBASE-8450:
-

HLog
 - hbase.regionserver.hlog.blocksize and dfs.block.size = 128M or 256M
 - io.bytes.per.checksum = 4096


> Update hbase-default.xml and general recommendations to better suit current 
> hw, h2, experience, etc.
> 
>
> Key: HBASE-8450
> URL: https://issues.apache.org/jira/browse/HBASE-8450
> Project: HBase
>  Issue Type: Task
>  Components: Usability
>Reporter: stack
>Priority: Critical
> Fix For: 0.95.1
>
> Attachments: 8450.txt
>
>
> This is a critical task we need to do before we release; review our defaults.
> On cursory review, there are configs in hbase-default.xml that no longer have 
> matching code; there are some that should be changed because we know better 
> now and others that should be amended because hardware and deploys are bigger 
> than they used to be.
> We could also move stuff around so that the must-edit stuff is near the top 
> (zk quorum config. is mid-way down the page) and beef up the descriptions -- 
> especially since these descriptions shine through in the refguide.
> Lastly, I notice that our tgz does not "include" an hbase-default.xml other 
> than the one bundled up in the jar.  Maybe we should make it more accessible.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-5930) Limits the amount of time an edit can live in the memstore.

2013-05-10 Thread Lars Hofhansl (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-5930?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655139#comment-13655139
 ] 

Lars Hofhansl commented on HBASE-5930:
--

Added release notes. I also would like add an option to disable this feature 
(for example by setting base.regionserver.optionalcacheflushinterval to a value 
<= 0).


> Limits the amount of time an edit can live in the memstore.
> ---
>
> Key: HBASE-5930
> URL: https://issues.apache.org/jira/browse/HBASE-5930
> Project: HBase
>  Issue Type: Improvement
>Reporter: Lars Hofhansl
>Assignee: Devaraj Das
> Fix For: 0.98.0, 0.94.8, 0.95.1
>
> Attachments: 5930-0.94-added-addendum.txt, 5930-0.94.txt, 
> 5930-1.patch, 5930-2.1.patch, 5930-2.2.patch, 5930-2.3.patch, 5930-2.4.patch, 
> 5930-track-oldest-sample.txt, 5930-wip.patch, HBASE-5930-ADD-0.patch, 
> hbase-5930-addendum2.patch, hbase-5930-test-execution.log
>
>
> A colleague of mine ran into an interesting issue.
> He inserted some data with the WAL disabled, which happened to fit in the 
> aggregate Memstores memory.
> Two weeks later he a had problem with the HDFS cluster, which caused the 
> region servers to abort. He found that his data was lost. Looking at the log 
> we found that the Memstores were not flushed at all during these two weeks.
> Should we have an option to flush memstores periodically. There are obvious 
> downsides to this, like many small storefiles, etc.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HBASE-5930) Limits the amount of time an edit can live in the memstore.

2013-05-10 Thread Lars Hofhansl (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-5930?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Hofhansl updated HBASE-5930:
-

Release Note: 
This feature limits the time that unflushed data will stay in the memstore.
By default a memstore will flush if data older than 1h (360ms) is present.

This can be configured via hbase.regionserver.optionalcacheflushinterval 
(default value is 360).


> Limits the amount of time an edit can live in the memstore.
> ---
>
> Key: HBASE-5930
> URL: https://issues.apache.org/jira/browse/HBASE-5930
> Project: HBase
>  Issue Type: Improvement
>Reporter: Lars Hofhansl
>Assignee: Devaraj Das
> Fix For: 0.98.0, 0.94.8, 0.95.1
>
> Attachments: 5930-0.94-added-addendum.txt, 5930-0.94.txt, 
> 5930-1.patch, 5930-2.1.patch, 5930-2.2.patch, 5930-2.3.patch, 5930-2.4.patch, 
> 5930-track-oldest-sample.txt, 5930-wip.patch, HBASE-5930-ADD-0.patch, 
> hbase-5930-addendum2.patch, hbase-5930-test-execution.log
>
>
> A colleague of mine ran into an interesting issue.
> He inserted some data with the WAL disabled, which happened to fit in the 
> aggregate Memstores memory.
> Two weeks later he a had problem with the HDFS cluster, which caused the 
> region servers to abort. He found that his data was lost. Looking at the log 
> we found that the Memstores were not flushed at all during these two weeks.
> Should we have an option to flush memstores periodically. There are obvious 
> downsides to this, like many small storefiles, etc.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-6134) Improvement for split-worker to speed up distributed log splitting

2013-05-10 Thread Lars Hofhansl (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-6134?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655137#comment-13655137
 ] 

Lars Hofhansl commented on HBASE-6134:
--

Patch probably went stale.
We never got closure in the 0.94 thing. [~enis], [~sershe], you want this in 
0.94? What's your confidence level in this patch?

> Improvement for split-worker to speed up distributed log splitting
> --
>
> Key: HBASE-6134
> URL: https://issues.apache.org/jira/browse/HBASE-6134
> Project: HBase
>  Issue Type: Improvement
>  Components: wal
>Reporter: chunhui shen
>Assignee: chunhui shen
>Priority: Critical
> Fix For: 0.95.0
>
> Attachments: 6134v4.patch, HBASE-6134.patch, HBASE-6134v2.patch, 
> HBASE-6134v3-92.patch, HBASE-6134v3.patch, HBASE-6134v4-94.patch, 
> HBASE-6134v4-94v2.patch, HBASE-6134v4.patch
>
>
> First,we do the test between local-master-splitting and 
> distributed-log-splitting
> Environment:34 hlog files, 5 regionservers,(after kill one, only 4 rs do ths 
> splitting work), 400 regions in one hlog file
> local-master-split:60s+
> distributed-log-splitting:165s+
> In fact, in our production environment, distributed-log-splitting also took 
> 60s with 30 regionservers for 34 hlog files (regionserver may be in high load)
> We found split-worker split one log file took about 20s
> (30ms~50ms per writer.close(); 10ms per create writers )
> I think we could do the improvement for this:
> Parallelizing the create and close writers in threads
> In the patch, change the logic for  distributed-log-splitting same as the 
> local-master-splitting and parallelizing the close in threads.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-8462) Custom timestamps should not be allowed to be negative

2013-05-10 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8462?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655132#comment-13655132
 ] 

Hadoop QA commented on HBASE-8462:
--

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12582762/hbase-8462_v2.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 6 new 
or modified tests.

{color:green}+1 hadoop1.0{color}.  The patch compiles against the hadoop 
1.0 profile.

{color:green}+1 hadoop2.0{color}.  The patch compiles against the hadoop 
2.0 profile.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 lineLengths{color}.  The patch does not introduce lines 
longer than 100

  {color:green}+1 site{color}.  The mvn site goal succeeds with this patch.

 {color:red}-1 core tests{color}.  The patch failed these unit tests:
   org.apache.hadoop.hbase.client.TestHCM

Test results: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5637//testReport/
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5637//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-client.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5637//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-examples.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5637//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-protocol.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5637//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-hadoop1-compat.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5637//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-prefix-tree.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5637//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-common.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5637//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-hadoop-compat.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5637//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-server.html
Console output: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5637//console

This message is automatically generated.

> Custom timestamps should not be allowed to be negative
> --
>
> Key: HBASE-8462
> URL: https://issues.apache.org/jira/browse/HBASE-8462
> Project: HBase
>  Issue Type: Bug
>  Components: Client
>Reporter: Enis Soztutar
>Assignee: Enis Soztutar
> Fix For: 0.98.0, 0.94.8, 0.95.1
>
> Attachments: hbase-8462_v1.patch, hbase-8462_v2.patch
>
>
> Client supplied timestamps should not be allowed to be negative, otherwise 
> unpredictable results will follow. Especially, since we are encoding the ts 
> using Bytes.Bytes(long), negative timestamps are sorted after positive ones. 
> Plus, the new PB messages define ts' as uint64. 
> Credit goes to Huned Lokhandwala for reporting this.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-8462) Custom timestamps should not be allowed to be negative

2013-05-10 Thread Ted Yu (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8462?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655130#comment-13655130
 ] 

Ted Yu commented on HBASE-8462:
---

{code}
+if (timestamp < 0) {
+  throw new IllegalArgumentException("Timestamp cannot be negative");
{code}
Giving users more clue would help them pinpoint the problem. Please include 
timestamp in the message.
{code}
+if (minStamp < 0 || maxStamp < 0) {
+  throw new IllegalArgumentException("Timestamp cannot be negative");
+}
{code}
Please be specific which timestamp is negative.

Please fill out Release Notes for this change.

Good job.

> Custom timestamps should not be allowed to be negative
> --
>
> Key: HBASE-8462
> URL: https://issues.apache.org/jira/browse/HBASE-8462
> Project: HBase
>  Issue Type: Bug
>  Components: Client
>Reporter: Enis Soztutar
>Assignee: Enis Soztutar
> Fix For: 0.98.0, 0.94.8, 0.95.1
>
> Attachments: hbase-8462_v1.patch, hbase-8462_v2.patch
>
>
> Client supplied timestamps should not be allowed to be negative, otherwise 
> unpredictable results will follow. Especially, since we are encoding the ts 
> using Bytes.Bytes(long), negative timestamps are sorted after positive ones. 
> Plus, the new PB messages define ts' as uint64. 
> Credit goes to Huned Lokhandwala for reporting this.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-8375) Streamline Table durability settings

2013-05-10 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8375?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655129#comment-13655129
 ] 

Hadoop QA commented on HBASE-8375:
--

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12582760/hbase-8375_v2.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 6 new 
or modified tests.

{color:green}+1 hadoop1.0{color}.  The patch compiles against the hadoop 
1.0 profile.

{color:green}+1 hadoop2.0{color}.  The patch compiles against the hadoop 
2.0 profile.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 lineLengths{color}.  The patch introduces lines longer than 
100

  {color:green}+1 site{color}.  The mvn site goal succeeds with this patch.

 {color:red}-1 core tests{color}.  The patch failed these unit tests:
   org.apache.hadoop.hbase.TestFullLogReconstruction

 {color:red}-1 core zombie tests{color}.  There are 1 zombie test(s):   
at 
org.apache.hadoop.hbase.client.TestHCM.testDeleteForZKConnLeak(TestHCM.java:855)

Test results: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5636//testReport/
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5636//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-prefix-tree.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5636//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-client.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5636//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-common.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5636//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-protocol.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5636//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-server.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5636//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-hadoop1-compat.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5636//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-examples.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5636//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-hadoop-compat.html
Console output: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5636//console

This message is automatically generated.

> Streamline Table durability settings
> 
>
> Key: HBASE-8375
> URL: https://issues.apache.org/jira/browse/HBASE-8375
> Project: HBase
>  Issue Type: Sub-task
>Reporter: Lars Hofhansl
>Assignee: Enis Soztutar
> Fix For: 0.98.0, 0.95.1
>
> Attachments: hbase-8375_v1.patch, hbase-8375_v2.patch
>
>
> HBASE-7801 introduces the notion of per mutation fine grained durability 
> settings.
> This issue is to consider and the discuss the same for the per table settings 
> (i.e. what would be used if the mutation indicates USE_DEFAULT). I propose 
> the following setting per table:
> * SKIP_WAL (i.e. an unlogged table)
> * ASYNC_WAL (the current deferred log flush)
> * SYNC_WAL (the current default)
> * FSYNC_WAL (for future uses of HDFS' hsync())

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HBASE-8530) Refine error message from ExportSnapshot when there is leftover snapshot in target cluster

2013-05-10 Thread Ted Yu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-8530?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu updated HBASE-8530:
--

Fix Version/s: 0.98.0
   Status: Patch Available  (was: Open)

> Refine error message from ExportSnapshot when there is leftover snapshot in 
> target cluster
> --
>
> Key: HBASE-8530
> URL: https://issues.apache.org/jira/browse/HBASE-8530
> Project: HBase
>  Issue Type: Bug
>Reporter: Ted Yu
>Assignee: Ted Yu
>Priority: Minor
> Fix For: 0.98.0
>
> Attachments: 8530-v1.txt
>
>
> ExportSnapshot performs the following validation check before starting export:
> {code}
> // Check if the snapshot already in-progress
> if (outputFs.exists(snapshotTmpDir)) {
>   System.err.println("A snapshot with the same name '" + snapshotName + 
> "' is in-progress");
>   return 1;
> }
> {code}
> The intention was to prevent concurrent snapshot export when snapshot with 
> the same name may be outstanding.
> Here is sample output:
> {code}
> hbase@ip-10-165-36-226 root$ hbase 
> org.apache.hadoop.hbase.snapshot.ExportSnapshot -snapshot snapone -copy-to 
> hdfs://10.12.113.22:8020/apps/hbase/data -mappers 1
> A snapshot with the same name 'snapone' is in-progress
> {code}
> We should note the actual path in the error message so that user can check 
> whether the snapshot with same name is still pending.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HBASE-8530) Refine error message from ExportSnapshot when there is leftover snapshot in target cluster

2013-05-10 Thread Ted Yu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-8530?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu updated HBASE-8530:
--

Attachment: 8530-v1.txt

> Refine error message from ExportSnapshot when there is leftover snapshot in 
> target cluster
> --
>
> Key: HBASE-8530
> URL: https://issues.apache.org/jira/browse/HBASE-8530
> Project: HBase
>  Issue Type: Bug
>Reporter: Ted Yu
>Assignee: Ted Yu
>Priority: Minor
> Attachments: 8530-v1.txt
>
>
> ExportSnapshot performs the following validation check before starting export:
> {code}
> // Check if the snapshot already in-progress
> if (outputFs.exists(snapshotTmpDir)) {
>   System.err.println("A snapshot with the same name '" + snapshotName + 
> "' is in-progress");
>   return 1;
> }
> {code}
> The intention was to prevent concurrent snapshot export when snapshot with 
> the same name may be outstanding.
> Here is sample output:
> {code}
> hbase@ip-10-165-36-226 root$ hbase 
> org.apache.hadoop.hbase.snapshot.ExportSnapshot -snapshot snapone -copy-to 
> hdfs://10.12.113.22:8020/apps/hbase/data -mappers 1
> A snapshot with the same name 'snapone' is in-progress
> {code}
> We should note the actual path in the error message so that user can check 
> whether the snapshot with same name is still pending.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HBASE-8530) Refine error message from ExportSnapshot when there is leftover snapshot in target cluster

2013-05-10 Thread Ted Yu (JIRA)
Ted Yu created HBASE-8530:
-

 Summary: Refine error message from ExportSnapshot when there is 
leftover snapshot in target cluster
 Key: HBASE-8530
 URL: https://issues.apache.org/jira/browse/HBASE-8530
 Project: HBase
  Issue Type: Bug
Reporter: Ted Yu
Assignee: Ted Yu
Priority: Minor


ExportSnapshot performs the following validation check before starting export:
{code}
// Check if the snapshot already in-progress
if (outputFs.exists(snapshotTmpDir)) {
  System.err.println("A snapshot with the same name '" + snapshotName + "' 
is in-progress");
  return 1;
}
{code}
The intention was to prevent concurrent snapshot export when snapshot with the 
same name may be outstanding.
Here is sample output:
{code}
hbase@ip-10-165-36-226 root$ hbase 
org.apache.hadoop.hbase.snapshot.ExportSnapshot -snapshot snapone -copy-to 
hdfs://10.12.113.22:8020/apps/hbase/data -mappers 1
A snapshot with the same name 'snapone' is in-progress
{code}
We should note the actual path in the error message so that user can check 
whether the snapshot with same name is still pending.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-8519) Backup master will never come up if primary master dies during initialization

2013-05-10 Thread Jerry He (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8519?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655117#comment-13655117
 ] 

Jerry He commented on HBASE-8519:
-

bq. I meant the reverse for the shutdown znode.

Yes, that all make sense now.

I am able to re-create the problem easily by killing the master right after 
issuing the stop-hbase. Then re-start the master before the ephemeral master 
znode goes away.

2013-05-10 18:20:17,690 INFO 
org.apache.hadoop.hbase.master.metrics.MasterMetrics: Initialized
2013-05-10 18:20:17,704 INFO 
org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper: Node /hbase/master 
already exists and this is not a retry
2013-05-10 18:20:17,705 INFO 
org.apache.hadoop.hbase.master.ActiveMasterManager: Adding ZNode for 
/hbase/backup-masters/xxx.com,6,1368235217501 in backup master directory
2013-05-10 18:20:17,711 INFO 
org.apache.hadoop.hbase.master.ActiveMasterManager: Current master has this 
master's address, xxx.com,6,1368234912934; master was restarted? Deleting 
node.
2013-05-10 18:20:17,711 DEBUG 
org.apache.hadoop.hbase.master.ActiveMasterManager: No master available. 
Notifying waiting threads
2013-05-10 18:20:17,711 INFO org.apache.hadoop.hbase.master.HMaster: Cluster 
went down before this master became active
2013-05-10 18:20:17,711 DEBUG org.apache.hadoop.hbase.master.HMaster: Stopping 
service threads


> Backup master will never come up if primary master dies during initialization
> -
>
> Key: HBASE-8519
> URL: https://issues.apache.org/jira/browse/HBASE-8519
> Project: HBase
>  Issue Type: Bug
>  Components: master
>Affects Versions: 0.94.7, 0.95.0
>Reporter: Jerry He
>Assignee: Jerry He
>Priority: Minor
> Fix For: 0.98.0
>
>
> The problem happens if primary master dies after becoming master but before 
> it completes initialization and calls clusterStatusTracker.setClusterUp(),
> The backup master will try to become the master, but will shutdown itself 
> promptly because it sees 'the cluster is not up'.
> This is the backup master log:
> 2013-05-09 15:08:05,568 INFO 
> org.apache.hadoop.hbase.master.metrics.MasterMetrics: Initialized
> 2013-05-09 15:08:05,573 DEBUG org.apache.hadoop.hbase.master.HMaster: HMaster 
> started in backup mode.  Stalling until master znode is written.
> 2013-05-09 15:08:05,589 INFO 
> org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper: Node /hbase/master 
> already exists and this is not a retry
> 2013-05-09 15:08:05,590 INFO 
> org.apache.hadoop.hbase.master.ActiveMasterManager: Adding ZNode for 
> /hbase/backup-masters/xxx.com,6,1368137285373 in backup master directory
> 2013-05-09 15:08:05,595 INFO 
> org.apache.hadoop.hbase.master.ActiveMasterManager: Another master is the 
> active master, xxx.com,6,1368137283107; waiting to become the next active 
> master
> 2013-05-09 15:09:45,006 DEBUG 
> org.apache.hadoop.hbase.master.ActiveMasterManager: No master available. 
> Notifying waiting threads
> 2013-05-09 15:09:45,006 INFO org.apache.hadoop.hbase.master.HMaster: Cluster 
> went down before this master became active
> 2013-05-09 15:09:45,006 DEBUG org.apache.hadoop.hbase.master.HMaster: 
> Stopping service threads
> 2013-05-09 15:09:45,006 INFO org.apache.hadoop.ipc.HBaseServer: Stopping 
> server on 6
>  
> In ActiveMasterManager::blockUntilBecomingActiveMaster()
> {code}
>   ..
>   if (!clusterStatusTracker.isClusterUp()) {
>   this.master.stop(
> "Cluster went down before this master became active");
> }
>   ..
> {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HBASE-8462) Custom timestamps should not be allowed to be negative

2013-05-10 Thread Enis Soztutar (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-8462?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Enis Soztutar updated HBASE-8462:
-

Attachment: hbase-8462_v2.patch

Two tests failed because they were using negative time ranges. v2 patch fixes 
those tests. 


> Custom timestamps should not be allowed to be negative
> --
>
> Key: HBASE-8462
> URL: https://issues.apache.org/jira/browse/HBASE-8462
> Project: HBase
>  Issue Type: Bug
>  Components: Client
>Reporter: Enis Soztutar
>Assignee: Enis Soztutar
> Fix For: 0.98.0, 0.94.8, 0.95.1
>
> Attachments: hbase-8462_v1.patch, hbase-8462_v2.patch
>
>
> Client supplied timestamps should not be allowed to be negative, otherwise 
> unpredictable results will follow. Especially, since we are encoding the ts 
> using Bytes.Bytes(long), negative timestamps are sorted after positive ones. 
> Plus, the new PB messages define ts' as uint64. 
> Credit goes to Huned Lokhandwala for reporting this.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-8450) Update hbase-default.xml and general recommendations to better suit current hw, h2, experience, etc.

2013-05-10 Thread Enis Soztutar (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8450?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655109#comment-13655109
 ] 

Enis Soztutar commented on HBASE-8450:
--

bq. Also are we adding all the configs to xml when it is getting added to code? 
I doubt this.. I have checked with 94 code base some time back and I was able 
to find some of the configs only referred in code but not in hbase-default.xml
I agree to the concern. I always saw this as a three tiered approach. (1)There 
are configs which are changed in every hbase-site. (2) There are some configs 
that you may or may not want to changed. (3) And there are configs that only if 
you know the code, and something crazy happens, you want to change it. 
For 1, and 2, they should live in hbase-default. for 3, I agree with Stack, 
that we should not pollute hbase-default too much. However, I have seen some 
(2) cases but the config was not in hbase-default. I think we should just fix 
those cases. 

> Update hbase-default.xml and general recommendations to better suit current 
> hw, h2, experience, etc.
> 
>
> Key: HBASE-8450
> URL: https://issues.apache.org/jira/browse/HBASE-8450
> Project: HBase
>  Issue Type: Task
>  Components: Usability
>Reporter: stack
>Priority: Critical
> Fix For: 0.95.1
>
> Attachments: 8450.txt
>
>
> This is a critical task we need to do before we release; review our defaults.
> On cursory review, there are configs in hbase-default.xml that no longer have 
> matching code; there are some that should be changed because we know better 
> now and others that should be amended because hardware and deploys are bigger 
> than they used to be.
> We could also move stuff around so that the must-edit stuff is near the top 
> (zk quorum config. is mid-way down the page) and beef up the descriptions -- 
> especially since these descriptions shine through in the refguide.
> Lastly, I notice that our tgz does not "include" an hbase-default.xml other 
> than the one bundled up in the jar.  Maybe we should make it more accessible.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HBASE-8375) Streamline Table durability settings

2013-05-10 Thread Enis Soztutar (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-8375?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Enis Soztutar updated HBASE-8375:
-

Attachment: hbase-8375_v2.patch

v2 from RB. 

> Streamline Table durability settings
> 
>
> Key: HBASE-8375
> URL: https://issues.apache.org/jira/browse/HBASE-8375
> Project: HBase
>  Issue Type: Sub-task
>Reporter: Lars Hofhansl
>Assignee: Enis Soztutar
> Fix For: 0.98.0, 0.95.1
>
> Attachments: hbase-8375_v1.patch, hbase-8375_v2.patch
>
>
> HBASE-7801 introduces the notion of per mutation fine grained durability 
> settings.
> This issue is to consider and the discuss the same for the per table settings 
> (i.e. what would be used if the mutation indicates USE_DEFAULT). I propose 
> the following setting per table:
> * SKIP_WAL (i.e. an unlogged table)
> * ASYNC_WAL (the current deferred log flush)
> * SYNC_WAL (the current default)
> * FSYNC_WAL (for future uses of HDFS' hsync())

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-6134) Improvement for split-worker to speed up distributed log splitting

2013-05-10 Thread Liang Lee (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-6134?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655106#comment-13655106
 ] 

Liang Lee commented on HBASE-6134:
--

Hi,when I apply the patch in Hbase-0.94.2 ,I find error in the 
HLogSplitter.java:
The object DistributedLogSplittingHelper can't find its defination,what's the 
matter?
Thanks

> Improvement for split-worker to speed up distributed log splitting
> --
>
> Key: HBASE-6134
> URL: https://issues.apache.org/jira/browse/HBASE-6134
> Project: HBase
>  Issue Type: Improvement
>  Components: wal
>Reporter: chunhui shen
>Assignee: chunhui shen
>Priority: Critical
> Fix For: 0.95.0
>
> Attachments: 6134v4.patch, HBASE-6134.patch, HBASE-6134v2.patch, 
> HBASE-6134v3-92.patch, HBASE-6134v3.patch, HBASE-6134v4-94.patch, 
> HBASE-6134v4-94v2.patch, HBASE-6134v4.patch
>
>
> First,we do the test between local-master-splitting and 
> distributed-log-splitting
> Environment:34 hlog files, 5 regionservers,(after kill one, only 4 rs do ths 
> splitting work), 400 regions in one hlog file
> local-master-split:60s+
> distributed-log-splitting:165s+
> In fact, in our production environment, distributed-log-splitting also took 
> 60s with 30 regionservers for 34 hlog files (regionserver may be in high load)
> We found split-worker split one log file took about 20s
> (30ms~50ms per writer.close(); 10ms per create writers )
> I think we could do the improvement for this:
> Parallelizing the create and close writers in threads
> In the patch, change the logic for  distributed-log-splitting same as the 
> local-master-splitting and parallelizing the close in threads.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-8520) TestIOFencing fails intermittently due to compaction kicking in too soon

2013-05-10 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8520?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655105#comment-13655105
 ] 

Hudson commented on HBASE-8520:
---

Integrated in HBase-TRUNK-on-Hadoop-2.0.0 #524 (See 
[https://builds.apache.org/job/HBase-TRUNK-on-Hadoop-2.0.0/524/])
HBASE-8520 TestIOFencing fails intermittently due to compaction kicking in 
too soon (Ted Yu) (Revision 1481226)

 Result = FAILURE
tedyu : 
Files : 
* 
/hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/TestIOFencing.java


> TestIOFencing fails intermittently due to compaction kicking in too soon
> 
>
> Key: HBASE-8520
> URL: https://issues.apache.org/jira/browse/HBASE-8520
> Project: HBase
>  Issue Type: Test
>Reporter: Ted Yu
>Assignee: Ted Yu
> Fix For: 0.98.0, 0.95.1
>
> Attachments: 8520-v1.txt
>
>
> Here is one example: 
> https://builds.apache.org/job/HBase-TRUNK/lastCompletedBuild/testReport/org.apache.hadoop.hbase/TestIOFencing/testFencingAroundCompactionAfterWALSync/
> It failed on this assertion:
> {code}
>   assertTrue(compactingRegion.countStoreFiles() > 1);
> {code}
> I searched for "Blocking compactions" but didn't find it in the test output.
> I think the compaction kicked in too soon, resulting in store file count 
> being 1.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-8517) Stochastic Loadbalancer isn't finding steady state on real clusters

2013-05-10 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8517?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655103#comment-13655103
 ] 

Hadoop QA commented on HBASE-8517:
--

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12582741/HBASE-8517-1.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 11 new 
or modified tests.

{color:green}+1 hadoop1.0{color}.  The patch compiles against the hadoop 
1.0 profile.

{color:green}+1 hadoop2.0{color}.  The patch compiles against the hadoop 
2.0 profile.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 lineLengths{color}.  The patch introduces lines longer than 
100

  {color:green}+1 site{color}.  The mvn site goal succeeds with this patch.

 {color:red}-1 core tests{color}.  The patch failed these unit tests:
 

 {color:red}-1 core zombie tests{color}.  There are 1 zombie test(s):   
at 
org.apache.hadoop.hbase.client.TestHCM.testDeleteForZKConnLeak(TestHCM.java:855)

Test results: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5634//testReport/
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5634//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-prefix-tree.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5634//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-client.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5634//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-common.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5634//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-protocol.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5634//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-server.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5634//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-hadoop1-compat.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5634//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-examples.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5634//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-hadoop-compat.html
Console output: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5634//console

This message is automatically generated.

> Stochastic Loadbalancer isn't finding steady state on real clusters
> ---
>
> Key: HBASE-8517
> URL: https://issues.apache.org/jira/browse/HBASE-8517
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 0.95.0
>Reporter: Elliott Clark
>Assignee: Elliott Clark
> Attachments: HBASE-8517-0.patch, HBASE-8517-1.patch
>
>
> I have a cluster that runs IT tests.  Last night after all tests were done I 
> noticed that the balancer was thrashing regions around.
> The number of regions on each machine is not correct.
> The balancer seems to value the cost of moving a region way too little.
> {code}
> 2013-05-09 16:34:58,920 DEBUG [IPC Server handler 4 on 6] 
> org.apache.hadoop.hbase.master.balancer.StochasticLoadBalancer: Finished 
> computing new load balance plan.  Computation took 5367ms to try 8910 
> different iterations.  Found a solution that moves 37 regions; Going from a 
> computed cost of 56.50254222730425 to a new cost of 11.214035466575254
> 2013-05-09 16:37:48,715 DEBUG [IPC Server handler 7 on 6] 
> org.apache.hadoop.hbase.master.balancer.StochasticLoadBalancer: Finished 
> computing new load balance plan.  Computation took 4735ms to try 8910 
> different iterations.  Found a solution that moves 38 regions; Going from a 
> computed cost of 56.612624531830996 to a new cost of 11.275763861636982
> 2013-05-09 16:38:11,398 DEBUG [IPC Server handler 6 on 6] 
> org.apache.hadoop.hbase.master.balancer.StochasticLoadBalancer: Finished 
> computing new load balance plan.  Computation took 4502ms to try 8910 
> different iterations.  Found a solution that moves 39 regions; Going from a 
> computed cost of 56.50048461413552 to a new cost of 11.225352339003237
> {code}
> Each of those balancer runs were tri

[jira] [Commented] (HBASE-7006) [MTTR] Improve Region Server Recovery Time - Distributed Log Replay

2013-05-10 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-7006?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655100#comment-13655100
 ] 

Hadoop QA commented on HBASE-7006:
--

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  
http://issues.apache.org/jira/secure/attachment/12582743/hbase-7006-combined-v8.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 31 new 
or modified tests.

{color:green}+1 hadoop1.0{color}.  The patch compiles against the hadoop 
1.0 profile.

{color:green}+1 hadoop2.0{color}.  The patch compiles against the hadoop 
2.0 profile.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 lineLengths{color}.  The patch introduces lines longer than 
100

  {color:green}+1 site{color}.  The mvn site goal succeeds with this patch.

{color:green}+1 core tests{color}.  The patch passed unit tests in .

Test results: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5635//testReport/
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5635//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-client.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5635//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-examples.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5635//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-protocol.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5635//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-hadoop1-compat.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5635//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-prefix-tree.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5635//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-common.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5635//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-hadoop-compat.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5635//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-server.html
Console output: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5635//console

This message is automatically generated.

> [MTTR] Improve Region Server Recovery Time - Distributed Log Replay
> ---
>
> Key: HBASE-7006
> URL: https://issues.apache.org/jira/browse/HBASE-7006
> Project: HBase
>  Issue Type: New Feature
>  Components: MTTR
>Reporter: stack
>Assignee: Jeffrey Zhong
>Priority: Critical
> Fix For: 0.95.1
>
> Attachments: hbase-7006-combined.patch, hbase-7006-combined-v1.patch, 
> hbase-7006-combined-v4.patch, hbase-7006-combined-v5.patch, 
> hbase-7006-combined-v6.patch, hbase-7006-combined-v7.patch, 
> hbase-7006-combined-v8.patch, LogSplitting Comparison.pdf, 
> ProposaltoimprovelogsplittingprocessregardingtoHBASE-7006-v2.pdf
>
>
> Just saw interesting issue where a cluster went down  hard and 30 nodes had 
> 1700 WALs to replay.  Replay took almost an hour.  It looks like it could run 
> faster that much of the time is spent zk'ing and nn'ing.
> Putting in 0.96 so it gets a look at least.  Can always punt.
> Distributed Log Replay Description:
> After a region server fails, we firstly assign a failed region to another 
> region server with recovering state marked in ZooKeeper. Then a 
> SplitLogWorker directly replays edits from WAL(Write-Ahead-Log)s of the 
> failed region server to the region after it's re-opened in the new location. 
> When a region is in recovering state, it can also accept writes but no 
> reads(including Append and Increment), region split or merge. 
> The feature piggybacks on existing distributed log splitting framework and 
> directly replay WAL edits to another region server instead of creating 
> recovered.edits files.
> The advantages over existing log splitting recovered edits implementation:
> 1) Eliminate the steps to write and read recovered.edits files. There could 
> be thousands of recovered.edits files are created and written concurrently 
> during a region server recovery. Many small random writes could degrade the 
>

[jira] [Commented] (HBASE-8528) [hadoop2] TestMultiTableInputFormat always fails on hadoop with YARN-129 applied

2013-05-10 Thread Gary Helmling (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8528?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655093#comment-13655093
 ] 

Gary Helmling commented on HBASE-8528:
--

bq. I like the yarn.is.minicluster setting approach because it is more direct. 
I'm wondering why this isn't the default when every minimrcluster is created.

This is set in the MiniMRCluster configuration automatically, but the 
Configuration instance seems to be a copy of any Configuration passed down to 
the MiniMRCluster constructor, so it isn't visible externally.  At the moment 
HBaseTestingUtility doesn't even pass down the Configuration instance, but even 
if it did it wouldn't help due to the above.

If you see any more transparent way of getting yarn.is.minicluster set 
automatically, that would be cool.

> [hadoop2] TestMultiTableInputFormat always fails on hadoop with YARN-129 
> applied
> 
>
> Key: HBASE-8528
> URL: https://issues.apache.org/jira/browse/HBASE-8528
> Project: HBase
>  Issue Type: Sub-task
>  Components: hadoop2, mapreduce, test
>Affects Versions: 0.95.0
>Reporter: Jonathan Hsieh
>Assignee: Jonathan Hsieh
> Fix For: 0.98.0, 0.95.1
>
> Attachments: hbase-8528_alt.patch, hbase-8528.patch
>
>
> TestMulitTableInputFormat always fails if YARN-129 is applied.  Its error 
> message isn't useful (assertion failed) -- but if you look at the 
> stderr/stdout/syslog files of the jobs in 
> hbase-server/target/org.apache.hadoop.mapred.MiniMRCluster* (minicluster data 
> dirs) you'll find this class not found exception:
> {code}
> Exception in thread "main" java.lang.NoClassDefFoundError: 
> org/apache/hadoop/mapreduce/v2/app/MRAppMaster
> Caused by: java.lang.ClassNotFoundException: 
> org.apache.hadoop.mapreduce.v2.app.MRAppMaster
> at java.net.URLClassLoader$1.run(URLClassLoader.java:202)
> at java.security.AccessController.doPrivileged(Native Method)
> at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
> at java.lang.ClassLoader.loadClass(ClassLoader.java:306)
> at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
> at java.lang.ClassLoader.loadClass(ClassLoader.java:247)
> Could not find the main class: 
> org.apache.hadoop.mapreduce.v2.app.MRAppMaster.  Program will exit.
> {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-5995) Fix and reenable TestLogRolling.testLogRollOnPipelineRestart

2013-05-10 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-5995?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655090#comment-13655090
 ] 

Hadoop QA commented on HBASE-5995:
--

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12582735/hbase-5995_v2.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 3 new 
or modified tests.

{color:green}+1 hadoop1.0{color}.  The patch compiles against the hadoop 
1.0 profile.

{color:green}+1 hadoop2.0{color}.  The patch compiles against the hadoop 
2.0 profile.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 lineLengths{color}.  The patch introduces lines longer than 
100

  {color:green}+1 site{color}.  The mvn site goal succeeds with this patch.

{color:green}+1 core tests{color}.  The patch passed unit tests in .

Test results: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5633//testReport/
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5633//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-protocol.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5633//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-client.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5633//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-examples.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5633//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-hadoop1-compat.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5633//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-prefix-tree.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5633//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-common.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5633//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-server.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5633//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-hadoop-compat.html
Console output: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5633//console

This message is automatically generated.

> Fix and reenable TestLogRolling.testLogRollOnPipelineRestart
> 
>
> Key: HBASE-5995
> URL: https://issues.apache.org/jira/browse/HBASE-5995
> Project: HBase
>  Issue Type: Sub-task
>  Components: test
>Reporter: stack
>Assignee: Enis Soztutar
>Priority: Blocker
> Fix For: 0.98.0, 0.95.1
>
> Attachments: hbase-5995_v1.patch, hbase-5995_v2.patch
>
>
> HBASE-5984 disabled this flakey test (See the issue for more).  This issue is 
> about getting it enabled again.  Made a blocker on 0.96.0 so it gets 
> attention.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-8520) TestIOFencing fails intermittently due to compaction kicking in too soon

2013-05-10 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8520?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655087#comment-13655087
 ] 

Hudson commented on HBASE-8520:
---

Integrated in hbase-0.95 #186 (See 
[https://builds.apache.org/job/hbase-0.95/186/])
HBASE-8520 TestIOFencing fails intermittently due to compaction kicking in 
too soon (Ted Yu) (Revision 1481227)

 Result = SUCCESS
tedyu : 
Files : 
* 
/hbase/branches/0.95/hbase-server/src/test/java/org/apache/hadoop/hbase/TestIOFencing.java


> TestIOFencing fails intermittently due to compaction kicking in too soon
> 
>
> Key: HBASE-8520
> URL: https://issues.apache.org/jira/browse/HBASE-8520
> Project: HBase
>  Issue Type: Test
>Reporter: Ted Yu
>Assignee: Ted Yu
> Fix For: 0.98.0, 0.95.1
>
> Attachments: 8520-v1.txt
>
>
> Here is one example: 
> https://builds.apache.org/job/HBase-TRUNK/lastCompletedBuild/testReport/org.apache.hadoop.hbase/TestIOFencing/testFencingAroundCompactionAfterWALSync/
> It failed on this assertion:
> {code}
>   assertTrue(compactingRegion.countStoreFiles() > 1);
> {code}
> I searched for "Blocking compactions" but didn't find it in the test output.
> I think the compaction kicked in too soon, resulting in store file count 
> being 1.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-8520) TestIOFencing fails intermittently due to compaction kicking in too soon

2013-05-10 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8520?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655086#comment-13655086
 ] 

Hudson commented on HBASE-8520:
---

Integrated in HBase-TRUNK #4110 (See 
[https://builds.apache.org/job/HBase-TRUNK/4110/])
HBASE-8520 TestIOFencing fails intermittently due to compaction kicking in 
too soon (Ted Yu) (Revision 1481226)

 Result = SUCCESS
tedyu : 
Files : 
* 
/hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/TestIOFencing.java


> TestIOFencing fails intermittently due to compaction kicking in too soon
> 
>
> Key: HBASE-8520
> URL: https://issues.apache.org/jira/browse/HBASE-8520
> Project: HBase
>  Issue Type: Test
>Reporter: Ted Yu
>Assignee: Ted Yu
> Fix For: 0.98.0, 0.95.1
>
> Attachments: 8520-v1.txt
>
>
> Here is one example: 
> https://builds.apache.org/job/HBase-TRUNK/lastCompletedBuild/testReport/org.apache.hadoop.hbase/TestIOFencing/testFencingAroundCompactionAfterWALSync/
> It failed on this assertion:
> {code}
>   assertTrue(compactingRegion.countStoreFiles() > 1);
> {code}
> I searched for "Blocking compactions" but didn't find it in the test output.
> I think the compaction kicked in too soon, resulting in store file count 
> being 1.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-6929) Publish Hbase 0.94 artifacts build against hadoop-2.0

2013-05-10 Thread Roshan Naik (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-6929?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655083#comment-13655083
 ] 

Roshan Naik commented on HBASE-6929:


these are not hadoop2 specific these are hbase0.95+ specific which comes in 
both hadoop1&2 flovaours. So it seems we will need to support the following 
combos...


1) building against hbase 0.94x  .. in this case there will be no -hadoop1 
suffix on hbase deps


2) building against hbase 0.95+  with hadoop 1 .. in this case the -hadoop1 
suffix will be used with the newer dependencies list


3) building against hbase 0.95+ with hadoop 2 .. in this case the -hadoop2 
suffix will be used with the newer dependency list



> Publish Hbase 0.94 artifacts build against hadoop-2.0
> -
>
> Key: HBASE-6929
> URL: https://issues.apache.org/jira/browse/HBASE-6929
> Project: HBase
>  Issue Type: Task
>  Components: build
>Affects Versions: 0.94.2
>Reporter: Enis Soztutar
> Attachments: 6929.txt, hbase-6929_v2.patch
>
>
> Downstream projects (flume, hive, pig, etc) depends on hbase, but since the 
> hbase binaries build with hadoop-2.0 are not pushed to maven, they cannot 
> depend on them. AFAIK, hadoop 1 and 2 are not binary compatible, so we should 
> also push hbase jars build with hadoop2.0 profile into maven, possibly with 
> version string like 0.94.2-hadoop2.0. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-8520) TestIOFencing fails intermittently due to compaction kicking in too soon

2013-05-10 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8520?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655082#comment-13655082
 ] 

Hudson commented on HBASE-8520:
---

Integrated in hbase-0.95-on-hadoop2 #94 (See 
[https://builds.apache.org/job/hbase-0.95-on-hadoop2/94/])
HBASE-8520 TestIOFencing fails intermittently due to compaction kicking in 
too soon (Ted Yu) (Revision 1481227)

 Result = FAILURE
tedyu : 
Files : 
* 
/hbase/branches/0.95/hbase-server/src/test/java/org/apache/hadoop/hbase/TestIOFencing.java


> TestIOFencing fails intermittently due to compaction kicking in too soon
> 
>
> Key: HBASE-8520
> URL: https://issues.apache.org/jira/browse/HBASE-8520
> Project: HBase
>  Issue Type: Test
>Reporter: Ted Yu
>Assignee: Ted Yu
> Fix For: 0.98.0, 0.95.1
>
> Attachments: 8520-v1.txt
>
>
> Here is one example: 
> https://builds.apache.org/job/HBase-TRUNK/lastCompletedBuild/testReport/org.apache.hadoop.hbase/TestIOFencing/testFencingAroundCompactionAfterWALSync/
> It failed on this assertion:
> {code}
>   assertTrue(compactingRegion.countStoreFiles() > 1);
> {code}
> I searched for "Blocking compactions" but didn't find it in the test output.
> I think the compaction kicked in too soon, resulting in store file count 
> being 1.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-6929) Publish Hbase 0.94 artifacts build against hadoop-2.0

2013-05-10 Thread Hari Shreedharan (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-6929?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655079#comment-13655079
 ] 

Hari Shreedharan commented on HBASE-6929:
-

Yep, we could define those in the hadoop-2 profile as dependencies.

> Publish Hbase 0.94 artifacts build against hadoop-2.0
> -
>
> Key: HBASE-6929
> URL: https://issues.apache.org/jira/browse/HBASE-6929
> Project: HBase
>  Issue Type: Task
>  Components: build
>Affects Versions: 0.94.2
>Reporter: Enis Soztutar
> Attachments: 6929.txt, hbase-6929_v2.patch
>
>
> Downstream projects (flume, hive, pig, etc) depends on hbase, but since the 
> hbase binaries build with hadoop-2.0 are not pushed to maven, they cannot 
> depend on them. AFAIK, hadoop 1 and 2 are not binary compatible, so we should 
> also push hbase jars build with hadoop2.0 profile into maven, possibly with 
> version string like 0.94.2-hadoop2.0. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-8375) Streamline Table durability settings

2013-05-10 Thread Enis Soztutar (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8375?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655078#comment-13655078
 ] 

Enis Soztutar commented on HBASE-8375:
--

Wow, totally forgot that there is a TestDurability class. Let me fix the test. 

> Streamline Table durability settings
> 
>
> Key: HBASE-8375
> URL: https://issues.apache.org/jira/browse/HBASE-8375
> Project: HBase
>  Issue Type: Sub-task
>Reporter: Lars Hofhansl
>Assignee: Enis Soztutar
> Fix For: 0.98.0, 0.95.1
>
> Attachments: hbase-8375_v1.patch
>
>
> HBASE-7801 introduces the notion of per mutation fine grained durability 
> settings.
> This issue is to consider and the discuss the same for the per table settings 
> (i.e. what would be used if the mutation indicates USE_DEFAULT). I propose 
> the following setting per table:
> * SKIP_WAL (i.e. an unlogged table)
> * ASYNC_WAL (the current deferred log flush)
> * SYNC_WAL (the current default)
> * FSYNC_WAL (for future uses of HDFS' hsync())

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-8375) Streamline Table durability settings

2013-05-10 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8375?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655075#comment-13655075
 ] 

Hadoop QA commented on HBASE-8375:
--

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12582727/hbase-8375_v1.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 6 new 
or modified tests.

{color:green}+1 hadoop1.0{color}.  The patch compiles against the hadoop 
1.0 profile.

{color:green}+1 hadoop2.0{color}.  The patch compiles against the hadoop 
2.0 profile.

{color:red}-1 javadoc{color}.  The javadoc tool appears to have generated 1 
warning messages.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 lineLengths{color}.  The patch introduces lines longer than 
100

  {color:green}+1 site{color}.  The mvn site goal succeeds with this patch.

 {color:red}-1 core tests{color}.  The patch failed these unit tests:
   org.apache.hadoop.hbase.regionserver.wal.TestDurability

 {color:red}-1 core zombie tests{color}.  There are 1 zombie test(s): 

Test results: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5632//testReport/
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5632//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-client.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5632//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-examples.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5632//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-protocol.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5632//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-hadoop1-compat.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5632//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-prefix-tree.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5632//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-common.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5632//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-hadoop-compat.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5632//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-server.html
Console output: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5632//console

This message is automatically generated.

> Streamline Table durability settings
> 
>
> Key: HBASE-8375
> URL: https://issues.apache.org/jira/browse/HBASE-8375
> Project: HBase
>  Issue Type: Sub-task
>Reporter: Lars Hofhansl
>Assignee: Enis Soztutar
> Fix For: 0.98.0, 0.95.1
>
> Attachments: hbase-8375_v1.patch
>
>
> HBASE-7801 introduces the notion of per mutation fine grained durability 
> settings.
> This issue is to consider and the discuss the same for the per table settings 
> (i.e. what would be used if the mutation indicates USE_DEFAULT). I propose 
> the following setting per table:
> * SKIP_WAL (i.e. an unlogged table)
> * ASYNC_WAL (the current deferred log flush)
> * SYNC_WAL (the current default)
> * FSYNC_WAL (for future uses of HDFS' hsync())

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-8528) [hadoop2] TestMultiTableInputFormat always fails on hadoop with YARN-129 applied

2013-05-10 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8528?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655074#comment-13655074
 ] 

Hadoop QA commented on HBASE-8528:
--

{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12582734/hbase-8528_alt.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 3 new 
or modified tests.

{color:green}+1 hadoop1.0{color}.  The patch compiles against the hadoop 
1.0 profile.

{color:green}+1 hadoop2.0{color}.  The patch compiles against the hadoop 
2.0 profile.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 lineLengths{color}.  The patch does not introduce lines 
longer than 100

  {color:green}+1 site{color}.  The mvn site goal succeeds with this patch.

{color:green}+1 core tests{color}.  The patch passed unit tests in .

Test results: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5631//testReport/
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5631//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-prefix-tree.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5631//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-client.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5631//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-common.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5631//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-protocol.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5631//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-server.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5631//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-hadoop1-compat.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5631//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-examples.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5631//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-hadoop-compat.html
Console output: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5631//console

This message is automatically generated.

> [hadoop2] TestMultiTableInputFormat always fails on hadoop with YARN-129 
> applied
> 
>
> Key: HBASE-8528
> URL: https://issues.apache.org/jira/browse/HBASE-8528
> Project: HBase
>  Issue Type: Sub-task
>  Components: hadoop2, mapreduce, test
>Affects Versions: 0.95.0
>Reporter: Jonathan Hsieh
>Assignee: Jonathan Hsieh
> Fix For: 0.98.0, 0.95.1
>
> Attachments: hbase-8528_alt.patch, hbase-8528.patch
>
>
> TestMulitTableInputFormat always fails if YARN-129 is applied.  Its error 
> message isn't useful (assertion failed) -- but if you look at the 
> stderr/stdout/syslog files of the jobs in 
> hbase-server/target/org.apache.hadoop.mapred.MiniMRCluster* (minicluster data 
> dirs) you'll find this class not found exception:
> {code}
> Exception in thread "main" java.lang.NoClassDefFoundError: 
> org/apache/hadoop/mapreduce/v2/app/MRAppMaster
> Caused by: java.lang.ClassNotFoundException: 
> org.apache.hadoop.mapreduce.v2.app.MRAppMaster
> at java.net.URLClassLoader$1.run(URLClassLoader.java:202)
> at java.security.AccessController.doPrivileged(Native Method)
> at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
> at java.lang.ClassLoader.loadClass(ClassLoader.java:306)
> at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
> at java.lang.ClassLoader.loadClass(ClassLoader.java:247)
> Could not find the main class: 
> org.apache.hadoop.mapreduce.v2.app.MRAppMaster.  Program will exit.
> {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-8528) [hadoop2] TestMultiTableInputFormat always fails on hadoop with YARN-129 applied

2013-05-10 Thread Ted Yu (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8528?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655071#comment-13655071
 ] 

Ted Yu commented on HBASE-8528:
---

+1 on Gary's patch.

> [hadoop2] TestMultiTableInputFormat always fails on hadoop with YARN-129 
> applied
> 
>
> Key: HBASE-8528
> URL: https://issues.apache.org/jira/browse/HBASE-8528
> Project: HBase
>  Issue Type: Sub-task
>  Components: hadoop2, mapreduce, test
>Affects Versions: 0.95.0
>Reporter: Jonathan Hsieh
>Assignee: Jonathan Hsieh
> Fix For: 0.98.0, 0.95.1
>
> Attachments: hbase-8528_alt.patch, hbase-8528.patch
>
>
> TestMulitTableInputFormat always fails if YARN-129 is applied.  Its error 
> message isn't useful (assertion failed) -- but if you look at the 
> stderr/stdout/syslog files of the jobs in 
> hbase-server/target/org.apache.hadoop.mapred.MiniMRCluster* (minicluster data 
> dirs) you'll find this class not found exception:
> {code}
> Exception in thread "main" java.lang.NoClassDefFoundError: 
> org/apache/hadoop/mapreduce/v2/app/MRAppMaster
> Caused by: java.lang.ClassNotFoundException: 
> org.apache.hadoop.mapreduce.v2.app.MRAppMaster
> at java.net.URLClassLoader$1.run(URLClassLoader.java:202)
> at java.security.AccessController.doPrivileged(Native Method)
> at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
> at java.lang.ClassLoader.loadClass(ClassLoader.java:306)
> at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
> at java.lang.ClassLoader.loadClass(ClassLoader.java:247)
> Could not find the main class: 
> org.apache.hadoop.mapreduce.v2.app.MRAppMaster.  Program will exit.
> {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Comment Edited] (HBASE-8517) Stochastic Loadbalancer isn't finding steady state on real clusters

2013-05-10 Thread Elliott Clark (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8517?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655067#comment-13655067
 ] 

Elliott Clark edited comment on HBASE-8517 at 5/11/13 12:23 AM:


bq.How was the constant 600 determined ?
That was the default before adding the scaling.

bq.Should it be related to cluster size (number of servers, number of regions) ?
Possibly but this works well, and number of servers should on a well run 
cluster be a proxy for number of regions; so it's not really needed in the 
computation.

bq.Looks like the return value from computeCost() is ignored.
Yes it is just used to output a good debug message.

bq.Is the addition of 9 needed when cluster.numMovedMetaRegions is 0 ?
Yes, because the possibility is there.  You want to scale based upon the 
highest possible cost.

  was (Author: eclark):
bq.How was the constant 600 determined ?
That was the default before adding the scaling.

bq.Should it be related to cluster size (number of servers, number of regions) ?
Possibly but this works well, and number of servers should on a well run 
cluster be a proxy for number of regions.

bq.Looks like the return value from computeCost() is ignored.
Yes it is just used to output a good debug message.

bq.Is the addition of 9 needed when cluster.numMovedMetaRegions is 0 ?
Yes, because the possibility is there.  You want to scale based upon the 
highest possible cost.
  
> Stochastic Loadbalancer isn't finding steady state on real clusters
> ---
>
> Key: HBASE-8517
> URL: https://issues.apache.org/jira/browse/HBASE-8517
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 0.95.0
>Reporter: Elliott Clark
>Assignee: Elliott Clark
> Attachments: HBASE-8517-0.patch, HBASE-8517-1.patch
>
>
> I have a cluster that runs IT tests.  Last night after all tests were done I 
> noticed that the balancer was thrashing regions around.
> The number of regions on each machine is not correct.
> The balancer seems to value the cost of moving a region way too little.
> {code}
> 2013-05-09 16:34:58,920 DEBUG [IPC Server handler 4 on 6] 
> org.apache.hadoop.hbase.master.balancer.StochasticLoadBalancer: Finished 
> computing new load balance plan.  Computation took 5367ms to try 8910 
> different iterations.  Found a solution that moves 37 regions; Going from a 
> computed cost of 56.50254222730425 to a new cost of 11.214035466575254
> 2013-05-09 16:37:48,715 DEBUG [IPC Server handler 7 on 6] 
> org.apache.hadoop.hbase.master.balancer.StochasticLoadBalancer: Finished 
> computing new load balance plan.  Computation took 4735ms to try 8910 
> different iterations.  Found a solution that moves 38 regions; Going from a 
> computed cost of 56.612624531830996 to a new cost of 11.275763861636982
> 2013-05-09 16:38:11,398 DEBUG [IPC Server handler 6 on 6] 
> org.apache.hadoop.hbase.master.balancer.StochasticLoadBalancer: Finished 
> computing new load balance plan.  Computation took 4502ms to try 8910 
> different iterations.  Found a solution that moves 39 regions; Going from a 
> computed cost of 56.50048461413552 to a new cost of 11.225352339003237
> {code}
> Each of those balancer runs were triggered when there was no load on the 
> cluster.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-8517) Stochastic Loadbalancer isn't finding steady state on real clusters

2013-05-10 Thread Elliott Clark (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8517?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655067#comment-13655067
 ] 

Elliott Clark commented on HBASE-8517:
--

bq.How was the constant 600 determined ?
That was the default before adding the scaling.

bq.Should it be related to cluster size (number of servers, number of regions) ?
Possibly but this works well, and number of servers should on a well run 
cluster be a proxy for number of regions.

bq.Looks like the return value from computeCost() is ignored.
Yes it is just used to output a good debug message.

bq.Is the addition of 9 needed when cluster.numMovedMetaRegions is 0 ?
Yes, because the possibility is there.  You want to scale based upon the 
highest possible cost.

> Stochastic Loadbalancer isn't finding steady state on real clusters
> ---
>
> Key: HBASE-8517
> URL: https://issues.apache.org/jira/browse/HBASE-8517
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 0.95.0
>Reporter: Elliott Clark
>Assignee: Elliott Clark
> Attachments: HBASE-8517-0.patch, HBASE-8517-1.patch
>
>
> I have a cluster that runs IT tests.  Last night after all tests were done I 
> noticed that the balancer was thrashing regions around.
> The number of regions on each machine is not correct.
> The balancer seems to value the cost of moving a region way too little.
> {code}
> 2013-05-09 16:34:58,920 DEBUG [IPC Server handler 4 on 6] 
> org.apache.hadoop.hbase.master.balancer.StochasticLoadBalancer: Finished 
> computing new load balance plan.  Computation took 5367ms to try 8910 
> different iterations.  Found a solution that moves 37 regions; Going from a 
> computed cost of 56.50254222730425 to a new cost of 11.214035466575254
> 2013-05-09 16:37:48,715 DEBUG [IPC Server handler 7 on 6] 
> org.apache.hadoop.hbase.master.balancer.StochasticLoadBalancer: Finished 
> computing new load balance plan.  Computation took 4735ms to try 8910 
> different iterations.  Found a solution that moves 38 regions; Going from a 
> computed cost of 56.612624531830996 to a new cost of 11.275763861636982
> 2013-05-09 16:38:11,398 DEBUG [IPC Server handler 6 on 6] 
> org.apache.hadoop.hbase.master.balancer.StochasticLoadBalancer: Finished 
> computing new load balance plan.  Computation took 4502ms to try 8910 
> different iterations.  Found a solution that moves 39 regions; Going from a 
> computed cost of 56.50048461413552 to a new cost of 11.225352339003237
> {code}
> Each of those balancer runs were triggered when there was no load on the 
> cluster.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-6929) Publish Hbase 0.94 artifacts build against hadoop-2.0

2013-05-10 Thread Roshan Naik (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-6929?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655066#comment-13655066
 ] 

Roshan Naik commented on HBASE-6929:


Hari:   it looks like flume needs these .. hbase-client, hbase-common 
(HBaseCommonTestingUtility.. at compile time),  hbase-server 
(HMaster,MiniZooKeeperCluster.. at compile time)  and hbase-server-test 
(HBaseTestingUtility..when running tests)

> Publish Hbase 0.94 artifacts build against hadoop-2.0
> -
>
> Key: HBASE-6929
> URL: https://issues.apache.org/jira/browse/HBASE-6929
> Project: HBase
>  Issue Type: Task
>  Components: build
>Affects Versions: 0.94.2
>Reporter: Enis Soztutar
> Attachments: 6929.txt, hbase-6929_v2.patch
>
>
> Downstream projects (flume, hive, pig, etc) depends on hbase, but since the 
> hbase binaries build with hadoop-2.0 are not pushed to maven, they cannot 
> depend on them. AFAIK, hadoop 1 and 2 are not binary compatible, so we should 
> also push hbase jars build with hadoop2.0 profile into maven, possibly with 
> version string like 0.94.2-hadoop2.0. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-8527) clean up code around compaction completion in HStore

2013-05-10 Thread Ted Yu (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8527?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655065#comment-13655065
 ] 

Ted Yu commented on HBASE-8527:
---

{code}
-  } finally {
-// We need the lock, as long as we are updating the storeFiles
-// or changing the memstore. Let us release it before calling
-// notifyChangeReadersObservers. See HBASE-4485 for a possible
-// deadlock scenario that could have happened if continue to hold
-// the lock.
-this.lock.writeLock().unlock();
{code}
I see the this.lock.writeLock().unlock() call is moved. Better move the comment 
as well.

> clean up code around compaction completion in HStore
> 
>
> Key: HBASE-8527
> URL: https://issues.apache.org/jira/browse/HBASE-8527
> Project: HBase
>  Issue Type: Improvement
>Reporter: Sergey Shelukhin
>Assignee: Sergey Shelukhin
>Priority: Minor
> Attachments: HBASE-8527-v0.patch
>
>
> The methods completeCompaction and it's caller are too long. Something I 
> changed while doing something else there; putting in a separate 
> easy-to-review JIRA to make the other future change smaller.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HBASE-8529) checkOpen is missing from multi, mutate, get and multiGet etc.

2013-05-10 Thread Jeffrey Zhong (JIRA)
Jeffrey Zhong created HBASE-8529:


 Summary: checkOpen is missing from multi, mutate, get and multiGet 
etc.
 Key: HBASE-8529
 URL: https://issues.apache.org/jira/browse/HBASE-8529
 Project: HBase
  Issue Type: Bug
Reporter: Jeffrey Zhong
Assignee: Jeffrey Zhong
Priority: Minor
 Fix For: 0.98.0, 0.95.1


I saw we have checkOpen in all those functions in 0.94 while they're missing 
from trunk. Does anyone know why?

For multi and mutate, if we don't call checkOpen we could flood our logs with 
bunch of "DFSOutputStream is closed" errors when we sync WAL.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-8528) [hadoop2] TestMultiTableInputFormat always fails on hadoop with YARN-129 applied

2013-05-10 Thread Jonathan Hsieh (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8528?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655063#comment-13655063
 ] 

Jonathan Hsieh commented on HBASE-8528:
---

[~ghelmling], I like the {{yarn.is.minicluster}} setting approach because it is 
more direct.  I'm wondering why this isn't the default when every minimrcluster 
is created.

I brought this up with our mr2 guy, [~sandyr] -- he's going to do a quick look 
and post here explaining why.




> [hadoop2] TestMultiTableInputFormat always fails on hadoop with YARN-129 
> applied
> 
>
> Key: HBASE-8528
> URL: https://issues.apache.org/jira/browse/HBASE-8528
> Project: HBase
>  Issue Type: Sub-task
>  Components: hadoop2, mapreduce, test
>Affects Versions: 0.95.0
>Reporter: Jonathan Hsieh
>Assignee: Jonathan Hsieh
> Fix For: 0.98.0, 0.95.1
>
> Attachments: hbase-8528_alt.patch, hbase-8528.patch
>
>
> TestMulitTableInputFormat always fails if YARN-129 is applied.  Its error 
> message isn't useful (assertion failed) -- but if you look at the 
> stderr/stdout/syslog files of the jobs in 
> hbase-server/target/org.apache.hadoop.mapred.MiniMRCluster* (minicluster data 
> dirs) you'll find this class not found exception:
> {code}
> Exception in thread "main" java.lang.NoClassDefFoundError: 
> org/apache/hadoop/mapreduce/v2/app/MRAppMaster
> Caused by: java.lang.ClassNotFoundException: 
> org.apache.hadoop.mapreduce.v2.app.MRAppMaster
> at java.net.URLClassLoader$1.run(URLClassLoader.java:202)
> at java.security.AccessController.doPrivileged(Native Method)
> at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
> at java.lang.ClassLoader.loadClass(ClassLoader.java:306)
> at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
> at java.lang.ClassLoader.loadClass(ClassLoader.java:247)
> Could not find the main class: 
> org.apache.hadoop.mapreduce.v2.app.MRAppMaster.  Program will exit.
> {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-8517) Stochastic Loadbalancer isn't finding steady state on real clusters

2013-05-10 Thread Ted Yu (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8517?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655053#comment-13655053
 ] 

Ted Yu commented on HBASE-8517:
---

{code}
+maxMoves = Math.max( (int) (cluster.numRegions * maxMovesPercent),
+600);
{code}
How was the constant 600 determined ?
Should it be related to cluster size (number of servers, number of regions) ?
{code}
 if (initCost > currentCost) {
+  computeCost(cluster, true);
{code}
Looks like the return value from computeCost() is ignored.
{code}
+return scale(0, cluster.numRegions+9, moveCost);
{code}
Is the addition of 9 needed when cluster.numMovedMetaRegions is 0 ?


> Stochastic Loadbalancer isn't finding steady state on real clusters
> ---
>
> Key: HBASE-8517
> URL: https://issues.apache.org/jira/browse/HBASE-8517
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 0.95.0
>Reporter: Elliott Clark
>Assignee: Elliott Clark
> Attachments: HBASE-8517-0.patch, HBASE-8517-1.patch
>
>
> I have a cluster that runs IT tests.  Last night after all tests were done I 
> noticed that the balancer was thrashing regions around.
> The number of regions on each machine is not correct.
> The balancer seems to value the cost of moving a region way too little.
> {code}
> 2013-05-09 16:34:58,920 DEBUG [IPC Server handler 4 on 6] 
> org.apache.hadoop.hbase.master.balancer.StochasticLoadBalancer: Finished 
> computing new load balance plan.  Computation took 5367ms to try 8910 
> different iterations.  Found a solution that moves 37 regions; Going from a 
> computed cost of 56.50254222730425 to a new cost of 11.214035466575254
> 2013-05-09 16:37:48,715 DEBUG [IPC Server handler 7 on 6] 
> org.apache.hadoop.hbase.master.balancer.StochasticLoadBalancer: Finished 
> computing new load balance plan.  Computation took 4735ms to try 8910 
> different iterations.  Found a solution that moves 38 regions; Going from a 
> computed cost of 56.612624531830996 to a new cost of 11.275763861636982
> 2013-05-09 16:38:11,398 DEBUG [IPC Server handler 6 on 6] 
> org.apache.hadoop.hbase.master.balancer.StochasticLoadBalancer: Finished 
> computing new load balance plan.  Computation took 4502ms to try 8910 
> different iterations.  Found a solution that moves 39 regions; Going from a 
> computed cost of 56.50048461413552 to a new cost of 11.225352339003237
> {code}
> Each of those balancer runs were triggered when there was no load on the 
> cluster.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-8154) HBaseTestingUtility.startMiniDFSClusterForTestHLog should handle "Address already in use" and retry automatically

2013-05-10 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8154?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655051#comment-13655051
 ] 

Hadoop QA commented on HBASE-8154:
--

{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12582720/8154-v4.txt
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 6 new 
or modified tests.

{color:green}+1 hadoop1.0{color}.  The patch compiles against the hadoop 
1.0 profile.

{color:green}+1 hadoop2.0{color}.  The patch compiles against the hadoop 
2.0 profile.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 lineLengths{color}.  The patch does not introduce lines 
longer than 100

  {color:green}+1 site{color}.  The mvn site goal succeeds with this patch.

{color:green}+1 core tests{color}.  The patch passed unit tests in .

Test results: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5630//testReport/
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5630//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-protocol.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5630//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-client.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5630//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-examples.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5630//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-hadoop1-compat.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5630//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-prefix-tree.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5630//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-common.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5630//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-server.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5630//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-hadoop-compat.html
Console output: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5630//console

This message is automatically generated.

> HBaseTestingUtility.startMiniDFSClusterForTestHLog should handle "Address 
> already in use" and retry automatically
> -
>
> Key: HBASE-8154
> URL: https://issues.apache.org/jira/browse/HBASE-8154
> Project: HBase
>  Issue Type: Test
>  Components: test
>Affects Versions: 0.98.0
>Reporter: Liang Xie
>Assignee: Ted Yu
> Attachments: 8154-v2.txt, 8154-v3.txt, 8154-v4.txt
>
>
> see 
> https://builds.apache.org/job/PreCommit-HBASE-Build/4906//testReport/org.apache.hadoop.hbase.regionserver.wal/TestHLog/testAppendClose/
> and 
> https://issues.apache.org/jira/browse/HBASE-7845?focusedCommentId=13607468&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-13607468

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-8519) Backup master will never come up if primary master dies during initialization

2013-05-10 Thread Jean-Daniel Cryans (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8519?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655047#comment-13655047
 ] 

Jean-Daniel Cryans commented on HBASE-8519:
---

bq. I think Jerry's description mentioned that 
clusterStatusTracker.isClusterUp() returned false, leading to backup master 
stopping.

Sorry, I meant the reverse for the shutdown znode. Which should also answer 
Jerry's second question.

bq. Can you give more details on the sequence of events?

Do a stop-hbase then kill the master just after that, you should get the same 
"Cluster went down before this master became active" when you restart.

> Backup master will never come up if primary master dies during initialization
> -
>
> Key: HBASE-8519
> URL: https://issues.apache.org/jira/browse/HBASE-8519
> Project: HBase
>  Issue Type: Bug
>  Components: master
>Affects Versions: 0.94.7, 0.95.0
>Reporter: Jerry He
>Assignee: Jerry He
>Priority: Minor
> Fix For: 0.98.0
>
>
> The problem happens if primary master dies after becoming master but before 
> it completes initialization and calls clusterStatusTracker.setClusterUp(),
> The backup master will try to become the master, but will shutdown itself 
> promptly because it sees 'the cluster is not up'.
> This is the backup master log:
> 2013-05-09 15:08:05,568 INFO 
> org.apache.hadoop.hbase.master.metrics.MasterMetrics: Initialized
> 2013-05-09 15:08:05,573 DEBUG org.apache.hadoop.hbase.master.HMaster: HMaster 
> started in backup mode.  Stalling until master znode is written.
> 2013-05-09 15:08:05,589 INFO 
> org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper: Node /hbase/master 
> already exists and this is not a retry
> 2013-05-09 15:08:05,590 INFO 
> org.apache.hadoop.hbase.master.ActiveMasterManager: Adding ZNode for 
> /hbase/backup-masters/xxx.com,6,1368137285373 in backup master directory
> 2013-05-09 15:08:05,595 INFO 
> org.apache.hadoop.hbase.master.ActiveMasterManager: Another master is the 
> active master, xxx.com,6,1368137283107; waiting to become the next active 
> master
> 2013-05-09 15:09:45,006 DEBUG 
> org.apache.hadoop.hbase.master.ActiveMasterManager: No master available. 
> Notifying waiting threads
> 2013-05-09 15:09:45,006 INFO org.apache.hadoop.hbase.master.HMaster: Cluster 
> went down before this master became active
> 2013-05-09 15:09:45,006 DEBUG org.apache.hadoop.hbase.master.HMaster: 
> Stopping service threads
> 2013-05-09 15:09:45,006 INFO org.apache.hadoop.ipc.HBaseServer: Stopping 
> server on 6
>  
> In ActiveMasterManager::blockUntilBecomingActiveMaster()
> {code}
>   ..
>   if (!clusterStatusTracker.isClusterUp()) {
>   this.master.stop(
> "Cluster went down before this master became active");
> }
>   ..
> {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HBASE-7006) [MTTR] Improve Region Server Recovery Time - Distributed Log Replay

2013-05-10 Thread Jeffrey Zhong (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-7006?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jeffrey Zhong updated HBASE-7006:
-

Attachment: hbase-7006-combined-v8.patch

V8 patch: Rebased and one line change add "checkOpen()" into replay command(I 
found "multi, mutate and multiGet" all missing checkOpen call. I'll file a 
separate JIRA on this). 

Summary on the JIRA:

All review comments are addressed. Unit tests passed and no issue is found so 
far from integration test IntegrationTestDataIngestWithChaosMonkey(will keeping 
running for a while). 

I think we can check it in now.

Next steps are:

1) Re-run performance tests to avoid any performance degradation from recent 
changes.
2) Add configuration to "disallow writes during recovery", which I 
intentionally leave it out from the combined patch.

Thanks.

> [MTTR] Improve Region Server Recovery Time - Distributed Log Replay
> ---
>
> Key: HBASE-7006
> URL: https://issues.apache.org/jira/browse/HBASE-7006
> Project: HBase
>  Issue Type: New Feature
>  Components: MTTR
>Reporter: stack
>Assignee: Jeffrey Zhong
>Priority: Critical
> Fix For: 0.95.1
>
> Attachments: hbase-7006-combined.patch, hbase-7006-combined-v1.patch, 
> hbase-7006-combined-v4.patch, hbase-7006-combined-v5.patch, 
> hbase-7006-combined-v6.patch, hbase-7006-combined-v7.patch, 
> hbase-7006-combined-v8.patch, LogSplitting Comparison.pdf, 
> ProposaltoimprovelogsplittingprocessregardingtoHBASE-7006-v2.pdf
>
>
> Just saw interesting issue where a cluster went down  hard and 30 nodes had 
> 1700 WALs to replay.  Replay took almost an hour.  It looks like it could run 
> faster that much of the time is spent zk'ing and nn'ing.
> Putting in 0.96 so it gets a look at least.  Can always punt.
> Distributed Log Replay Description:
> After a region server fails, we firstly assign a failed region to another 
> region server with recovering state marked in ZooKeeper. Then a 
> SplitLogWorker directly replays edits from WAL(Write-Ahead-Log)s of the 
> failed region server to the region after it's re-opened in the new location. 
> When a region is in recovering state, it can also accept writes but no 
> reads(including Append and Increment), region split or merge. 
> The feature piggybacks on existing distributed log splitting framework and 
> directly replay WAL edits to another region server instead of creating 
> recovered.edits files.
> The advantages over existing log splitting recovered edits implementation:
> 1) Eliminate the steps to write and read recovered.edits files. There could 
> be thousands of recovered.edits files are created and written concurrently 
> during a region server recovery. Many small random writes could degrade the 
> overall system performance.
> 2) Allow writes even when a region is in recovering state. It only takes 
> seconds for a failed over region to accept writes again. 
> The feature can be disabled by setting hbase.master.distributed.log.replay to 
> false(by default is true)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-8519) Backup master will never come up if primary master dies during initialization

2013-05-10 Thread Jerry He (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8519?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655040#comment-13655040
 ] 

Jerry He commented on HBASE-8519:
-

bq. If you kill the master while it's shutting down the cluster and you try to 
restart HBase, it will have the same error and will go down (and the cluster 
won't even come up).
Can you give more details on the sequence of events?
bq. All you need is the master and the shutdown znodes in place when starting 
HMaster.
If the shutdown/clusterStateZNode znode exists, we will be ok thru this code, 
correct?
{code}
  ..
  if (!clusterStatusTracker.isClusterUp()) {
  this.master.stop(
"Cluster went down before this master became active");
}
  ..
{code}
The master znode is ephemeral, and the clusterStateZNode is persistent.

> Backup master will never come up if primary master dies during initialization
> -
>
> Key: HBASE-8519
> URL: https://issues.apache.org/jira/browse/HBASE-8519
> Project: HBase
>  Issue Type: Bug
>  Components: master
>Affects Versions: 0.94.7, 0.95.0
>Reporter: Jerry He
>Assignee: Jerry He
>Priority: Minor
> Fix For: 0.98.0
>
>
> The problem happens if primary master dies after becoming master but before 
> it completes initialization and calls clusterStatusTracker.setClusterUp(),
> The backup master will try to become the master, but will shutdown itself 
> promptly because it sees 'the cluster is not up'.
> This is the backup master log:
> 2013-05-09 15:08:05,568 INFO 
> org.apache.hadoop.hbase.master.metrics.MasterMetrics: Initialized
> 2013-05-09 15:08:05,573 DEBUG org.apache.hadoop.hbase.master.HMaster: HMaster 
> started in backup mode.  Stalling until master znode is written.
> 2013-05-09 15:08:05,589 INFO 
> org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper: Node /hbase/master 
> already exists and this is not a retry
> 2013-05-09 15:08:05,590 INFO 
> org.apache.hadoop.hbase.master.ActiveMasterManager: Adding ZNode for 
> /hbase/backup-masters/xxx.com,6,1368137285373 in backup master directory
> 2013-05-09 15:08:05,595 INFO 
> org.apache.hadoop.hbase.master.ActiveMasterManager: Another master is the 
> active master, xxx.com,6,1368137283107; waiting to become the next active 
> master
> 2013-05-09 15:09:45,006 DEBUG 
> org.apache.hadoop.hbase.master.ActiveMasterManager: No master available. 
> Notifying waiting threads
> 2013-05-09 15:09:45,006 INFO org.apache.hadoop.hbase.master.HMaster: Cluster 
> went down before this master became active
> 2013-05-09 15:09:45,006 DEBUG org.apache.hadoop.hbase.master.HMaster: 
> Stopping service threads
> 2013-05-09 15:09:45,006 INFO org.apache.hadoop.ipc.HBaseServer: Stopping 
> server on 6
>  
> In ActiveMasterManager::blockUntilBecomingActiveMaster()
> {code}
>   ..
>   if (!clusterStatusTracker.isClusterUp()) {
>   this.master.stop(
> "Cluster went down before this master became active");
> }
>   ..
> {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HBASE-8517) Stochastic Loadbalancer isn't finding steady state on real clusters

2013-05-10 Thread Elliott Clark (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-8517?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Elliott Clark updated HBASE-8517:
-

Attachment: HBASE-8517-1.patch

Without the long line.

> Stochastic Loadbalancer isn't finding steady state on real clusters
> ---
>
> Key: HBASE-8517
> URL: https://issues.apache.org/jira/browse/HBASE-8517
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 0.95.0
>Reporter: Elliott Clark
>Assignee: Elliott Clark
> Attachments: HBASE-8517-0.patch, HBASE-8517-1.patch
>
>
> I have a cluster that runs IT tests.  Last night after all tests were done I 
> noticed that the balancer was thrashing regions around.
> The number of regions on each machine is not correct.
> The balancer seems to value the cost of moving a region way too little.
> {code}
> 2013-05-09 16:34:58,920 DEBUG [IPC Server handler 4 on 6] 
> org.apache.hadoop.hbase.master.balancer.StochasticLoadBalancer: Finished 
> computing new load balance plan.  Computation took 5367ms to try 8910 
> different iterations.  Found a solution that moves 37 regions; Going from a 
> computed cost of 56.50254222730425 to a new cost of 11.214035466575254
> 2013-05-09 16:37:48,715 DEBUG [IPC Server handler 7 on 6] 
> org.apache.hadoop.hbase.master.balancer.StochasticLoadBalancer: Finished 
> computing new load balance plan.  Computation took 4735ms to try 8910 
> different iterations.  Found a solution that moves 38 regions; Going from a 
> computed cost of 56.612624531830996 to a new cost of 11.275763861636982
> 2013-05-09 16:38:11,398 DEBUG [IPC Server handler 6 on 6] 
> org.apache.hadoop.hbase.master.balancer.StochasticLoadBalancer: Finished 
> computing new load balance plan.  Computation took 4502ms to try 8910 
> different iterations.  Found a solution that moves 39 regions; Going from a 
> computed cost of 56.50048461413552 to a new cost of 11.225352339003237
> {code}
> Each of those balancer runs were triggered when there was no load on the 
> cluster.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-8519) Backup master will never come up if primary master dies during initialization

2013-05-10 Thread Ted Yu (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8519?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655032#comment-13655032
 ] 

Ted Yu commented on HBASE-8519:
---

bq. Finally, this jira, you have both
I think Jerry's description mentioned that clusterStatusTracker.isClusterUp() 
returned false, leading to backup master stopping.

> Backup master will never come up if primary master dies during initialization
> -
>
> Key: HBASE-8519
> URL: https://issues.apache.org/jira/browse/HBASE-8519
> Project: HBase
>  Issue Type: Bug
>  Components: master
>Affects Versions: 0.94.7, 0.95.0
>Reporter: Jerry He
>Assignee: Jerry He
>Priority: Minor
> Fix For: 0.98.0
>
>
> The problem happens if primary master dies after becoming master but before 
> it completes initialization and calls clusterStatusTracker.setClusterUp(),
> The backup master will try to become the master, but will shutdown itself 
> promptly because it sees 'the cluster is not up'.
> This is the backup master log:
> 2013-05-09 15:08:05,568 INFO 
> org.apache.hadoop.hbase.master.metrics.MasterMetrics: Initialized
> 2013-05-09 15:08:05,573 DEBUG org.apache.hadoop.hbase.master.HMaster: HMaster 
> started in backup mode.  Stalling until master znode is written.
> 2013-05-09 15:08:05,589 INFO 
> org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper: Node /hbase/master 
> already exists and this is not a retry
> 2013-05-09 15:08:05,590 INFO 
> org.apache.hadoop.hbase.master.ActiveMasterManager: Adding ZNode for 
> /hbase/backup-masters/xxx.com,6,1368137285373 in backup master directory
> 2013-05-09 15:08:05,595 INFO 
> org.apache.hadoop.hbase.master.ActiveMasterManager: Another master is the 
> active master, xxx.com,6,1368137283107; waiting to become the next active 
> master
> 2013-05-09 15:09:45,006 DEBUG 
> org.apache.hadoop.hbase.master.ActiveMasterManager: No master available. 
> Notifying waiting threads
> 2013-05-09 15:09:45,006 INFO org.apache.hadoop.hbase.master.HMaster: Cluster 
> went down before this master became active
> 2013-05-09 15:09:45,006 DEBUG org.apache.hadoop.hbase.master.HMaster: 
> Stopping service threads
> 2013-05-09 15:09:45,006 INFO org.apache.hadoop.ipc.HBaseServer: Stopping 
> server on 6
>  
> In ActiveMasterManager::blockUntilBecomingActiveMaster()
> {code}
>   ..
>   if (!clusterStatusTracker.isClusterUp()) {
>   this.master.stop(
> "Cluster went down before this master became active");
> }
>   ..
> {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-8420) Port HBASE-6874 Implement prefetching for scanners from 0.89-fb

2013-05-10 Thread Jimmy Xiang (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8420?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655030#comment-13655030
 ] 

Jimmy Xiang commented on HBASE-8420:


The configurable max prefetched result size is set to 512 M by default. 
Whenever a scanner is called to fetch data, we will check if the caller thread 
is a prefetcher thread based on the thread name.  If it is a prefetcher thread, 
the result wire size is added to global atomic long variable (a counter).  
After the prefetched result is retrieved, the same amount is reduced from the 
counter.  If the counter value is over the pre-configured threshold, the next 
prefetching will not be scheduled.  This is a rough control.  We don't do micro 
control, i.e. stop a prefetcher thread in the middle.

> Port  HBASE-6874  Implement prefetching for scanners from 0.89-fb
> -
>
> Key: HBASE-8420
> URL: https://issues.apache.org/jira/browse/HBASE-8420
> Project: HBase
>  Issue Type: Improvement
>Reporter: Jimmy Xiang
>Assignee: Jimmy Xiang
> Attachments: 0.94-8420_v1.patch, 0.94-8420_v2.patch, 
> trunk-8420_v1.patch, trunk-8420_v2.patch, trunk-8420_v3.patch
>
>
> This should help scanner performance.  We should have it in trunk.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-6929) Publish Hbase 0.94 artifacts build against hadoop-2.0

2013-05-10 Thread Enis Soztutar (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-6929?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655026#comment-13655026
 ] 

Enis Soztutar commented on HBASE-6929:
--

I think Flume needs at least the hbase-server tests jar to use the minicluster. 
We currently do not have hbase-minicluster kind of module. 

> Publish Hbase 0.94 artifacts build against hadoop-2.0
> -
>
> Key: HBASE-6929
> URL: https://issues.apache.org/jira/browse/HBASE-6929
> Project: HBase
>  Issue Type: Task
>  Components: build
>Affects Versions: 0.94.2
>Reporter: Enis Soztutar
> Attachments: 6929.txt, hbase-6929_v2.patch
>
>
> Downstream projects (flume, hive, pig, etc) depends on hbase, but since the 
> hbase binaries build with hadoop-2.0 are not pushed to maven, they cannot 
> depend on them. AFAIK, hadoop 1 and 2 are not binary compatible, so we should 
> also push hbase jars build with hadoop2.0 profile into maven, possibly with 
> version string like 0.94.2-hadoop2.0. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-8527) clean up code around compaction completion in HStore

2013-05-10 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8527?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655020#comment-13655020
 ] 

Hadoop QA commented on HBASE-8527:
--

{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12582707/HBASE-8527-v0.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 3 new 
or modified tests.

{color:green}+1 hadoop1.0{color}.  The patch compiles against the hadoop 
1.0 profile.

{color:green}+1 hadoop2.0{color}.  The patch compiles against the hadoop 
2.0 profile.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 lineLengths{color}.  The patch does not introduce lines 
longer than 100

  {color:green}+1 site{color}.  The mvn site goal succeeds with this patch.

{color:green}+1 core tests{color}.  The patch passed unit tests in .

Test results: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5628//testReport/
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5628//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-prefix-tree.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5628//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-client.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5628//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-common.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5628//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-protocol.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5628//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-server.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5628//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-hadoop1-compat.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5628//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-examples.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5628//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-hadoop-compat.html
Console output: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5628//console

This message is automatically generated.

> clean up code around compaction completion in HStore
> 
>
> Key: HBASE-8527
> URL: https://issues.apache.org/jira/browse/HBASE-8527
> Project: HBase
>  Issue Type: Improvement
>Reporter: Sergey Shelukhin
>Assignee: Sergey Shelukhin
>Priority: Minor
> Attachments: HBASE-8527-v0.patch
>
>
> The methods completeCompaction and it's caller are too long. Something I 
> changed while doing something else there; putting in a separate 
> easy-to-review JIRA to make the other future change smaller.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-8517) Stochastic Loadbalancer isn't finding steady state on real clusters

2013-05-10 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8517?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655017#comment-13655017
 ] 

Hadoop QA commented on HBASE-8517:
--

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12582702/HBASE-8517-0.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 11 new 
or modified tests.

{color:green}+1 hadoop1.0{color}.  The patch compiles against the hadoop 
1.0 profile.

{color:green}+1 hadoop2.0{color}.  The patch compiles against the hadoop 
2.0 profile.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 lineLengths{color}.  The patch introduces lines longer than 
100

  {color:green}+1 site{color}.  The mvn site goal succeeds with this patch.

{color:green}+1 core tests{color}.  The patch passed unit tests in .

Test results: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5629//testReport/
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5629//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-client.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5629//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-examples.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5629//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-protocol.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5629//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-hadoop1-compat.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5629//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-prefix-tree.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5629//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-common.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5629//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-hadoop-compat.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5629//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-server.html
Console output: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5629//console

This message is automatically generated.

> Stochastic Loadbalancer isn't finding steady state on real clusters
> ---
>
> Key: HBASE-8517
> URL: https://issues.apache.org/jira/browse/HBASE-8517
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 0.95.0
>Reporter: Elliott Clark
>Assignee: Elliott Clark
> Attachments: HBASE-8517-0.patch
>
>
> I have a cluster that runs IT tests.  Last night after all tests were done I 
> noticed that the balancer was thrashing regions around.
> The number of regions on each machine is not correct.
> The balancer seems to value the cost of moving a region way too little.
> {code}
> 2013-05-09 16:34:58,920 DEBUG [IPC Server handler 4 on 6] 
> org.apache.hadoop.hbase.master.balancer.StochasticLoadBalancer: Finished 
> computing new load balance plan.  Computation took 5367ms to try 8910 
> different iterations.  Found a solution that moves 37 regions; Going from a 
> computed cost of 56.50254222730425 to a new cost of 11.214035466575254
> 2013-05-09 16:37:48,715 DEBUG [IPC Server handler 7 on 6] 
> org.apache.hadoop.hbase.master.balancer.StochasticLoadBalancer: Finished 
> computing new load balance plan.  Computation took 4735ms to try 8910 
> different iterations.  Found a solution that moves 38 regions; Going from a 
> computed cost of 56.612624531830996 to a new cost of 11.275763861636982
> 2013-05-09 16:38:11,398 DEBUG [IPC Server handler 6 on 6] 
> org.apache.hadoop.hbase.master.balancer.StochasticLoadBalancer: Finished 
> computing new load balance plan.  Computation took 4502ms to try 8910 
> different iterations.  Found a solution that moves 39 regions; Going from a 
> computed cost of 56.50048461413552 to a new cost of 11.225352339003237
> {code}
> Each of those balancer runs were triggered when there was no load on the 
> cluster.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more infor

[jira] [Updated] (HBASE-5995) Fix and reenable TestLogRolling.testLogRollOnPipelineRestart

2013-05-10 Thread Enis Soztutar (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-5995?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Enis Soztutar updated HBASE-5995:
-

Attachment: hbase-5995_v2.patch

Here is a second attempt. This patch also fixes a condition where we the hdfs 
output stream is closed because of errors, but we try to continuously do 
hflush() before closing the stream, although it is already closed. 
FSDataOutputStream does not have an isClosed() kind of API, or throw a special 
exception, so I had to parse the exception msg (which is ugly I admit). 

In hadoop2, we check the GS's of the block and replica to reason about the file 
length, which is why we get the error. In hadoop1, it seems that we do not do 
that, and except the replica length with checking GS (I might be wrong on this 
though, haven't confirmed it with hdfs folks). 

> Fix and reenable TestLogRolling.testLogRollOnPipelineRestart
> 
>
> Key: HBASE-5995
> URL: https://issues.apache.org/jira/browse/HBASE-5995
> Project: HBase
>  Issue Type: Sub-task
>  Components: test
>Reporter: stack
>Assignee: Enis Soztutar
>Priority: Blocker
> Fix For: 0.98.0, 0.95.1
>
> Attachments: hbase-5995_v1.patch, hbase-5995_v2.patch
>
>
> HBASE-5984 disabled this flakey test (See the issue for more).  This issue is 
> about getting it enabled again.  Made a blocker on 0.96.0 so it gets 
> attention.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-6929) Publish Hbase 0.94 artifacts build against hadoop-2.0

2013-05-10 Thread Hari Shreedharan (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-6929?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655007#comment-13655007
 ] 

Hari Shreedharan commented on HBASE-6929:
-

Roshan: we should not need anything more than the HBase client jars (I believe 
- not sure where the minicluster classes for tests are and what we'd need to 
handle that though). So we should have to do only what we do hadoop-common v/s 
hadoop-core.

> Publish Hbase 0.94 artifacts build against hadoop-2.0
> -
>
> Key: HBASE-6929
> URL: https://issues.apache.org/jira/browse/HBASE-6929
> Project: HBase
>  Issue Type: Task
>  Components: build
>Affects Versions: 0.94.2
>Reporter: Enis Soztutar
> Attachments: 6929.txt, hbase-6929_v2.patch
>
>
> Downstream projects (flume, hive, pig, etc) depends on hbase, but since the 
> hbase binaries build with hadoop-2.0 are not pushed to maven, they cannot 
> depend on them. AFAIK, hadoop 1 and 2 are not binary compatible, so we should 
> also push hbase jars build with hadoop2.0 profile into maven, possibly with 
> version string like 0.94.2-hadoop2.0. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HBASE-8528) [hadoop2] TestMultiTableInputFormat always fails on hadoop with YARN-129 applied

2013-05-10 Thread Gary Helmling (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-8528?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gary Helmling updated HBASE-8528:
-

Attachment: hbase-8528_alt.patch

Here's an alternate patch for trunk that sets yarn.is.minicluster.  
TestMultiTableInputFormat passes for me with this change.

> [hadoop2] TestMultiTableInputFormat always fails on hadoop with YARN-129 
> applied
> 
>
> Key: HBASE-8528
> URL: https://issues.apache.org/jira/browse/HBASE-8528
> Project: HBase
>  Issue Type: Sub-task
>  Components: hadoop2, mapreduce, test
>Affects Versions: 0.95.0
>Reporter: Jonathan Hsieh
>Assignee: Jonathan Hsieh
> Fix For: 0.98.0, 0.95.1
>
> Attachments: hbase-8528_alt.patch, hbase-8528.patch
>
>
> TestMulitTableInputFormat always fails if YARN-129 is applied.  Its error 
> message isn't useful (assertion failed) -- but if you look at the 
> stderr/stdout/syslog files of the jobs in 
> hbase-server/target/org.apache.hadoop.mapred.MiniMRCluster* (minicluster data 
> dirs) you'll find this class not found exception:
> {code}
> Exception in thread "main" java.lang.NoClassDefFoundError: 
> org/apache/hadoop/mapreduce/v2/app/MRAppMaster
> Caused by: java.lang.ClassNotFoundException: 
> org.apache.hadoop.mapreduce.v2.app.MRAppMaster
> at java.net.URLClassLoader$1.run(URLClassLoader.java:202)
> at java.security.AccessController.doPrivileged(Native Method)
> at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
> at java.lang.ClassLoader.loadClass(ClassLoader.java:306)
> at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
> at java.lang.ClassLoader.loadClass(ClassLoader.java:247)
> Could not find the main class: 
> org.apache.hadoop.mapreduce.v2.app.MRAppMaster.  Program will exit.
> {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-6929) Publish Hbase 0.94 artifacts build against hadoop-2.0

2013-05-10 Thread Roshan Naik (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-6929?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13654998#comment-13654998
 ] 

Roshan Naik commented on HBASE-6929:


there will be a little complexity ..  

Hbase jar has been split up in 0.95.. so in addition to the hadoop2 profile
flume (and others such prjs) will now additionally need a profile for hbase0.95.

the complexity is that neither the hadoop2 profile nor the hbase0.95+ profile 
can independently conjure up the full name and dependencies list for hbase.  

let me give it a try.. will get back.


> Publish Hbase 0.94 artifacts build against hadoop-2.0
> -
>
> Key: HBASE-6929
> URL: https://issues.apache.org/jira/browse/HBASE-6929
> Project: HBase
>  Issue Type: Task
>  Components: build
>Affects Versions: 0.94.2
>Reporter: Enis Soztutar
> Attachments: 6929.txt, hbase-6929_v2.patch
>
>
> Downstream projects (flume, hive, pig, etc) depends on hbase, but since the 
> hbase binaries build with hadoop-2.0 are not pushed to maven, they cannot 
> depend on them. AFAIK, hadoop 1 and 2 are not binary compatible, so we should 
> also push hbase jars build with hadoop2.0 profile into maven, possibly with 
> version string like 0.94.2-hadoop2.0. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-8528) [hadoop2] TestMultiTableInputFormat always fails on hadoop with YARN-129 applied

2013-05-10 Thread Gary Helmling (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8528?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13654997#comment-13654997
 ] 

Gary Helmling commented on HBASE-8528:
--

Coincidentally, I've just been looking at the exact same problem when running 
against 2.0.4-alpha.  Setting {{yarn.is.minicluster}} to "true" in the 
configuration solves the problem and in effect does the same as your patch (at 
least in 2.0.4-alpha).

> [hadoop2] TestMultiTableInputFormat always fails on hadoop with YARN-129 
> applied
> 
>
> Key: HBASE-8528
> URL: https://issues.apache.org/jira/browse/HBASE-8528
> Project: HBase
>  Issue Type: Sub-task
>  Components: hadoop2, mapreduce, test
>Affects Versions: 0.95.0
>Reporter: Jonathan Hsieh
>Assignee: Jonathan Hsieh
> Fix For: 0.98.0, 0.95.1
>
> Attachments: hbase-8528.patch
>
>
> TestMulitTableInputFormat always fails if YARN-129 is applied.  Its error 
> message isn't useful (assertion failed) -- but if you look at the 
> stderr/stdout/syslog files of the jobs in 
> hbase-server/target/org.apache.hadoop.mapred.MiniMRCluster* (minicluster data 
> dirs) you'll find this class not found exception:
> {code}
> Exception in thread "main" java.lang.NoClassDefFoundError: 
> org/apache/hadoop/mapreduce/v2/app/MRAppMaster
> Caused by: java.lang.ClassNotFoundException: 
> org.apache.hadoop.mapreduce.v2.app.MRAppMaster
> at java.net.URLClassLoader$1.run(URLClassLoader.java:202)
> at java.security.AccessController.doPrivileged(Native Method)
> at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
> at java.lang.ClassLoader.loadClass(ClassLoader.java:306)
> at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
> at java.lang.ClassLoader.loadClass(ClassLoader.java:247)
> Could not find the main class: 
> org.apache.hadoop.mapreduce.v2.app.MRAppMaster.  Program will exit.
> {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HBASE-8528) [hadoop2] TestMultiTableInputFormat always fails on hadoop with YARN-129 applied

2013-05-10 Thread Jonathan Hsieh (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-8528?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jonathan Hsieh updated HBASE-8528:
--

Status: Patch Available  (was: Open)

> [hadoop2] TestMultiTableInputFormat always fails on hadoop with YARN-129 
> applied
> 
>
> Key: HBASE-8528
> URL: https://issues.apache.org/jira/browse/HBASE-8528
> Project: HBase
>  Issue Type: Sub-task
>  Components: hadoop2, mapreduce, test
>Affects Versions: 0.95.0
>Reporter: Jonathan Hsieh
>Assignee: Jonathan Hsieh
> Fix For: 0.98.0, 0.95.1
>
> Attachments: hbase-8528.patch
>
>
> TestMulitTableInputFormat always fails if YARN-129 is applied.  Its error 
> message isn't useful (assertion failed) -- but if you look at the 
> stderr/stdout/syslog files of the jobs in 
> hbase-server/target/org.apache.hadoop.mapred.MiniMRCluster* (minicluster data 
> dirs) you'll find this class not found exception:
> {code}
> Exception in thread "main" java.lang.NoClassDefFoundError: 
> org/apache/hadoop/mapreduce/v2/app/MRAppMaster
> Caused by: java.lang.ClassNotFoundException: 
> org.apache.hadoop.mapreduce.v2.app.MRAppMaster
> at java.net.URLClassLoader$1.run(URLClassLoader.java:202)
> at java.security.AccessController.doPrivileged(Native Method)
> at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
> at java.lang.ClassLoader.loadClass(ClassLoader.java:306)
> at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
> at java.lang.ClassLoader.loadClass(ClassLoader.java:247)
> Could not find the main class: 
> org.apache.hadoop.mapreduce.v2.app.MRAppMaster.  Program will exit.
> {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HBASE-8375) Streamline Table durability settings

2013-05-10 Thread Enis Soztutar (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-8375?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Enis Soztutar updated HBASE-8375:
-

Release Note: Added Durability setting for table with 
HTableDescriptor.setDurability() API. Deprecated 
HTableDescriptor.setDeferredLogFlush(), and shell usage in favor of the new 
durability setting. Now, durability can be configured per table and per 
mutation. 

> Streamline Table durability settings
> 
>
> Key: HBASE-8375
> URL: https://issues.apache.org/jira/browse/HBASE-8375
> Project: HBase
>  Issue Type: Sub-task
>Reporter: Lars Hofhansl
>Assignee: Enis Soztutar
> Fix For: 0.98.0, 0.95.1
>
> Attachments: hbase-8375_v1.patch
>
>
> HBASE-7801 introduces the notion of per mutation fine grained durability 
> settings.
> This issue is to consider and the discuss the same for the per table settings 
> (i.e. what would be used if the mutation indicates USE_DEFAULT). I propose 
> the following setting per table:
> * SKIP_WAL (i.e. an unlogged table)
> * ASYNC_WAL (the current deferred log flush)
> * SYNC_WAL (the current default)
> * FSYNC_WAL (for future uses of HDFS' hsync())

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-8528) [hadoop2] TestMultiTableInputFormat always fails on hadoop with YARN-129 applied

2013-05-10 Thread Jonathan Hsieh (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8528?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13654983#comment-13654983
 ] 

Jonathan Hsieh commented on HBASE-8528:
---

Somehow the simplification of the yarn classpaths in YARN-129 doesn't include 
as much as before.  So we add unit tests's classpath into to the 
yarn.application.classpath so that the spawned children have the MR2 jars 
necessary to run.

> [hadoop2] TestMultiTableInputFormat always fails on hadoop with YARN-129 
> applied
> 
>
> Key: HBASE-8528
> URL: https://issues.apache.org/jira/browse/HBASE-8528
> Project: HBase
>  Issue Type: Sub-task
>  Components: hadoop2, mapreduce, test
>Affects Versions: 0.95.0
>Reporter: Jonathan Hsieh
>Assignee: Jonathan Hsieh
> Fix For: 0.98.0, 0.95.1
>
> Attachments: hbase-8528.patch
>
>
> TestMulitTableInputFormat always fails if YARN-129 is applied.  Its error 
> message isn't useful (assertion failed) -- but if you look at the 
> stderr/stdout/syslog files of the jobs in 
> hbase-server/target/org.apache.hadoop.mapred.MiniMRCluster* (minicluster data 
> dirs) you'll find this class not found exception:
> {code}
> Exception in thread "main" java.lang.NoClassDefFoundError: 
> org/apache/hadoop/mapreduce/v2/app/MRAppMaster
> Caused by: java.lang.ClassNotFoundException: 
> org.apache.hadoop.mapreduce.v2.app.MRAppMaster
> at java.net.URLClassLoader$1.run(URLClassLoader.java:202)
> at java.security.AccessController.doPrivileged(Native Method)
> at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
> at java.lang.ClassLoader.loadClass(ClassLoader.java:306)
> at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
> at java.lang.ClassLoader.loadClass(ClassLoader.java:247)
> Could not find the main class: 
> org.apache.hadoop.mapreduce.v2.app.MRAppMaster.  Program will exit.
> {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HBASE-8528) [hadoop2] TestMultiTableInputFormat always fails on hadoop with YARN-129 applied

2013-05-10 Thread Jonathan Hsieh (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-8528?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jonathan Hsieh updated HBASE-8528:
--

Issue Type: Sub-task  (was: Bug)
Parent: HBASE-6891

> [hadoop2] TestMultiTableInputFormat always fails on hadoop with YARN-129 
> applied
> 
>
> Key: HBASE-8528
> URL: https://issues.apache.org/jira/browse/HBASE-8528
> Project: HBase
>  Issue Type: Sub-task
>  Components: hadoop2, mapreduce, test
>Affects Versions: 0.95.0
>Reporter: Jonathan Hsieh
>Assignee: Jonathan Hsieh
> Fix For: 0.98.0, 0.95.1
>
> Attachments: hbase-8528.patch
>
>
> TestMulitTableInputFormat always fails if YARN-129 is applied.  Its error 
> message isn't useful (assertion failed) -- but if you look at the 
> stderr/stdout/syslog files of the jobs in 
> hbase-server/target/org.apache.hadoop.mapred.MiniMRCluster* (minicluster data 
> dirs) you'll find this class not found exception:
> {code}
> Exception in thread "main" java.lang.NoClassDefFoundError: 
> org/apache/hadoop/mapreduce/v2/app/MRAppMaster
> Caused by: java.lang.ClassNotFoundException: 
> org.apache.hadoop.mapreduce.v2.app.MRAppMaster
> at java.net.URLClassLoader$1.run(URLClassLoader.java:202)
> at java.security.AccessController.doPrivileged(Native Method)
> at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
> at java.lang.ClassLoader.loadClass(ClassLoader.java:306)
> at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
> at java.lang.ClassLoader.loadClass(ClassLoader.java:247)
> Could not find the main class: 
> org.apache.hadoop.mapreduce.v2.app.MRAppMaster.  Program will exit.
> {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HBASE-8375) Streamline Table durability settings

2013-05-10 Thread Enis Soztutar (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-8375?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Enis Soztutar updated HBASE-8375:
-

Status: Patch Available  (was: Open)

> Streamline Table durability settings
> 
>
> Key: HBASE-8375
> URL: https://issues.apache.org/jira/browse/HBASE-8375
> Project: HBase
>  Issue Type: Sub-task
>Reporter: Lars Hofhansl
>Assignee: Enis Soztutar
> Fix For: 0.95.2
>
> Attachments: hbase-8375_v1.patch
>
>
> HBASE-7801 introduces the notion of per mutation fine grained durability 
> settings.
> This issue is to consider and the discuss the same for the per table settings 
> (i.e. what would be used if the mutation indicates USE_DEFAULT). I propose 
> the following setting per table:
> * SKIP_WAL (i.e. an unlogged table)
> * ASYNC_WAL (the current deferred log flush)
> * SYNC_WAL (the current default)
> * FSYNC_WAL (for future uses of HDFS' hsync())

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HBASE-8375) Streamline Table durability settings

2013-05-10 Thread Enis Soztutar (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-8375?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Enis Soztutar updated HBASE-8375:
-

Fix Version/s: (was: 0.95.2)
   0.95.1
   0.98.0

> Streamline Table durability settings
> 
>
> Key: HBASE-8375
> URL: https://issues.apache.org/jira/browse/HBASE-8375
> Project: HBase
>  Issue Type: Sub-task
>Reporter: Lars Hofhansl
>Assignee: Enis Soztutar
> Fix For: 0.98.0, 0.95.1
>
> Attachments: hbase-8375_v1.patch
>
>
> HBASE-7801 introduces the notion of per mutation fine grained durability 
> settings.
> This issue is to consider and the discuss the same for the per table settings 
> (i.e. what would be used if the mutation indicates USE_DEFAULT). I propose 
> the following setting per table:
> * SKIP_WAL (i.e. an unlogged table)
> * ASYNC_WAL (the current deferred log flush)
> * SYNC_WAL (the current default)
> * FSYNC_WAL (for future uses of HDFS' hsync())

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-8375) Streamline Table durability settings

2013-05-10 Thread Enis Soztutar (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8375?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13654972#comment-13654972
 ] 

Enis Soztutar commented on HBASE-8375:
--

Put up a patch at https://reviews.apache.org/r/11064/. Reviews welcome, no rush 
:)
The patch does what the jira description talks about.  

> Streamline Table durability settings
> 
>
> Key: HBASE-8375
> URL: https://issues.apache.org/jira/browse/HBASE-8375
> Project: HBase
>  Issue Type: Sub-task
>Reporter: Lars Hofhansl
>Assignee: Enis Soztutar
> Fix For: 0.95.2
>
> Attachments: hbase-8375_v1.patch
>
>
> HBASE-7801 introduces the notion of per mutation fine grained durability 
> settings.
> This issue is to consider and the discuss the same for the per table settings 
> (i.e. what would be used if the mutation indicates USE_DEFAULT). I propose 
> the following setting per table:
> * SKIP_WAL (i.e. an unlogged table)
> * ASYNC_WAL (the current deferred log flush)
> * SYNC_WAL (the current default)
> * FSYNC_WAL (for future uses of HDFS' hsync())

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HBASE-8375) Streamline Table durability settings

2013-05-10 Thread Enis Soztutar (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-8375?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Enis Soztutar updated HBASE-8375:
-

Attachment: hbase-8375_v1.patch

> Streamline Table durability settings
> 
>
> Key: HBASE-8375
> URL: https://issues.apache.org/jira/browse/HBASE-8375
> Project: HBase
>  Issue Type: Sub-task
>Reporter: Lars Hofhansl
>Assignee: Enis Soztutar
> Fix For: 0.95.2
>
> Attachments: hbase-8375_v1.patch
>
>
> HBASE-7801 introduces the notion of per mutation fine grained durability 
> settings.
> This issue is to consider and the discuss the same for the per table settings 
> (i.e. what would be used if the mutation indicates USE_DEFAULT). I propose 
> the following setting per table:
> * SKIP_WAL (i.e. an unlogged table)
> * ASYNC_WAL (the current deferred log flush)
> * SYNC_WAL (the current default)
> * FSYNC_WAL (for future uses of HDFS' hsync())

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HBASE-8528) [hadoop2] TestMultiTableInputFormat always fails on hadoop with YARN-129 applied

2013-05-10 Thread Jonathan Hsieh (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-8528?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jonathan Hsieh updated HBASE-8528:
--

Attachment: hbase-8528.patch

Attached patch fixes problem for TestMultiTableInputFormat.  Waiting for QA bot 
to see if it affects and other tests.

> [hadoop2] TestMultiTableInputFormat always fails on hadoop with YARN-129 
> applied
> 
>
> Key: HBASE-8528
> URL: https://issues.apache.org/jira/browse/HBASE-8528
> Project: HBase
>  Issue Type: Bug
>  Components: hadoop2, mapreduce, test
>Affects Versions: 0.95.0
>Reporter: Jonathan Hsieh
>Assignee: Jonathan Hsieh
> Fix For: 0.98.0, 0.95.1
>
> Attachments: hbase-8528.patch
>
>
> TestMulitTableInputFormat always fails if YARN-129 is applied.  Its error 
> message isn't useful (assertion failed) -- but if you look at the 
> stderr/stdout/syslog files of the jobs in 
> hbase-server/target/org.apache.hadoop.mapred.MiniMRCluster* (minicluster data 
> dirs) you'll find this class not found exception:
> {code}
> Exception in thread "main" java.lang.NoClassDefFoundError: 
> org/apache/hadoop/mapreduce/v2/app/MRAppMaster
> Caused by: java.lang.ClassNotFoundException: 
> org.apache.hadoop.mapreduce.v2.app.MRAppMaster
> at java.net.URLClassLoader$1.run(URLClassLoader.java:202)
> at java.security.AccessController.doPrivileged(Native Method)
> at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
> at java.lang.ClassLoader.loadClass(ClassLoader.java:306)
> at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
> at java.lang.ClassLoader.loadClass(ClassLoader.java:247)
> Could not find the main class: 
> org.apache.hadoop.mapreduce.v2.app.MRAppMaster.  Program will exit.
> {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HBASE-8528) [hadoop2] TestMultiTableInputFormat always fails on hadoop with YARN-129 applied

2013-05-10 Thread Jonathan Hsieh (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-8528?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jonathan Hsieh updated HBASE-8528:
--

Description: 
TestMulitTableInputFormat always fails if YARN-129 is applied.  Its error 
message isn't useful (assertion failed) -- but if you look at the 
stderr/stdout/syslog files of the jobs in 
hbase-server/target/org.apache.hadoop.mapred.MiniMRCluster* (minicluster data 
dirs) you'll find this class not found exception:

{code}
Exception in thread "main" java.lang.NoClassDefFoundError: 
org/apache/hadoop/mapreduce/v2/app/MRAppMaster
Caused by: java.lang.ClassNotFoundException: 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster
at java.net.URLClassLoader$1.run(URLClassLoader.java:202)
at java.security.AccessController.doPrivileged(Native Method)
at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
at java.lang.ClassLoader.loadClass(ClassLoader.java:306)
at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
at java.lang.ClassLoader.loadClass(ClassLoader.java:247)
Could not find the main class: org.apache.hadoop.mapreduce.v2.app.MRAppMaster.  
Program will exit.
{code}

  was:
TestMulitTableInputFormat always fails.  Its error message isn't useful -- but 
if you look at the stderr/stdout/syslog files of the jobs in 
hbase-server/target/org.apache.hadoop.mapred.MiniMRCluster* (minicluster data 
dirs) you'll find this class not found exception:

{code}
Exception in thread "main" java.lang.NoClassDefFoundError: 
org/apache/hadoop/mapreduce/v2/app/MRAppMaster
Caused by: java.lang.ClassNotFoundException: 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster
at java.net.URLClassLoader$1.run(URLClassLoader.java:202)
at java.security.AccessController.doPrivileged(Native Method)
at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
at java.lang.ClassLoader.loadClass(ClassLoader.java:306)
at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
at java.lang.ClassLoader.loadClass(ClassLoader.java:247)
Could not find the main class: org.apache.hadoop.mapreduce.v2.app.MRAppMaster.  
Program will exit.
{code}


> [hadoop2] TestMultiTableInputFormat always fails on hadoop with YARN-129 
> applied
> 
>
> Key: HBASE-8528
> URL: https://issues.apache.org/jira/browse/HBASE-8528
> Project: HBase
>  Issue Type: Bug
>  Components: hadoop2, mapreduce, test
>Affects Versions: 0.95.0
>Reporter: Jonathan Hsieh
>Assignee: Jonathan Hsieh
> Fix For: 0.98.0, 0.95.1
>
>
> TestMulitTableInputFormat always fails if YARN-129 is applied.  Its error 
> message isn't useful (assertion failed) -- but if you look at the 
> stderr/stdout/syslog files of the jobs in 
> hbase-server/target/org.apache.hadoop.mapred.MiniMRCluster* (minicluster data 
> dirs) you'll find this class not found exception:
> {code}
> Exception in thread "main" java.lang.NoClassDefFoundError: 
> org/apache/hadoop/mapreduce/v2/app/MRAppMaster
> Caused by: java.lang.ClassNotFoundException: 
> org.apache.hadoop.mapreduce.v2.app.MRAppMaster
> at java.net.URLClassLoader$1.run(URLClassLoader.java:202)
> at java.security.AccessController.doPrivileged(Native Method)
> at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
> at java.lang.ClassLoader.loadClass(ClassLoader.java:306)
> at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
> at java.lang.ClassLoader.loadClass(ClassLoader.java:247)
> Could not find the main class: 
> org.apache.hadoop.mapreduce.v2.app.MRAppMaster.  Program will exit.
> {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HBASE-8528) [hadoop2] TestMultiTableInputFormat always fails on hadoop with YARN-129 applied

2013-05-10 Thread Jonathan Hsieh (JIRA)
Jonathan Hsieh created HBASE-8528:
-

 Summary: [hadoop2] TestMultiTableInputFormat always fails on 
hadoop with YARN-129 applied
 Key: HBASE-8528
 URL: https://issues.apache.org/jira/browse/HBASE-8528
 Project: HBase
  Issue Type: Bug
  Components: hadoop2, mapreduce, test
Affects Versions: 0.95.0
Reporter: Jonathan Hsieh
Assignee: Jonathan Hsieh
 Fix For: 0.98.0, 0.95.1


TestMulitTableInputFormat always fails.  Its error message isn't useful -- but 
if you look at the stderr/stdout/syslog files of the jobs in 
hbase-server/target/org.apache.hadoop.mapred.MiniMRCluster* (minicluster data 
dirs) you'll find this class not found exception:

{code}
Exception in thread "main" java.lang.NoClassDefFoundError: 
org/apache/hadoop/mapreduce/v2/app/MRAppMaster
Caused by: java.lang.ClassNotFoundException: 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster
at java.net.URLClassLoader$1.run(URLClassLoader.java:202)
at java.security.AccessController.doPrivileged(Native Method)
at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
at java.lang.ClassLoader.loadClass(ClassLoader.java:306)
at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
at java.lang.ClassLoader.loadClass(ClassLoader.java:247)
Could not find the main class: org.apache.hadoop.mapreduce.v2.app.MRAppMaster.  
Program will exit.
{code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-8483) HConnectionManager can leak ZooKeeper connections when using deleteStaleConnection

2013-05-10 Thread Ted Yu (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8483?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13654964#comment-13654964
 ] 

Ted Yu commented on HBASE-8483:
---

In 
https://builds.apache.org/job/HBase-TRUNK/4109/testReport/org.apache.hadoop.hbase.client/TestHCM/testDeleteForZKConnLeak/
 :
{code}
java.lang.AssertionError: Live zookeeper in closed connection
at org.junit.Assert.fail(Assert.java:88)
at 
org.apache.hadoop.hbase.client.TestHCM.testDeleteForZKConnLeak(TestHCM.java:851)
{code}

> HConnectionManager can leak ZooKeeper connections when using 
> deleteStaleConnection
> --
>
> Key: HBASE-8483
> URL: https://issues.apache.org/jira/browse/HBASE-8483
> Project: HBase
>  Issue Type: Bug
>  Components: Client
>Affects Versions: 0.94.4
>Reporter: Eric Yu
>Priority: Critical
> Fix For: 0.98.0, 0.94.8, 0.95.1
>
> Attachments: HBASE-8483-0.94.patch, HBASE-8483.patch, 
> LeakZKConnections.java
>
>
> If one thread calls deleteStaleConnection while other threads are using 
> connection, can leak ZK connections.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HBASE-8154) HBaseTestingUtility.startMiniDFSClusterForTestHLog should handle "Address already in use" and retry automatically

2013-05-10 Thread Ted Yu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-8154?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu updated HBASE-8154:
--

Attachment: 8154-v4.txt

> HBaseTestingUtility.startMiniDFSClusterForTestHLog should handle "Address 
> already in use" and retry automatically
> -
>
> Key: HBASE-8154
> URL: https://issues.apache.org/jira/browse/HBASE-8154
> Project: HBase
>  Issue Type: Test
>  Components: test
>Affects Versions: 0.98.0
>Reporter: Liang Xie
>Assignee: Ted Yu
> Attachments: 8154-v2.txt, 8154-v3.txt, 8154-v4.txt
>
>
> see 
> https://builds.apache.org/job/PreCommit-HBASE-Build/4906//testReport/org.apache.hadoop.hbase.regionserver.wal/TestHLog/testAppendClose/
> and 
> https://issues.apache.org/jira/browse/HBASE-7845?focusedCommentId=13607468&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-13607468

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HBASE-8154) HBaseTestingUtility.startMiniDFSClusterForTestHLog should handle "Address already in use" and retry automatically

2013-05-10 Thread Ted Yu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-8154?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu updated HBASE-8154:
--

Attachment: (was: 8154-v4.txt)

> HBaseTestingUtility.startMiniDFSClusterForTestHLog should handle "Address 
> already in use" and retry automatically
> -
>
> Key: HBASE-8154
> URL: https://issues.apache.org/jira/browse/HBASE-8154
> Project: HBase
>  Issue Type: Test
>  Components: test
>Affects Versions: 0.98.0
>Reporter: Liang Xie
>Assignee: Ted Yu
> Attachments: 8154-v2.txt, 8154-v3.txt, 8154-v4.txt
>
>
> see 
> https://builds.apache.org/job/PreCommit-HBASE-Build/4906//testReport/org.apache.hadoop.hbase.regionserver.wal/TestHLog/testAppendClose/
> and 
> https://issues.apache.org/jira/browse/HBASE-7845?focusedCommentId=13607468&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-13607468

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HBASE-8154) HBaseTestingUtility.startMiniDFSClusterForTestHLog should handle "Address already in use" and retry automatically

2013-05-10 Thread Ted Yu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-8154?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu updated HBASE-8154:
--

Attachment: (was: 8154-v4.txt)

> HBaseTestingUtility.startMiniDFSClusterForTestHLog should handle "Address 
> already in use" and retry automatically
> -
>
> Key: HBASE-8154
> URL: https://issues.apache.org/jira/browse/HBASE-8154
> Project: HBase
>  Issue Type: Test
>  Components: test
>Affects Versions: 0.98.0
>Reporter: Liang Xie
>Assignee: Ted Yu
> Attachments: 8154-v2.txt, 8154-v3.txt, 8154-v4.txt
>
>
> see 
> https://builds.apache.org/job/PreCommit-HBASE-Build/4906//testReport/org.apache.hadoop.hbase.regionserver.wal/TestHLog/testAppendClose/
> and 
> https://issues.apache.org/jira/browse/HBASE-7845?focusedCommentId=13607468&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-13607468

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-8520) TestIOFencing fails intermittently due to compaction kicking in too soon

2013-05-10 Thread Ted Yu (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8520?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13654955#comment-13654955
 ] 

Ted Yu commented on HBASE-8520:
---

Integrated to 0.95 and trunk.

> TestIOFencing fails intermittently due to compaction kicking in too soon
> 
>
> Key: HBASE-8520
> URL: https://issues.apache.org/jira/browse/HBASE-8520
> Project: HBase
>  Issue Type: Test
>Reporter: Ted Yu
>Assignee: Ted Yu
> Fix For: 0.98.0, 0.95.1
>
> Attachments: 8520-v1.txt
>
>
> Here is one example: 
> https://builds.apache.org/job/HBase-TRUNK/lastCompletedBuild/testReport/org.apache.hadoop.hbase/TestIOFencing/testFencingAroundCompactionAfterWALSync/
> It failed on this assertion:
> {code}
>   assertTrue(compactingRegion.countStoreFiles() > 1);
> {code}
> I searched for "Blocking compactions" but didn't find it in the test output.
> I think the compaction kicked in too soon, resulting in store file count 
> being 1.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-8420) Port HBASE-6874 Implement prefetching for scanners from 0.89-fb

2013-05-10 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8420?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13654937#comment-13654937
 ] 

Hadoop QA commented on HBASE-8420:
--

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12582686/trunk-8420_v3.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 12 new 
or modified tests.

{color:green}+1 hadoop1.0{color}.  The patch compiles against the hadoop 
1.0 profile.

{color:green}+1 hadoop2.0{color}.  The patch compiles against the hadoop 
2.0 profile.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 lineLengths{color}.  The patch introduces lines longer than 
100

  {color:green}+1 site{color}.  The mvn site goal succeeds with this patch.

 {color:red}-1 core tests{color}.  The patch failed these unit tests:
 

 {color:red}-1 core zombie tests{color}.  There are 2 zombie test(s):   
at 
org.apache.hadoop.hbase.client.TestHCM.testDeleteForZKConnLeak(TestHCM.java:837)

Test results: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5626//testReport/
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5626//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-prefix-tree.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5626//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-client.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5626//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-common.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5626//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-protocol.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5626//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-server.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5626//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-hadoop1-compat.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5626//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-examples.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5626//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-hadoop-compat.html
Console output: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5626//console

This message is automatically generated.

> Port  HBASE-6874  Implement prefetching for scanners from 0.89-fb
> -
>
> Key: HBASE-8420
> URL: https://issues.apache.org/jira/browse/HBASE-8420
> Project: HBase
>  Issue Type: Improvement
>Reporter: Jimmy Xiang
>Assignee: Jimmy Xiang
> Attachments: 0.94-8420_v1.patch, 0.94-8420_v2.patch, 
> trunk-8420_v1.patch, trunk-8420_v2.patch, trunk-8420_v3.patch
>
>
> This should help scanner performance.  We should have it in trunk.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-6929) Publish Hbase 0.94 artifacts build against hadoop-2.0

2013-05-10 Thread Hari Shreedharan (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-6929?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13654934#comment-13654934
 ] 

Hari Shreedharan commented on HBASE-6929:
-

[~saint@gmail.com], This should work for Flume. We have different mvn 
profiles for hadoop-1 and hadoop-2. So we can just change the hadoop-2 profile 
to pull in the correct artifact (we already do this for 
hadoop-common/hadoop-core).

> Publish Hbase 0.94 artifacts build against hadoop-2.0
> -
>
> Key: HBASE-6929
> URL: https://issues.apache.org/jira/browse/HBASE-6929
> Project: HBase
>  Issue Type: Task
>  Components: build
>Affects Versions: 0.94.2
>Reporter: Enis Soztutar
> Attachments: 6929.txt, hbase-6929_v2.patch
>
>
> Downstream projects (flume, hive, pig, etc) depends on hbase, but since the 
> hbase binaries build with hadoop-2.0 are not pushed to maven, they cannot 
> depend on them. AFAIK, hadoop 1 and 2 are not binary compatible, so we should 
> also push hbase jars build with hadoop2.0 profile into maven, possibly with 
> version string like 0.94.2-hadoop2.0. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HBASE-8517) Stochastic Loadbalancer isn't finding steady state on real clusters

2013-05-10 Thread Elliott Clark (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-8517?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Elliott Clark updated HBASE-8517:
-

Affects Version/s: 0.95.0
   Status: Patch Available  (was: Open)

> Stochastic Loadbalancer isn't finding steady state on real clusters
> ---
>
> Key: HBASE-8517
> URL: https://issues.apache.org/jira/browse/HBASE-8517
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 0.95.0
>Reporter: Elliott Clark
>Assignee: Elliott Clark
> Attachments: HBASE-8517-0.patch
>
>
> I have a cluster that runs IT tests.  Last night after all tests were done I 
> noticed that the balancer was thrashing regions around.
> The number of regions on each machine is not correct.
> The balancer seems to value the cost of moving a region way too little.
> {code}
> 2013-05-09 16:34:58,920 DEBUG [IPC Server handler 4 on 6] 
> org.apache.hadoop.hbase.master.balancer.StochasticLoadBalancer: Finished 
> computing new load balance plan.  Computation took 5367ms to try 8910 
> different iterations.  Found a solution that moves 37 regions; Going from a 
> computed cost of 56.50254222730425 to a new cost of 11.214035466575254
> 2013-05-09 16:37:48,715 DEBUG [IPC Server handler 7 on 6] 
> org.apache.hadoop.hbase.master.balancer.StochasticLoadBalancer: Finished 
> computing new load balance plan.  Computation took 4735ms to try 8910 
> different iterations.  Found a solution that moves 38 regions; Going from a 
> computed cost of 56.612624531830996 to a new cost of 11.275763861636982
> 2013-05-09 16:38:11,398 DEBUG [IPC Server handler 6 on 6] 
> org.apache.hadoop.hbase.master.balancer.StochasticLoadBalancer: Finished 
> computing new load balance plan.  Computation took 4502ms to try 8910 
> different iterations.  Found a solution that moves 39 regions; Going from a 
> computed cost of 56.50048461413552 to a new cost of 11.225352339003237
> {code}
> Each of those balancer runs were triggered when there was no load on the 
> cluster.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-8154) HBaseTestingUtility.startMiniDFSClusterForTestHLog should handle "Address already in use" and retry automatically

2013-05-10 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8154?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13654927#comment-13654927
 ] 

Hadoop QA commented on HBASE-8154:
--

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12582688/8154-v4.txt
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 6 new 
or modified tests.

{color:green}+1 hadoop1.0{color}.  The patch compiles against the hadoop 
1.0 profile.

{color:green}+1 hadoop2.0{color}.  The patch compiles against the hadoop 
2.0 profile.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 lineLengths{color}.  The patch does not introduce lines 
longer than 100

  {color:green}+1 site{color}.  The mvn site goal succeeds with this patch.

 {color:red}-1 core tests{color}.  The patch failed these unit tests:
   org.apache.hadoop.hbase.regionserver.wal.TestHLog

 {color:red}-1 core zombie tests{color}.  There are 1 zombie test(s): 

Test results: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5627//testReport/
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5627//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-client.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5627//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-examples.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5627//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-protocol.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5627//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-hadoop1-compat.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5627//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-prefix-tree.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5627//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-common.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5627//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-hadoop-compat.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5627//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-server.html
Console output: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5627//console

This message is automatically generated.

> HBaseTestingUtility.startMiniDFSClusterForTestHLog should handle "Address 
> already in use" and retry automatically
> -
>
> Key: HBASE-8154
> URL: https://issues.apache.org/jira/browse/HBASE-8154
> Project: HBase
>  Issue Type: Test
>  Components: test
>Affects Versions: 0.98.0
>Reporter: Liang Xie
>Assignee: Ted Yu
> Attachments: 8154-v2.txt, 8154-v3.txt, 8154-v4.txt, 8154-v4.txt
>
>
> see 
> https://builds.apache.org/job/PreCommit-HBASE-Build/4906//testReport/org.apache.hadoop.hbase.regionserver.wal/TestHLog/testAppendClose/
> and 
> https://issues.apache.org/jira/browse/HBASE-7845?focusedCommentId=13607468&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-13607468

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HBASE-8527) clean up code around compaction completion in HStore

2013-05-10 Thread Sergey Shelukhin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-8527?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Shelukhin updated HBASE-8527:


Status: Patch Available  (was: Open)

> clean up code around compaction completion in HStore
> 
>
> Key: HBASE-8527
> URL: https://issues.apache.org/jira/browse/HBASE-8527
> Project: HBase
>  Issue Type: Improvement
>Reporter: Sergey Shelukhin
>Assignee: Sergey Shelukhin
>Priority: Minor
> Attachments: HBASE-8527-v0.patch
>
>
> The methods completeCompaction and it's caller are too long. Something I 
> changed while doing something else there; putting in a separate 
> easy-to-review JIRA to make the other future change smaller.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HBASE-8527) clean up code around compaction completion in HStore

2013-05-10 Thread Sergey Shelukhin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-8527?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Shelukhin updated HBASE-8527:


Attachment: HBASE-8527-v0.patch

Move some stuff into methods, mostly

> clean up code around compaction completion in HStore
> 
>
> Key: HBASE-8527
> URL: https://issues.apache.org/jira/browse/HBASE-8527
> Project: HBase
>  Issue Type: Improvement
>Reporter: Sergey Shelukhin
>Assignee: Sergey Shelukhin
>Priority: Minor
> Attachments: HBASE-8527-v0.patch
>
>
> The methods completeCompaction and it's caller are too long. Something I 
> changed while doing something else there; putting in a separate 
> easy-to-review JIRA to make the other future change smaller.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HBASE-8517) Stochastic Loadbalancer isn't finding steady state on real clusters

2013-05-10 Thread Elliott Clark (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-8517?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Elliott Clark updated HBASE-8517:
-

Attachment: HBASE-8517-0.patch

Here's a patch that solves the issue I was seeing on a cluster.

Most of the issue was upping the move cost so that all cost functions below it 
couldn't overcome the cost of a move.  In other words the balancer will find 
the best moves when the region load is our of wack, but it won't move just to 
make locality better.

Then I also took a crack at making this thing work on larger clusters.  

To do that I made it closer to simulated annealing.  It picks either a random 
server, or the most loaded server.  This means that more often we have a chance 
to work on the worst balanced server.

Then I also made the algorithm prefer to move regions off of the over loaded 
servers.

Then I made the max number of moves scale with cluster size. 

> Stochastic Loadbalancer isn't finding steady state on real clusters
> ---
>
> Key: HBASE-8517
> URL: https://issues.apache.org/jira/browse/HBASE-8517
> Project: HBase
>  Issue Type: Bug
>Reporter: Elliott Clark
>Assignee: Elliott Clark
> Attachments: HBASE-8517-0.patch
>
>
> I have a cluster that runs IT tests.  Last night after all tests were done I 
> noticed that the balancer was thrashing regions around.
> The number of regions on each machine is not correct.
> The balancer seems to value the cost of moving a region way too little.
> {code}
> 2013-05-09 16:34:58,920 DEBUG [IPC Server handler 4 on 6] 
> org.apache.hadoop.hbase.master.balancer.StochasticLoadBalancer: Finished 
> computing new load balance plan.  Computation took 5367ms to try 8910 
> different iterations.  Found a solution that moves 37 regions; Going from a 
> computed cost of 56.50254222730425 to a new cost of 11.214035466575254
> 2013-05-09 16:37:48,715 DEBUG [IPC Server handler 7 on 6] 
> org.apache.hadoop.hbase.master.balancer.StochasticLoadBalancer: Finished 
> computing new load balance plan.  Computation took 4735ms to try 8910 
> different iterations.  Found a solution that moves 38 regions; Going from a 
> computed cost of 56.612624531830996 to a new cost of 11.275763861636982
> 2013-05-09 16:38:11,398 DEBUG [IPC Server handler 6 on 6] 
> org.apache.hadoop.hbase.master.balancer.StochasticLoadBalancer: Finished 
> computing new load balance plan.  Computation took 4502ms to try 8910 
> different iterations.  Found a solution that moves 39 regions; Going from a 
> computed cost of 56.50048461413552 to a new cost of 11.225352339003237
> {code}
> Each of those balancer runs were triggered when there was no load on the 
> cluster.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-7006) [MTTR] Improve Region Server Recovery Time - Distributed Log Replay

2013-05-10 Thread stack (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-7006?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13654874#comment-13654874
 ] 

stack commented on HBASE-7006:
--

Where are we w/ this [~jeffreyz]?  What is your sense?  Can you summarize.  Are 
all the reviews up in rb addressed in this v7?

> [MTTR] Improve Region Server Recovery Time - Distributed Log Replay
> ---
>
> Key: HBASE-7006
> URL: https://issues.apache.org/jira/browse/HBASE-7006
> Project: HBase
>  Issue Type: New Feature
>  Components: MTTR
>Reporter: stack
>Assignee: Jeffrey Zhong
>Priority: Critical
> Fix For: 0.95.1
>
> Attachments: hbase-7006-combined.patch, hbase-7006-combined-v1.patch, 
> hbase-7006-combined-v4.patch, hbase-7006-combined-v5.patch, 
> hbase-7006-combined-v6.patch, hbase-7006-combined-v7.patch, LogSplitting 
> Comparison.pdf, 
> ProposaltoimprovelogsplittingprocessregardingtoHBASE-7006-v2.pdf
>
>
> Just saw interesting issue where a cluster went down  hard and 30 nodes had 
> 1700 WALs to replay.  Replay took almost an hour.  It looks like it could run 
> faster that much of the time is spent zk'ing and nn'ing.
> Putting in 0.96 so it gets a look at least.  Can always punt.
> Distributed Log Replay Description:
> After a region server fails, we firstly assign a failed region to another 
> region server with recovering state marked in ZooKeeper. Then a 
> SplitLogWorker directly replays edits from WAL(Write-Ahead-Log)s of the 
> failed region server to the region after it's re-opened in the new location. 
> When a region is in recovering state, it can also accept writes but no 
> reads(including Append and Increment), region split or merge. 
> The feature piggybacks on existing distributed log splitting framework and 
> directly replay WAL edits to another region server instead of creating 
> recovered.edits files.
> The advantages over existing log splitting recovered edits implementation:
> 1) Eliminate the steps to write and read recovered.edits files. There could 
> be thousands of recovered.edits files are created and written concurrently 
> during a region server recovery. Many small random writes could degrade the 
> overall system performance.
> 2) Allow writes even when a region is in recovering state. It only takes 
> seconds for a failed over region to accept writes again. 
> The feature can be disabled by setting hbase.master.distributed.log.replay to 
> false(by default is true)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-6929) Publish Hbase 0.94 artifacts build against hadoop-2.0

2013-05-10 Thread stack (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-6929?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13654870#comment-13654870
 ] 

stack commented on HBASE-6929:
--

[~roshan_naik] Here is our attempt: 
https://repository.apache.org/content/groups/snapshots/org/apache/hbase/hbase-client/
  What do you think?  Will it work for you?  We publish two artifacts, one 
named w/ a hadoop1 suffix and the other with a hadoop2.   Would be interested 
in feedback.  What I have so far is that this scheme will not work (our Elliott 
has it our hadoop dependency is resolved before we have a chance to influence 
it -- profiles).

> Publish Hbase 0.94 artifacts build against hadoop-2.0
> -
>
> Key: HBASE-6929
> URL: https://issues.apache.org/jira/browse/HBASE-6929
> Project: HBase
>  Issue Type: Task
>  Components: build
>Affects Versions: 0.94.2
>Reporter: Enis Soztutar
> Attachments: 6929.txt, hbase-6929_v2.patch
>
>
> Downstream projects (flume, hive, pig, etc) depends on hbase, but since the 
> hbase binaries build with hadoop-2.0 are not pushed to maven, they cannot 
> depend on them. AFAIK, hadoop 1 and 2 are not binary compatible, so we should 
> also push hbase jars build with hadoop2.0 profile into maven, possibly with 
> version string like 0.94.2-hadoop2.0. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Comment Edited] (HBASE-8420) Port HBASE-6874 Implement prefetching for scanners from 0.89-fb

2013-05-10 Thread Jimmy Xiang (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8420?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13654866#comment-13654866
 ] 

Jimmy Xiang edited comment on HBASE-8420 at 5/10/13 9:30 PM:
-

Did some quick scan testing with YCSB on a same cluster: 1 master node + 4 
region server nodes.  All are the same except the hbase jar file.
Total rows: 3,000,000.  Scan ops count: 3,000 (not that many).

With prefetching (v2):

throughput: 36.9 ops/sec, avg latency: 27.0 ms, 95th: 48 ms, 99th: 69 ms

Without prefetching (without the patch at all):

throughput: 32.8 ops/sec, avg latency: 30.4 ms, 95th: 50 ms, 99th: 227 ms

On average, the latency is about 10% shorter.


  was (Author: jxiang):
Did some quest scan testing with YCSB on a same cluster: 1 master node + 4 
region server nodes.  All are the same except the hbase jar file.
Total rows: 3,000,000.  Scan ops count: 3,000 (not that many).

With prefetching (v2):

throughput: 36.9 ops/sec, avg latency: 27.0 ms, 95th: 48 ms, 99th: 69 ms

Without prefetching (without the patch at all):

throughput: 32.8 ops/sec, avg latency: 30.4 ms, 95th: 50 ms, 99th: 227 ms

On average, the latency is about 10% shorter.

  
> Port  HBASE-6874  Implement prefetching for scanners from 0.89-fb
> -
>
> Key: HBASE-8420
> URL: https://issues.apache.org/jira/browse/HBASE-8420
> Project: HBase
>  Issue Type: Improvement
>Reporter: Jimmy Xiang
>Assignee: Jimmy Xiang
> Attachments: 0.94-8420_v1.patch, 0.94-8420_v2.patch, 
> trunk-8420_v1.patch, trunk-8420_v2.patch, trunk-8420_v3.patch
>
>
> This should help scanner performance.  We should have it in trunk.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-8420) Port HBASE-6874 Implement prefetching for scanners from 0.89-fb

2013-05-10 Thread Jimmy Xiang (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8420?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13654866#comment-13654866
 ] 

Jimmy Xiang commented on HBASE-8420:


Did some quest scan testing with YCSB on a same cluster: 1 master node + 4 
region server nodes.  All are the same except the hbase jar file.
Total rows: 3,000,000.  Scan ops count: 3,000 (not that many).

With prefetching (v2):

throughput: 36.9 ops/sec, avg latency: 27.0 ms, 95th: 48 ms, 99th: 69 ms

Without prefetching (without the patch at all):

throughput: 32.8 ops/sec, avg latency: 30.4 ms, 95th: 50 ms, 99th: 227 ms

On average, the latency is about 10% shorter.


> Port  HBASE-6874  Implement prefetching for scanners from 0.89-fb
> -
>
> Key: HBASE-8420
> URL: https://issues.apache.org/jira/browse/HBASE-8420
> Project: HBase
>  Issue Type: Improvement
>Reporter: Jimmy Xiang
>Assignee: Jimmy Xiang
> Attachments: 0.94-8420_v1.patch, 0.94-8420_v2.patch, 
> trunk-8420_v1.patch, trunk-8420_v2.patch, trunk-8420_v3.patch
>
>
> This should help scanner performance.  We should have it in trunk.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-8519) Backup master will never come up if primary master dies during initialization

2013-05-10 Thread Jean-Daniel Cryans (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8519?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13654859#comment-13654859
 ] 

Jean-Daniel Cryans commented on HBASE-8519:
---

What this jira describes is not the only failure mode. If you kill the master 
while it's shutting down the cluster and you try to restart HBase, it will have 
the same error and will go down (and the cluster won't even come up).

All you need is the master and the shutdown znodes in place when starting 
HMaster.

There's a weird misconnection here:

 # If you start the cluster on a clean ZK, you won't find either znodes so you 
just start.
 # If you start the cluster and the shutdown znode exists but not the master 
znode, *you just clean it*:
{code}
// Set the cluster as up.  If new RSs, they'll be waiting on this before
// going ahead with their startup.
boolean wasUp = this.clusterStatusTracker.isClusterUp();
if (!wasUp) this.clusterStatusTracker.setClusterUp();
{code}
 # If you start the cluster and the master znode exists but not the shutdown, 
you are a backup master.
 # Finally, this jira, you have both so you assume the cluster is shutting down 
and you were meant to be a backup cluster.

The tricky part here is that 2 and 4 are almost the same, except that you want 
to handle the case where the whole cluster is shutting down while you're 
waiting. Maybe we could check if both znodes are there before waiting and then 
recognize the situation as being a start and not a stop, but I'm sure that's 
going to screw someone at some point too...

Comments?

> Backup master will never come up if primary master dies during initialization
> -
>
> Key: HBASE-8519
> URL: https://issues.apache.org/jira/browse/HBASE-8519
> Project: HBase
>  Issue Type: Bug
>  Components: master
>Affects Versions: 0.94.7, 0.95.0
>Reporter: Jerry He
>Assignee: Jerry He
>Priority: Minor
> Fix For: 0.98.0
>
>
> The problem happens if primary master dies after becoming master but before 
> it completes initialization and calls clusterStatusTracker.setClusterUp(),
> The backup master will try to become the master, but will shutdown itself 
> promptly because it sees 'the cluster is not up'.
> This is the backup master log:
> 2013-05-09 15:08:05,568 INFO 
> org.apache.hadoop.hbase.master.metrics.MasterMetrics: Initialized
> 2013-05-09 15:08:05,573 DEBUG org.apache.hadoop.hbase.master.HMaster: HMaster 
> started in backup mode.  Stalling until master znode is written.
> 2013-05-09 15:08:05,589 INFO 
> org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper: Node /hbase/master 
> already exists and this is not a retry
> 2013-05-09 15:08:05,590 INFO 
> org.apache.hadoop.hbase.master.ActiveMasterManager: Adding ZNode for 
> /hbase/backup-masters/xxx.com,6,1368137285373 in backup master directory
> 2013-05-09 15:08:05,595 INFO 
> org.apache.hadoop.hbase.master.ActiveMasterManager: Another master is the 
> active master, xxx.com,6,1368137283107; waiting to become the next active 
> master
> 2013-05-09 15:09:45,006 DEBUG 
> org.apache.hadoop.hbase.master.ActiveMasterManager: No master available. 
> Notifying waiting threads
> 2013-05-09 15:09:45,006 INFO org.apache.hadoop.hbase.master.HMaster: Cluster 
> went down before this master became active
> 2013-05-09 15:09:45,006 DEBUG org.apache.hadoop.hbase.master.HMaster: 
> Stopping service threads
> 2013-05-09 15:09:45,006 INFO org.apache.hadoop.ipc.HBaseServer: Stopping 
> server on 6
>  
> In ActiveMasterManager::blockUntilBecomingActiveMaster()
> {code}
>   ..
>   if (!clusterStatusTracker.isClusterUp()) {
>   this.master.stop(
> "Cluster went down before this master became active");
> }
>   ..
> {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HBASE-8527) clean up code around compaction completion in HStore

2013-05-10 Thread Sergey Shelukhin (JIRA)
Sergey Shelukhin created HBASE-8527:
---

 Summary: clean up code around compaction completion in HStore
 Key: HBASE-8527
 URL: https://issues.apache.org/jira/browse/HBASE-8527
 Project: HBase
  Issue Type: Improvement
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
Priority: Minor


The methods completeCompaction and it's caller are too long. Something I 
changed while doing something else there; putting in a separate easy-to-review 
JIRA to make the other future change smaller.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HBASE-8154) HBaseTestingUtility.startMiniDFSClusterForTestHLog should handle "Address already in use" and retry automatically

2013-05-10 Thread Ted Yu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-8154?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu updated HBASE-8154:
--

Attachment: 8154-v4.txt

> HBaseTestingUtility.startMiniDFSClusterForTestHLog should handle "Address 
> already in use" and retry automatically
> -
>
> Key: HBASE-8154
> URL: https://issues.apache.org/jira/browse/HBASE-8154
> Project: HBase
>  Issue Type: Test
>  Components: test
>Affects Versions: 0.98.0
>Reporter: Liang Xie
>Assignee: Ted Yu
> Attachments: 8154-v2.txt, 8154-v3.txt, 8154-v4.txt, 8154-v4.txt
>
>
> see 
> https://builds.apache.org/job/PreCommit-HBASE-Build/4906//testReport/org.apache.hadoop.hbase.regionserver.wal/TestHLog/testAppendClose/
> and 
> https://issues.apache.org/jira/browse/HBASE-7845?focusedCommentId=13607468&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-13607468

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HBASE-8420) Port HBASE-6874 Implement prefetching for scanners from 0.89-fb

2013-05-10 Thread Jimmy Xiang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-8420?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jimmy Xiang updated HBASE-8420:
---

Status: Patch Available  (was: Open)

Patch v3 added configuration to control the size of max number of prefetcher 
threads, and max prefetched result size.

> Port  HBASE-6874  Implement prefetching for scanners from 0.89-fb
> -
>
> Key: HBASE-8420
> URL: https://issues.apache.org/jira/browse/HBASE-8420
> Project: HBase
>  Issue Type: Improvement
>Reporter: Jimmy Xiang
>Assignee: Jimmy Xiang
> Attachments: 0.94-8420_v1.patch, 0.94-8420_v2.patch, 
> trunk-8420_v1.patch, trunk-8420_v2.patch, trunk-8420_v3.patch
>
>
> This should help scanner performance.  We should have it in trunk.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HBASE-8420) Port HBASE-6874 Implement prefetching for scanners from 0.89-fb

2013-05-10 Thread Jimmy Xiang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-8420?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jimmy Xiang updated HBASE-8420:
---

Attachment: trunk-8420_v3.patch

> Port  HBASE-6874  Implement prefetching for scanners from 0.89-fb
> -
>
> Key: HBASE-8420
> URL: https://issues.apache.org/jira/browse/HBASE-8420
> Project: HBase
>  Issue Type: Improvement
>Reporter: Jimmy Xiang
>Assignee: Jimmy Xiang
> Attachments: 0.94-8420_v1.patch, 0.94-8420_v2.patch, 
> trunk-8420_v1.patch, trunk-8420_v2.patch, trunk-8420_v3.patch
>
>
> This should help scanner performance.  We should have it in trunk.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-7380) [replication] When transferring queues, check if the peer still exists before copying the znodes

2013-05-10 Thread Jean-Daniel Cryans (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-7380?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13654820#comment-13654820
 ] 

Jean-Daniel Cryans commented on HBASE-7380:
---

I posted a patch yesterday on rb: https://reviews.apache.org/r/11014/

It's much different from the one I posted for 0.94 as I saw that it was wrong.

> [replication] When transferring queues, check if the peer still exists before 
> copying the znodes
> 
>
> Key: HBASE-7380
> URL: https://issues.apache.org/jira/browse/HBASE-7380
> Project: HBase
>  Issue Type: Bug
>  Components: Replication, Usability
>Affects Versions: 0.94.3
>Reporter: Jean-Daniel Cryans
>Assignee: Jean-Daniel Cryans
> Fix For: 0.95.1
>
> Attachments: HBASE-7380-0.94.patch
>
>
> Right now it's a pain if you remove a peer and still have rogue queues 
> because they get moved on and on and on. NodeFailoverWorker needs to run the 
> check:
> bq. if (!zkHelper.getPeerClusters().containsKey(src.getPeerClusterId())) {
> before this:
> bq. SortedMap> newQueues = 
> zkHelper.copyQueuesFromRS(rsZnode);
> And test.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HBASE-8526) Backport HBASE-5667 into 0.94

2013-05-10 Thread Jean-Marc Spaggiari (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-8526?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jean-Marc Spaggiari updated HBASE-8526:
---

Attachment: (was: HBASE-8526-v0-0.94.patch)

> Backport HBASE-5667 into 0.94
> -
>
> Key: HBASE-8526
> URL: https://issues.apache.org/jira/browse/HBASE-8526
> Project: HBase
>  Issue Type: Improvement
>  Components: Filters
>Affects Versions: 0.94.7
>Reporter: Jean-Marc Spaggiari
>Assignee: Jean-Marc Spaggiari
>Priority: Minor
>
> RegexStringComparator supports java.util.regex.Pattern flags
> https://issues.apache.org/jira/browse/HBASE-5667

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-8526) Backport HBASE-5667 into 0.94

2013-05-10 Thread Jean-Marc Spaggiari (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8526?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13654793#comment-13654793
 ] 

Jean-Marc Spaggiari commented on HBASE-8526:


1) I took the wrong patch as the source
2) There is some backward wire compatibility issues...

Working on that.

> Backport HBASE-5667 into 0.94
> -
>
> Key: HBASE-8526
> URL: https://issues.apache.org/jira/browse/HBASE-8526
> Project: HBase
>  Issue Type: Improvement
>  Components: Filters
>Affects Versions: 0.94.7
>Reporter: Jean-Marc Spaggiari
>Assignee: Jean-Marc Spaggiari
>Priority: Minor
>
> RegexStringComparator supports java.util.regex.Pattern flags
> https://issues.apache.org/jira/browse/HBASE-5667

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-8154) HBaseTestingUtility.startMiniDFSClusterForTestHLog should handle "Address already in use" and retry automatically

2013-05-10 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8154?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13654783#comment-13654783
 ] 

Hadoop QA commented on HBASE-8154:
--

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12582660/8154-v4.txt
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 6 new 
or modified tests.

{color:green}+1 hadoop1.0{color}.  The patch compiles against the hadoop 
1.0 profile.

{color:green}+1 hadoop2.0{color}.  The patch compiles against the hadoop 
2.0 profile.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 lineLengths{color}.  The patch does not introduce lines 
longer than 100

  {color:green}+1 site{color}.  The mvn site goal succeeds with this patch.

 {color:red}-1 core tests{color}.  The patch failed these unit tests:
 

 {color:red}-1 core zombie tests{color}.  There are 1 zombie test(s): 

Test results: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5624//testReport/
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5624//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-protocol.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5624//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-client.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5624//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-examples.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5624//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-hadoop1-compat.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5624//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-prefix-tree.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5624//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-common.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5624//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-server.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5624//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-hadoop-compat.html
Console output: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5624//console

This message is automatically generated.

> HBaseTestingUtility.startMiniDFSClusterForTestHLog should handle "Address 
> already in use" and retry automatically
> -
>
> Key: HBASE-8154
> URL: https://issues.apache.org/jira/browse/HBASE-8154
> Project: HBase
>  Issue Type: Test
>  Components: test
>Affects Versions: 0.98.0
>Reporter: Liang Xie
>Assignee: Ted Yu
> Attachments: 8154-v2.txt, 8154-v3.txt, 8154-v4.txt
>
>
> see 
> https://builds.apache.org/job/PreCommit-HBASE-Build/4906//testReport/org.apache.hadoop.hbase.regionserver.wal/TestHLog/testAppendClose/
> and 
> https://issues.apache.org/jira/browse/HBASE-7845?focusedCommentId=13607468&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-13607468

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-7006) [MTTR] Improve Region Server Recovery Time - Distributed Log Replay

2013-05-10 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-7006?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13654782#comment-13654782
 ] 

Hadoop QA commented on HBASE-7006:
--

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  
http://issues.apache.org/jira/secure/attachment/12582663/hbase-7006-combined-v7.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 31 new 
or modified tests.

{color:green}+1 hadoop1.0{color}.  The patch compiles against the hadoop 
1.0 profile.

{color:green}+1 hadoop2.0{color}.  The patch compiles against the hadoop 
2.0 profile.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 lineLengths{color}.  The patch introduces lines longer than 
100

  {color:green}+1 site{color}.  The mvn site goal succeeds with this patch.

 {color:red}-1 core tests{color}.  The patch failed these unit tests:
   org.apache.hadoop.hbase.client.TestHCM

Test results: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5625//testReport/
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5625//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-client.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5625//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-examples.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5625//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-protocol.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5625//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-hadoop1-compat.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5625//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-prefix-tree.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5625//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-common.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5625//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-hadoop-compat.html
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5625//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-server.html
Console output: 
https://builds.apache.org/job/PreCommit-HBASE-Build/5625//console

This message is automatically generated.

> [MTTR] Improve Region Server Recovery Time - Distributed Log Replay
> ---
>
> Key: HBASE-7006
> URL: https://issues.apache.org/jira/browse/HBASE-7006
> Project: HBase
>  Issue Type: New Feature
>  Components: MTTR
>Reporter: stack
>Assignee: Jeffrey Zhong
>Priority: Critical
> Fix For: 0.95.1
>
> Attachments: hbase-7006-combined.patch, hbase-7006-combined-v1.patch, 
> hbase-7006-combined-v4.patch, hbase-7006-combined-v5.patch, 
> hbase-7006-combined-v6.patch, hbase-7006-combined-v7.patch, LogSplitting 
> Comparison.pdf, 
> ProposaltoimprovelogsplittingprocessregardingtoHBASE-7006-v2.pdf
>
>
> Just saw interesting issue where a cluster went down  hard and 30 nodes had 
> 1700 WALs to replay.  Replay took almost an hour.  It looks like it could run 
> faster that much of the time is spent zk'ing and nn'ing.
> Putting in 0.96 so it gets a look at least.  Can always punt.
> Distributed Log Replay Description:
> After a region server fails, we firstly assign a failed region to another 
> region server with recovering state marked in ZooKeeper. Then a 
> SplitLogWorker directly replays edits from WAL(Write-Ahead-Log)s of the 
> failed region server to the region after it's re-opened in the new location. 
> When a region is in recovering state, it can also accept writes but no 
> reads(including Append and Increment), region split or merge. 
> The feature piggybacks on existing distributed log splitting framework and 
> directly replay WAL edits to another region server instead of creating 
> recovered.edits files.
> The advantages over existing log splitting recovered edits implementation:
> 1) Eliminate the steps to write and read recovered.edits files. There could 
> be thousands of recovered.edits files are created and written concurrently 
> during a region server recovery. Many small r

[jira] [Commented] (HBASE-8525) Use sleep multilier when choosing sinks in ReplicationSource

2013-05-10 Thread Jean-Daniel Cryans (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8525?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13654750#comment-13654750
 ] 

Jean-Daniel Cryans commented on HBASE-8525:
---

2 comments:

- Should this.sleepForRetries be reset after this?
- Before the patch we were sleeping even if finding sinks. Now we might wait 
even more time if we spin a bit and then find them. 

> Use sleep multilier when choosing sinks in ReplicationSource
> 
>
> Key: HBASE-8525
> URL: https://issues.apache.org/jira/browse/HBASE-8525
> Project: HBase
>  Issue Type: Bug
>Reporter: Lars Hofhansl
>Priority: Trivial
> Fix For: 0.98.0, 0.94.8, 0.95.1
>
> Attachments: 8525-0.94.txt
>
>
> Currently we see this every second. Filling up the log:
> {code}
> 2013-05-10 18:36:00,766 INFO org.apache.zookeeper.ClientCnxn: Opening socket 
> connection to server ist6-mnds1-2-sfm.ops.sfdc.net/10.224.156.197:2181. Will 
> not attempt to authenticate using SASL (Unable to locate a login 
> configuration)
> 2013-05-10 18:36:00,767 WARN org.apache.zookeeper.ClientCnxn: Session 0x0 for 
> server null, unexpected error, closing socket connection and attempting 
> reconnect
> java.net.ConnectException: Connection refused
>   at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
>   at 
> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:599)
>   at 
> org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:350)
>   at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1068)
> 2013-05-10 18:36:01,868 INFO org.apache.zookeeper.ClientCnxn: Opening socket 
> connection to server ist6-mnds1-4-sfm.ops.sfdc.net/10.224.156.199:2181. Will 
> not attempt to authenticate using SASL (Unable to locate a login 
> configuration)
> 2013-05-10 18:36:01,870 WARN org.apache.zookeeper.ClientCnxn: Session 0x0 for 
> server null, unexpected error, closing socket connection and attempting 
> reconnect
> java.net.ConnectException: Connection refused
>   at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
>   at 
> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:599)
>   at 
> org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:350)
>   at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1068)
> 2013-05-10 18:36:01,971 INFO org.apache.zookeeper.ClientCnxn: Opening socket 
> connection to server ist6-mnds1-3-sfm.ops.sfdc.net/10.224.156.198:2181. Will 
> not attempt to authenticate using SASL (Unable to locate a login 
> configuration)
> {code}
> Patch is trivial.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HBASE-8355) BaseRegionObserver#pre(Compact|Flush|Store)ScannerOpen returns null

2013-05-10 Thread Jesse Yates (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-8355?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jesse Yates updated HBASE-8355:
---

Status: Open  (was: Patch Available)

> BaseRegionObserver#pre(Compact|Flush|Store)ScannerOpen returns null
> ---
>
> Key: HBASE-8355
> URL: https://issues.apache.org/jira/browse/HBASE-8355
> Project: HBase
>  Issue Type: Bug
>  Components: Coprocessors
>Affects Versions: 0.98.0, 0.94.8, 0.95.1
>Reporter: Andrew Purtell
>Assignee: Jesse Yates
> Attachments: 8355-0.94.patch, 8355.patch, hbase-8355-trunk-v0.patch, 
> hbase-8355-trunk-v1.patch
>
>
> As pointed out in https://github.com/forcedotcom/phoenix/pull/131, 
> BaseRegionObserver#preCompactScannerOpen returns null by default, which hoses 
> any coprocessors down the line, making override of this method mandatory. The 
> fix is trivial, patch coming momentarily.
> Update:
> This same behavior is present in the Flush and Store versions of the same 
> method - this should all be moved to the proposed 'return passed scanner' 
> default behavior.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HBASE-8355) BaseRegionObserver#pre(Compact|Flush|Store)ScannerOpen returns null

2013-05-10 Thread Jesse Yates (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-8355?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jesse Yates updated HBASE-8355:
---

Status: Patch Available  (was: Open)

> BaseRegionObserver#pre(Compact|Flush|Store)ScannerOpen returns null
> ---
>
> Key: HBASE-8355
> URL: https://issues.apache.org/jira/browse/HBASE-8355
> Project: HBase
>  Issue Type: Bug
>  Components: Coprocessors
>Affects Versions: 0.98.0, 0.94.8, 0.95.1
>Reporter: Andrew Purtell
>Assignee: Jesse Yates
> Attachments: 8355-0.94.patch, 8355.patch, hbase-8355-trunk-v0.patch, 
> hbase-8355-trunk-v1.patch
>
>
> As pointed out in https://github.com/forcedotcom/phoenix/pull/131, 
> BaseRegionObserver#preCompactScannerOpen returns null by default, which hoses 
> any coprocessors down the line, making override of this method mandatory. The 
> fix is trivial, patch coming momentarily.
> Update:
> This same behavior is present in the Flush and Store versions of the same 
> method - this should all be moved to the proposed 'return passed scanner' 
> default behavior.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-6929) Publish Hbase 0.94 artifacts build against hadoop-2.0

2013-05-10 Thread Roshan Naik (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-6929?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13654744#comment-13654744
 ] 

Roshan Naik commented on HBASE-6929:


checking to see if hbase is still planning to have hadoop2 binaries published 
for 0.96 ?

> Publish Hbase 0.94 artifacts build against hadoop-2.0
> -
>
> Key: HBASE-6929
> URL: https://issues.apache.org/jira/browse/HBASE-6929
> Project: HBase
>  Issue Type: Task
>  Components: build
>Affects Versions: 0.94.2
>Reporter: Enis Soztutar
> Attachments: 6929.txt, hbase-6929_v2.patch
>
>
> Downstream projects (flume, hive, pig, etc) depends on hbase, but since the 
> hbase binaries build with hadoop-2.0 are not pushed to maven, they cannot 
> depend on them. AFAIK, hadoop 1 and 2 are not binary compatible, so we should 
> also push hbase jars build with hadoop2.0 profile into maven, possibly with 
> version string like 0.94.2-hadoop2.0. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-8154) HBaseTestingUtility.startMiniDFSClusterForTestHLog should handle "Address already in use" and retry automatically

2013-05-10 Thread Jeffrey Zhong (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-8154?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13654741#comment-13654741
 ] 

Jeffrey Zhong commented on HBASE-8154:
--

+1 on the v4 patch. Thanks!

> HBaseTestingUtility.startMiniDFSClusterForTestHLog should handle "Address 
> already in use" and retry automatically
> -
>
> Key: HBASE-8154
> URL: https://issues.apache.org/jira/browse/HBASE-8154
> Project: HBase
>  Issue Type: Test
>  Components: test
>Affects Versions: 0.98.0
>Reporter: Liang Xie
>Assignee: Ted Yu
> Attachments: 8154-v2.txt, 8154-v3.txt, 8154-v4.txt
>
>
> see 
> https://builds.apache.org/job/PreCommit-HBASE-Build/4906//testReport/org.apache.hadoop.hbase.regionserver.wal/TestHLog/testAppendClose/
> and 
> https://issues.apache.org/jira/browse/HBASE-7845?focusedCommentId=13607468&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-13607468

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HBASE-8525) Use sleep multilier when choosing sinks in ReplicationSource

2013-05-10 Thread Lars Hofhansl (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-8525?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Hofhansl updated HBASE-8525:
-

Attachment: 8525-0.94.txt

Simple patch. A bit more complicated as we want to percolate an 
InterruptedException up to the called in this case (so that the thread can be 
ended).

> Use sleep multilier when choosing sinks in ReplicationSource
> 
>
> Key: HBASE-8525
> URL: https://issues.apache.org/jira/browse/HBASE-8525
> Project: HBase
>  Issue Type: Bug
>Reporter: Lars Hofhansl
>Priority: Trivial
> Fix For: 0.98.0, 0.94.8, 0.95.1
>
> Attachments: 8525-0.94.txt
>
>
> Currently we see this every second. Filling up the log:
> {code}
> 2013-05-10 18:36:00,766 INFO org.apache.zookeeper.ClientCnxn: Opening socket 
> connection to server ist6-mnds1-2-sfm.ops.sfdc.net/10.224.156.197:2181. Will 
> not attempt to authenticate using SASL (Unable to locate a login 
> configuration)
> 2013-05-10 18:36:00,767 WARN org.apache.zookeeper.ClientCnxn: Session 0x0 for 
> server null, unexpected error, closing socket connection and attempting 
> reconnect
> java.net.ConnectException: Connection refused
>   at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
>   at 
> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:599)
>   at 
> org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:350)
>   at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1068)
> 2013-05-10 18:36:01,868 INFO org.apache.zookeeper.ClientCnxn: Opening socket 
> connection to server ist6-mnds1-4-sfm.ops.sfdc.net/10.224.156.199:2181. Will 
> not attempt to authenticate using SASL (Unable to locate a login 
> configuration)
> 2013-05-10 18:36:01,870 WARN org.apache.zookeeper.ClientCnxn: Session 0x0 for 
> server null, unexpected error, closing socket connection and attempting 
> reconnect
> java.net.ConnectException: Connection refused
>   at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
>   at 
> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:599)
>   at 
> org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:350)
>   at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1068)
> 2013-05-10 18:36:01,971 INFO org.apache.zookeeper.ClientCnxn: Opening socket 
> connection to server ist6-mnds1-3-sfm.ops.sfdc.net/10.224.156.198:2181. Will 
> not attempt to authenticate using SASL (Unable to locate a login 
> configuration)
> {code}
> Patch is trivial.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


  1   2   >