[
https://issues.apache.org/jira/browse/HBASE-14575?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15007915#comment-15007915
]
Ted Yu commented on HBASE-14575:
--------------------------------
>From
>https://builds.apache.org/job/PreCommit-HBASE-Build/16539//testReport/org.apache.hadoop.hbase.regionserver/TestHRegionServerBulkLoad/testAtomicBulkLoad_1_/
> :
{code}
java.lang.RuntimeException: Deferred
at
org.apache.hadoop.hbase.MultithreadedTestUtil$TestContext.checkException(MultithreadedTestUtil.java:78)
at
org.apache.hadoop.hbase.MultithreadedTestUtil$TestContext.stop(MultithreadedTestUtil.java:105)
at
org.apache.hadoop.hbase.regionserver.TestHRegionServerBulkLoad.runAtomicBulkloadTest(TestHRegionServerBulkLoad.java:378)
at
org.apache.hadoop.hbase.regionserver.TestHRegionServerBulkLoad.testAtomicBulkLoad(TestHRegionServerBulkLoad.java:353)
Caused by: org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed
after attempts=36, exceptions:
Tue Nov 17 01:33:37 UTC 2015, null, java.net.SocketTimeoutException:
callTimeout=60000, callDuration=77968: row '' on table 'atomicBulkLoad' at
region=atomicBulkLoad,,
1447723922183.4b02772004db4147b2144c3820b9182a.,
hostname=pietas.apache.org,41039,1447723918534, seqNum=2
at
org.apache.hadoop.hbase.client.RpcRetryingCallerWithReadReplicas.throwEnrichedException(RpcRetryingCallerWithReadReplicas.java:269)
at
org.apache.hadoop.hbase.client.ScannerCallableWithReplicas.call(ScannerCallableWithReplicas.java:195)
at
org.apache.hadoop.hbase.client.ScannerCallableWithReplicas.call(ScannerCallableWithReplicas.java:59)
at
org.apache.hadoop.hbase.client.RpcRetryingCallerImpl.callWithoutRetries(RpcRetryingCallerImpl.java:185)
at org.apache.hadoop.hbase.client.ClientScanner.call(ClientScanner.java:311)
at
org.apache.hadoop.hbase.client.ClientScanner.nextScanner(ClientScanner.java:286)
at
org.apache.hadoop.hbase.client.ClientScanner.initializeScannerInConstruction(ClientScanner.java:162)
at
org.apache.hadoop.hbase.client.ClientScanner.<init>(ClientScanner.java:155)
at
org.apache.hadoop.hbase.client.ClientSimpleScanner.<init>(ClientSimpleScanner.java:40)
at org.apache.hadoop.hbase.client.HTable.getScanner(HTable.java:378)
at
org.apache.hadoop.hbase.regionserver.TestHRegionServerBulkLoad$AtomicScanReader.doAnAction(TestHRegionServerBulkLoad.java:283)
at
org.apache.hadoop.hbase.MultithreadedTestUtil$RepeatingTestThread.doWork(MultithreadedTestUtil.java:147)
at
org.apache.hadoop.hbase.MultithreadedTestUtil$TestThread.run(MultithreadedTestUtil.java:123)
Caused by: java.net.SocketTimeoutException: callTimeout=60000,
callDuration=77968: row '' on table 'atomicBulkLoad' at
region=atomicBulkLoad,,1447723922183.
4b02772004db4147b2144c3820b9182a.,
hostname=pietas.apache.org,41039,1447723918534, seqNum=2
at
org.apache.hadoop.hbase.client.RpcRetryingCallerImpl.callWithRetries(RpcRetryingCallerImpl.java:151)
at
org.apache.hadoop.hbase.client.ResultBoundedCompletionService$QueueingFuture.run(ResultBoundedCompletionService.java:64)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:744)
Caused by: org.apache.hadoop.hbase.ipc.CallTimeoutException: callId=6337,
method=Scan, rpcTimeout=10000, param {region { type: REGION_NAME value:
"atomicBulkLoad,,1447723922183. 4b02772004db4147b2144c3820b9182a." } scan {
column { family: "family_0000" } column { family: "family_0001" } column {
family: "family_0002" } column { family: "family_0003" } column { family:
"family_0004" } column { family: "family_0005" } column { family: "family_0006"
} column { family: "family_0007" } column { family: "family_0008" } column {
family: "family_0009" } max_versions: 1 cache_blocks: true max_result_size:
2097152 caching: 2147483647 } number_of_rows: 0 close_scanner: false
client_handles_partials: true client_handles_heartbeats: true
track_scan_metrics: false}
at org.apache.hadoop.hbase.ipc.AsyncRpcClient.call(AsyncRpcClient.java:257)
at
org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:217)
at
org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:295)
at
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$BlockingStub.scan(ClientProtos.java:33872)
at
org.apache.hadoop.hbase.client.ScannerCallable.openScanner(ScannerCallable.java:390)
at
org.apache.hadoop.hbase.client.ScannerCallable.call(ScannerCallable.java:199)
{code}
> Reduce scope of compactions holding region lock
> -----------------------------------------------
>
> Key: HBASE-14575
> URL: https://issues.apache.org/jira/browse/HBASE-14575
> Project: HBase
> Issue Type: Sub-task
> Components: Compaction, regionserver
> Reporter: Nick Dimiduk
> Assignee: Nick Dimiduk
> Attachments: 14575-test-case.patch, 14575-v1.patch, 14575-v2.patch,
> 14575-v3.patch, 14575-v4.patch, 14575-v5.patch, 14575-v6.txt, 14575-v6.txt,
> 14575.v00.patch
>
>
> Per [~devaraj]'s idea on parent issue, let's see if we can reduce the scope
> of critical section under which compactions hold the region read lock.
> Here is summary from parent issue:
> Another idea is we can reduce the scope of when the read lock is held during
> compaction. In theory the compactor only needs a region read lock while
> deciding what files to compact and at the time of committing the compaction.
> We're protected from the case of region close events because compactions are
> checking (every 10k bytes written) if the store has been closed in order to
> abort in such a case.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)