[ 
https://issues.apache.org/jira/browse/HBASE-16857?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Soldatov updated HBASE-16857:
------------------------------------
    Status: Patch Available  (was: Open)

> RateLimiter may fails during parallel scan execution
> ----------------------------------------------------
>
>                 Key: HBASE-16857
>                 URL: https://issues.apache.org/jira/browse/HBASE-16857
>             Project: HBase
>          Issue Type: Bug
>    Affects Versions: 1.1.2
>         Environment: hbase.quota.enabled=true 
> hbase.quota.refresh.period=5000 
>            Reporter: Sergey Soldatov
>            Assignee: Sergey Soldatov
>         Attachments: HBASE-16857.patch
>
>
> Steps to reproduce using phoenix (that's the easiest way to run a lot of 
> parallel scans):
> 1. Create table:
> {code}
> create table "abc" (id bigint not null primary key, name varchar) 
> salt_buckets=50; 
> {code}
> 2. set quota from hbase shell:
> {code}
> set_quota TYPE => THROTTLE, TABLE => 'abc', LIMIT => '10G/sec' 
> {code}
> 3. in phoenix run 
> {code}
>  select * from "abc"; 
> {code}
> It will fail with ThrottlingException.
> Sometimes it requires to run it several times to reproduce.
> That happens because the logic in DefaultOperationQuota. First we run 
> limiter.checkQuota where we may change available to Long.MAX_VALUE, after 
> that we run limiter.grabQuota where we reduce available for 1000 (is it scan 
> overhead or what?) and in close() we adding this 1000 back. 
> When number of parallel scans are executing there is a chance that one of the 
> threads run limiter.checkQuota right before the second run close(). We get 
> overflow and as the result available value becomes negative, so during the 
> next check we just fail. 
> This behavior was introduced in HBASE-13686.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to