You need to upgrade.
You hit a known bug.
> On Apr 29, 2015, at 2:27 AM, Dejan Menges <[email protected]> wrote:
>
> Using currently 0.98.0 (or Hortonworks 2.1.10, with bunch of patches that
> fixed all my questions in last two months here on the list :))
>
> Thanks, going to take a look.
>
> On Wed, Apr 29, 2015 at 6:21 AM, Ted Yu <[email protected]> wrote:
>
>> Which hbase release are you using ?
>>
>> Please take a look at StochasticLoadBalancer#TableSkewCostFunction
>> You can increase the weight for
>> "hbase.master.balancer.stochastic.tableSkewCost"
>>
>> Cheers
>>
>> On Tue, Apr 28, 2015 at 3:04 AM, Dejan Menges <[email protected]>
>> wrote:
>>
>>> And one more follow up - I know about hbase.regions.slop which is in our
>>> case default 0.2
>>>
>>> So in this specific scenario, one table, having in total 15 region
>> servers,
>>> one table having in total 225 regions, how to avoid some region servers
>>> serving 10 regions, and some 29 in this case (at least not 40 anymore)?
>>>
>>> On Tue, Apr 28, 2015 at 11:22 AM Dejan Menges <[email protected]>
>>> wrote:
>>>
>>>> Hi,
>>>>
>>>> We have a HBase cluster with multiple tables in it. We hit today
>>>> interesting issue - one of our jobs started failing with
>>>> OutOfOrderScannerNextException.
>>>>
>>>> As this job was basically reading data from only one specific table
>> from
>>>> one specific cluster and doing something with it, I checked table
>>>> properties, and saw that it's, mostly nicely said, very imbalanced
>>> regions
>>>> related. For example, one node had 40 regions hosted, another one had
>>> only
>>>> 8. After I took down this node with 40 regions and forced it to
>> rebalance
>>>> regions that way, job passed fine.
>>>>
>>>> However, when I took it back, the same node got 30 regions this time,
>>>> what's still in the best case 40% more than other node with biggest
>>> amount
>>>> of regions for the same table.
>>>>
>>>> Is there any way to balance this?
>>>>
>>>> Thanks a lot,
>>>> Dejan
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>
>>