Even still… eventually the data will be distributed equally across the cluster. 

What’s happening with the HDFS balancer?  Is that heterogenous or homogenous in 
terms of storage? 


> On Mar 19, 2015, at 2:50 PM, Nick Dimiduk <[email protected]> wrote:
> 
> Seems more fantasy than fact, I'm afraid. The default load balancer [0]
> takes store file size into account, but has no concept of capacity. It
> doesn't know that nodes in a heterogenous environment have different
> capacity.
> 
> This would be a good feature to add though.
> 
> [0]:
> https://github.com/apache/hbase/blob/branch-1.0/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.java
> 
> On Tue, Mar 17, 2015 at 7:26 AM, Ted Tuttle <[email protected]> wrote:
> 
>> Hello-
>> 
>> Sometime back I asked a question about introducing new nodes w/ more
>> storage that existing nodes.  I was told at the time that HBase will not be
>> able to utilize the additional storage; I assumed at the time that regions
>> are allocated to nodes in something like a round-robin fashion and the node
>> with the least storage sets the limit for how much each node can utilize.
>> 
>> My question this time around has to do with nodes w/ unequal numbers of
>> volumes: Does HBase allocate regions based on nodes or volumes on the
>> nodes?  I am hoping I can add a node with 8 volumes totaling 8X TB and all
>> the volumes will be filled.  This even though legacy nodes have 5 volumes
>> and total storage of 5X TB.
>> 
>> Fact or fantasy?
>> 
>> Thanks,
>> Ted
>> 
>> 

The opinions expressed here are mine, while they may reflect a cognitive 
thought, that is purely accidental. 
Use at your own risk. 
Michael Segel
michael_segel (AT) hotmail.com





Reply via email to