Even still… eventually the data will be distributed equally across the cluster.
What’s happening with the HDFS balancer? Is that heterogenous or homogenous in terms of storage? > On Mar 19, 2015, at 2:50 PM, Nick Dimiduk <[email protected]> wrote: > > Seems more fantasy than fact, I'm afraid. The default load balancer [0] > takes store file size into account, but has no concept of capacity. It > doesn't know that nodes in a heterogenous environment have different > capacity. > > This would be a good feature to add though. > > [0]: > https://github.com/apache/hbase/blob/branch-1.0/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.java > > On Tue, Mar 17, 2015 at 7:26 AM, Ted Tuttle <[email protected]> wrote: > >> Hello- >> >> Sometime back I asked a question about introducing new nodes w/ more >> storage that existing nodes. I was told at the time that HBase will not be >> able to utilize the additional storage; I assumed at the time that regions >> are allocated to nodes in something like a round-robin fashion and the node >> with the least storage sets the limit for how much each node can utilize. >> >> My question this time around has to do with nodes w/ unequal numbers of >> volumes: Does HBase allocate regions based on nodes or volumes on the >> nodes? I am hoping I can add a node with 8 volumes totaling 8X TB and all >> the volumes will be filled. This even though legacy nodes have 5 volumes >> and total storage of 5X TB. >> >> Fact or fantasy? >> >> Thanks, >> Ted >> >> The opinions expressed here are mine, while they may reflect a cognitive thought, that is purely accidental. Use at your own risk. Michael Segel michael_segel (AT) hotmail.com
