Hi Lucas Thanks a lot for your help i got what i required, really appreciate you for giving your time on the issue.
On Tue, Feb 18, 2014 at 3:56 PM, Lukas Nalezenec < [email protected]> wrote: > Hi, > Add this import: > import org.apache.hadoop.hbase.HServerLoad; > > And change names of classes: > ServerLoad -> HServerLoad > RegionLoad -> HServerLoad.RegionLoad > > Lukas > > > On 18.2.2014 11:19, Vikram Singh Chandel wrote: > >> Hi Lucas >> As you said that RegionSizeCalculator is developed on top of 0.94, the >> class has interdependencies vig >> import org.apache.hadoop.hbase.RegionLoad; >> import org.apache.hadoop.hbase.ServerLoad; >> >> unable to find these classes in 0.94.X >> >> are these classes available in 0.94 under some other package, >> >> >> >> On Tue, Feb 11, 2014 at 3:12 PM, Lukas Nalezenec < >> [email protected]> wrote: >> >> Hi, >>> >>> I am hbase newbie, maybe there is simpler solution but this will work. I >>> tried estimating size using HDFS but it is not best solution(see link >>> [1]); >>> >>> You dont need to work with TableSplits., look at class >>> org.apache.hadoop.hbase.util.RegionSizeCalculator. >>> It can do what you need. Create instance of this class, than call method >>> getRegionSizeMap() and sum all values in map. Note that the size contains >>> only storeFile sizes, not memStore sizes. >>> If you need customize behaviour of this class, just copy the code and >>> change it. >>> >>> This class will be in version 0.98 but it was developed on 0.94 - it will >>> work but you will have to change some java imports. >>> >>> >>> [1] >>> https://issues.apache.org/jira/browse/HBASE-10413? >>> focusedCommentId=13889745&page=com.atlassian.jira. >>> plugin.system.issuetabpanels:comment-tabpanel#comment-13889745 >>> >>> Lukas >>> >>> >>> >>> On 11.2.2014 08:14, Vikram Singh Chandel wrote: >>> >>> Hi Lukas >>>> >>>> the table split constructor expects startRow, endRow and location we >>>> won't >>>> be having info about any of these. >>>> Moreover we require table size as a whole, not split size. >>>> >>>> We will use the table size to look for a threshold breach in metadata >>>> table, if breached we have to trigger a delete operation on to the >>>> table(of >>>> which threshold is breached) to delete LRU records until table size is >>>> within limit (~ 50-60Gb) >>>> >>>> >>>> On Mon, Feb 10, 2014 at 6:01 PM, Vikram Singh Chandel < >>>> [email protected]> wrote: >>>> >>>> Hi >>>> >>>>> The requirement is to get the hbase table size (using API) and have to >>>>> save this size for each table in a metadata table . >>>>> >>>>> i tried hdfs command to check table size but need api method (if >>>>> available) >>>>> >>>>> Hadoop fs -du -h hdfs:// >>>>> >>>>> >>>>> >>>>> Thanks >>>>> >>>>> -- >>>>> *Regards* >>>>> >>>>> *VIKRAM SINGH CHANDEL* >>>>> >>>>> >>>>> Please do not print this email unless it is absolutely >>>>> necessary,Reduce. >>>>> Reuse. Recycle. Save our planet. >>>>> >>>>> >>>>> >>>> >> > -- *Regards* *VIKRAM SINGH CHANDEL* Please do not print this email unless it is absolutely necessary,Reduce. Reuse. Recycle. Save our planet.
