Hi, Our Biz logic is to scan 31 hbase tables sequentially then join these data in the local machine. We dispatch these logic with different parameter to our computing cluster. So a lot of connections will connect to HBase almost concurrently, It took a thundred seconds to get a few rows. I wonder the overhead is by the following code. HTable table = new HTable(config, Bytes.toBytes(TableName));
It will scan META table while we new a HTable, right? (META region server got a lot of requests I saw on the web console) Splitting META into two region, does it work? Because all my clients query it with the same table name at the same time, it seems no chance to load balance it by Splitting META ! Is it true? If I can query META once than broastcasting it through my way to my clients, then they can direct access to region server in order to avoid META bottleneck. Is that possible? Any suggestions? Thanks Fleming Chiu(邱宏明) 707-6128 y_823...@tsmc.com 週一無肉日吃素救地球(Meat Free Monday Taiwan) Ryan Rawson <ryano...@gmail.c To: hbase-user@hadoop.apache.org om> cc: (bcc: Y_823910/TSMC) Subject: Re: MATA load balance 2010/02/03 04:32 PM Please respond to hbase-user A little birdy told me that META performance can potentially degrade with a high # of store files, so try to major_compact '.META.' first. Secondly, yes META can be a bottleneck, but it should serve out of ram nearly constantly. Combined with longer lived clients, this should mitigate things somewhat. One option is to use a long lived gateway process, eg: thrift, which will amortize the cost of the META lookup over many small client connections. This is what I do with PHP, and it works well. -ryan 2010/2/3 <y_823...@tsmc.com>: > Hi, > Our cluster with 3 zookeepers, 10 region servers, 19 data nodes. > Each machine has 4 core cpu, 12G ram. > There are 1322 regions in our cluster now. > We fired up to 3000 hbase client in parallel to fetch hbase data for > distributed computing. > Despite of Htable just one time visit to MATA table; there is only one > server with the MATA information, > it seems a bottleneck while I fired so many clients at the same time. > Any suggestions? > > Fleming Chiu(邱宏明) > 707-6128 > y_823...@tsmc.com > 週一無肉日吃素救地球(Meat Free Monday Taiwan) > > > --------------------------------------------------------------------------- > TSMC PROPERTY > This email communication (and any attachments) is proprietary information > for the sole use of its > intended recipient. Any unauthorized review, use or distribution by anyone > other than the intended > recipient is strictly prohibited. If you are not the intended recipient, > please notify the sender by > replying to this email, and then delete this email and any copies of it > immediately. Thank you. > --------------------------------------------------------------------------- > > > > --------------------------------------------------------------------------- TSMC PROPERTY This email communication (and any attachments) is proprietary information for the sole use of its intended recipient. Any unauthorized review, use or distribution by anyone other than the intended recipient is strictly prohibited. If you are not the intended recipient, please notify the sender by replying to this email, and then delete this email and any copies of it immediately. Thank you. ---------------------------------------------------------------------------