I think in this case, writing data to HDFS or HFile directly (for subsequent bulk loading) is the best option. HBase will never compete in write speed with HDFS.
Best regards, Vladimir Rodionov Principal Platform Engineer Carrier IQ, www.carrieriq.com e-mail: vrodio...@carrieriq.com ________________________________________ From: Ted Yu [yuzhih...@gmail.com] Sent: Saturday, January 04, 2014 2:33 PM To: user@hbase.apache.org Subject: Re: Hbase Performance Issue There're 8 items under: http://hbase.apache.org/book.html#perf.writing I guess you have through all of them :-) On Sat, Jan 4, 2014 at 1:34 PM, Akhtar Muhammad Din <akhtar.m...@gmail.com>wrote: > Thanks guys for your precious time. > Vladimir, as Ted rightly said i want to improve write performance currently > (of course i want to read data as fast as possible later on) > Kevin, my current understanding of bulk load is that you generate > StoreFiles and later load through a command line program. I dont want to do > any manual step. Our system is getting data after every 15 minutes, so > requirement is to automate it through client API completely. > > Confidentiality Notice: The information contained in this message, including any attachments hereto, may be confidential and is intended to be read only by the individual or entity to whom this message is addressed. If the reader of this message is not the intended recipient or an agent or designee of the intended recipient, please note that any review, use, disclosure or distribution of this message or its attachments, in any form, is strictly prohibited. If you have received this message in error, please immediately notify the sender and/or notificati...@carrieriq.com and delete or destroy any copy of this message and its attachments.