Can you tell us which version of HBase you are using ? Thanks
On Sat, Jan 12, 2013 at 10:57 AM, Asaf Mesika <[email protected]> wrote: > Most time is spent reading from Store file and not on network transfer time > of Increment objects. > > Sent from my iPhone > > On 12 בינו 2013, at 17:40, Anoop John <[email protected]> wrote: > > Hi > Can you check with using API HTable#batch()? Here you can batch a > number of increments for many rows in just one RPC call. Might help you to > reduce the net time taken. Good luck. > > -Anoop- > > On Sat, Jan 12, 2013 at 4:07 PM, kiran <[email protected]> > wrote: > > Hi, > > > My usecase is I need to increment 1 million rows with in 15 mins. I tried > > two approaches but none of the yielded results. > > > I have used HTable.increment, but is not getting completed in the specified > > time. I tried multi-threading also but it is very costly. I have also > > implemented get and put as other alternative, but that approach is also not > > getting completed in 15 mins. > > > Can I use any low level implementation like using "Store or HRegionServer" > > to increment 1 million rows. I know the table splits, and region servers > > serving them, and rows which fall into table splits. I suspect the major > > concern as network I/O rather than processing with the above two > > approaches. > > > -- > > Thank you > > Kiran Sarvabhotla > > > -----Even a correct decision is wrong when it is taken late >
