Weird indeed, even after the WAL was rolled 4 times (theoretically 256MB of data) I don't even see a flush request... although you're running at INFO level instead of DEBUG. Could you switch that and send us just the full log.
Thanks a lot! J-D On Wed, Sep 29, 2010 at 4:25 AM, Andrey Stepachev <[email protected]> wrote: > Hi all, > > I'm stuck. I can't insert any valuable peace of data into hbase. > > Data is something around ~20mil rows (20G). I try to insert them into > nondistributed hbase with 4 parallel jobs. > MR job run until all memory given to hbase is exhaused and then > hbase produces hprof file. As profiler shows, all memory accumulated > in MemStore.kvset. > I don't understand, why hbase doesn't block untill flush memstore. > The same if I give hbase 6Gb or RAM. > > 6GB gc log http://paste.ubuntu.com/502577/ > > hadoop: 0.20.2+320 > hbase: stumbleupon-20100830 > > nothing in hbase-site.xml (except hbase.rootdir and zookeeper.quorum). > > > > Andrey. >
