The fix is here: http://pastebin.com/zuL23e0U
We're going to do a push to github later today, along with other patches that require more testing. J-D On Wed, Sep 29, 2010 at 10:54 AM, Andrey Stepachev <[email protected]> wrote: > wow. i'll wait. thanks for reply. > > 2010/9/29 Jean-Daniel Cryans <[email protected]>: >> Ok I found the bug, I think it's only in our distro. >> >> Stay tuned! >> >> J-D >> >> On Wed, Sep 29, 2010 at 9:26 AM, Jean-Daniel Cryans <[email protected]> >> wrote: >>> Weird indeed, even after the WAL was rolled 4 times (theoretically >>> 256MB of data) I don't even see a flush request... although you're >>> running at INFO level instead of DEBUG. Could you switch that and send >>> us just the full log. >>> >>> Thanks a lot! >>> >>> J-D >>> >>> On Wed, Sep 29, 2010 at 4:25 AM, Andrey Stepachev <[email protected]> wrote: >>>> Hi all, >>>> >>>> I'm stuck. I can't insert any valuable peace of data into hbase. >>>> >>>> Data is something around ~20mil rows (20G). I try to insert them into >>>> nondistributed hbase with 4 parallel jobs. >>>> MR job run until all memory given to hbase is exhaused and then >>>> hbase produces hprof file. As profiler shows, all memory accumulated >>>> in MemStore.kvset. >>>> I don't understand, why hbase doesn't block untill flush memstore. >>>> The same if I give hbase 6Gb or RAM. >>>> >>>> 6GB gc log http://paste.ubuntu.com/502577/ >>>> >>>> hadoop: 0.20.2+320 >>>> hbase: stumbleupon-20100830 >>>> >>>> nothing in hbase-site.xml (except hbase.rootdir and zookeeper.quorum). >>>> >>>> >>>> >>>> Andrey. >>>> >>> >> >
