Thanx once more. Now works fine.
2010/9/29 Jean-Daniel Cryans <[email protected]>: > The fix is here: http://pastebin.com/zuL23e0U > > We're going to do a push to github later today, along with other > patches that require more testing. > > J-D > > On Wed, Sep 29, 2010 at 10:54 AM, Andrey Stepachev <[email protected]> wrote: >> wow. i'll wait. thanks for reply. >> >> 2010/9/29 Jean-Daniel Cryans <[email protected]>: >>> Ok I found the bug, I think it's only in our distro. >>> >>> Stay tuned! >>> >>> J-D >>> >>> On Wed, Sep 29, 2010 at 9:26 AM, Jean-Daniel Cryans <[email protected]> >>> wrote: >>>> Weird indeed, even after the WAL was rolled 4 times (theoretically >>>> 256MB of data) I don't even see a flush request... although you're >>>> running at INFO level instead of DEBUG. Could you switch that and send >>>> us just the full log. >>>> >>>> Thanks a lot! >>>> >>>> J-D >>>> >>>> On Wed, Sep 29, 2010 at 4:25 AM, Andrey Stepachev <[email protected]> wrote: >>>>> Hi all, >>>>> >>>>> I'm stuck. I can't insert any valuable peace of data into hbase. >>>>> >>>>> Data is something around ~20mil rows (20G). I try to insert them into >>>>> nondistributed hbase with 4 parallel jobs. >>>>> MR job run until all memory given to hbase is exhaused and then >>>>> hbase produces hprof file. As profiler shows, all memory accumulated >>>>> in MemStore.kvset. >>>>> I don't understand, why hbase doesn't block untill flush memstore. >>>>> The same if I give hbase 6Gb or RAM. >>>>> >>>>> 6GB gc log http://paste.ubuntu.com/502577/ >>>>> >>>>> hadoop: 0.20.2+320 >>>>> hbase: stumbleupon-20100830 >>>>> >>>>> nothing in hbase-site.xml (except hbase.rootdir and zookeeper.quorum). >>>>> >>>>> >>>>> >>>>> Andrey. >>>>> >>>> >>> >> >
