wow. i'll wait. thanks for reply.
2010/9/29 Jean-Daniel Cryans <[email protected]>: > Ok I found the bug, I think it's only in our distro. > > Stay tuned! > > J-D > > On Wed, Sep 29, 2010 at 9:26 AM, Jean-Daniel Cryans <[email protected]> > wrote: >> Weird indeed, even after the WAL was rolled 4 times (theoretically >> 256MB of data) I don't even see a flush request... although you're >> running at INFO level instead of DEBUG. Could you switch that and send >> us just the full log. >> >> Thanks a lot! >> >> J-D >> >> On Wed, Sep 29, 2010 at 4:25 AM, Andrey Stepachev <[email protected]> wrote: >>> Hi all, >>> >>> I'm stuck. I can't insert any valuable peace of data into hbase. >>> >>> Data is something around ~20mil rows (20G). I try to insert them into >>> nondistributed hbase with 4 parallel jobs. >>> MR job run until all memory given to hbase is exhaused and then >>> hbase produces hprof file. As profiler shows, all memory accumulated >>> in MemStore.kvset. >>> I don't understand, why hbase doesn't block untill flush memstore. >>> The same if I give hbase 6Gb or RAM. >>> >>> 6GB gc log http://paste.ubuntu.com/502577/ >>> >>> hadoop: 0.20.2+320 >>> hbase: stumbleupon-20100830 >>> >>> nothing in hbase-site.xml (except hbase.rootdir and zookeeper.quorum). >>> >>> >>> >>> Andrey. >>> >> >
