An entire mutation needs to fit in memory several times, so you should not attempt to push in a single mutation larger than a 100MB unless you have a lot of memory in your tserver/logger.
And while I'm at it, large keys will create large indexes, so try to keep your (row,cf,cq,cv) under 100K. -Eric On Thu, Aug 23, 2012 at 4:37 PM, Cardon, Tejay E <[email protected]>wrote: > In my case I’ll be doing a document based index store (like the > wikisearch example), but my documents may be as large as several GB. I > just wanted to pick the collective brain of the group to see if I’m walking > into a major headache. If it’s never been tried before, then I’ll give it > a shot and report back.**** > > > Tejay**** > > ** ** > > *From:* William Slacum [mailto:[email protected]] > *Sent:* Thursday, August 23, 2012 2:07 PM > *To:* [email protected] > *Subject:* EXTERNAL: Re: Large files in Accumulo**** > > ** ** > > Are these RFiles as a whole? I know at some point HBase needed to have > entire rows fit into memory; Accumulo does not have this restriction.**** > > On Thu, Aug 23, 2012 at 12:55 PM, Cardon, Tejay E <[email protected]> > wrote:**** > > Alright, this one’s a quick question. I’ve been told that HBase does not > perform well if large (> 100MB) files are stored in it). Does Accumulo > have similar trouble? If so, can it be overcome by storing the large files > in their own locality group?**** > > **** > > Thanks,**** > > Tejay**** > > ** ** >
