The effective block size would be >= 50K for your schema. Cheers
On Mon, Feb 24, 2014 at 12:38 PM, Varun Sharma <[email protected]> wrote: > Hi Ted, > > I am not concerned about wide rows here. My schema has only 1 column in it > but it has a "value" of 50-100K bytes. The block size is configured to be > 32K bytes. How does that work in practice - does it mean that the effective > block size is upwards of 50K ? > > Varun > > > On Mon, Feb 24, 2014 at 10:07 AM, Ted Yu <[email protected]> wrote: > > > Cycling old bits: > > > > > > > http://search-hadoop.com/m/DHED4v7stT1/larger+HFile+block+size+for+very+wide+row&subj=larger+HFile+block+size+for+very+wide+row+ > > > > > > On Mon, Feb 24, 2014 at 11:51 AM, Varun Sharma <[email protected]> > > wrote: > > > > > Hi, > > > > > > What happens if my block size is 32K while the cells are 50K. Do Hfile > > > blocks round up to 50K or are values split across blocks ? Also how > does > > > this play with the block cache ? > > > > > > Thanks > > > Varun > > > > > >
