The effective block size would be >= 50K for your schema.

Cheers


On Mon, Feb 24, 2014 at 12:38 PM, Varun Sharma <[email protected]> wrote:

> Hi Ted,
>
> I am not concerned about wide rows here. My schema has only 1 column in it
> but it has a "value" of 50-100K bytes. The block size is configured to be
> 32K bytes. How does that work in practice - does it mean that the effective
> block size is upwards of 50K ?
>
> Varun
>
>
> On Mon, Feb 24, 2014 at 10:07 AM, Ted Yu <[email protected]> wrote:
>
> > Cycling old bits:
> >
> >
> >
> http://search-hadoop.com/m/DHED4v7stT1/larger+HFile+block+size+for+very+wide+row&subj=larger+HFile+block+size+for+very+wide+row+
> >
> >
> > On Mon, Feb 24, 2014 at 11:51 AM, Varun Sharma <[email protected]>
> > wrote:
> >
> > > Hi,
> > >
> > > What happens if my block size is 32K while the cells are 50K. Do Hfile
> > > blocks round up to 50K or are values split across blocks ? Also how
> does
> > > this play with the block cache ?
> > >
> > > Thanks
> > > Varun
> > >
> >
>

Reply via email to