Re: Doubt

2017-02-14 Thread gaurhari dass
thanks a lot it is really nice.

I didn't have any idea about this.

Thanks
Gaurhari

On Tue, Feb 14, 2017 at 3:32 PM, Ted Yu  wrote:

> All the major NoSQL projects are indexed by http://search-hadoop.com/
>
> You can easily find relevant information there.
>
> FYI
>
> On Tue, Feb 14, 2017 at 7:28 AM, gaurhari dass 
> wrote:
>
> > Actually I am not using HBase right now at large scale.
> >
> > I am using Mongodb mainly so to keep myself updated I follow these
> > discussions to learn about issues so I post it on my profile so that I
> can
> > have easy access to those issues.
> >
> > Thanks
> > Gaurhari
> >
> > On Tue, Feb 14, 2017 at 3:15 PM, Ted Yu  wrote:
> >
> > > [image: Boxbe]  Automatic Cleanup:
> keep
> > > last 1 emails (yuzhih...@gmail.com) Edit rule
> > >  > com%2Fcleanup%3Fkey%3DTGqVsGxY5rwRFnibssg44ywsKdnF
> > LlvN7oL%252BZGzAtdc%253D%26token%3Dn%252FzPWvUmHmaveE0%
> > 252FIOALy5uUs37HKIhENDj5Blu3WvnYK%252FnGhG%252FVlR4Sj%
> > 252BlEr5a9BtoaVIWeQdGT%252B2eB2dXJDAV0dXkeIYhzbEyBV7sXFv%
> > 252BDwGyFMXMAO3Oy66qytBf3u3OkHvnhQFo%253D&tc_serial=28885329001&tc_rand=
> > 196581&utm_source=stf&utm_medium=email&utm_campaign=
> > ANNO_CLEANUP_EDIT&utm_content=001>
> > > | Delete rule
> > >  > com%2Fcleanup%3Fkey%3DTGqVsGxY5rwRFnibssg44ywsKdnF
> > LlvN7oL%252BZGzAtdc%253D%26token%3Dn%252FzPWvUmHmaveE0%
> > 252FIOALy5uUs37HKIhENDj5Blu3WvnYK%252FnGhG%252FVlR4Sj%
> > 252BlEr5a9BtoaVIWeQdGT%252B2eB2dXJDAV0dXkeIYhzbEyBV7sXFv%
> > 252BDwGyFMXMAO3Oy66qytBf3u3OkHvnhQFo%253D&tc_serial=28885329001&tc_rand=
> > 196581&utm_source=stf&utm_medium=email&utm_campaign=
> > ANNO_CLEANUP_EDIT&utm_content=001>
> > > | Mark important
> > >  > com%2Fcleanup%3Fkey%3DTGqVsGxY5rwRFnibssg44ywsKdnF
> > LlvN7oL%252BZGzAtdc%253D%26token%3Dn%252FzPWvUmHmaveE0%
> > 252FIOALy5uUs37HKIhENDj5Blu3WvnYK%252FnGhG%252FVlR4Sj%
> > 252BlEr5a9BtoaVIWeQdGT%252B2eB2dXJDAV0dXkeIYhzbEyBV7sXFv%
> > 252BDwGyFMXMAO3Oy66qytBf3u3OkHvnhQFo%253D%26important%
> > 3Dtrue%26emlId%3D56028479123&tc_serial=28885329001&tc_rand=
> > 196581&utm_source=stf&utm_medium=email&utm_campaign=
> > ANNO_CLEANUP_EDIT&utm_content=001>
> > >
> > > Do you use hbase (I assume you do) ?
> > >
> > > Why not share your own experience (bugs you encountered, tips for
> > > good operations practice, etc) ?
> > >
> > > Thanks
> > >
> > > On Tue, Feb 14, 2017 at 5:52 AM, gaurhari dass  >
> > > wrote:
> > >
> > > > I was not sure by sharing some one's email id publicly so I am
> > confused.
> > > >
> > > > As I  don't want to get into legal trouble.
> > > >
> > > > Thanks
> > > > Gaurhari
> > > >
> > > > On Tue, Feb 14, 2017 at 1:41 PM, Ted Yu  wrote:
> > > >
> > > > > The previous thread from mailing list is on-going.
> > > > >
> > > > > I would expect proper edit of the thread, including attribution to
> > > > > participants.
> > > > > Otherwise the post doesn't have much value since prior conversation
> > is
> > > > > publicly available.
> > > > >
> > > > > > On Feb 14, 2017, at 5:31 AM, gaurhari dass <
> gaurharid...@gmail.com
> > >
> > > > > wrote:
> > > > > >
> > > > > > Hi Yu,
> > > > > >
> > > > > > No there is no permission required.
> > > > > >
> > > > > > https://www.linkedin.com/pulse/hbase-read-write-
> > > > > performance-conversation-gaurhari-dass
> > > > > >
> > > > > > @Richard yes it is that conversation i copied from current
> emails.
> > > > > >
> > > > > > if there is any issue then let me know or should I include link.
> > > > > >
> > > > > >
> > > > > >
> > > > > > Thanks
> > > > > > Gaurhari
> > > > > >
> > > > > >
> > > > > >
> > > > > >> On Tue, Feb 14, 2017 at 12:30 PM, Ted Yu 
> > > wrote:
> > > > > >>
> > > > > >> [image: Boxbe]  Automatic
> > Cleanup:
> > > > keep
> > > > > >> last 1 emails (yuzhih...@gmail.com) Edit rule
> > > > > >>  > > > > com%2Fcleanup%3Fkey%3Dn5F%252FWrvrtfr4JtKVYnoPZ1SoaKtmC8
> > > > > mmR%252BzFiiUP57k%253D%26token%3DDUvFf4%
> 252B0DZ9ynrpiyl5%252FXFJ9x%
> > > > > 252FwkY9UD5xRjLEU9FjM3TaObJLdwRr9HKKk6xH%252BNpF9%252BZ76IWvgV%
> > > > > 252BgrntNS8LRAoeKvDPBvmwZsd3xHQ329mVZh6QMhaGEwjPVq9pa%
> > > > > 252BTcpWjKZBzM7Q%253D&tc_serial=28883268824&tc_rand=
> > > > > 1790716340&utm_source=stf&utm_medium=email&utm_campaign=
> > > > > ANNO_CLEANUP_EDIT&utm_content=001>
> > > > > >> | Delete rule
> > > > > >>  > > > > com%2Fcleanup%3Fkey%3Dn5F%252FWrvrtfr4JtKVYnoPZ1SoaKtmC8
> > > > > mmR%252BzFiiUP57k%253D%26token%3DDUvFf4%
> 252B0DZ9ynrpiyl5%252FXFJ9x%
> > > > > 252FwkY9UD5xRjLEU9FjM3TaObJLdwRr9HKKk6xH%252BNpF9%252BZ76IWvgV%
> > > > > 252BgrntNS8LRAoeKvDPBvmwZsd3xHQ329mVZh6QMhaGEwjPVq9pa%
> > > > > 252BTcpWjKZBzM7Q%253D&tc_serial=28883268824&tc_rand=
> > > > > 1790716340&utm_

Re: Doubt

2017-02-14 Thread Ted Yu
All the major NoSQL projects are indexed by http://search-hadoop.com/

You can easily find relevant information there.

FYI

On Tue, Feb 14, 2017 at 7:28 AM, gaurhari dass 
wrote:

> Actually I am not using HBase right now at large scale.
>
> I am using Mongodb mainly so to keep myself updated I follow these
> discussions to learn about issues so I post it on my profile so that I can
> have easy access to those issues.
>
> Thanks
> Gaurhari
>
> On Tue, Feb 14, 2017 at 3:15 PM, Ted Yu  wrote:
>
> > [image: Boxbe]  Automatic Cleanup: keep
> > last 1 emails (yuzhih...@gmail.com) Edit rule
> >  com%2Fcleanup%3Fkey%3DTGqVsGxY5rwRFnibssg44ywsKdnF
> LlvN7oL%252BZGzAtdc%253D%26token%3Dn%252FzPWvUmHmaveE0%
> 252FIOALy5uUs37HKIhENDj5Blu3WvnYK%252FnGhG%252FVlR4Sj%
> 252BlEr5a9BtoaVIWeQdGT%252B2eB2dXJDAV0dXkeIYhzbEyBV7sXFv%
> 252BDwGyFMXMAO3Oy66qytBf3u3OkHvnhQFo%253D&tc_serial=28885329001&tc_rand=
> 196581&utm_source=stf&utm_medium=email&utm_campaign=
> ANNO_CLEANUP_EDIT&utm_content=001>
> > | Delete rule
> >  com%2Fcleanup%3Fkey%3DTGqVsGxY5rwRFnibssg44ywsKdnF
> LlvN7oL%252BZGzAtdc%253D%26token%3Dn%252FzPWvUmHmaveE0%
> 252FIOALy5uUs37HKIhENDj5Blu3WvnYK%252FnGhG%252FVlR4Sj%
> 252BlEr5a9BtoaVIWeQdGT%252B2eB2dXJDAV0dXkeIYhzbEyBV7sXFv%
> 252BDwGyFMXMAO3Oy66qytBf3u3OkHvnhQFo%253D&tc_serial=28885329001&tc_rand=
> 196581&utm_source=stf&utm_medium=email&utm_campaign=
> ANNO_CLEANUP_EDIT&utm_content=001>
> > | Mark important
> >  com%2Fcleanup%3Fkey%3DTGqVsGxY5rwRFnibssg44ywsKdnF
> LlvN7oL%252BZGzAtdc%253D%26token%3Dn%252FzPWvUmHmaveE0%
> 252FIOALy5uUs37HKIhENDj5Blu3WvnYK%252FnGhG%252FVlR4Sj%
> 252BlEr5a9BtoaVIWeQdGT%252B2eB2dXJDAV0dXkeIYhzbEyBV7sXFv%
> 252BDwGyFMXMAO3Oy66qytBf3u3OkHvnhQFo%253D%26important%
> 3Dtrue%26emlId%3D56028479123&tc_serial=28885329001&tc_rand=
> 196581&utm_source=stf&utm_medium=email&utm_campaign=
> ANNO_CLEANUP_EDIT&utm_content=001>
> >
> > Do you use hbase (I assume you do) ?
> >
> > Why not share your own experience (bugs you encountered, tips for
> > good operations practice, etc) ?
> >
> > Thanks
> >
> > On Tue, Feb 14, 2017 at 5:52 AM, gaurhari dass 
> > wrote:
> >
> > > I was not sure by sharing some one's email id publicly so I am
> confused.
> > >
> > > As I  don't want to get into legal trouble.
> > >
> > > Thanks
> > > Gaurhari
> > >
> > > On Tue, Feb 14, 2017 at 1:41 PM, Ted Yu  wrote:
> > >
> > > > The previous thread from mailing list is on-going.
> > > >
> > > > I would expect proper edit of the thread, including attribution to
> > > > participants.
> > > > Otherwise the post doesn't have much value since prior conversation
> is
> > > > publicly available.
> > > >
> > > > > On Feb 14, 2017, at 5:31 AM, gaurhari dass  >
> > > > wrote:
> > > > >
> > > > > Hi Yu,
> > > > >
> > > > > No there is no permission required.
> > > > >
> > > > > https://www.linkedin.com/pulse/hbase-read-write-
> > > > performance-conversation-gaurhari-dass
> > > > >
> > > > > @Richard yes it is that conversation i copied from current emails.
> > > > >
> > > > > if there is any issue then let me know or should I include link.
> > > > >
> > > > >
> > > > >
> > > > > Thanks
> > > > > Gaurhari
> > > > >
> > > > >
> > > > >
> > > > >> On Tue, Feb 14, 2017 at 12:30 PM, Ted Yu 
> > wrote:
> > > > >>
> > > > >> [image: Boxbe]  Automatic
> Cleanup:
> > > keep
> > > > >> last 1 emails (yuzhih...@gmail.com) Edit rule
> > > > >>  > > > com%2Fcleanup%3Fkey%3Dn5F%252FWrvrtfr4JtKVYnoPZ1SoaKtmC8
> > > > mmR%252BzFiiUP57k%253D%26token%3DDUvFf4%252B0DZ9ynrpiyl5%252FXFJ9x%
> > > > 252FwkY9UD5xRjLEU9FjM3TaObJLdwRr9HKKk6xH%252BNpF9%252BZ76IWvgV%
> > > > 252BgrntNS8LRAoeKvDPBvmwZsd3xHQ329mVZh6QMhaGEwjPVq9pa%
> > > > 252BTcpWjKZBzM7Q%253D&tc_serial=28883268824&tc_rand=
> > > > 1790716340&utm_source=stf&utm_medium=email&utm_campaign=
> > > > ANNO_CLEANUP_EDIT&utm_content=001>
> > > > >> | Delete rule
> > > > >>  > > > com%2Fcleanup%3Fkey%3Dn5F%252FWrvrtfr4JtKVYnoPZ1SoaKtmC8
> > > > mmR%252BzFiiUP57k%253D%26token%3DDUvFf4%252B0DZ9ynrpiyl5%252FXFJ9x%
> > > > 252FwkY9UD5xRjLEU9FjM3TaObJLdwRr9HKKk6xH%252BNpF9%252BZ76IWvgV%
> > > > 252BgrntNS8LRAoeKvDPBvmwZsd3xHQ329mVZh6QMhaGEwjPVq9pa%
> > > > 252BTcpWjKZBzM7Q%253D&tc_serial=28883268824&tc_rand=
> > > > 1790716340&utm_source=stf&utm_medium=email&utm_campaign=
> > > > ANNO_CLEANUP_EDIT&utm_content=001>
> > > > >> | Mark important
> > > > >>  > > > com%2Fcleanup%3Fkey%3Dn5F%252FWrvrtfr4JtKVYnoPZ1SoaKtmC8
> > > > mmR%252BzFiiUP57k%253D%26token%3DDUvFf4%252B0DZ9ynrpiyl5%252FXFJ9x%
> > > > 252FwkY9UD5xRjLEU9FjM3TaObJLdwRr9HKKk6xH%252BNpF9%252BZ76IWvgV%
> > > > 252BgrntNS8LRAoe

Re: Doubt

2017-02-14 Thread gaurhari dass
Actually I am not using HBase right now at large scale.

I am using Mongodb mainly so to keep myself updated I follow these
discussions to learn about issues so I post it on my profile so that I can
have easy access to those issues.

Thanks
Gaurhari

On Tue, Feb 14, 2017 at 3:15 PM, Ted Yu  wrote:

> [image: Boxbe]  Automatic Cleanup: keep
> last 1 emails (yuzhih...@gmail.com) Edit rule
> 
> | Delete rule
> 
> | Mark important
> 
>
> Do you use hbase (I assume you do) ?
>
> Why not share your own experience (bugs you encountered, tips for
> good operations practice, etc) ?
>
> Thanks
>
> On Tue, Feb 14, 2017 at 5:52 AM, gaurhari dass 
> wrote:
>
> > I was not sure by sharing some one's email id publicly so I am confused.
> >
> > As I  don't want to get into legal trouble.
> >
> > Thanks
> > Gaurhari
> >
> > On Tue, Feb 14, 2017 at 1:41 PM, Ted Yu  wrote:
> >
> > > The previous thread from mailing list is on-going.
> > >
> > > I would expect proper edit of the thread, including attribution to
> > > participants.
> > > Otherwise the post doesn't have much value since prior conversation is
> > > publicly available.
> > >
> > > > On Feb 14, 2017, at 5:31 AM, gaurhari dass 
> > > wrote:
> > > >
> > > > Hi Yu,
> > > >
> > > > No there is no permission required.
> > > >
> > > > https://www.linkedin.com/pulse/hbase-read-write-
> > > performance-conversation-gaurhari-dass
> > > >
> > > > @Richard yes it is that conversation i copied from current emails.
> > > >
> > > > if there is any issue then let me know or should I include link.
> > > >
> > > >
> > > >
> > > > Thanks
> > > > Gaurhari
> > > >
> > > >
> > > >
> > > >> On Tue, Feb 14, 2017 at 12:30 PM, Ted Yu 
> wrote:
> > > >>
> > > >> [image: Boxbe]  Automatic Cleanup:
> > keep
> > > >> last 1 emails (yuzhih...@gmail.com) Edit rule
> > > >>  > > com%2Fcleanup%3Fkey%3Dn5F%252FWrvrtfr4JtKVYnoPZ1SoaKtmC8
> > > mmR%252BzFiiUP57k%253D%26token%3DDUvFf4%252B0DZ9ynrpiyl5%252FXFJ9x%
> > > 252FwkY9UD5xRjLEU9FjM3TaObJLdwRr9HKKk6xH%252BNpF9%252BZ76IWvgV%
> > > 252BgrntNS8LRAoeKvDPBvmwZsd3xHQ329mVZh6QMhaGEwjPVq9pa%
> > > 252BTcpWjKZBzM7Q%253D&tc_serial=28883268824&tc_rand=
> > > 1790716340&utm_source=stf&utm_medium=email&utm_campaign=
> > > ANNO_CLEANUP_EDIT&utm_content=001>
> > > >> | Delete rule
> > > >>  > > com%2Fcleanup%3Fkey%3Dn5F%252FWrvrtfr4JtKVYnoPZ1SoaKtmC8
> > > mmR%252BzFiiUP57k%253D%26token%3DDUvFf4%252B0DZ9ynrpiyl5%252FXFJ9x%
> > > 252FwkY9UD5xRjLEU9FjM3TaObJLdwRr9HKKk6xH%252BNpF9%252BZ76IWvgV%
> > > 252BgrntNS8LRAoeKvDPBvmwZsd3xHQ329mVZh6QMhaGEwjPVq9pa%
> > > 252BTcpWjKZBzM7Q%253D&tc_serial=28883268824&tc_rand=
> > > 1790716340&utm_source=stf&utm_medium=email&utm_campaign=
> > > ANNO_CLEANUP_EDIT&utm_content=001>
> > > >> | Mark important
> > > >>  > > com%2Fcleanup%3Fkey%3Dn5F%252FWrvrtfr4JtKVYnoPZ1SoaKtmC8
> > > mmR%252BzFiiUP57k%253D%26token%3DDUvFf4%252B0DZ9ynrpiyl5%252FXFJ9x%
> > > 252FwkY9UD5xRjLEU9FjM3TaObJLdwRr9HKKk6xH%252BNpF9%252BZ76IWvgV%
> > > 252BgrntNS8LRAoeKvDPBvmwZsd3xHQ329mVZh6QMhaGEwjPVq9pa%
> > > 252BTcpWjKZBzM7Q%253D%26important%3Dtrue%26emlId%
> > 3D56023041068&tc_serial=
> > > 28883268824&tc_rand=1790716340&utm_source=stf&utm_
> > > medium=email&utm_campaign=ANNO_CLEANUP_EDIT&utm_content=001>
> > > >>
> > > >> Clicking on both links directed me to:
> > > >>
> > > >> https://www.linkedin.com/post/new
> > > >>
> > > >> Do the pages require read permission ?
> > > >>
> > > >> On Tue, F

Re: Doubt

2017-02-14 Thread Ted Yu
Do you use hbase (I assume you do) ?

Why not share your own experience (bugs you encountered, tips for
good operations practice, etc) ?

Thanks

On Tue, Feb 14, 2017 at 5:52 AM, gaurhari dass 
wrote:

> I was not sure by sharing some one's email id publicly so I am confused.
>
> As I  don't want to get into legal trouble.
>
> Thanks
> Gaurhari
>
> On Tue, Feb 14, 2017 at 1:41 PM, Ted Yu  wrote:
>
> > The previous thread from mailing list is on-going.
> >
> > I would expect proper edit of the thread, including attribution to
> > participants.
> > Otherwise the post doesn't have much value since prior conversation is
> > publicly available.
> >
> > > On Feb 14, 2017, at 5:31 AM, gaurhari dass 
> > wrote:
> > >
> > > Hi Yu,
> > >
> > > No there is no permission required.
> > >
> > > https://www.linkedin.com/pulse/hbase-read-write-
> > performance-conversation-gaurhari-dass
> > >
> > > @Richard yes it is that conversation i copied from current emails.
> > >
> > > if there is any issue then let me know or should I include link.
> > >
> > >
> > >
> > > Thanks
> > > Gaurhari
> > >
> > >
> > >
> > >> On Tue, Feb 14, 2017 at 12:30 PM, Ted Yu  wrote:
> > >>
> > >> [image: Boxbe]  Automatic Cleanup:
> keep
> > >> last 1 emails (yuzhih...@gmail.com) Edit rule
> > >>  > com%2Fcleanup%3Fkey%3Dn5F%252FWrvrtfr4JtKVYnoPZ1SoaKtmC8
> > mmR%252BzFiiUP57k%253D%26token%3DDUvFf4%252B0DZ9ynrpiyl5%252FXFJ9x%
> > 252FwkY9UD5xRjLEU9FjM3TaObJLdwRr9HKKk6xH%252BNpF9%252BZ76IWvgV%
> > 252BgrntNS8LRAoeKvDPBvmwZsd3xHQ329mVZh6QMhaGEwjPVq9pa%
> > 252BTcpWjKZBzM7Q%253D&tc_serial=28883268824&tc_rand=
> > 1790716340&utm_source=stf&utm_medium=email&utm_campaign=
> > ANNO_CLEANUP_EDIT&utm_content=001>
> > >> | Delete rule
> > >>  > com%2Fcleanup%3Fkey%3Dn5F%252FWrvrtfr4JtKVYnoPZ1SoaKtmC8
> > mmR%252BzFiiUP57k%253D%26token%3DDUvFf4%252B0DZ9ynrpiyl5%252FXFJ9x%
> > 252FwkY9UD5xRjLEU9FjM3TaObJLdwRr9HKKk6xH%252BNpF9%252BZ76IWvgV%
> > 252BgrntNS8LRAoeKvDPBvmwZsd3xHQ329mVZh6QMhaGEwjPVq9pa%
> > 252BTcpWjKZBzM7Q%253D&tc_serial=28883268824&tc_rand=
> > 1790716340&utm_source=stf&utm_medium=email&utm_campaign=
> > ANNO_CLEANUP_EDIT&utm_content=001>
> > >> | Mark important
> > >>  > com%2Fcleanup%3Fkey%3Dn5F%252FWrvrtfr4JtKVYnoPZ1SoaKtmC8
> > mmR%252BzFiiUP57k%253D%26token%3DDUvFf4%252B0DZ9ynrpiyl5%252FXFJ9x%
> > 252FwkY9UD5xRjLEU9FjM3TaObJLdwRr9HKKk6xH%252BNpF9%252BZ76IWvgV%
> > 252BgrntNS8LRAoeKvDPBvmwZsd3xHQ329mVZh6QMhaGEwjPVq9pa%
> > 252BTcpWjKZBzM7Q%253D%26important%3Dtrue%26emlId%
> 3D56023041068&tc_serial=
> > 28883268824&tc_rand=1790716340&utm_source=stf&utm_
> > medium=email&utm_campaign=ANNO_CLEANUP_EDIT&utm_content=001>
> > >>
> > >> Clicking on both links directed me to:
> > >>
> > >> https://www.linkedin.com/post/new
> > >>
> > >> Do the pages require read permission ?
> > >>
> > >> On Tue, Feb 14, 2017 at 1:46 AM, gaurhari dass <
> gaurharid...@gmail.com>
> > >> wrote:
> > >>
> > >>> Hi
> > >>>
> > >>> I want to post like this
> > >>>
> > >>> https://www.linkedin.com/post/edit/hbase-read-write-
> > >>> performance-conversation-gaurhari-dass
> > >>>
> > >>> https://www.linkedin.com/post/edit/hbase-performance-
> > >>> improvements-gaurhari-dass
> > >>>
> > >>> it is just for my reference as well can be helpful to others if it is
> > ok.
> > >>>
> > >>> Thanks
> > >>> Gaurhari
> > >>>
> >  On Tue, Feb 14, 2017 at 6:27 AM, Yu Li  wrote:
> > 
> >  [image: Boxbe]  This message is
> > >> eligible
> >  for Automatic Cleanup! (car...@gmail.com) Add cleanup rule
> >   > >>> com%2Fcleanup%3Fkey%3Dlo0h5jSW%252B4%252FJDQc5MGAN%252Fm6wlgiwZwd%
> > >>> 252Fg%252BsPLxAODrs%253D%26token%3DtcsooRleiqYIyi33PWCX0y%
> > >>> 252BERL88Oa80r1utgWgCgYPz7LFe6D0RdRDy4m%
> 252BmfgI1OPas8eDVfAU3GCEFxpeB6
> > >>> cWVAplUMXEpvD4gZ%252FMYuYE25tzifmD6E7XFNwZ7KX3X
> > >> nWM7mLgh1Ns%253D&tc_serial=
> > >>> 28877655010&tc_rand=1497511706&utm_source=stf&utm_
> > >>> medium=email&utm_campaign=ANNO_CLEANUP_ADD&utm_content=001>
> >  | More info
> >   > >>> cleanup?tc_serial=28877655010&tc_rand=1497511706&utm_source=
> > >>> stf&utm_medium=email&utm_campaign=ANNO_CLEANUP_ADD&utm_content=001>
> > 
> >  Would like to hear more details, wherever it will be posted (smile).
> > 
> >  Best Regards,
> >  Yu
> > 
> > > On 14 February 2017 at 06:06, Stack  wrote:
> > >
> > > You might consider adding list of issues and solutions to the hbase
> > > reference guide?
> > > Yours,
> > > S
> > >
> > > On Fri, Feb 10, 2017 at 1:34 AM, gaurhari dass <
> > >> gaurharid...@gmail.com
> > 
> > > wrote:
> > >
> > >> Hi ,
> > >>
> > >> I always follow pro

Re: Doubt

2017-02-14 Thread gaurhari dass
I was not sure by sharing some one's email id publicly so I am confused.

As I  don't want to get into legal trouble.

Thanks
Gaurhari

On Tue, Feb 14, 2017 at 1:41 PM, Ted Yu  wrote:

> The previous thread from mailing list is on-going.
>
> I would expect proper edit of the thread, including attribution to
> participants.
> Otherwise the post doesn't have much value since prior conversation is
> publicly available.
>
> > On Feb 14, 2017, at 5:31 AM, gaurhari dass 
> wrote:
> >
> > Hi Yu,
> >
> > No there is no permission required.
> >
> > https://www.linkedin.com/pulse/hbase-read-write-
> performance-conversation-gaurhari-dass
> >
> > @Richard yes it is that conversation i copied from current emails.
> >
> > if there is any issue then let me know or should I include link.
> >
> >
> >
> > Thanks
> > Gaurhari
> >
> >
> >
> >> On Tue, Feb 14, 2017 at 12:30 PM, Ted Yu  wrote:
> >>
> >> [image: Boxbe]  Automatic Cleanup: keep
> >> last 1 emails (yuzhih...@gmail.com) Edit rule
> >>  com%2Fcleanup%3Fkey%3Dn5F%252FWrvrtfr4JtKVYnoPZ1SoaKtmC8
> mmR%252BzFiiUP57k%253D%26token%3DDUvFf4%252B0DZ9ynrpiyl5%252FXFJ9x%
> 252FwkY9UD5xRjLEU9FjM3TaObJLdwRr9HKKk6xH%252BNpF9%252BZ76IWvgV%
> 252BgrntNS8LRAoeKvDPBvmwZsd3xHQ329mVZh6QMhaGEwjPVq9pa%
> 252BTcpWjKZBzM7Q%253D&tc_serial=28883268824&tc_rand=
> 1790716340&utm_source=stf&utm_medium=email&utm_campaign=
> ANNO_CLEANUP_EDIT&utm_content=001>
> >> | Delete rule
> >>  com%2Fcleanup%3Fkey%3Dn5F%252FWrvrtfr4JtKVYnoPZ1SoaKtmC8
> mmR%252BzFiiUP57k%253D%26token%3DDUvFf4%252B0DZ9ynrpiyl5%252FXFJ9x%
> 252FwkY9UD5xRjLEU9FjM3TaObJLdwRr9HKKk6xH%252BNpF9%252BZ76IWvgV%
> 252BgrntNS8LRAoeKvDPBvmwZsd3xHQ329mVZh6QMhaGEwjPVq9pa%
> 252BTcpWjKZBzM7Q%253D&tc_serial=28883268824&tc_rand=
> 1790716340&utm_source=stf&utm_medium=email&utm_campaign=
> ANNO_CLEANUP_EDIT&utm_content=001>
> >> | Mark important
> >>  com%2Fcleanup%3Fkey%3Dn5F%252FWrvrtfr4JtKVYnoPZ1SoaKtmC8
> mmR%252BzFiiUP57k%253D%26token%3DDUvFf4%252B0DZ9ynrpiyl5%252FXFJ9x%
> 252FwkY9UD5xRjLEU9FjM3TaObJLdwRr9HKKk6xH%252BNpF9%252BZ76IWvgV%
> 252BgrntNS8LRAoeKvDPBvmwZsd3xHQ329mVZh6QMhaGEwjPVq9pa%
> 252BTcpWjKZBzM7Q%253D%26important%3Dtrue%26emlId%3D56023041068&tc_serial=
> 28883268824&tc_rand=1790716340&utm_source=stf&utm_
> medium=email&utm_campaign=ANNO_CLEANUP_EDIT&utm_content=001>
> >>
> >> Clicking on both links directed me to:
> >>
> >> https://www.linkedin.com/post/new
> >>
> >> Do the pages require read permission ?
> >>
> >> On Tue, Feb 14, 2017 at 1:46 AM, gaurhari dass 
> >> wrote:
> >>
> >>> Hi
> >>>
> >>> I want to post like this
> >>>
> >>> https://www.linkedin.com/post/edit/hbase-read-write-
> >>> performance-conversation-gaurhari-dass
> >>>
> >>> https://www.linkedin.com/post/edit/hbase-performance-
> >>> improvements-gaurhari-dass
> >>>
> >>> it is just for my reference as well can be helpful to others if it is
> ok.
> >>>
> >>> Thanks
> >>> Gaurhari
> >>>
>  On Tue, Feb 14, 2017 at 6:27 AM, Yu Li  wrote:
> 
>  [image: Boxbe]  This message is
> >> eligible
>  for Automatic Cleanup! (car...@gmail.com) Add cleanup rule
>   >>> com%2Fcleanup%3Fkey%3Dlo0h5jSW%252B4%252FJDQc5MGAN%252Fm6wlgiwZwd%
> >>> 252Fg%252BsPLxAODrs%253D%26token%3DtcsooRleiqYIyi33PWCX0y%
> >>> 252BERL88Oa80r1utgWgCgYPz7LFe6D0RdRDy4m%252BmfgI1OPas8eDVfAU3GCEFxpeB6
> >>> cWVAplUMXEpvD4gZ%252FMYuYE25tzifmD6E7XFNwZ7KX3X
> >> nWM7mLgh1Ns%253D&tc_serial=
> >>> 28877655010&tc_rand=1497511706&utm_source=stf&utm_
> >>> medium=email&utm_campaign=ANNO_CLEANUP_ADD&utm_content=001>
>  | More info
>   >>> cleanup?tc_serial=28877655010&tc_rand=1497511706&utm_source=
> >>> stf&utm_medium=email&utm_campaign=ANNO_CLEANUP_ADD&utm_content=001>
> 
>  Would like to hear more details, wherever it will be posted (smile).
> 
>  Best Regards,
>  Yu
> 
> > On 14 February 2017 at 06:06, Stack  wrote:
> >
> > You might consider adding list of issues and solutions to the hbase
> > reference guide?
> > Yours,
> > S
> >
> > On Fri, Feb 10, 2017 at 1:34 AM, gaurhari dass <
> >> gaurharid...@gmail.com
> 
> > wrote:
> >
> >> Hi ,
> >>
> >> I always follow problems users facing with hbase and solutions
>  provided.
> >>
> >> these problems and solutions are always helpful to me.
> >>
> >> I want to know if it is ok if I share these problems and solutions
> >> on
>  my
> >> linkedin articles as it is.
> >>
> >> Thanks
> >> Gaurhari
> >>
> >>
>


Re: Doubt

2017-02-14 Thread Ted Yu
The previous thread from mailing list is on-going. 

I would expect proper edit of the thread, including attribution to 
participants. 
Otherwise the post doesn't have much value since prior conversation is publicly 
available. 

> On Feb 14, 2017, at 5:31 AM, gaurhari dass  wrote:
> 
> Hi Yu,
> 
> No there is no permission required.
> 
> https://www.linkedin.com/pulse/hbase-read-write-performance-conversation-gaurhari-dass
> 
> @Richard yes it is that conversation i copied from current emails.
> 
> if there is any issue then let me know or should I include link.
> 
> 
> 
> Thanks
> Gaurhari
> 
> 
> 
>> On Tue, Feb 14, 2017 at 12:30 PM, Ted Yu  wrote:
>> 
>> [image: Boxbe]  Automatic Cleanup: keep
>> last 1 emails (yuzhih...@gmail.com) Edit rule
>> 
>> | Delete rule
>> 
>> | Mark important
>> 
>> 
>> Clicking on both links directed me to:
>> 
>> https://www.linkedin.com/post/new
>> 
>> Do the pages require read permission ?
>> 
>> On Tue, Feb 14, 2017 at 1:46 AM, gaurhari dass 
>> wrote:
>> 
>>> Hi
>>> 
>>> I want to post like this
>>> 
>>> https://www.linkedin.com/post/edit/hbase-read-write-
>>> performance-conversation-gaurhari-dass
>>> 
>>> https://www.linkedin.com/post/edit/hbase-performance-
>>> improvements-gaurhari-dass
>>> 
>>> it is just for my reference as well can be helpful to others if it is ok.
>>> 
>>> Thanks
>>> Gaurhari
>>> 
 On Tue, Feb 14, 2017 at 6:27 AM, Yu Li  wrote:
 
 [image: Boxbe]  This message is
>> eligible
 for Automatic Cleanup! (car...@gmail.com) Add cleanup rule
 >> com%2Fcleanup%3Fkey%3Dlo0h5jSW%252B4%252FJDQc5MGAN%252Fm6wlgiwZwd%
>>> 252Fg%252BsPLxAODrs%253D%26token%3DtcsooRleiqYIyi33PWCX0y%
>>> 252BERL88Oa80r1utgWgCgYPz7LFe6D0RdRDy4m%252BmfgI1OPas8eDVfAU3GCEFxpeB6
>>> cWVAplUMXEpvD4gZ%252FMYuYE25tzifmD6E7XFNwZ7KX3X
>> nWM7mLgh1Ns%253D&tc_serial=
>>> 28877655010&tc_rand=1497511706&utm_source=stf&utm_
>>> medium=email&utm_campaign=ANNO_CLEANUP_ADD&utm_content=001>
 | More info
 >> cleanup?tc_serial=28877655010&tc_rand=1497511706&utm_source=
>>> stf&utm_medium=email&utm_campaign=ANNO_CLEANUP_ADD&utm_content=001>
 
 Would like to hear more details, wherever it will be posted (smile).
 
 Best Regards,
 Yu
 
> On 14 February 2017 at 06:06, Stack  wrote:
> 
> You might consider adding list of issues and solutions to the hbase
> reference guide?
> Yours,
> S
> 
> On Fri, Feb 10, 2017 at 1:34 AM, gaurhari dass <
>> gaurharid...@gmail.com
 
> wrote:
> 
>> Hi ,
>> 
>> I always follow problems users facing with hbase and solutions
 provided.
>> 
>> these problems and solutions are always helpful to me.
>> 
>> I want to know if it is ok if I share these problems and solutions
>> on
 my
>> linkedin articles as it is.
>> 
>> Thanks
>> Gaurhari
>> 
>> 


Re: Doubt

2017-02-14 Thread gaurhari dass
Hi Yu,

No there is no permission required.

https://www.linkedin.com/pulse/hbase-read-write-performance-conversation-gaurhari-dass

@Richard yes it is that conversation i copied from current emails.

if there is any issue then let me know or should I include link.



Thanks
Gaurhari



On Tue, Feb 14, 2017 at 12:30 PM, Ted Yu  wrote:

> [image: Boxbe]  Automatic Cleanup: keep
> last 1 emails (yuzhih...@gmail.com) Edit rule
> 
> | Delete rule
> 
> | Mark important
> 
>
> Clicking on both links directed me to:
>
> https://www.linkedin.com/post/new
>
> Do the pages require read permission ?
>
> On Tue, Feb 14, 2017 at 1:46 AM, gaurhari dass 
> wrote:
>
> > Hi
> >
> > I want to post like this
> >
> > https://www.linkedin.com/post/edit/hbase-read-write-
> > performance-conversation-gaurhari-dass
> >
> > https://www.linkedin.com/post/edit/hbase-performance-
> > improvements-gaurhari-dass
> >
> > it is just for my reference as well can be helpful to others if it is ok.
> >
> > Thanks
> > Gaurhari
> >
> > On Tue, Feb 14, 2017 at 6:27 AM, Yu Li  wrote:
> >
> > > [image: Boxbe]  This message is
> eligible
> > > for Automatic Cleanup! (car...@gmail.com) Add cleanup rule
> > >  > com%2Fcleanup%3Fkey%3Dlo0h5jSW%252B4%252FJDQc5MGAN%252Fm6wlgiwZwd%
> > 252Fg%252BsPLxAODrs%253D%26token%3DtcsooRleiqYIyi33PWCX0y%
> > 252BERL88Oa80r1utgWgCgYPz7LFe6D0RdRDy4m%252BmfgI1OPas8eDVfAU3GCEFxpeB6
> > cWVAplUMXEpvD4gZ%252FMYuYE25tzifmD6E7XFNwZ7KX3X
> nWM7mLgh1Ns%253D&tc_serial=
> > 28877655010&tc_rand=1497511706&utm_source=stf&utm_
> > medium=email&utm_campaign=ANNO_CLEANUP_ADD&utm_content=001>
> > > | More info
> > >  > cleanup?tc_serial=28877655010&tc_rand=1497511706&utm_source=
> > stf&utm_medium=email&utm_campaign=ANNO_CLEANUP_ADD&utm_content=001>
> > >
> > > Would like to hear more details, wherever it will be posted (smile).
> > >
> > > Best Regards,
> > > Yu
> > >
> > > On 14 February 2017 at 06:06, Stack  wrote:
> > >
> > > > You might consider adding list of issues and solutions to the hbase
> > > > reference guide?
> > > > Yours,
> > > > S
> > > >
> > > > On Fri, Feb 10, 2017 at 1:34 AM, gaurhari dass <
> gaurharid...@gmail.com
> > >
> > > > wrote:
> > > >
> > > > > Hi ,
> > > > >
> > > > > I always follow problems users facing with hbase and solutions
> > > provided.
> > > > >
> > > > > these problems and solutions are always helpful to me.
> > > > >
> > > > > I want to know if it is ok if I share these problems and solutions
> on
> > > my
> > > > > linkedin articles as it is.
> > > > >
> > > > > Thanks
> > > > > Gaurhari
> > > > >
> > > >
> > >
> > >
> >
>
>


Re: Doubt

2017-02-14 Thread Richard Startin
I took a look at 
https://www.linkedin.com/pulse/hbase-read-write-performance-conversation-gaurhari-dass?trk=prof-post


Looks like an unattributed copy of 
<http://apache-hbase.679495.n3.nabble.com/HBase-Performance-td4086182.html> 
http://apache-hbase.679495.n3.nabble.com/HBase-Performance-td4086182.html#a4086185


https://richardstartin.com/



From: Ted Yu 
Sent: 14 February 2017 12:30
To: user@hbase.apache.org
Subject: Re: Doubt

Clicking on both links directed me to:

https://www.linkedin.com/post/new

Do the pages require read permission ?

On Tue, Feb 14, 2017 at 1:46 AM, gaurhari dass 
wrote:

> Hi
>
> I want to post like this
>
> https://www.linkedin.com/post/edit/hbase-read-write-
> performance-conversation-gaurhari-dass
>
> https://www.linkedin.com/post/edit/hbase-performance-
> improvements-gaurhari-dass
>
> it is just for my reference as well can be helpful to others if it is ok.
>
> Thanks
> Gaurhari
>
> On Tue, Feb 14, 2017 at 6:27 AM, Yu Li  wrote:
>
> > [image: Boxbe] <https://www.boxbe.com/overview> This message is eligible
> > for Automatic Cleanup! (car...@gmail.com) Add cleanup rule
> > <https://www.boxbe.com/popup?url=https%3A%2F%2Fwww.boxbe.
> com%2Fcleanup%3Fkey%3Dlo0h5jSW%252B4%252FJDQc5MGAN%252Fm6wlgiwZwd%
> 252Fg%252BsPLxAODrs%253D%26token%3DtcsooRleiqYIyi33PWCX0y%
> 252BERL88Oa80r1utgWgCgYPz7LFe6D0RdRDy4m%252BmfgI1OPas8eDVfAU3GCEFxpeB6
> cWVAplUMXEpvD4gZ%252FMYuYE25tzifmD6E7XFNwZ7KX3XnWM7mLgh1Ns%253D&tc_serial=
> 28877655010&tc_rand=1497511706&utm_source=stf&utm_
> medium=email&utm_campaign=ANNO_CLEANUP_ADD&utm_content=001>
> > | More info
> > <http://blog.boxbe.com/general/boxbe-automatic-
> cleanup?tc_serial=28877655010&tc_rand=1497511706&utm_source=
> stf&utm_medium=email&utm_campaign=ANNO_CLEANUP_ADD&utm_content=001>
> >
> > Would like to hear more details, wherever it will be posted (smile).
> >
> > Best Regards,
> > Yu
> >
> > On 14 February 2017 at 06:06, Stack  wrote:
> >
> > > You might consider adding list of issues and solutions to the hbase
> > > reference guide?
> > > Yours,
> > > S
> > >
> > > On Fri, Feb 10, 2017 at 1:34 AM, gaurhari dass  >
> > > wrote:
> > >
> > > > Hi ,
> > > >
> > > > I always follow problems users facing with hbase and solutions
> > provided.
> > > >
> > > > these problems and solutions are always helpful to me.
> > > >
> > > > I want to know if it is ok if I share these problems and solutions on
> > my
> > > > linkedin articles as it is.
> > > >
> > > > Thanks
> > > > Gaurhari
> > > >
> > >
> >
> >
>


Re: Doubt

2017-02-14 Thread Ted Yu
Clicking on both links directed me to:

https://www.linkedin.com/post/new

Do the pages require read permission ?

On Tue, Feb 14, 2017 at 1:46 AM, gaurhari dass 
wrote:

> Hi
>
> I want to post like this
>
> https://www.linkedin.com/post/edit/hbase-read-write-
> performance-conversation-gaurhari-dass
>
> https://www.linkedin.com/post/edit/hbase-performance-
> improvements-gaurhari-dass
>
> it is just for my reference as well can be helpful to others if it is ok.
>
> Thanks
> Gaurhari
>
> On Tue, Feb 14, 2017 at 6:27 AM, Yu Li  wrote:
>
> > [image: Boxbe]  This message is eligible
> > for Automatic Cleanup! (car...@gmail.com) Add cleanup rule
> >  com%2Fcleanup%3Fkey%3Dlo0h5jSW%252B4%252FJDQc5MGAN%252Fm6wlgiwZwd%
> 252Fg%252BsPLxAODrs%253D%26token%3DtcsooRleiqYIyi33PWCX0y%
> 252BERL88Oa80r1utgWgCgYPz7LFe6D0RdRDy4m%252BmfgI1OPas8eDVfAU3GCEFxpeB6
> cWVAplUMXEpvD4gZ%252FMYuYE25tzifmD6E7XFNwZ7KX3XnWM7mLgh1Ns%253D&tc_serial=
> 28877655010&tc_rand=1497511706&utm_source=stf&utm_
> medium=email&utm_campaign=ANNO_CLEANUP_ADD&utm_content=001>
> > | More info
> >  cleanup?tc_serial=28877655010&tc_rand=1497511706&utm_source=
> stf&utm_medium=email&utm_campaign=ANNO_CLEANUP_ADD&utm_content=001>
> >
> > Would like to hear more details, wherever it will be posted (smile).
> >
> > Best Regards,
> > Yu
> >
> > On 14 February 2017 at 06:06, Stack  wrote:
> >
> > > You might consider adding list of issues and solutions to the hbase
> > > reference guide?
> > > Yours,
> > > S
> > >
> > > On Fri, Feb 10, 2017 at 1:34 AM, gaurhari dass  >
> > > wrote:
> > >
> > > > Hi ,
> > > >
> > > > I always follow problems users facing with hbase and solutions
> > provided.
> > > >
> > > > these problems and solutions are always helpful to me.
> > > >
> > > > I want to know if it is ok if I share these problems and solutions on
> > my
> > > > linkedin articles as it is.
> > > >
> > > > Thanks
> > > > Gaurhari
> > > >
> > >
> >
> >
>


Re: Doubt

2017-02-14 Thread gaurhari dass
Hi

I want to post like this

https://www.linkedin.com/post/edit/hbase-read-write-performance-conversation-gaurhari-dass

https://www.linkedin.com/post/edit/hbase-performance-improvements-gaurhari-dass

it is just for my reference as well can be helpful to others if it is ok.

Thanks
Gaurhari

On Tue, Feb 14, 2017 at 6:27 AM, Yu Li  wrote:

> [image: Boxbe]  This message is eligible
> for Automatic Cleanup! (car...@gmail.com) Add cleanup rule
> 
> | More info
> 
>
> Would like to hear more details, wherever it will be posted (smile).
>
> Best Regards,
> Yu
>
> On 14 February 2017 at 06:06, Stack  wrote:
>
> > You might consider adding list of issues and solutions to the hbase
> > reference guide?
> > Yours,
> > S
> >
> > On Fri, Feb 10, 2017 at 1:34 AM, gaurhari dass 
> > wrote:
> >
> > > Hi ,
> > >
> > > I always follow problems users facing with hbase and solutions
> provided.
> > >
> > > these problems and solutions are always helpful to me.
> > >
> > > I want to know if it is ok if I share these problems and solutions on
> my
> > > linkedin articles as it is.
> > >
> > > Thanks
> > > Gaurhari
> > >
> >
>
>


Re: Doubt

2017-02-13 Thread Yu Li
Would like to hear more details, wherever it will be posted (smile).

Best Regards,
Yu

On 14 February 2017 at 06:06, Stack  wrote:

> You might consider adding list of issues and solutions to the hbase
> reference guide?
> Yours,
> S
>
> On Fri, Feb 10, 2017 at 1:34 AM, gaurhari dass 
> wrote:
>
> > Hi ,
> >
> > I always follow problems users facing with hbase and solutions provided.
> >
> > these problems and solutions are always helpful to me.
> >
> > I want to know if it is ok if I share these problems and solutions on my
> > linkedin articles as it is.
> >
> > Thanks
> > Gaurhari
> >
>


Re: Doubt

2017-02-13 Thread Stack
You might consider adding list of issues and solutions to the hbase
reference guide?
Yours,
S

On Fri, Feb 10, 2017 at 1:34 AM, gaurhari dass 
wrote:

> Hi ,
>
> I always follow problems users facing with hbase and solutions provided.
>
> these problems and solutions are always helpful to me.
>
> I want to know if it is ok if I share these problems and solutions on my
> linkedin articles as it is.
>
> Thanks
> Gaurhari
>


Re: Doubt

2017-02-12 Thread gaurhari dass
thanks

On Fri, Feb 10, 2017 at 9:44 PM, Josh Elser  wrote:

> [image: Boxbe]  This message is eligible
> for Automatic Cleanup! (els...@apache.org) Add cleanup rule
> 
> | More info
> 
>
> Gaurhari,
>
> This is a publicly-archived mailing list. As such, I don't believe anyone
> is going to be upset with intellectual property concerns.
>
> If anything, you could take a look at the ASF's trademark policy[1]. This
> likely covers any sensitive issues.
>
> [1] http://www.apache.org/foundation/marks/
>
> gaurhari dass wrote:
>
>> Hi ,
>>
>> I always follow problems users facing with hbase and solutions provided.
>>
>> these problems and solutions are always helpful to me.
>>
>> I want to know if it is ok if I share these problems and solutions on my
>> linkedin articles as it is.
>>
>> Thanks
>> Gaurhari
>>
>>
>


Re: Doubt

2017-02-10 Thread Josh Elser

Gaurhari,

This is a publicly-archived mailing list. As such, I don't believe 
anyone is going to be upset with intellectual property concerns.


If anything, you could take a look at the ASF's trademark policy[1]. 
This likely covers any sensitive issues.


[1] http://www.apache.org/foundation/marks/

gaurhari dass wrote:

Hi ,

I always follow problems users facing with hbase and solutions provided.

these problems and solutions are always helpful to me.

I want to know if it is ok if I share these problems and solutions on my
linkedin articles as it is.

Thanks
Gaurhari



Doubt

2017-02-10 Thread gaurhari dass
Hi ,

I always follow problems users facing with hbase and solutions provided.

these problems and solutions are always helpful to me.

I want to know if it is ok if I share these problems and solutions on my
linkedin articles as it is.

Thanks
Gaurhari


Re: Doubt in Scanner.next()

2017-01-11 Thread Ted Yu
As refguide states, hbase.client.scanner.caching works
with hbase.client.scanner.max.result.size to try and use the network
efficiently.

Make sure the release you use is 1.1.0+ which had important bug fixes
w.r.t. max result size.

On Wed, Jan 11, 2017 at 9:46 AM, Josh Elser  wrote:

> Behind the scenes, the ClientScanner is buffering results from the
> previous RPC. Ignoring multiple RegionServers for now, the caching value
> denotes the number of records that were fetched by the ClientScanner in an
> RPC. When the buffered results are consumed by your client, a new RPC will
> be made to fetch another 100 rows.
>
> Thus, larger scanner caching values reduces the number of RPCs the
> ClietnScanner object will make; however, it increases the length of the RPC
> itself (as it takes longer to aggregate a larger number of rows to return).
>
> Rajeshkumar J wrote:
>
>> I have hbase.client.scanner.caching as 100.  I am scanning a table. For
>> instance say we have 500 rows matching for the current scan. When I give
>> the statement ResultScanner.next() what will happen? whether it will
>> return
>> 100 rows for each next operation or something else
>>
>>
>> Thanks
>>
>>


Re: Doubt in Scanner.next()

2017-01-11 Thread Josh Elser
Behind the scenes, the ClientScanner is buffering results from the 
previous RPC. Ignoring multiple RegionServers for now, the caching value 
denotes the number of records that were fetched by the ClientScanner in 
an RPC. When the buffered results are consumed by your client, a new RPC 
will be made to fetch another 100 rows.


Thus, larger scanner caching values reduces the number of RPCs the 
ClietnScanner object will make; however, it increases the length of the 
RPC itself (as it takes longer to aggregate a larger number of rows to 
return).


Rajeshkumar J wrote:

I have hbase.client.scanner.caching as 100.  I am scanning a table. For
instance say we have 500 rows matching for the current scan. When I give
the statement ResultScanner.next() what will happen? whether it will return
100 rows for each next operation or something else


Thanks



Doubt in Scanner.next()

2017-01-11 Thread Rajeshkumar J
I have hbase.client.scanner.caching as 100.  I am scanning a table. For
instance say we have 500 rows matching for the current scan. When I give
the statement ResultScanner.next() what will happen? whether it will return
100 rows for each next operation or something else


Thanks


Re: Hbase master selection doubt

2015-06-27 Thread Shushant Arora
By strictly consistent I mean - all clients should see same data at any
time in different sessions.

Say a client C1 was connected to follower F1, And F1 was few seconds behind
the leader. And client C2 connects to F2 which is in sync with Leader . Now
C1 and C2 will see different data under root dir say(/master ) is visible
to C2 not to C1. Till F1 comes in sync with Leader.

On Sat, Jun 27, 2015 at 8:23 PM, Shushant Arora 
wrote:

> Zookeeper is Sequential Consistency
> Updates from a client will be applied in the order that they were sent.
>
> On Sat, Jun 27, 2015 at 8:18 PM, Ted Yu  wrote:
>
>> bq. non strictly consistency of Zookeeper
>>
>> Can you elaborate on what the above means ?
>>
>> please read this:
>>
>> http://zookeeper.apache.org/doc/trunk/zookeeperProgrammers.html#ch_zkGuarantees
>>
>> Cheers
>>
>> On Sat, Jun 27, 2015 at 7:20 AM, Shushant Arora <
>> shushantaror...@gmail.com>
>> wrote:
>>
>> > How Hbase uses Zookeeper for Master selection and region server failure
>> > detection when Zookeeper is not strictly consistent.
>> >
>> > Say In Hbase Master selection process, how does a node is 100 % sure
>> that a
>> > master is created ? Does it has to create the /master node and that node
>> > already exists will thow node exists excpetion .  Since only by reading
>> (ls
>> > /) . It may get stale data and gets node does not exists.but in actual
>> > /master was present.
>> >
>> > Does there any issue with non strictly consistency of Zookeeper for
>> Hbase?
>> >
>>
>
>


Re: Hbase master selection doubt

2015-06-27 Thread Shushant Arora
Zookeeper is Sequential Consistency
Updates from a client will be applied in the order that they were sent.

On Sat, Jun 27, 2015 at 8:18 PM, Ted Yu  wrote:

> bq. non strictly consistency of Zookeeper
>
> Can you elaborate on what the above means ?
>
> please read this:
>
> http://zookeeper.apache.org/doc/trunk/zookeeperProgrammers.html#ch_zkGuarantees
>
> Cheers
>
> On Sat, Jun 27, 2015 at 7:20 AM, Shushant Arora  >
> wrote:
>
> > How Hbase uses Zookeeper for Master selection and region server failure
> > detection when Zookeeper is not strictly consistent.
> >
> > Say In Hbase Master selection process, how does a node is 100 % sure
> that a
> > master is created ? Does it has to create the /master node and that node
> > already exists will thow node exists excpetion .  Since only by reading
> (ls
> > /) . It may get stale data and gets node does not exists.but in actual
> > /master was present.
> >
> > Does there any issue with non strictly consistency of Zookeeper for
> Hbase?
> >
>


Re: Hbase master selection doubt

2015-06-27 Thread Ted Yu
bq. non strictly consistency of Zookeeper

Can you elaborate on what the above means ?

please read this:
http://zookeeper.apache.org/doc/trunk/zookeeperProgrammers.html#ch_zkGuarantees

Cheers

On Sat, Jun 27, 2015 at 7:20 AM, Shushant Arora 
wrote:

> How Hbase uses Zookeeper for Master selection and region server failure
> detection when Zookeeper is not strictly consistent.
>
> Say In Hbase Master selection process, how does a node is 100 % sure that a
> master is created ? Does it has to create the /master node and that node
> already exists will thow node exists excpetion .  Since only by reading (ls
> /) . It may get stale data and gets node does not exists.but in actual
> /master was present.
>
> Does there any issue with non strictly consistency of Zookeeper for Hbase?
>


Hbase master selection doubt

2015-06-27 Thread Shushant Arora
How Hbase uses Zookeeper for Master selection and region server failure
detection when Zookeeper is not strictly consistent.

Say In Hbase Master selection process, how does a node is 100 % sure that a
master is created ? Does it has to create the /master node and that node
already exists will thow node exists excpetion .  Since only by reading (ls
/) . It may get stale data and gets node does not exists.but in actual
/master was present.

Does there any issue with non strictly consistency of Zookeeper for Hbase?


Re: Regarding a doubt I am having for HBase

2015-03-12 Thread Wilm Schumacher
Hi,

I would like to add a question: Why do you need the ID in the first
place? The hash seems to be generated by another source, thus is
imutable. But is this true for the ID, too? If not, why not using only
the hash?

Best wishes,

Wilm

Am 10.03.2015 um 21:40 schrieb Alex Baranau:
> CCing HBase's user ML.
>
> Could you give an example of the row key and example of two different
> queries you are making to better understand your case?
>
> Thank you,
>
> Alex Baranau
> --
> http://cdap.io - open source framework to build and run data applications
> on Hadoop & HBase
>
>
> On Mon, Mar 9, 2015 at 9:00 AM, Jaspreet Singh  wrote:
>
>>  Hi Alex,
>>
>>
>>  Thanks a lot for the response!!! The data I have is inform of hashes and
>> id, every id is related to a cookie hashed data so for the id part the
>> solution that you said would work good but in case hashes it would not be
>> possible to mention a stoprow based on increment. Also I have millions of
>> rows so duplicating the row means i end up having double of what I have
>> right now.And ofcourse you can share this thread to the hbase mailing list.
>> Let me know if you get any idea about what to do with this hash based data.
>>
>>
>>  Jaspreet Singh
>>  ------
>> *From:* Alex Baranau 
>> *Sent:* Thursday, March 5, 2015 2:56 PM
>> *To:* Jaspreet Singh
>> *Subject:* Re: Regarding a doubt I am having for HBase
>>
>>   Hi Jaspreet,
>>
>>  Do you see this time when you fetch by first field or by second? How do
>> you construct your scan? In particular, what are start and stop keys and
>> filters used?
>>
>>  For scan by first filed the simple prefix scan should work. For scan by
>> second field, you will have to do filtering unless you can denormalize your
>> data and create separate index to scan.
>>
>>  In first case you may be able to use fast forwarding in filter while
>> scanning. E.g. use fuzzy row filter if the first field is of fixed size.
>> Depending on your case this may help speedup scanning. Otherwise, you may
>> consider implementing custom fast-forwarding filter.
>>
>>  If denormalizing is an option, you could store the record twice, with
>> second_first key format in addition. And use prefix scanning again.
>>
>>  Which one works best for you?
>>
>>  Also: can I CC the hbase mailing list to the thread - people are amazing
>> there and will be happy to provide help too :)
>>
>>  Alex
>>
>> On Tue, Mar 3, 2015 at 12:21 PM, Jaspreet Singh 
>> wrote:
>>
>>>  Hi Alex,
>>>
>>>
>>>  I was trying to look up some composite key related questions and
>>> noticed that you are pro when it comes HBase related questions :) I read
>>> many of your blogs, but I am still confused about the doubt I have. I have
>>> a composite primary key in the form of first field_second field. I want to
>>> scan my table giving the value of first field or second field and get the
>>> results. I used the rowfilter for this but the time I am getting to fetch
>>> the row is too high approx 27 seconds ( the number of rows is in millions).
>>> I want to achieve something near 2 seconds or even less than that, can you
>>> suggest me what I should do for this.
>>>
>>> Thank You
>>>
>>>
>>>  Jaspreet Singh
>>>
>>> Clarte.co
>>>
>>



Re: Regarding a doubt I am having for HBase

2015-03-10 Thread Alex Baranau
CCing HBase's user ML.

Could you give an example of the row key and example of two different
queries you are making to better understand your case?

Thank you,

Alex Baranau
--
http://cdap.io - open source framework to build and run data applications
on Hadoop & HBase


On Mon, Mar 9, 2015 at 9:00 AM, Jaspreet Singh  wrote:

>  Hi Alex,
>
>
>  Thanks a lot for the response!!! The data I have is inform of hashes and
> id, every id is related to a cookie hashed data so for the id part the
> solution that you said would work good but in case hashes it would not be
> possible to mention a stoprow based on increment. Also I have millions of
> rows so duplicating the row means i end up having double of what I have
> right now.And ofcourse you can share this thread to the hbase mailing list.
> Let me know if you get any idea about what to do with this hash based data.
>
>
>  Jaspreet Singh
>  --
> *From:* Alex Baranau 
> *Sent:* Thursday, March 5, 2015 2:56 PM
> *To:* Jaspreet Singh
> *Subject:* Re: Regarding a doubt I am having for HBase
>
>   Hi Jaspreet,
>
>  Do you see this time when you fetch by first field or by second? How do
> you construct your scan? In particular, what are start and stop keys and
> filters used?
>
>  For scan by first filed the simple prefix scan should work. For scan by
> second field, you will have to do filtering unless you can denormalize your
> data and create separate index to scan.
>
>  In first case you may be able to use fast forwarding in filter while
> scanning. E.g. use fuzzy row filter if the first field is of fixed size.
> Depending on your case this may help speedup scanning. Otherwise, you may
> consider implementing custom fast-forwarding filter.
>
>  If denormalizing is an option, you could store the record twice, with
> second_first key format in addition. And use prefix scanning again.
>
>  Which one works best for you?
>
>  Also: can I CC the hbase mailing list to the thread - people are amazing
> there and will be happy to provide help too :)
>
>  Alex
>
> On Tue, Mar 3, 2015 at 12:21 PM, Jaspreet Singh 
> wrote:
>
>>
>>  Hi Alex,
>>
>>
>>  I was trying to look up some composite key related questions and
>> noticed that you are pro when it comes HBase related questions :) I read
>> many of your blogs, but I am still confused about the doubt I have. I have
>> a composite primary key in the form of first field_second field. I want to
>> scan my table giving the value of first field or second field and get the
>> results. I used the rowfilter for this but the time I am getting to fetch
>> the row is too high approx 27 seconds ( the number of rows is in millions).
>> I want to achieve something near 2 seconds or even less than that, can you
>> suggest me what I should do for this.
>>
>> Thank You
>>
>>
>>  Jaspreet Singh
>>
>> Clarte.co
>>
>
>


Re: small hbase doubt

2013-07-13 Thread Ted Yu
Not yet, to my knowledge.

On Sat, Jul 13, 2013 at 10:57 AM, Asaf Mesika  wrote:

>  I ment to ask if this type of compression is supported over RPC calls in
> 0.95?
>
> On Saturday, July 13, 2013, Ted Yu wrote:
>
> > bq. Do you think prefix compression can also be utilized here?
> >
> > In your use case, prefix compression would help in reducing bandwidth
> > consumption.
> >
> > On Thu, Jul 11, 2013 at 9:11 PM, Asaf Mesika  >
> > wrote:
> >
> > > Do you think prefix compression can also be utilized here? In our use
> > case
> > > we sent a list of Put of counters in which the key is quite long and
> the
> > > keys are quite similar to one another. This can save bandwidth.
> > >
> > > On Friday, July 12, 2013, Ted Yu wrote:
> > >
> > > > Right.
> > > >
> > > > Take a look at http://hbase.apache.org/book.html#d2617e13654 and
> > section
> > > > J.4.3.2
> > > >
> > > > On Thu, Jul 11, 2013 at 9:01 PM, Asaf Mesika  
> > > >
> > > > wrote:
> > > >
> > > > > I thought that in 0.95 ProtoBuf provides RPC compression, no?
> > > > >
> > > > > On Friday, July 12, 2013, Alok Singh Mahor wrote:
> > > > >
> > > > > > To Jean :
> > > > > > Thanks for replying. well could you please elaborate your
> > answer..and
> > > > by
> > > > > > that 'query' ..i meant can anyone clear my doubt :-)
> > > > > >
> > > > > > To Doug:
> > > > > > Thanks for replying. but then how LZO improves efficiency of
> > network
> > > > > > bandwidth when getting data from remote server...? what's that?
> > > > > >
> > > > > >
> > > > > > On Thu, Jul 11, 2013 at 6:54 PM, Jean-Marc Spaggiari <
> > > > > > jean-m...@spaggiari.org  
> > > wrote:
> > > > > >
> > > > > > > Hi Alok,
> > > > > > >
> > > > > > > What do you mean by "query"?
> > > > > > >
> > > > > > > Gets are done based on the key. And snappy and LZO are used to
> > > > compress
> > > > > > the
> > > > > > > value. So only when a row feet your needs HBase will
> decrompress
> > > the
> > > > > > value
> > > > > > > and send it back to you...
> > > > > > >
> > > > > > > Does it reply to your question?
> > > > > > >
> > > > > > > JM
> > > > > > >
> > > > > > > 2013/7/11 Alok Singh Mahor 
> > >
> > > > > > >
> > > > > > > > Hello everyone,
> > > > > > > > could anyone tell me small query?
> > > > > > > >
> > > > > > > > Does Hbase decompress data before executing query or it
> execute
> > > > > queries
> > > > > > > on
> > > > > > > > compressed data? and how snappy and lzo actually behave ?
> > > > > > > >
> > > > > > > > thanks
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
>
>
> --
> Sent from Gmail Mobile
>


Re: small hbase doubt

2013-07-13 Thread Asaf Mesika
 I ment to ask if this type of compression is supported over RPC calls in
0.95?

On Saturday, July 13, 2013, Ted Yu wrote:

> bq. Do you think prefix compression can also be utilized here?
>
> In your use case, prefix compression would help in reducing bandwidth
> consumption.
>
> On Thu, Jul 11, 2013 at 9:11 PM, Asaf Mesika 
> >
> wrote:
>
> > Do you think prefix compression can also be utilized here? In our use
> case
> > we sent a list of Put of counters in which the key is quite long and the
> > keys are quite similar to one another. This can save bandwidth.
> >
> > On Friday, July 12, 2013, Ted Yu wrote:
> >
> > > Right.
> > >
> > > Take a look at http://hbase.apache.org/book.html#d2617e13654 and
> section
> > > J.4.3.2
> > >
> > > On Thu, Jul 11, 2013 at 9:01 PM, Asaf Mesika 
> > > 
> > >
> > > wrote:
> > >
> > > > I thought that in 0.95 ProtoBuf provides RPC compression, no?
> > > >
> > > > On Friday, July 12, 2013, Alok Singh Mahor wrote:
> > > >
> > > > > To Jean :
> > > > > Thanks for replying. well could you please elaborate your
> answer..and
> > > by
> > > > > that 'query' ..i meant can anyone clear my doubt :-)
> > > > >
> > > > > To Doug:
> > > > > Thanks for replying. but then how LZO improves efficiency of
> network
> > > > > bandwidth when getting data from remote server...? what's that?
> > > > >
> > > > >
> > > > > On Thu, Jul 11, 2013 at 6:54 PM, Jean-Marc Spaggiari <
> > > > > jean-m...@spaggiari.org  
> > wrote:
> > > > >
> > > > > > Hi Alok,
> > > > > >
> > > > > > What do you mean by "query"?
> > > > > >
> > > > > > Gets are done based on the key. And snappy and LZO are used to
> > > compress
> > > > > the
> > > > > > value. So only when a row feet your needs HBase will decrompress
> > the
> > > > > value
> > > > > > and send it back to you...
> > > > > >
> > > > > > Does it reply to your question?
> > > > > >
> > > > > > JM
> > > > > >
> > > > > > 2013/7/11 Alok Singh Mahor 
> >
> > > > > >
> > > > > > > Hello everyone,
> > > > > > > could anyone tell me small query?
> > > > > > >
> > > > > > > Does Hbase decompress data before executing query or it execute
> > > > queries
> > > > > > on
> > > > > > > compressed data? and how snappy and lzo actually behave ?
> > > > > > >
> > > > > > > thanks
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
>


-- 
Sent from Gmail Mobile


Re: small hbase doubt

2013-07-12 Thread Ted Yu
bq. Do you think prefix compression can also be utilized here?

In your use case, prefix compression would help in reducing bandwidth
consumption.

On Thu, Jul 11, 2013 at 9:11 PM, Asaf Mesika  wrote:

> Do you think prefix compression can also be utilized here? In our use case
> we sent a list of Put of counters in which the key is quite long and the
> keys are quite similar to one another. This can save bandwidth.
>
> On Friday, July 12, 2013, Ted Yu wrote:
>
> > Right.
> >
> > Take a look at http://hbase.apache.org/book.html#d2617e13654 and section
> > J.4.3.2
> >
> > On Thu, Jul 11, 2013 at 9:01 PM, Asaf Mesika  >
> > wrote:
> >
> > > I thought that in 0.95 ProtoBuf provides RPC compression, no?
> > >
> > > On Friday, July 12, 2013, Alok Singh Mahor wrote:
> > >
> > > > To Jean :
> > > > Thanks for replying. well could you please elaborate your answer..and
> > by
> > > > that 'query' ..i meant can anyone clear my doubt :-)
> > > >
> > > > To Doug:
> > > > Thanks for replying. but then how LZO improves efficiency of network
> > > > bandwidth when getting data from remote server...? what's that?
> > > >
> > > >
> > > > On Thu, Jul 11, 2013 at 6:54 PM, Jean-Marc Spaggiari <
> > > > jean-m...@spaggiari.org  > wrote:
> > > >
> > > > > Hi Alok,
> > > > >
> > > > > What do you mean by "query"?
> > > > >
> > > > > Gets are done based on the key. And snappy and LZO are used to
> > compress
> > > > the
> > > > > value. So only when a row feet your needs HBase will decrompress
> the
> > > > value
> > > > > and send it back to you...
> > > > >
> > > > > Does it reply to your question?
> > > > >
> > > > > JM
> > > > >
> > > > > 2013/7/11 Alok Singh Mahor 
> > > > > >
> > > > >
> > > > > > Hello everyone,
> > > > > > could anyone tell me small query?
> > > > > >
> > > > > > Does Hbase decompress data before executing query or it execute
> > > queries
> > > > > on
> > > > > > compressed data? and how snappy and lzo actually behave ?
> > > > > >
> > > > > > thanks
> > > > > >
> > > > >
> > > >
> > >
> >
>


Re: small hbase doubt

2013-07-11 Thread Asaf Mesika
Do you think prefix compression can also be utilized here? In our use case
we sent a list of Put of counters in which the key is quite long and the
keys are quite similar to one another. This can save bandwidth.

On Friday, July 12, 2013, Ted Yu wrote:

> Right.
>
> Take a look at http://hbase.apache.org/book.html#d2617e13654 and section
> J.4.3.2
>
> On Thu, Jul 11, 2013 at 9:01 PM, Asaf Mesika 
> >
> wrote:
>
> > I thought that in 0.95 ProtoBuf provides RPC compression, no?
> >
> > On Friday, July 12, 2013, Alok Singh Mahor wrote:
> >
> > > To Jean :
> > > Thanks for replying. well could you please elaborate your answer..and
> by
> > > that 'query' ..i meant can anyone clear my doubt :-)
> > >
> > > To Doug:
> > > Thanks for replying. but then how LZO improves efficiency of network
> > > bandwidth when getting data from remote server...? what's that?
> > >
> > >
> > > On Thu, Jul 11, 2013 at 6:54 PM, Jean-Marc Spaggiari <
> > > jean-m...@spaggiari.org  > wrote:
> > >
> > > > Hi Alok,
> > > >
> > > > What do you mean by "query"?
> > > >
> > > > Gets are done based on the key. And snappy and LZO are used to
> compress
> > > the
> > > > value. So only when a row feet your needs HBase will decrompress the
> > > value
> > > > and send it back to you...
> > > >
> > > > Does it reply to your question?
> > > >
> > > > JM
> > > >
> > > > 2013/7/11 Alok Singh Mahor  > > > >
> > > >
> > > > > Hello everyone,
> > > > > could anyone tell me small query?
> > > > >
> > > > > Does Hbase decompress data before executing query or it execute
> > queries
> > > > on
> > > > > compressed data? and how snappy and lzo actually behave ?
> > > > >
> > > > > thanks
> > > > >
> > > >
> > >
> >
>


Re: small hbase doubt

2013-07-11 Thread Ted Yu
Right.

Take a look at http://hbase.apache.org/book.html#d2617e13654 and section
J.4.3.2

On Thu, Jul 11, 2013 at 9:01 PM, Asaf Mesika  wrote:

> I thought that in 0.95 ProtoBuf provides RPC compression, no?
>
> On Friday, July 12, 2013, Alok Singh Mahor wrote:
>
> > To Jean :
> > Thanks for replying. well could you please elaborate your answer..and by
> > that 'query' ..i meant can anyone clear my doubt :-)
> >
> > To Doug:
> > Thanks for replying. but then how LZO improves efficiency of network
> > bandwidth when getting data from remote server...? what's that?
> >
> >
> > On Thu, Jul 11, 2013 at 6:54 PM, Jean-Marc Spaggiari <
> > jean-m...@spaggiari.org > wrote:
> >
> > > Hi Alok,
> > >
> > > What do you mean by "query"?
> > >
> > > Gets are done based on the key. And snappy and LZO are used to compress
> > the
> > > value. So only when a row feet your needs HBase will decrompress the
> > value
> > > and send it back to you...
> > >
> > > Does it reply to your question?
> > >
> > > JM
> > >
> > > 2013/7/11 Alok Singh Mahor >
> > >
> > > > Hello everyone,
> > > > could anyone tell me small query?
> > > >
> > > > Does Hbase decompress data before executing query or it execute
> queries
> > > on
> > > > compressed data? and how snappy and lzo actually behave ?
> > > >
> > > > thanks
> > > >
> > >
> >
>


Re: small hbase doubt

2013-07-11 Thread Asaf Mesika
I thought that in 0.95 ProtoBuf provides RPC compression, no?

On Friday, July 12, 2013, Alok Singh Mahor wrote:

> To Jean :
> Thanks for replying. well could you please elaborate your answer..and by
> that 'query' ..i meant can anyone clear my doubt :-)
>
> To Doug:
> Thanks for replying. but then how LZO improves efficiency of network
> bandwidth when getting data from remote server...? what's that?
>
>
> On Thu, Jul 11, 2013 at 6:54 PM, Jean-Marc Spaggiari <
> jean-m...@spaggiari.org > wrote:
>
> > Hi Alok,
> >
> > What do you mean by "query"?
> >
> > Gets are done based on the key. And snappy and LZO are used to compress
> the
> > value. So only when a row feet your needs HBase will decrompress the
> value
> > and send it back to you...
> >
> > Does it reply to your question?
> >
> > JM
> >
> > 2013/7/11 Alok Singh Mahor >
> >
> > > Hello everyone,
> > > could anyone tell me small query?
> > >
> > > Does Hbase decompress data before executing query or it execute queries
> > on
> > > compressed data? and how snappy and lzo actually behave ?
> > >
> > > thanks
> > >
> >
>


Re: small hbase doubt

2013-07-11 Thread Alok Singh Mahor
To Jean :
Thanks for replying. well could you please elaborate your answer..and by
that 'query' ..i meant can anyone clear my doubt :-)

To Doug:
Thanks for replying. but then how LZO improves efficiency of network
bandwidth when getting data from remote server...? what's that?


On Thu, Jul 11, 2013 at 6:54 PM, Jean-Marc Spaggiari <
jean-m...@spaggiari.org> wrote:

> Hi Alok,
>
> What do you mean by "query"?
>
> Gets are done based on the key. And snappy and LZO are used to compress the
> value. So only when a row feet your needs HBase will decrompress the value
> and send it back to you...
>
> Does it reply to your question?
>
> JM
>
> 2013/7/11 Alok Singh Mahor 
>
> > Hello everyone,
> > could anyone tell me small query?
> >
> > Does Hbase decompress data before executing query or it execute queries
> on
> > compressed data? and how snappy and lzo actually behave ?
> >
> > thanks
> >
>


Re: small hbase doubt

2013-07-11 Thread Doug Meil

Compression only applies to data on disk.  Over the wire (I.E., RS to
client) is uncompressed.






On 7/11/13 9:24 AM, "Jean-Marc Spaggiari"  wrote:

>Hi Alok,
>
>What do you mean by "query"?
>
>Gets are done based on the key. And snappy and LZO are used to compress
>the
>value. So only when a row feet your needs HBase will decrompress the value
>and send it back to you...
>
>Does it reply to your question?
>
>JM
>
>2013/7/11 Alok Singh Mahor 
>
>> Hello everyone,
>> could anyone tell me small query?
>>
>> Does Hbase decompress data before executing query or it execute queries
>>on
>> compressed data? and how snappy and lzo actually behave ?
>>
>> thanks
>>




Re: small hbase doubt

2013-07-11 Thread Jean-Marc Spaggiari
Hi Alok,

What do you mean by "query"?

Gets are done based on the key. And snappy and LZO are used to compress the
value. So only when a row feet your needs HBase will decrompress the value
and send it back to you...

Does it reply to your question?

JM

2013/7/11 Alok Singh Mahor 

> Hello everyone,
> could anyone tell me small query?
>
> Does Hbase decompress data before executing query or it execute queries on
> compressed data? and how snappy and lzo actually behave ?
>
> thanks
>


small hbase doubt

2013-07-11 Thread Alok Singh Mahor
Hello everyone,
could anyone tell me small query?

Does Hbase decompress data before executing query or it execute queries on
compressed data? and how snappy and lzo actually behave ?

thanks


Re: Doubt Regading HLogs

2013-05-17 Thread Nicolas Liochon
Yes, it's by design.
The last log file the one beeing written by HBase The safe option is to
wait for this file to be closed by HBase.  As Yong said, you can change the
roll parameter if you want it to be terminated sooner, but changing this
parameter impacts the hdfs namenode load.10 minutes is likely safer than 3
seconds.


On Fri, May 17, 2013 at 1:40 PM, yonghu  wrote:

> In this situation, you can set the
>
> > 
> >
> > hbase.regionserver.
> logroll.period
> >
> > 360
> >
> > 
>
> to a short value, let's say 3000 and then you can see your log file with
> current size after 3 seconds.
>
> To Nicolas,
>
> I guess he wants somehow to analyze the HLog.
>
> regards!
>
> Yong
>
>
>
> On Fri, May 17, 2013 at 1:27 PM, Rishabh Agrawal <
> rishabh.agra...@impetus.co.in> wrote:
>
> > Thanks Nicolas,
> >
> > When will  this file be finalized. Is it time bound? Or it will be always
> > be zero for last one (even if it contains the data)
> >
> > -Original Message-
> > From: Nicolas Liochon [mailto:nkey...@gmail.com]
> > Sent: Friday, May 17, 2013 4:39 PM
> > To: user
> > Subject: Re: Doubt Regading HLogs
> >
> > That's HDFS.
> >
> > When a file is currently written, the size is not known, as the write is
> > in progress. So the namenode reports a size of zero (more exactly, it
> does
> > not take into account the hdfs block beeing written when it calculates
> the
> > size). When you read, you go to the datanode owning the data, so you see
> > the real content as it is at the time of reading.
> >
> > btw, why do you want to read the HLog?
> >
> >
> > On Fri, May 17, 2013 at 12:53 PM, Rishabh Agrawal <
> > rishabh.agra...@impetus.co.in> wrote:
> >
> > >  Hello,
> > >
> > > I am working with Hlogs of Hbase and I have this doubt that HDFS shows
> > > size of last log file as zero. But when I open it I see data in it.
> > > When I add extra data a new file with zero size is created and
> > > previous HLog file gets its size.  This thing applies to each region
> > > server.  Following is the purged screen shot of same:
> > >
> > >
> > >
> > > I have set following parameters in hbase-site.xml for logs:
> > >
> > > 
> > >
> > > hbase.regionserver.logroll.period
> > >
> > > 360
> > >
> > > 
> > >
> > > 
> > >
> > >hbase.master.logcleaner.ttl
> > >
> > >60480
> > >
> > > 
> > >
> > > 
> > >
> > >hbase.regionserver.optionallogflushinterval
> > >
> > >3000
> > >
> > > 
> > >
> > >
> > >
> > >
> > >
> > > I plan to read log files for some validation work, Please guide me
> > > through this behavior of Hbase.
> > >
> > >
> > >
> > >
> > >
> > > Thanks and Regards
> > >
> > > Rishabh Agrawal
> > >
> > > Software Engineer
> > >
> > > Impetus Infotech (India) Pvt. Ltd.
> > >
> > > (O) +91.731.426.9300 x4526
> > >
> > > (M) +91.812.026.2722
> > >
> > > www.impetus.com
> > >
> > >
> > >
> > > --
> > >
> > >
> > >
> > >
> > >
> > >
> > > NOTE: This message may contain information that is confidential,
> > > proprietary, privileged or otherwise protected by law. The message is
> > > intended solely for the named addressee. If received in error, please
> > > destroy and notify the sender. Any use of this email is prohibited
> > > when received in error. Impetus does not represent, warrant and/or
> > > guarantee, that the integrity of this communication has been
> > > maintained nor that the communication is free of errors, virus,
> > interception or interference.
> > >
> >
> > 
> >
> >
> >
> >
> >
> >
> > NOTE: This message may contain information that is confidential,
> > proprietary, privileged or otherwise protected by law. The message is
> > intended solely for the named addressee. If received in error, please
> > destroy and notify the sender. Any use of this email is prohibited when
> > received in error. Impetus does not represent, warrant and/or guarantee,
> > that the integrity of this communication has been maintained nor that the
> > communication is free of errors, virus, interception or interference.
> >
>


Re: Doubt Regading HLogs

2013-05-17 Thread yonghu
In this situation, you can set the

> 
>
> hbase.regionserver.
logroll.period
>
> 360
>
> 

to a short value, let's say 3000 and then you can see your log file with
current size after 3 seconds.

To Nicolas,

I guess he wants somehow to analyze the HLog.

regards!

Yong



On Fri, May 17, 2013 at 1:27 PM, Rishabh Agrawal <
rishabh.agra...@impetus.co.in> wrote:

> Thanks Nicolas,
>
> When will  this file be finalized. Is it time bound? Or it will be always
> be zero for last one (even if it contains the data)
>
> -Original Message-
> From: Nicolas Liochon [mailto:nkey...@gmail.com]
> Sent: Friday, May 17, 2013 4:39 PM
> To: user
> Subject: Re: Doubt Regading HLogs
>
> That's HDFS.
>
> When a file is currently written, the size is not known, as the write is
> in progress. So the namenode reports a size of zero (more exactly, it does
> not take into account the hdfs block beeing written when it calculates the
> size). When you read, you go to the datanode owning the data, so you see
> the real content as it is at the time of reading.
>
> btw, why do you want to read the HLog?
>
>
> On Fri, May 17, 2013 at 12:53 PM, Rishabh Agrawal <
> rishabh.agra...@impetus.co.in> wrote:
>
> >  Hello,
> >
> > I am working with Hlogs of Hbase and I have this doubt that HDFS shows
> > size of last log file as zero. But when I open it I see data in it.
> > When I add extra data a new file with zero size is created and
> > previous HLog file gets its size.  This thing applies to each region
> > server.  Following is the purged screen shot of same:
> >
> >
> >
> > I have set following parameters in hbase-site.xml for logs:
> >
> > 
> >
> > hbase.regionserver.logroll.period
> >
> > 360
> >
> > 
> >
> > 
> >
> >hbase.master.logcleaner.ttl
> >
> >60480
> >
> > 
> >
> > 
> >
> >hbase.regionserver.optionallogflushinterval
> >
> >3000
> >
> > 
> >
> >
> >
> >
> >
> > I plan to read log files for some validation work, Please guide me
> > through this behavior of Hbase.
> >
> >
> >
> >
> >
> > Thanks and Regards
> >
> > Rishabh Agrawal
> >
> > Software Engineer
> >
> > Impetus Infotech (India) Pvt. Ltd.
> >
> > (O) +91.731.426.9300 x4526
> >
> > (M) +91.812.026.2722
> >
> > www.impetus.com
> >
> >
> >
> > --
> >
> >
> >
> >
> >
> >
> > NOTE: This message may contain information that is confidential,
> > proprietary, privileged or otherwise protected by law. The message is
> > intended solely for the named addressee. If received in error, please
> > destroy and notify the sender. Any use of this email is prohibited
> > when received in error. Impetus does not represent, warrant and/or
> > guarantee, that the integrity of this communication has been
> > maintained nor that the communication is free of errors, virus,
> interception or interference.
> >
>
> 
>
>
>
>
>
>
> NOTE: This message may contain information that is confidential,
> proprietary, privileged or otherwise protected by law. The message is
> intended solely for the named addressee. If received in error, please
> destroy and notify the sender. Any use of this email is prohibited when
> received in error. Impetus does not represent, warrant and/or guarantee,
> that the integrity of this communication has been maintained nor that the
> communication is free of errors, virus, interception or interference.
>


RE: Doubt Regading HLogs

2013-05-17 Thread Rishabh Agrawal
Thanks Nicolas,

When will  this file be finalized. Is it time bound? Or it will be always be 
zero for last one (even if it contains the data)

-Original Message-
From: Nicolas Liochon [mailto:nkey...@gmail.com]
Sent: Friday, May 17, 2013 4:39 PM
To: user
Subject: Re: Doubt Regading HLogs

That's HDFS.

When a file is currently written, the size is not known, as the write is in 
progress. So the namenode reports a size of zero (more exactly, it does not 
take into account the hdfs block beeing written when it calculates the size). 
When you read, you go to the datanode owning the data, so you see the real 
content as it is at the time of reading.

btw, why do you want to read the HLog?


On Fri, May 17, 2013 at 12:53 PM, Rishabh Agrawal < 
rishabh.agra...@impetus.co.in> wrote:

>  Hello,
>
> I am working with Hlogs of Hbase and I have this doubt that HDFS shows
> size of last log file as zero. But when I open it I see data in it.
> When I add extra data a new file with zero size is created and
> previous HLog file gets its size.  This thing applies to each region
> server.  Following is the purged screen shot of same:
>
>
>
> I have set following parameters in hbase-site.xml for logs:
>
> 
>
> hbase.regionserver.logroll.period
>
> 360
>
> 
>
> 
>
>hbase.master.logcleaner.ttl
>
>60480
>
> 
>
> 
>
>hbase.regionserver.optionallogflushinterval
>
>3000
>
> 
>
>
>
>
>
> I plan to read log files for some validation work, Please guide me
> through this behavior of Hbase.
>
>
>
>
>
> Thanks and Regards
>
> Rishabh Agrawal
>
> Software Engineer
>
> Impetus Infotech (India) Pvt. Ltd.
>
> (O) +91.731.426.9300 x4526
>
> (M) +91.812.026.2722
>
> www.impetus.com
>
>
>
> --
>
>
>
>
>
>
> NOTE: This message may contain information that is confidential,
> proprietary, privileged or otherwise protected by law. The message is
> intended solely for the named addressee. If received in error, please
> destroy and notify the sender. Any use of this email is prohibited
> when received in error. Impetus does not represent, warrant and/or
> guarantee, that the integrity of this communication has been
> maintained nor that the communication is free of errors, virus, interception 
> or interference.
>








NOTE: This message may contain information that is confidential, proprietary, 
privileged or otherwise protected by law. The message is intended solely for 
the named addressee. If received in error, please destroy and notify the 
sender. Any use of this email is prohibited when received in error. Impetus 
does not represent, warrant and/or guarantee, that the integrity of this 
communication has been maintained nor that the communication is free of errors, 
virus, interception or interference.


Re: Doubt Regading HLogs

2013-05-17 Thread Nicolas Liochon
That's HDFS.

When a file is currently written, the size is not known, as the write is in
progress. So the namenode reports a size of zero (more exactly, it does not
take into account the hdfs block beeing written when it calculates the
size). When you read, you go to the datanode owning the data, so you see
the real content as it is at the time of reading.

btw, why do you want to read the HLog?


On Fri, May 17, 2013 at 12:53 PM, Rishabh Agrawal <
rishabh.agra...@impetus.co.in> wrote:

>  Hello,
>
> I am working with Hlogs of Hbase and I have this doubt that HDFS shows
> size of last log file as zero. But when I open it I see data in it. When I
> add extra data a new file with zero size is created and previous HLog file
> gets its size.  This thing applies to each region server.  Following is the
> purged screen shot of same:
>
>
>
> I have set following parameters in hbase-site.xml for logs:
>
> 
>
> hbase.regionserver.logroll.period
>
> 360
>
> 
>
> 
>
>hbase.master.logcleaner.ttl
>
>60480
>
> 
>
> 
>
>hbase.regionserver.optionallogflushinterval
>
>3000
>
> 
>
>
>
>
>
> I plan to read log files for some validation work, Please guide me through
> this behavior of Hbase.
>
>
>
>
>
> Thanks and Regards
>
> Rishabh Agrawal
>
> Software Engineer
>
> Impetus Infotech (India) Pvt. Ltd.
>
> (O) +91.731.426.9300 x4526
>
> (M) +91.812.026.2722
>
> www.impetus.com
>
>
>
> --
>
>
>
>
>
>
> NOTE: This message may contain information that is confidential,
> proprietary, privileged or otherwise protected by law. The message is
> intended solely for the named addressee. If received in error, please
> destroy and notify the sender. Any use of this email is prohibited when
> received in error. Impetus does not represent, warrant and/or guarantee,
> that the integrity of this communication has been maintained nor that the
> communication is free of errors, virus, interception or interference.
>


RE: Doubt Regading HLogs

2013-05-17 Thread Rishabh Agrawal
Is it  a bug or part of design. It seems more of a design to me. Can  someone 
guide me through the purpose of this feature.

Thanks
Rishabh

From: Rishabh Agrawal
Sent: Friday, May 17, 2013 4:24 PM
To: user@hbase.apache.org
Subject: Doubt Regading HLogs

Hello,
I am working with Hlogs of Hbase and I have this doubt that HDFS shows size of 
last log file as zero. But when I open it I see data in it. When I add extra 
data a new file with zero size is created and previous HLog file gets its size. 
 This thing applies to each region server.  Following is the purged screen shot 
of same:
[cid:image001.jpg@01CE531A.E89188B0]

I have set following parameters in hbase-site.xml for logs:

hbase.regionserver.logroll.period
360


   hbase.master.logcleaner.ttl
   60480


   hbase.regionserver.optionallogflushinterval
   3000



I plan to read log files for some validation work, Please guide me through this 
behavior of Hbase.


Thanks and Regards
Rishabh Agrawal
Software Engineer
Impetus Infotech (India) Pvt. Ltd.
(O) +91.731.426.9300 x4526
(M) +91.812.026.2722
www.impetus.com









NOTE: This message may contain information that is confidential, proprietary, 
privileged or otherwise protected by law. The message is intended solely for 
the named addressee. If received in error, please destroy and notify the 
sender. Any use of this email is prohibited when received in error. Impetus 
does not represent, warrant and/or guarantee, that the integrity of this 
communication has been maintained nor that the communication is free of errors, 
virus, interception or interference.








NOTE: This message may contain information that is confidential, proprietary, 
privileged or otherwise protected by law. The message is intended solely for 
the named addressee. If received in error, please destroy and notify the 
sender. Any use of this email is prohibited when received in error. Impetus 
does not represent, warrant and/or guarantee, that the integrity of this 
communication has been maintained nor that the communication is free of errors, 
virus, interception or interference.


Re: doubt reg rowkey alter

2012-12-30 Thread Stack
On Sun, Dec 30, 2012 at 10:00 AM, Manoj Babu  wrote:

> Hi All,
>
> How to modify the rowkey of a existing row in HBase table.
> Is it possible to modify with alter table command?
>

Copy original row's content to a new row with new key.  Delete the old.

The alter table command works on tables and their attributes, not on rows.

St.Ack


Re: HBase Key Design : Doubt

2012-10-11 Thread Jean-Marc Spaggiari
No, you're right.

But if you just want to keep "500" as the value, you just have to set
the number of version to 1 for your table...

If you just want to keep 100, then you can insert with a revert
timestamp, so the last cell inserted will be hidden by the previous
one.

JM

2012/10/11, Narayanan K :
> Hi,
>
> I have 2 column families A and B in table T1.
>
> put 'T1', 'R1', 'A:qualf1',100
> put 'T1', R1', 'B:qualf2', 200
>
> As per my understanding the above is one row and one single version each
> for the 2 column families.
>
> If I do a put 'T1', 'R1', 'A:qualf1', 500, then there is another version
> for the rowkey pertaining to the combination {R1, A, qualf1}
>
> Please correct me if I am wrong.
>
> Regards,
> Narayanan
>
> On Thu, Oct 11, 2012 at 1:02 AM, Doug Meil
> wrote:
>
>>
>> Correct.
>>
>> If you do 2 Puts for row key A-B-C-D on different days, the second Put
>> logically replaces the first and the earlier Put becomes a previous
>> version.  Unless you specifically want older versions, you won't get them
>> in either Gets or Scans.
>>
>> Definitely want to read thisŠ
>>
>> http://hbase.apache.org/book.html#datamodel
>>
>> See this for more information about they internal KeyValue structure.
>>
>> http://hbase.apache.org/book.html#regions.arch
>> 9.7.5.4. KeyValue
>>
>>
>> Older versions are kept around as long as the table descriptor says so
>> (e.g., max versions).  See the StoreFile and Compactions entries in the
>> RefGuide for more information on the internals.
>>
>>
>>
>>
>> On 10/10/12 3:24 PM, "Jerry Lam"  wrote:
>>
>> >correct me if I'm wrong. The version applies to the individual cell (ie.
>> >row key, column family and column qualifier) not (row key, column
>> > family).
>> >
>> >
>> >On Wed, Oct 10, 2012 at 3:13 PM, Narayanan K 
>> >wrote:
>> >
>> >> Hi all,
>> >>
>> >> I have a usecase wherein I need to find the unique of some things in
>> >>HBase
>> >> across dates.
>> >>
>> >> Say, on 1st Oct, A-B-C-D appeared, hence I insert a row with rowkey :
>> >> A-B-C-D.
>> >> On 2nd Oct, I get the same value A-B-C-D and I don't want to
>> >> redundantly
>> >> store the row again with a new rowkey - A-B-C-D for 2nd Oct
>> >> i.e I will not want to have 20121001-A-B-C-D and 20121002-A-B-C-D as 2
>> >> rowkeys in the table.
>> >>
>> >> Eg: If I have 1st Oct , 2nd Oct as 2 column families and if number of
>> >> versions are set to 1, only 1 row will be present in for both the
>> >> dates
>> >> having rowkey A-B-C-D.
>> >> Hence if I need to find unique number of times A-B-C-D appeared during
>> >>Oct
>> >> 1 and Oct 2, I just need to take rowcount of the row A-B-C-D by
>> >>filtering
>> >> over the 2 column families.
>> >> Similarly, if we have 10  date column families, and I need to scan
>> >> only
>> >>for
>> >> 2 dates, then it scans only those store files having the specified
>> >>column
>> >> families. This will make scanning faster.
>> >>
>> >> But here the design problem is that I cant add more column families to
>> >>the
>> >> table each day.
>> >>
>> >> I would need to store data every day and I read that HBase doesnt work
>> >>well
>> >> with more than 3 column families.
>> >>
>> >> The other option is to have one single column family and store dates
>> >> as
>> >> qualifiers : date:d1, date:d2 But here if there are 30 date
>> >>qualifiers
>> >> under date column family, to scan a single date qualifier or may be
>> >>range
>> >> of 2-3 dates will have to scan through the entire data of all d1 to
>> >> d30
>> >> qualifiers in the date column family which would be slower compared to
>> >> having separate column families for the each date..
>> >>
>> >> Please share your thoughts on this. Also any alternate design
>> >>suggestions
>> >> you might have.
>> >>
>> >> Regards,
>> >> Narayanan
>> >>
>>
>>
>>
>


Re: HBase Key Design : Doubt

2012-10-11 Thread Narayanan K
Hi,

I have 2 column families A and B in table T1.

put 'T1', 'R1', 'A:qualf1',100
put 'T1', R1', 'B:qualf2', 200

As per my understanding the above is one row and one single version each
for the 2 column families.

If I do a put 'T1', 'R1', 'A:qualf1', 500, then there is another version
for the rowkey pertaining to the combination {R1, A, qualf1}

Please correct me if I am wrong.

Regards,
Narayanan

On Thu, Oct 11, 2012 at 1:02 AM, Doug Meil wrote:

>
> Correct.
>
> If you do 2 Puts for row key A-B-C-D on different days, the second Put
> logically replaces the first and the earlier Put becomes a previous
> version.  Unless you specifically want older versions, you won't get them
> in either Gets or Scans.
>
> Definitely want to read thisŠ
>
> http://hbase.apache.org/book.html#datamodel
>
> See this for more information about they internal KeyValue structure.
>
> http://hbase.apache.org/book.html#regions.arch
> 9.7.5.4. KeyValue
>
>
> Older versions are kept around as long as the table descriptor says so
> (e.g., max versions).  See the StoreFile and Compactions entries in the
> RefGuide for more information on the internals.
>
>
>
>
> On 10/10/12 3:24 PM, "Jerry Lam"  wrote:
>
> >correct me if I'm wrong. The version applies to the individual cell (ie.
> >row key, column family and column qualifier) not (row key, column family).
> >
> >
> >On Wed, Oct 10, 2012 at 3:13 PM, Narayanan K 
> >wrote:
> >
> >> Hi all,
> >>
> >> I have a usecase wherein I need to find the unique of some things in
> >>HBase
> >> across dates.
> >>
> >> Say, on 1st Oct, A-B-C-D appeared, hence I insert a row with rowkey :
> >> A-B-C-D.
> >> On 2nd Oct, I get the same value A-B-C-D and I don't want to redundantly
> >> store the row again with a new rowkey - A-B-C-D for 2nd Oct
> >> i.e I will not want to have 20121001-A-B-C-D and 20121002-A-B-C-D as 2
> >> rowkeys in the table.
> >>
> >> Eg: If I have 1st Oct , 2nd Oct as 2 column families and if number of
> >> versions are set to 1, only 1 row will be present in for both the dates
> >> having rowkey A-B-C-D.
> >> Hence if I need to find unique number of times A-B-C-D appeared during
> >>Oct
> >> 1 and Oct 2, I just need to take rowcount of the row A-B-C-D by
> >>filtering
> >> over the 2 column families.
> >> Similarly, if we have 10  date column families, and I need to scan only
> >>for
> >> 2 dates, then it scans only those store files having the specified
> >>column
> >> families. This will make scanning faster.
> >>
> >> But here the design problem is that I cant add more column families to
> >>the
> >> table each day.
> >>
> >> I would need to store data every day and I read that HBase doesnt work
> >>well
> >> with more than 3 column families.
> >>
> >> The other option is to have one single column family and store dates as
> >> qualifiers : date:d1, date:d2 But here if there are 30 date
> >>qualifiers
> >> under date column family, to scan a single date qualifier or may be
> >>range
> >> of 2-3 dates will have to scan through the entire data of all d1 to d30
> >> qualifiers in the date column family which would be slower compared to
> >> having separate column families for the each date..
> >>
> >> Please share your thoughts on this. Also any alternate design
> >>suggestions
> >> you might have.
> >>
> >> Regards,
> >> Narayanan
> >>
>
>
>


Re: HBase Key Design : Doubt

2012-10-10 Thread Doug Meil

Correct.

If you do 2 Puts for row key A-B-C-D on different days, the second Put
logically replaces the first and the earlier Put becomes a previous
version.  Unless you specifically want older versions, you won't get them
in either Gets or Scans.

Definitely want to read thisŠ

http://hbase.apache.org/book.html#datamodel

See this for more information about they internal KeyValue structure.

http://hbase.apache.org/book.html#regions.arch
9.7.5.4. KeyValue


Older versions are kept around as long as the table descriptor says so
(e.g., max versions).  See the StoreFile and Compactions entries in the
RefGuide for more information on the internals.




On 10/10/12 3:24 PM, "Jerry Lam"  wrote:

>correct me if I'm wrong. The version applies to the individual cell (ie.
>row key, column family and column qualifier) not (row key, column family).
>
>
>On Wed, Oct 10, 2012 at 3:13 PM, Narayanan K 
>wrote:
>
>> Hi all,
>>
>> I have a usecase wherein I need to find the unique of some things in
>>HBase
>> across dates.
>>
>> Say, on 1st Oct, A-B-C-D appeared, hence I insert a row with rowkey :
>> A-B-C-D.
>> On 2nd Oct, I get the same value A-B-C-D and I don't want to redundantly
>> store the row again with a new rowkey - A-B-C-D for 2nd Oct
>> i.e I will not want to have 20121001-A-B-C-D and 20121002-A-B-C-D as 2
>> rowkeys in the table.
>>
>> Eg: If I have 1st Oct , 2nd Oct as 2 column families and if number of
>> versions are set to 1, only 1 row will be present in for both the dates
>> having rowkey A-B-C-D.
>> Hence if I need to find unique number of times A-B-C-D appeared during
>>Oct
>> 1 and Oct 2, I just need to take rowcount of the row A-B-C-D by
>>filtering
>> over the 2 column families.
>> Similarly, if we have 10  date column families, and I need to scan only
>>for
>> 2 dates, then it scans only those store files having the specified
>>column
>> families. This will make scanning faster.
>>
>> But here the design problem is that I cant add more column families to
>>the
>> table each day.
>>
>> I would need to store data every day and I read that HBase doesnt work
>>well
>> with more than 3 column families.
>>
>> The other option is to have one single column family and store dates as
>> qualifiers : date:d1, date:d2 But here if there are 30 date
>>qualifiers
>> under date column family, to scan a single date qualifier or may be
>>range
>> of 2-3 dates will have to scan through the entire data of all d1 to d30
>> qualifiers in the date column family which would be slower compared to
>> having separate column families for the each date..
>>
>> Please share your thoughts on this. Also any alternate design
>>suggestions
>> you might have.
>>
>> Regards,
>> Narayanan
>>




Re: HBase Key Design : Doubt

2012-10-10 Thread Jerry Lam
correct me if I'm wrong. The version applies to the individual cell (ie.
row key, column family and column qualifier) not (row key, column family).


On Wed, Oct 10, 2012 at 3:13 PM, Narayanan K  wrote:

> Hi all,
>
> I have a usecase wherein I need to find the unique of some things in HBase
> across dates.
>
> Say, on 1st Oct, A-B-C-D appeared, hence I insert a row with rowkey :
> A-B-C-D.
> On 2nd Oct, I get the same value A-B-C-D and I don't want to redundantly
> store the row again with a new rowkey - A-B-C-D for 2nd Oct
> i.e I will not want to have 20121001-A-B-C-D and 20121002-A-B-C-D as 2
> rowkeys in the table.
>
> Eg: If I have 1st Oct , 2nd Oct as 2 column families and if number of
> versions are set to 1, only 1 row will be present in for both the dates
> having rowkey A-B-C-D.
> Hence if I need to find unique number of times A-B-C-D appeared during Oct
> 1 and Oct 2, I just need to take rowcount of the row A-B-C-D by filtering
> over the 2 column families.
> Similarly, if we have 10  date column families, and I need to scan only for
> 2 dates, then it scans only those store files having the specified column
> families. This will make scanning faster.
>
> But here the design problem is that I cant add more column families to the
> table each day.
>
> I would need to store data every day and I read that HBase doesnt work well
> with more than 3 column families.
>
> The other option is to have one single column family and store dates as
> qualifiers : date:d1, date:d2 But here if there are 30 date qualifiers
> under date column family, to scan a single date qualifier or may be range
> of 2-3 dates will have to scan through the entire data of all d1 to d30
> qualifiers in the date column family which would be slower compared to
> having separate column families for the each date..
>
> Please share your thoughts on this. Also any alternate design suggestions
> you might have.
>
> Regards,
> Narayanan
>


HBase Key Design : Doubt

2012-10-10 Thread Narayanan K
Hi all,

I have a usecase wherein I need to find the unique of some things in HBase
across dates.

Say, on 1st Oct, A-B-C-D appeared, hence I insert a row with rowkey :
A-B-C-D.
On 2nd Oct, I get the same value A-B-C-D and I don't want to redundantly
store the row again with a new rowkey - A-B-C-D for 2nd Oct
i.e I will not want to have 20121001-A-B-C-D and 20121002-A-B-C-D as 2
rowkeys in the table.

Eg: If I have 1st Oct , 2nd Oct as 2 column families and if number of
versions are set to 1, only 1 row will be present in for both the dates
having rowkey A-B-C-D.
Hence if I need to find unique number of times A-B-C-D appeared during Oct
1 and Oct 2, I just need to take rowcount of the row A-B-C-D by filtering
over the 2 column families.
Similarly, if we have 10  date column families, and I need to scan only for
2 dates, then it scans only those store files having the specified column
families. This will make scanning faster.

But here the design problem is that I cant add more column families to the
table each day.

I would need to store data every day and I read that HBase doesnt work well
with more than 3 column families.

The other option is to have one single column family and store dates as
qualifiers : date:d1, date:d2 But here if there are 30 date qualifiers
under date column family, to scan a single date qualifier or may be range
of 2-3 dates will have to scan through the entire data of all d1 to d30
qualifiers in the date column family which would be slower compared to
having separate column families for the each date..

Please share your thoughts on this. Also any alternate design suggestions
you might have.

Regards,
Narayanan


Re: Doubt in performance tuning

2012-09-10 Thread Michael Segel
Well, 

Lets actually skip a few rounds of questions... and start from the beginning. 

What does your physical cluster look like? 

On Sep 10, 2012, at 12:40 PM, Ramasubramanian 
 wrote:

> Hi,
> Will be helpful if u say specific things to  look into. Pls help
> 
> Regards,
> Rams
> 
> On 10-Sep-2012, at 10:40 PM, Stack  wrote:
> 
>> On Mon, Sep 10, 2012 at 9:58 AM, Ramasubramanian
>>  wrote:
>>> Hi,
>>> 
>>> Currently it takes 11 odd minutes to load 1.2 million record into hbase 
>>> from hdfs. Can u pls share some tips to do the same in few seconds?
>>> 
>>> We tried doing this in both pig script and in pentaho. Both are taking 11 
>>> odd minutes.
>>> 
>> 
>> You have had a look at http://hbase.apache.org/book.html#performance?
>> St.Ack
> 



Re: Doubt in performance tuning

2012-09-10 Thread Ramasubramanian
Hi,
Will be helpful if u say specific things to  look into. Pls help

Regards,
Rams

On 10-Sep-2012, at 10:40 PM, Stack  wrote:

> On Mon, Sep 10, 2012 at 9:58 AM, Ramasubramanian
>  wrote:
>> Hi,
>> 
>> Currently it takes 11 odd minutes to load 1.2 million record into hbase from 
>> hdfs. Can u pls share some tips to do the same in few seconds?
>> 
>> We tried doing this in both pig script and in pentaho. Both are taking 11 
>> odd minutes.
>> 
> 
> You have had a look at http://hbase.apache.org/book.html#performance?
> St.Ack


Re: Doubt in performance tuning

2012-09-10 Thread Stack
On Mon, Sep 10, 2012 at 9:58 AM, Ramasubramanian
 wrote:
> Hi,
>
> Currently it takes 11 odd minutes to load 1.2 million record into hbase from 
> hdfs. Can u pls share some tips to do the same in few seconds?
>
> We tried doing this in both pig script and in pentaho. Both are taking 11 odd 
> minutes.
>

You have had a look at http://hbase.apache.org/book.html#performance?
St.Ack


Doubt in performance tuning

2012-09-10 Thread Ramasubramanian
Hi,

Currently it takes 11 odd minutes to load 1.2 million record into hbase from 
hdfs. Can u pls share some tips to do the same in few seconds?

We tried doing this in both pig script and in pentaho. Both are taking 11 odd 
minutes. 

Regards,
Rams

Re: Doubt at hbase configuration

2012-03-30 Thread Jean-Daniel Cryans
The timeout is on the client side, not the server side.

J-D

On Fri, Mar 30, 2012 at 12:11 AM, Balaji k  wrote:
> Hi,
>          I was running coprocessor AggregationProtocol, it got socket
> timeout exception. Somebody said to set  "hbase.rpc.timeout" to larger
> value.
> I added the following  in hbase.site.xml
> 
>  hbase.rpc.timeout
>  300
> 
>
> but it was not working. The samething i tried dynamically in my code
>
>        Configuration conf = HBaseConfiguration.create();
>        conf.setLong("hbase.rpc.timeout", 600);
>        AggregationClient agg = new AggregationClient(conf);
>        Scan scan = new Scan();
>        scan.addColumn(Bytes.toBytes("f1l"), Bytes.toBytes("c1"));
>        scan.setCaching(500);
>        scan.setCacheBlocks(false);
>        ColumnInterpreter ci = null;
>        long a = agg.rowCount(Bytes.toBytes(""), ci, scan);
>        System.out.println(a);
>
> now it works fine. What is the problem writing in hbase.site.xml ?
> where i went wrong?
>
>
> --
> Regards,
> Balaji,K


Doubt at hbase configuration

2012-03-30 Thread Balaji k
Hi,
  I was running coprocessor AggregationProtocol, it got socket
timeout exception. Somebody said to set  "hbase.rpc.timeout" to larger
value.
I added the following  in hbase.site.xml

  hbase.rpc.timeout
  300


but it was not working. The samething i tried dynamically in my code

Configuration conf = HBaseConfiguration.create();
conf.setLong("hbase.rpc.timeout", 600);
AggregationClient agg = new AggregationClient(conf);
Scan scan = new Scan();
scan.addColumn(Bytes.toBytes("f1l"), Bytes.toBytes("c1"));
scan.setCaching(500);
scan.setCacheBlocks(false);
ColumnInterpreter ci = null;
long a = agg.rowCount(Bytes.toBytes(""), ci, scan);
System.out.println(a);

now it works fine. What is the problem writing in hbase.site.xml ?
where i went wrong?


-- 
Regards,
Balaji,K


Re: schema doubt

2011-09-15 Thread Rita
Each file is about 12k to 6k.

Inserting wont be an issue just the access. I would like to access them
quickly.

Not sure what the proper key should be. The file name is ok, but just
wondering if there is anything more I can be doing to leverage hbase.


On Thu, Sep 15, 2011 at 9:24 AM, Akash Ashok  wrote:

> Also could you tell how small these files are ? If they are way less than
> 64MB default HDFS block size you'd want to splice them before running a
> MapReduce.
>
> Cheers,
> Akash A
>
> On Thu, Sep 15, 2011 at 6:02 PM, Joey Echeverria 
> wrote:
>
> > It sounds lik you're planning to use the HBase shell to insert all of
> > this data. If that's correct, I'd recommend against it. I would write
> > a simple MapReduce program to insert the data instead. You could run a
> > map-only job that reads in the files and writes each one as a row in
> > HBase. WIth the java APIs you can write the raw bytes pretty easily.
> >
> > -Joey
> >
> > On Thu, Sep 15, 2011 at 7:56 AM, Rita  wrote:
> > > I have many small files (close to 1 million) and I was thinking of
> > creating
> > > a key value pair for them. The file name can be the key and the content
> > can
> > > be value.
> > >
> > > Would it be better if I do a base64 on the content and load it to hbase
> > or
> > > try to echo the content for hbase shell?
> > >
> > > Has anyone done something similar to this?
> > >
> > >
> > >
> > > --
> > > --- Get your facts first, then you can distort them as you please.--
> > >
> >
> >
> >
> > --
> > Joseph Echeverria
> > Cloudera, Inc.
> > 443.305.9434
> >
>



-- 
--- Get your facts first, then you can distort them as you please.--


Re: schema doubt

2011-09-15 Thread Akash Ashok
Also could you tell how small these files are ? If they are way less than
64MB default HDFS block size you'd want to splice them before running a
MapReduce.

Cheers,
Akash A

On Thu, Sep 15, 2011 at 6:02 PM, Joey Echeverria  wrote:

> It sounds lik you're planning to use the HBase shell to insert all of
> this data. If that's correct, I'd recommend against it. I would write
> a simple MapReduce program to insert the data instead. You could run a
> map-only job that reads in the files and writes each one as a row in
> HBase. WIth the java APIs you can write the raw bytes pretty easily.
>
> -Joey
>
> On Thu, Sep 15, 2011 at 7:56 AM, Rita  wrote:
> > I have many small files (close to 1 million) and I was thinking of
> creating
> > a key value pair for them. The file name can be the key and the content
> can
> > be value.
> >
> > Would it be better if I do a base64 on the content and load it to hbase
> or
> > try to echo the content for hbase shell?
> >
> > Has anyone done something similar to this?
> >
> >
> >
> > --
> > --- Get your facts first, then you can distort them as you please.--
> >
>
>
>
> --
> Joseph Echeverria
> Cloudera, Inc.
> 443.305.9434
>


Re: schema doubt

2011-09-15 Thread Joey Echeverria
It sounds lik you're planning to use the HBase shell to insert all of
this data. If that's correct, I'd recommend against it. I would write
a simple MapReduce program to insert the data instead. You could run a
map-only job that reads in the files and writes each one as a row in
HBase. WIth the java APIs you can write the raw bytes pretty easily.

-Joey

On Thu, Sep 15, 2011 at 7:56 AM, Rita  wrote:
> I have many small files (close to 1 million) and I was thinking of creating
> a key value pair for them. The file name can be the key and the content can
> be value.
>
> Would it be better if I do a base64 on the content and load it to hbase or
> try to echo the content for hbase shell?
>
> Has anyone done something similar to this?
>
>
>
> --
> --- Get your facts first, then you can distort them as you please.--
>



-- 
Joseph Echeverria
Cloudera, Inc.
443.305.9434


RE: Doubt in backup master ?

2011-08-09 Thread Ramkrishna S Vasudevan

hi Shanmuganathan,

I am currently using ./bin/hbase-daemon.sh start master to start both active
and backup master.

so once the active goes down automatically the backup becomes master and the
RS will connect to it.(Hope this part is working fine).
I have not used in the way that you have mentioned ./bin/hbase-daemon.sh
start master --backup.

So i would suggest that you can now use the same command
./bin/hbase-daemon.sh start master for both.

By the time if there is any problem in ./bin/hbase-daemon.sh start master
--backup we could raise a JIRA for that and fix it.

Regards
Ram




-Original Message-
From: shanmuganathan.r [mailto:shanmuganatha...@zohocorp.com] 
Sent: Tuesday, August 09, 2011 12:47 PM
To: user
Subject: Doubt in backup master ?



Hi Ram,


  Now I am using the HBase 0.90.3 version.


When I am using the following command for the backup master means that
working properly
./bin/hbase-daemon.sh start master



The log printed in my backup master is 


2011-08-09 12:31:55,156 INFO org.apache.hadoop.hbase.metrics: MetricsString
added: version
2011-08-09 12:31:55,156 INFO org.apache.hadoop.hbase.metrics: new MBeanInfo
2011-08-09 12:31:55,157 INFO org.apache.hadoop.hbase.metrics: new MBeanInfo
2011-08-09 12:31:55,157 INFO
org.apache.hadoop.hbase.master.metrics.MasterMetrics: Initialized
2011-08-09 12:31:55,186 INFO
org.apache.hadoop.hbase.master.ActiveMasterManager: Another master is the
active master, rohinis.zohocorpin.com:6; waiting to become the next
active master



---


If I am using 
./bin/hbase-daemon.sh start master --backup


  The Region servers are not assigned to the backup master. 




The log printed in my backup master is


2011-08-09 12:19:19,041 INFO org.apache.hadoop.hbase.metrics: new MBeanInfo
2011-08-09 12:19:19,041 INFO
org.apache.hadoop.hbase.master.metrics.MasterMetrics: Initialized
2011-08-09 12:19:19,041 DEBUG org.apache.hadoop.hbase.master.HMaster:
HMaster started in backup mode.  Stalling until master znode is written.
2011-08-09 12:19:19,041 DEBUG org.apache.hadoop.hbase.master.HMaster:
Waiting for master address ZNode to be written (Also watching cluster state
node)



---


The log printed in my RS is


2011-08-09 12:23:03,918 DEBUG
org.apache.hadoop.hbase.io.hfile.LruBlockCache: LRU Stats: total=955.33 KB,
free=197.79 MB, max=198.73 MB, blocks=1, accesses=76, hits=75,
hitRatio=98.68%%, cachingAccesses=76, cachingHits=75,
cachingHitsRatio=98.68%%, evictions=0, evicted=0, evictedPerRun=NaN
2011-08-09 12:23:08,452 WARN
org.apache.hadoop.hbase.regionserver.HRegionServer: Unable to connect to
master. Retrying. Error was:
java.net.ConnectException: Connection refused
at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
at
sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:592)
at
org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:2
06)
at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:404)
at
org.apache.hadoop.hbase.ipc.HBaseClient$Connection.setupIOstreams(HBaseClien
t.java:328)
at
org.apache.hadoop.hbase.ipc.HBaseClient.getConnection(HBaseClient.java:883)
at
org.apache.hadoop.hbase.ipc.HBaseClient.call(HBaseClient.java:750)
at
org.apache.hadoop.hbase.ipc.HBaseRPC$Invoker.invoke(HBaseRPC.java:257)
at $Proxy5.getProtocolVersion(Unknown Source)
at org.apache.hadoop.hbase.ipc.HBaseRPC.getProxy(HBaseRPC.java:419)
at org.apache.hadoop.hbase.ipc.HBaseRPC.getProxy(HBaseRPC.java:393)
at org.apache.hadoop.hbase.ipc.HBaseRPC.getProxy(HBaseRPC.java:444)
at
org.apache.hadoop.hbase.ipc.HBaseRPC.waitForProxy(HBaseRPC.java:349)
at
org.apache.hadoop.hbase.regionserver.HRegionServer.getMaster(HRegionServer.j
ava:1442)
at
org.apache.hadoop.hbase.regionserver.HRegionServer.tryRegionServerReport(HRe
gionServer.java:741)
at
org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:59
4)
at java.lang.Thread.run(Thread.java:636)
2011-08-09 12:23:11,453 DEBUG
org.apache.hadoop.hbase.regionserver.HRegionServer: No master found, will
retry
2011-08-09 12:23:14,453 DEBUG
org.apache.hadoop.hbase.regionserver.HRegionServer: No master found, will
retry
2011-08-09 12:23:17,453 DEBUG
org.apache.hadoop.hbase.regionserver.HRegionServer: No master found, will
retry
2011-08-09 12:23:20,454 DEBUG
org.apache.hadoop.hbase.regionserver.HRegionServer: No master found, will
retry
2011-08-09 12:23:23,454 DEBUG
org.apache.hadoop.hbase.regionserver.HRegionServer: No master found, will
retry
2011-08-09 12:23:26,454 DEBUG
org.apache.hadoop.hbase.regionserver.HRegionServer: No master found, will
retry


.
.
.
---


My doubt is


Are the above two commands are same for this backup master configuration ?


Can I use the First

Doubt in backup master ?

2011-08-09 Thread shanmuganathan.r


Hi Ram,


  Now I am using the HBase 0.90.3 version.


When I am using the following command for the backup master means that working 
properly
./bin/hbase-daemon.sh start master



The log printed in my backup master is 


2011-08-09 12:31:55,156 INFO org.apache.hadoop.hbase.metrics: MetricsString 
added: version
2011-08-09 12:31:55,156 INFO org.apache.hadoop.hbase.metrics: new MBeanInfo
2011-08-09 12:31:55,157 INFO org.apache.hadoop.hbase.metrics: new MBeanInfo
2011-08-09 12:31:55,157 INFO 
org.apache.hadoop.hbase.master.metrics.MasterMetrics: Initialized
2011-08-09 12:31:55,186 INFO 
org.apache.hadoop.hbase.master.ActiveMasterManager: Another master is the 
active master, rohinis.zohocorpin.com:6; waiting to become the next active 
master



---


If I am using 
./bin/hbase-daemon.sh start master --backup


  The Region servers are not assigned to the backup master. 




The log printed in my backup master is


2011-08-09 12:19:19,041 INFO org.apache.hadoop.hbase.metrics: new MBeanInfo
2011-08-09 12:19:19,041 INFO 
org.apache.hadoop.hbase.master.metrics.MasterMetrics: Initialized
2011-08-09 12:19:19,041 DEBUG org.apache.hadoop.hbase.master.HMaster: HMaster 
started in backup mode.  Stalling until master znode is written.
2011-08-09 12:19:19,041 DEBUG org.apache.hadoop.hbase.master.HMaster: Waiting 
for master address ZNode to be written (Also watching cluster state node)



---


The log printed in my RS is


2011-08-09 12:23:03,918 DEBUG org.apache.hadoop.hbase.io.hfile.LruBlockCache: 
LRU Stats: total=955.33 KB, free=197.79 MB, max=198.73 MB, blocks=1, 
accesses=76, hits=75, hitRatio=98.68%%, cachingAccesses=76, cachingHits=75, 
cachingHitsRatio=98.68%%, evictions=0, evicted=0, evictedPerRun=NaN
2011-08-09 12:23:08,452 WARN 
org.apache.hadoop.hbase.regionserver.HRegionServer: Unable to connect to 
master. Retrying. Error was:
java.net.ConnectException: Connection refused
at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
at 
sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:592)
at 
org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:404)
at 
org.apache.hadoop.hbase.ipc.HBaseClient$Connection.setupIOstreams(HBaseClient.java:328)
at 
org.apache.hadoop.hbase.ipc.HBaseClient.getConnection(HBaseClient.java:883)
at org.apache.hadoop.hbase.ipc.HBaseClient.call(HBaseClient.java:750)
at 
org.apache.hadoop.hbase.ipc.HBaseRPC$Invoker.invoke(HBaseRPC.java:257)
at $Proxy5.getProtocolVersion(Unknown Source)
at org.apache.hadoop.hbase.ipc.HBaseRPC.getProxy(HBaseRPC.java:419)
at org.apache.hadoop.hbase.ipc.HBaseRPC.getProxy(HBaseRPC.java:393)
at org.apache.hadoop.hbase.ipc.HBaseRPC.getProxy(HBaseRPC.java:444)
at org.apache.hadoop.hbase.ipc.HBaseRPC.waitForProxy(HBaseRPC.java:349)
at 
org.apache.hadoop.hbase.regionserver.HRegionServer.getMaster(HRegionServer.java:1442)
at 
org.apache.hadoop.hbase.regionserver.HRegionServer.tryRegionServerReport(HRegionServer.java:741)
at 
org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:594)
at java.lang.Thread.run(Thread.java:636)
2011-08-09 12:23:11,453 DEBUG 
org.apache.hadoop.hbase.regionserver.HRegionServer: No master found, will retry
2011-08-09 12:23:14,453 DEBUG 
org.apache.hadoop.hbase.regionserver.HRegionServer: No master found, will retry
2011-08-09 12:23:17,453 DEBUG 
org.apache.hadoop.hbase.regionserver.HRegionServer: No master found, will retry
2011-08-09 12:23:20,454 DEBUG 
org.apache.hadoop.hbase.regionserver.HRegionServer: No master found, will retry
2011-08-09 12:23:23,454 DEBUG 
org.apache.hadoop.hbase.regionserver.HRegionServer: No master found, will retry
2011-08-09 12:23:26,454 DEBUG 
org.apache.hadoop.hbase.regionserver.HRegionServer: No master found, will retry


.
.
.
---


My doubt is


Are the above two commands are same for this backup master configuration ?


Can I use the First command for the Multi master configuration ?


If the second one is the only command for Multi master configuration, than how 
can i solve this problem ?



Regards,

Shanmuganathan








RE: Doubt in hbase installation?

2011-07-27 Thread Buttler, David
You probably don't need the data directory in hbase-site.xml, but the rest look 
important.  If you have an independent zookeeper installation, hbase still 
needs to know how to connect to it.

Dave


-Original Message-
From: shanmuganathan.r [mailto:shanmuganatha...@zohocorp.com] 
Sent: Wednesday, July 27, 2011 2:08 AM
To: user@hbase.apache.org
Subject: Doubt in hbase installation?


Hi All,

I have some doubt in the installation of hbase in fully distributed 
mode. 
In my configuration the hbase does not manage the zookeeper. I added zoo.cfg in 
zookeeper configuration file . Also specified some configuration in the 
hbase-site.xml file.
Are the two configurations required for my zookeeper or any one is enough?


My zoo.cfg file 


tickTime=2000
dataDir=/home/hadoop/zookeeperdata
clientPort=2181
initLimit=5
syncLimit=2
server.1=192.168.111.101:2888:3888
server.2=192.168.111.102:2888:3888
server.3=192.168.111.103:2888:3888


hbase-site.xml


  <property>
<name>hbase.rootdir</name>
<value>hdfs://shanmuganathanr:54310/hbase</value>
  </property>


  <property>
<name>hbase.cluster.distributed</name>
<value>true</value>
  </property>


  <property>
<name>hbase.zookeeper.property.clientPort</name>
<value>2181</value>
<description>Property from ZooKeeper’s config zoo.cfg.The port at 
which the clients will connect.
</description>
  </property>


  <property>
<name>hbase.zookeeper.property.dataDir</name>
<value>/home/hadoop/zookeeperdata</value>
<description>Property from ZooKeeper’s config zoo.cfg. The 
directory where the snapshot is stored.
</description>
  </property>


  <property>
<name>hbase.master.port</name>
<value>6</value>
  </property>


  <property>
<name>hbase.zookeeper.quorum</name>

<value>192.168.111.101,192.168.111.102,192.168.111.103</value>
  </property>




Regards


Shanmuganathan


Doubt in hbase installation?

2011-07-27 Thread shanmuganathan.r

Hi All,

I have some doubt in the installation of hbase in fully distributed 
mode. 
In my configuration the hbase does not manage the zookeeper. I added zoo.cfg in 
zookeeper configuration file . Also specified some configuration in the 
hbase-site.xml file.
Are the two configurations required for my zookeeper or any one is enough?


My zoo.cfg file 


tickTime=2000
dataDir=/home/hadoop/zookeeperdata
clientPort=2181
initLimit=5
syncLimit=2
server.1=192.168.111.101:2888:3888
server.2=192.168.111.102:2888:3888
server.3=192.168.111.103:2888:3888


hbase-site.xml


  <property>
<name>hbase.rootdir</name>
<value>hdfs://shanmuganathanr:54310/hbase</value>
  </property>


  <property>
<name>hbase.cluster.distributed</name>
<value>true</value>
  </property>


  <property>
<name>hbase.zookeeper.property.clientPort</name>
<value>2181</value>
<description>Property from ZooKeeper’s config zoo.cfg.The port at 
which the clients will connect.
</description>
  </property>


  <property>
<name>hbase.zookeeper.property.dataDir</name>
<value>/home/hadoop/zookeeperdata</value>
<description>Property from ZooKeeper’s config zoo.cfg. The 
directory where the snapshot is stored.
</description>
  </property>


  <property>
<name>hbase.master.port</name>
<value>6</value>
  </property>


  <property>
<name>hbase.zookeeper.quorum</name>

<value>192.168.111.101,192.168.111.102,192.168.111.103</value>
  </property>




Regards


Shanmuganathan