You can easily store that much data as long as you don't have small files,
which is typically why people turn to federation.

-Dima

On Friday, August 19, 2016, Alexandr Porunov <alexandr.poru...@gmail.com>
wrote:

> We are talking about facebook. So, there are 25 TB per month. 15 billion
> messages with 1024 bytes and 120 billion messages with 100 bytes per month.
>
> I thought that they used only hbase to handle such a huge data If they used
> their own implementation of hbase then I haven't questions.
>
> Sincerely,
> Alexandr
>
> On Sat, Aug 20, 2016 at 1:39 AM, Dima Spivak <dspi...@cloudera.com
> <javascript:;>> wrote:
>
> > I'd +1 what Vladimir says. How much data (in TBs/PBs) and how many files
> > are we talking about here? I'd say that use cases that benefit from HBase
> > don't tend to hit the kind of HDFS file limits that federation seeks to
> > address.
> >
> > -Dima
> >
> > On Fri, Aug 19, 2016 at 2:19 PM, Vladimir Rodionov <
> vladrodio...@gmail.com <javascript:;>
> > >
> > wrote:
> >
> > > FB has its own "federation". It is a proprietary code, I presume.
> > >
> > > -Vladimir
> > >
> > >
> > > On Fri, Aug 19, 2016 at 1:22 PM, Alexandr Porunov <
> > > alexandr.poru...@gmail.com <javascript:;>> wrote:
> > >
> > > > No. There isn't. But I want to figure out how to configure that type
> of
> > > > cluster in the case if there is particular reason. How facebook can
> > > handle
> > > > such a huge amount of ops without federation? I don't think that they
> > > just
> > > > have one namenode server and one standby namenode server. It isn't
> > > > possible. I am sure that they use federation.
> > > >
> > > > On Fri, Aug 19, 2016 at 10:08 PM, Vladimir Rodionov <
> > > > vladrodio...@gmail.com <javascript:;>>
> > > > wrote:
> > > >
> > > > > >> I am not sure how to do it but I have to configure federated
> > cluster
> > > > > with
> > > > > >> hbase to store huge amount of messages (client to client) (40%
> > > writes,
> > > > > 60%
> > > > > >> reads).
> > > > >
> > > > > Any particular reason for federated cluster? How huge is huge
> amount
> > > and
> > > > > what is the message size?
> > > > >
> > > > > -Vladimir
> > > > >
> > > > > On Fri, Aug 19, 2016 at 11:57 AM, Dima Spivak <
> dspi...@cloudera.com <javascript:;>>
> > > > > wrote:
> > > > >
> > > > > > As far as I know, HBase doesn't support spreading tables across
> > > > > namespaces;
> > > > > > you'd have to point it at one namenode at a time. I've heard of
> > > people
> > > > > > trying to run multiple HBase instances in order to get access to
> > all
> > > > > their
> > > > > > HDFS data, but it doesn't tend to be much fun.
> > > > > >
> > > > > > -Dima
> > > > > >
> > > > > > On Fri, Aug 19, 2016 at 11:51 AM, Alexandr Porunov <
> > > > > > alexandr.poru...@gmail.com <javascript:;>> wrote:
> > > > > >
> > > > > > > Hello,
> > > > > > >
> > > > > > > I am not sure how to do it but I have to configure federated
> > > cluster
> > > > > with
> > > > > > > hbase to store huge amount of messages (client to client) (40%
> > > > writes,
> > > > > > 60%
> > > > > > > reads). Does somebody have any idea or examples how to
> configure
> > > it?
> > > > > > >
> > > > > > > Of course we can configure hdfs in a federated mode but as for
> me
> > > it
> > > > > > isn't
> > > > > > > suitable for hbase. If we want to save message from client 1 to
> > > > client
> > > > > 2
> > > > > > in
> > > > > > > the hbase cluster then how hbase know in which namespace it
> have
> > to
> > > > > save
> > > > > > > it? Which namenode will be responsible for that message? How we
> > can
> > > > > read
> > > > > > > client messages?
> > > > > > >
> > > > > > > Give me any ideas, please
> > > > > > >
> > > > > > > Sincerely,
> > > > > > > Alexandr
> > > > > > >
> > > > > >
> > > > > >
> > > > > >
> > > > > > --
> > > > > > -Dima
> > > > > >
> > > > >
> > > >
> > >
> >
> >
> >
> > --
> > -Dima
> >
>


-- 
-Dima

Reply via email to