[
https://issues.apache.org/jira/browse/HBASE-11165?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14120516#comment-14120516
]
Mikhail Antonov commented on HBASE-11165:
-----------------------------------------
[~toffer] thanks! I'd be really curious to look at those numbers.
Is the NN you mentioned with 250M files is solely dedicated to HBase
installation? I mean, could the assumption be made that the HBase cluster with
1M or large regions consumes about 250M of files in HDFS, so roughly 250 files
/ per region, or would it be too bold assumption?
[~eclark] so if we take as a baseline that (num of files) >> (num regions), I
wonder how close to NN limits we are? I mean, if we're talking about case with
10M regions (or even 50M), with the same ratio of region-to-files, 10M regions
would give us 2.5B files in HDFS? How close is that to HDFS limits?
> Scaling so cluster can host 1M regions and beyond (50M regions?)
> ----------------------------------------------------------------
>
> Key: HBASE-11165
> URL: https://issues.apache.org/jira/browse/HBASE-11165
> Project: HBase
> Issue Type: Brainstorming
> Reporter: stack
> Attachments: HBASE-11165.zip, Region Scalability test.pdf,
> zk_less_assignment_comparison_2.pdf
>
>
> This discussion issue comes out of "Co-locate Meta And Master HBASE-10569"
> and comments on the doc posted there.
> A user -- our Francis Liu -- needs to be able to scale a cluster to do 1M
> regions maybe even 50M later. This issue is about discussing how we will do
> that (or if not 50M on a cluster, how otherwise we can attain same end).
> More detail to follow.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)