Hi folks - is there a recommended way of estimating HBase HDFS usage for a
new environment?

We have a DEV HBase cluster in place, and from this, I'm trying to estimate
the specs of our not-yet-built PROD environment.  One of the variables we're
considering is HBase usage of HDFS.  What I've just tried is to calculate an
average bytes/record ratio by using "hadoop dfs -du /hbase", and dividing by
the number of records/table.  But this ignores any kind of fixed overhead,
so I have concerns about it.

Is there a better way?

Norbert

Reply via email to