Hello,
Will enabling compression on Hdfs effect the performance of cluster? In
processing jobs?

Regards,
Ouch Whisper
01010101010
On Jan 11, 2013 12:25 PM, "Dhaval Shah" <[email protected]> wrote:

>
> Also depending on compression type chosen it might take less disk space
>
>
> ------------------------------
> On Fri 11 Jan, 2013 3:53 PM IST Mesika, Asaf wrote:
>
> >130 GB raw data will take in HBase since it adds the family name,
> qualifier and timestamp to each value, so it can even be 150GB. You can
> check it exactly, by loading only one row with one column and see how much
> it takes on the HDFS file system (run compaction first).
> >
> >Next, you 5 times that since you have 5 times replication, so 5x150=750GB
> >
> >On Jan 11, 2013, at 5:07 AM, Panshul Whisper wrote:
> >
> >> Hello,
> >>
> >> I have a 5 node hadoop cluster and a fully distributed Hbase setup on
> the
> >> cluster with 130 GB of HDFS space avaialble. HDFS replication is set to
> 5.
> >>
> >> I have a total of 115 GB of JSON files that need to be loaded into the
> >> Hbase database and then they have to processed.
> >>
> >> So is the available HDFS space sufficient for the operations??
> considering
> >> the replication and all factors?
> >> or should I increase the space and by how much?
> >>
> >> Thanking You,
> >>
> >> --
> >> Regards,
> >> Ouch Whisper
> >> 010101010101
> >
>
>

Reply via email to