Hi,

I use the default replication factor 3 here, the cluster has 10 nodes, each
of my datanode has 8 hard disks.  If one of the nodes is down because of
hardware failure, i.e. the 8 hard disks will no longer be available
immediately during the down time of this machine, does it mean that I will
have data lost? (8 hard disks >  3 replicated)

Or what would be the maximum number of servers that are allowed to be down
without data lost here?

Regards
Arthur

On Wednesday, December 17, 2014, Harshit Mathur <[email protected]>
wrote:

> Hi Arthur,
>
> In HDFS there will be block level replication, In case of total failure of
> a datanode the lost blocks will get under replicated hence the namenode
> will create copy of these under replicated blocks on some other datanode.
>
> BR,
> Harshit
>
> On Wed, Dec 17, 2014 at 11:35 AM, [email protected]
> <javascript:_e(%7B%7D,'cvml','[email protected]');> <
> [email protected]
> <javascript:_e(%7B%7D,'cvml','[email protected]');>> wrote:
>>
>> Hi,
>>
>> If each of  my datanode servers has 8 hard disks (a 10-node cluster) and
>> I use the default replication factor of 3, how will Hadoop handle it when a
>> datanode with total hardware failure suddenly?
>>
>> Regards
>> Arthur
>>
>
>
>
> --
> Harshit Mathur
>

Reply via email to