Hi, I use the default replication factor 3 here, the cluster has 10 nodes, each of my datanode has 8 hard disks. If one of the nodes is down because of hardware failure, i.e. the 8 hard disks will no longer be available immediately during the down time of this machine, does it mean that I will have data lost? (8 hard disks > 3 replicated)
Or what would be the maximum number of servers that are allowed to be down without data lost here? Regards Arthur On Wednesday, December 17, 2014, Harshit Mathur <[email protected]> wrote: > Hi Arthur, > > In HDFS there will be block level replication, In case of total failure of > a datanode the lost blocks will get under replicated hence the namenode > will create copy of these under replicated blocks on some other datanode. > > BR, > Harshit > > On Wed, Dec 17, 2014 at 11:35 AM, [email protected] > <javascript:_e(%7B%7D,'cvml','[email protected]');> < > [email protected] > <javascript:_e(%7B%7D,'cvml','[email protected]');>> wrote: >> >> Hi, >> >> If each of my datanode servers has 8 hard disks (a 10-node cluster) and >> I use the default replication factor of 3, how will Hadoop handle it when a >> datanode with total hardware failure suddenly? >> >> Regards >> Arthur >> > > > > -- > Harshit Mathur >
