[ https://issues.apache.org/jira/browse/HADOOP-2094?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12560050#action_12560050 ]
Raghu Angadi commented on HADOOP-2094: -------------------------------------- Better (but much more complicated) policy could be to select random partition from least loaded disks. > DFS should not use round robin policy in determing on which volume (file > system partition) to allocate for the next block > -------------------------------------------------------------------------------------------------------------------------- > > Key: HADOOP-2094 > URL: https://issues.apache.org/jira/browse/HADOOP-2094 > Project: Hadoop > Issue Type: Improvement > Components: dfs > Reporter: Runping Qi > Assignee: dhruba borthakur > Attachments: randomDatanodePartition.patch > > > When multiple file system partitions are configured for the data storage of a > data node, > it uses a strict round robin policy to decide which partition to use for > writing the next block. > This may result in anormaly cases in which the blocks of a file are not > evenly distributed across > the partitions. For example, when we use distcp to copy files with each node > have 4 mappers running concurrently, > those 4 mappers are writing to DFS at about the same rate. Thus, it is > possible that the 4 mappers write out > blocks interleavingly. If there are 4 file system partitions configured for > the local data node, it is possible that each mapper will > continue to write its blocks on to the same file system partition. > A simple random placement policy will avoid such anormaly cases, and does not > have any obvious drawbacks. > -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.