[ 
https://issues.apache.org/jira/browse/HDFS-12946?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16662643#comment-16662643
 ] 

Xiao Chen commented on HDFS-12946:
----------------------------------

I had more thoughts into this. How about we extract the logic of 
{{ErasureCodingClusterSetupVerifier}} util, then use it in both the hdfs-client 
and hdfs sides? This means we don't need to add any RPCs, and be able to use 
the logic to calculate in NN (from internal state) and at the client (using 
existing APIs to get policies and DN stats). JMX would still be great to have.

> Add a tool to check rack configuration against EC policies
> ----------------------------------------------------------
>
>                 Key: HDFS-12946
>                 URL: https://issues.apache.org/jira/browse/HDFS-12946
>             Project: Hadoop HDFS
>          Issue Type: Improvement
>          Components: erasure-coding
>            Reporter: Xiao Chen
>            Assignee: Kitti Nanasi
>            Priority: Major
>         Attachments: HDFS-12946.01.patch, HDFS-12946.02.patch, 
> HDFS-12946.03.patch, HDFS-12946.04.fsck.patch
>
>
> From testing we have seen setups with problematic racks / datanodes that 
> would not suffice basic EC usages. These are usually found out only after the 
> tests failed.
> We should provide a way to check this beforehand.
> Some scenarios:
> - not enough datanodes compared to EC policy's highest data+parity number
> - not enough racks to satisfy BPPRackFaultTolerant
> - highly uneven racks to satisfy BPPRackFaultTolerant
> - highly uneven racks (so that BPP's considerLoad logic may exclude some busy 
> nodes on the rack, resulting in #2)



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to