Kitti Nanasi created HDFS-14187: ----------------------------------- Summary: Make warning message more clear when there are not enough data nodes for EC write Key: HDFS-14187 URL: https://issues.apache.org/jira/browse/HDFS-14187 Project: Hadoop HDFS Issue Type: Improvement Components: erasure-coding Affects Versions: 3.1.1 Reporter: Kitti Nanasi Assignee: Kitti Nanasi
When setting an erasure coding policy for which there are not enough racks or data nodes, write will fail with the following message: {code:java} [root@oks-upgrade6727-1 ~]# sudo -u systest hdfs dfs -mkdir /user/systest/testdir [root@oks-upgrade6727-1 ~]# sudo -u hdfs hdfs ec -setPolicy -path /user/systest/testdir Set default erasure coding policy on /user/systest/testdir [root@oks-upgrade6727-1 ~]# sudo -u systest hdfs dfs -put /tmp/file1 /user/systest/testdir 18/11/12 05:41:26 WARN hdfs.DFSOutputStream: Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Not enough datanodes? Exclude nodes=[] 18/11/12 05:41:26 WARN hdfs.DFSOutputStream: Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Not enough datanodes? Exclude nodes=[] 18/11/12 05:41:26 WARN hdfs.DFSOutputStream: Block group <1> failed to write 2 blocks. It's at high risk of losing data. {code} I suggest to log a more descriptive message suggesting to use hdfs ec -verifyCluster command to verify the cluster setup against the ec policies. -- This message was sent by Atlassian JIRA (v7.6.3#76005) --------------------------------------------------------------------- To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org