[ 
https://issues.apache.org/jira/browse/HDFS-14669?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16894453#comment-16894453
 ] 

qiang Liu commented on HDFS-14669:
----------------------------------

after doing some tests and anylize, I think if we successfully create some 
files with countent when there is only one datanode, the datanode blocks must 
has been created too. which means doing scan check multiple times is not 
necessary.

[~templedf] you introduced this multiple check logic in HDFS-13819 , if my 
anylize is wrong or missing something, just let me know.**

if everything is ok, I will submit a pach latter remove the use of 
GenericTestUtils.waitFor

> TestDirectoryScanner#testDirectoryScannerInFederatedCluster fails 
> intermittently in trunk
> -----------------------------------------------------------------------------------------
>
>                 Key: HDFS-14669
>                 URL: https://issues.apache.org/jira/browse/HDFS-14669
>             Project: Hadoop HDFS
>          Issue Type: Test
>          Components: datanode
>    Affects Versions: 3.2.0
>         Environment: env free
>            Reporter: qiang Liu
>            Assignee: qiang Liu
>            Priority: Minor
>              Labels: scanner, test
>         Attachments: HDFS-14669-trunk-001.patch, HDFS-14669-trunk.002.patch
>
>
> org.apache.hadoop.hdfs.server.datanode.TestDirectoryScanner#testDirectoryScannerInFederatedCluster
>  radomlly Failes because of write files of the same name, meaning intent to 
> write 2 files but  2 files are the same name, witch cause a race condition of 
> datanode delete block and the scan action count block.
>  
> Ref :: 
> [https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1207/testReport/junit/org.apache.hadoop.hdfs.server.datanode/TestDirectoryScanner/testDirectoryScannerInFederatedCluster/]
>  



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

---------------------------------------------------------------------
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org

Reply via email to