[
https://issues.apache.org/jira/browse/HDFS-8581?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
tongshiquan updated HDFS-8581:
------------------------------
Description:
If one directory such as "/result" exists about 200000 files, then when execute
"hdfs dfs -count /", the result will go wrong. For all directories whose name
after "/result", file num will not be included.
My cluster see as snapshot, "/result_1433858936" is the directory exist huge
files, and files in "/sparkJobHistory", "/tmp", "/user" are not included
was:
If one directory such as "/result" exists about 200000 files, then when execute
"hdfs dfs -count /", the result will go wrong. For all directories whose name
after "/result", file num will not be included.
Here is my cluster:
> count cmd calculate wrong when huge files exist in one folder
> -------------------------------------------------------------
>
> Key: HDFS-8581
> URL: https://issues.apache.org/jira/browse/HDFS-8581
> Project: Hadoop HDFS
> Issue Type: Bug
> Components: HDFS
> Reporter: tongshiquan
> Assignee: J.Andreina
> Priority: Minor
>
> If one directory such as "/result" exists about 200000 files, then when
> execute "hdfs dfs -count /", the result will go wrong. For all directories
> whose name after "/result", file num will not be included.
> My cluster see as snapshot, "/result_1433858936" is the directory exist huge
> files, and files in "/sparkJobHistory", "/tmp", "/user" are not included
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)