Hi, I am new to hdfs management. Hope you guys help me on this. I have a 12-node cluster, up to last week, I found the space usage of hdfs was approaching to 75%, then I remove tens of thousands the files, which are quite small ones, after deletion, i use *hadop fs -du / *, it shows only 11TB space are used. *32296810 hdfs://hadoop-name01/hadoop 3506 hdfs://hadoop-name01/hbase 66,380,418,656 hdfs://hadoop-name01/tmp 10,280,770,071,939 hdfs://hadoop-name01/user*
but when I tried to use *hadoop fs -df /*, it said there are still 68% of space are used. *Filesystem Size Used Avail Use% / 70767414312960 48673246151277 18472473866240 68%* I am puzzled here since du said only 11T are used, but in df said 44T are used. Also in dfsadmin report, 44TB is reported as used, any reason some space(*33TB*) are missing after my deletion? Any light shed on me are highly welcomed! *$ hadoop dfsadmin -report* *Configured Capacity: 70767414312960 (64.36 TB) Present Capacity: 67143178552151 (61.07 TB) DFS Remaining: 18416113192960 (16.75 TB) DFS Used: 48727065359191 (44.32 TB) DFS Used%: 72.57% Under replicated blocks: 0 Blocks with corrupt replicas: 0 Missing blocks: 0* Regards, Peter Li