0

I'm doing some tests on Hadoop and the HDFS is almost full after those tests. So I deleted a lot of test files recursively but the released space is not as much as I expected.

Running hadoop fs -du -s / command, it reports 876447242289. Means about 0.8 TB. But in Name Node UI, it reports 12 TB still in used.

Configured Capacity  :   19.89 TB
DFS Used         :   12.15 TB
Non DFS Used     :   801.66 GB
DFS Remaining    :   6.95 TB
DFS Used%        :   61.11 %
DFS Remaining%   :   34.96 %
Block Pool Used  :   12.15 TB
Block Pool Used% :   61.11 %

I Ran `df -h. in each data node and I found the same result as Name Node UI.

I checked the .Trash folder and nothing is there. Looks like many files are removed from Name Node, but still not deleted in Data Node. Can anyone help on this?

Manjunath Ballur
  • 6,287
  • 3
  • 37
  • 48
avhacker
  • 667
  • 1
  • 9
  • 20

1 Answers1

0

I found that the space will be released gradually. After 8 hours, all the spaces that were held by deleted files are released.

avhacker
  • 667
  • 1
  • 9
  • 20