Deleting large files is time-consuming, and a large part of the time is spent in f2fs_invalidate_blocks() ->down_write(sit_info->sentry_lock) and up_write().
If some blocks are continuous and belong to the same segment, we can process these blocks at the same time. This can reduce the number of calls to the down_write() and the up_write(), thereby improving the overall speed of doing truncate. Test steps: Set the CPU and DDR frequencies to the maximum. dd if=/dev/random of=./test.txt bs=1M count=100000 sync rm test.txt Time Comparison of rm: original optimization ratio 7.17s 3.27s 54.39% Hi, currently I have only optimized the f2fs doing truncate route, and other functions using f2fs_invalidate_blocks() are not taken into consideration. So new function f2fs_invalidate_compress_pages_range() and check_f2fs_invalidate_consecutive_blocks() are not general functions. Is this modification acceptable? Yi Sun (2): f2fs: introduce update_sit_entry_for_release() f2fs: introduce f2fs_invalidate_consecutive_blocks() for truncate fs/f2fs/compress.c | 14 ++++++ fs/f2fs/f2fs.h | 5 ++ fs/f2fs/file.c | 34 ++++++++++++- fs/f2fs/segment.c | 116 +++++++++++++++++++++++++++++++-------------- 4 files changed, 133 insertions(+), 36 deletions(-) -- 2.25.1 _______________________________________________ Linux-f2fs-devel mailing list Linux-f2fs-devel@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/linux-f2fs-devel