On Mon, Sep 28, 2015 at 10:59:44AM -0700, Jaegeuk Kim <jaeg...@kernel.org> wrote: > In order to verify this also, could you retrieve the following logs?
First thing, the allocation-failure-on-mount is still in the backported 3.18 f2fs module. If it's supposed to be gone in that version, it's not working: http://ue.tst.eu/a1bc4796012bd7191ab2ada566d4cd22.txt And here are traces and descriptions. The traces all start directly after mount, my test script is http://data.plan9.de/f2fstest (event tracing is cool btw., thanks for showing me :) ################ -s1, f2fs git ############################################## /opt/f2fs-tools/sbin/mkfs.f2fs -lTEST -s1 -t0 -a0 /dev/vg_test/test mount -t f2fs -onoatime,flush_merge,no_heap /dev/vg_test/test /mnt For the fist ~120GB, performance was solid (100MB/s+), but much worse than stock 3.18.21 (with -s64!). 3.1.8.21 regularly reached >190MB/s regularly (at least near the beginning of the disk) then was idle in between writes, as the source wasn't fast enough to keep up. With the backport, tar was almost never idle, and if, then not for long, so it could just keep up. (Just keeping up with the read speed of a 6-disk raid is very good, but I know f2fs can do much better :) At the 122GB mark, it started to slow down, being consistently <100MB/s At 127GB, it was <<20MB/s, and I stopped. Most of the time, the test was write-I/O-bound. http://data.plan9.de/f2fs.s1.trace.xz ################ -s64, f2fs 3.18.21 ######################################### As contrast I then did a test with the original f2fs module, and -s64. Throughput was up to 202MB/s, almost continously. At the 100GB mark, it slowed down to maybe 170MB/s peak, which might well be the speed of the platters. I stopped at 217GB. I have a 12GB mbuffer between the read-tar and the write-tar, configured to write minimum bursts of ~120MB. At no time was the buffer filled at >2%, while with the -s1, f2fs git case, it was basically always >2%. The trace includes a few minutes after tar was stopped. http://data.plan9.de/f2fs.s64.3.18.trace.xz ################ -s64, f2fs git ############################################# The direct equivalent of the previous test, but with f2fs git. Almost from the very beginning, it was often write-bound, but could still keep up. At around 70GB, it mostly stopped being able to keep up, and the read tar overtook the write tar. At 139GB, performance degraded to <2MB/s. I stopped at 147GB. So mostly, behaviour was the same as with -s1, excedpt it took longer to slow down. http://data.plan9.de/f2fs.s64.trace.xz ################ -s20, f2fs git ############################################# By special request, here is the test with -s20. Surprisingly, this stopped being able to cope at the 40GB mark, but I didn't wait very long after the previous test, maybe that influenced it. I stopped at 63GB. http://data.plan9.de/f2fs.s20.trace.xz ############################################################################# I hope to find time to look at these traces myself later this day. -- The choice of a Deliantra, the free code+content MORPG -----==- _GNU_ http://www.deliantra.net ----==-- _ generation ---==---(_)__ __ ____ __ Marc Lehmann --==---/ / _ \/ // /\ \/ / schm...@schmorp.de -=====/_/_//_/\_,_/ /_/\_\ ------------------------------------------------------------------------------ _______________________________________________ Linux-f2fs-devel mailing list Linux-f2fs-devel@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/linux-f2fs-devel