On Mon, Aug 10, 2015 at 10:53:32PM +0200, Marc Lehmann wrote: > On Mon, Aug 10, 2015 at 01:31:06PM -0700, Jaegeuk Kim <jaeg...@kernel.org> > wrote: > > I'm very interested in trying f2fs on SMR drives too. > > I also think that several characteristics of SMR drives are very similar > > with > > flash drives. > > Indeed, but of course there isn't an exact match for any characteristic. > Also, in the end, drive-managed SMR drives will suck somewhat with any > filesystem (note that nilfs performs very badly, even thought it should be > better than anything else till the drive is completely full).
IMO, it's similar to flash drives too. Indeed, I believe host-managed SMR/flash drives are likely to show much better performance than drive-managed ones. However, I think there are many HW constraints inside the storage not to move forward to it easily. > Now, looking at the characteristics of f2fs, it could be a good match for > any rotational media, too, since it writes linearly and can defragment. At > least for desktop or similar loads (where files usually aren't randomly > written, but mostly replaced and rarely appended). Possible, but not much different from other filesystems. :) > The only crucial ability it would need to have is to be able to free large > chunks for rewriting, which should be in f2fs as well. > > So at this time, what I apparently need is mkfs.f2fs -s128 instead of -s7. I wrote a patch to fix the document. Sorry about that. > Unfortunately, I probably can't make these tests immediately, and they do > take some days to run, but hopefully I cna repeatmy experiments next week. > > > - over 4TB storage space case > > fsck limits could well have been the issue for my first big filesystem, > but not the second (which was only 128G in size to be able to utilize it > within a reasonable time). > > > - inline_dentry mount option; I'm still working on extent_cache for v4.3 too > > I only enabled mount options other than noatime for the 128G filesystem, > so it might well have cauzsed the trouble with it. Okay, so I think it'd be good to start with: - noatime,inline_xattr,inline_data,flush_merge,extent_cache. And you can control defragementation through /sys/fs/f2fs/[DEV]/gc_[min|max|no]_sleep_time > Another thing that will seriously hamper adoption of these drives is the > 32000 limit on hardlinks - I am hard pressed to find any large file tree > here that doesn't have places with of 40000 subdirs somewhere, but I guess > on a 32GB phone flash storage, this was less of a concern. Looking at a glance, it'll be no problme to increase as 64k. Let me check again. > In any case, if f2fs turns out to be workable, it will become the fs of > choice for me for my archival uses, and maybe even more, and I then have > to somehow cope with that limit. > > > In your logs, I suspect some fsck.f2fs bugs in a large storage case. > > In order to confirm that, could you use the latest f2fs-tools from: > > http://git.kernel.org/cgit/linux/kernel/git/jaegeuk/f2fs-tools.git > > Will do so. > > Is there a repository for out-of-tree module builds for f2fs? It seems > kernels 3.17.x to 4.1 (at least) have a kernel bug making reads to these SMR > drives unstable (https://bugzilla.kernel.org/show_bug.cgi?id=93581), so I > will have to test with a relatively old kernel or play too many tricks. What kernel version do you prefer? I've been maintaining f2fs for v3.10 mainly. http://git.kernel.org/cgit/linux/kernel/git/jaegeuk/f2fs.git/log/?h=linux-3.10 Thanks, > And I suspect from glancing over patches (And mount options) that there > have been quite some improvements in f2fs since 3.16 days. > > > And, if possible, could you share some experiences when you didn't fill up > > the > > partition to 100%? If there is no problem, we can nicely focus on ENOSPC > > only. > > My experience was that f2fs wrote at nearly maximum I/O speed of the drives. > In fact, I couldn't saturate the bandwidth except when writing small files, > because the 8 drive source raid using xfs was not able to read files quickly > enough. After writing an initial tree of >2TB > > Directory reading and mass stat seemed to be considerably slower and take > more time directly afterwards. I don't know if that is something that > balancing can fix (or improve), but I am not overly concerned about that, > as the difference to e.g. xfs is not that big (roughly a factor of two), > and thes eoperations are too slow for me on any device, so I usually put a > dm-cache in front of such storage devices. > > I don't think that I have more useful data to report - if I used 14MB > sections, performance would predictably suck, so the real test is still > outstanding. Stay tuned, and thanks for your reply! > > -- > The choice of a Deliantra, the free code+content MORPG > -----==- _GNU_ http://www.deliantra.net > ----==-- _ generation > ---==---(_)__ __ ____ __ Marc Lehmann > --==---/ / _ \/ // /\ \/ / schm...@schmorp.de > -=====/_/_//_/\_,_/ /_/\_\ ------------------------------------------------------------------------------ _______________________________________________ Linux-f2fs-devel mailing list Linux-f2fs-devel@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/linux-f2fs-devel