On Apr 18, 2007, at 6:44 AM, Yaniv Aknin wrote: > Hello, > > I'd like to plan a storage solution for a system currently in > production. > > The system's storage is based on code which writes many files to > the file system, with overall storage needs currently around 40TB > and expected to reach hundreds of TBs. The average file size of the > system is ~100K, which translates to ~500 million files today, and > billions of files in the future. This storage is accessed over NFS > by a rack of 40 Linux blades, and is mostly read-only (99% of the > activity is reads). While I realize calling this sub-optimal system > design is probably an understatement, the design of the system is > beyond my control and isn't likely to change in the near future. > > The system's current storage is based on 4 VxFS filesystems, > created on SVM meta-devices each ~10TB in size. A 2-node Sun > Cluster serves the filesystems, 2 filesystems per node. Each of the > filesystems undergoes growfs as more storage is made available. > We're looking for an alternative solution, in an attempt to improve > performance and ability to recover from disasters (fsck on 2^42 > files isn't practical, and I'm getting pretty worried due to this > fact - even the smallest filesystem inconsistency will leave me > lots of useless bits). > > Question is - does anyone here have experience with large ZFS > filesystems with many small-files? Is it practical to base such a > solution on a few (8) zpools, each with single large filesystem in it?
hey Yaniv, Why not 1 pool? That's what we usually recommend (you can have 8 filesystems on top of the 1 pool if you need to). eric