On Apr 18, 2007, at 6:44 AM, Yaniv Aknin wrote:

> Hello,
>
> I'd like to plan a storage solution for a system currently in  
> production.
>
> The system's storage is based on code which writes many files to  
> the file system, with overall storage needs currently around 40TB  
> and expected to reach hundreds of TBs. The average file size of the  
> system is ~100K, which translates to ~500 million files today, and  
> billions of files in the future. This storage is accessed over NFS  
> by a rack of 40 Linux blades, and is mostly read-only (99% of the  
> activity is reads). While I realize calling this sub-optimal system  
> design is probably an understatement, the design of the system is  
> beyond my control and isn't likely to change in the near future.
>
> The system's current storage is based on 4 VxFS filesystems,  
> created on SVM meta-devices each ~10TB in size. A 2-node Sun  
> Cluster serves the filesystems, 2 filesystems per node. Each of the  
> filesystems undergoes growfs as more storage is made available.  
> We're looking for an alternative solution, in an attempt to improve  
> performance and ability to recover from disasters (fsck on 2^42  
> files isn't practical, and I'm getting pretty worried due to this  
> fact - even the smallest filesystem inconsistency will leave me  
> lots of useless bits).
>
> Question is - does anyone here have experience with large ZFS  
> filesystems with many small-files? Is it practical to base such a  
> solution on a few (8) zpools, each with single large filesystem in it?

hey Yaniv,

Why not 1 pool?  That's what we usually recommend (you can have 8  
filesystems on top of the 1 pool if you need to).

eric


Reply via email to