Re: ZFS for Cyrus IMAP storage
On Fri, May 04, 2007 at 06:35:41PM -0500, Gary Mills wrote: There was a question earlier regarding ZFS for Cyrus IMAP storage. We recently converted to that filesystem here. I'm extremely pleased with it. Our server has about 30,000 users with over 200,000 mailboxes. It peaks at about 1900 IMAP sessions. It currently has 1 TB of storage with about 300 GB in use. The server is a Sun T2000 with 16 GB of memory, running Solaris 10. It uses iSCSI to provide two 500 GB devices from our Netapp filer. Disk redundancy is all on the Netapp because it's currently superior to that provided by ZFS. I see that I need to clarify this statement. ZFS' disk redundancy, which is excellent, is not used in this configuration. ZFS' disk management, which also can't be used in this configuration, is currently incomplete. Disk management on the Netapp works very nicely. -- -Gary Mills--Unix Support--U of M Academic Computing and Networking- Cyrus Home Page: http://cyrusimap.web.cmu.edu/ Cyrus Wiki/FAQ: http://cyrusimap.web.cmu.edu/twiki List Archives/Info: http://asg.web.cmu.edu/cyrus/mailing-list.html
Re: ZFS for Cyrus IMAP storage
I originally brought up the ZFS question. We seem to have arrived at a similar solution after much experimentation. Meaning using ZFS for the things it does well already, and leveraging proven hardware to fill in the weak spots. I have a pair of Sun 3510FC arrays we have exported 2 RAID5 LUNs (5 disks each) with one on primary and the other on second controller. This is to exploit the active-active controller feature of the 3510FC. We are also doing multipathing through a pair of SAN fabric switches. On top of that we then use ZFS to join a LUN from each array into a mirror pair, and then add the other pair as well. I guess you could call it RAID 5+1+0. This architecture allows us to add more storage to the pool while online, by adding more 3510FC array pairs to the pool. Performance in benchmarking (Bonnie++ etc.) has shown to be little different from turning them into JBOD and doing everything with ZFS. Behavior is more predictable to me since I know that the 3510 firmware knows how to rebuild a RAID5 set using the assigned spare drive in that array. With ZFS I see no way to specify which disk is assigned as spare to a particular set of disks, which could mean a spare is pulled from another array. It's pretty nifty to be able to walk into the machine room and flip off the power to an entire array and things keep working without a blip. It's not the most efficient usage of disk space but with performance safety this promising for an EMAIL SERVER it will definitely be welcome. I dread the idea of silent data corruption or long fsck time on a 1+ TB mail spool which ZFS should save us from. I have atime=off and compression=on. Our setup is slightly different from yours in that we are clustering 2 T2000 with 8GB RAM each, and we are currently setting up Solaris Cluster 3.2 software in failover configuration so we can patch without downtime. Thanks for the idea about daily snapshots for recovering recent data, I like that idea a lot. I'll tinker around with it I wonder if there'd be much penalty to upping the snapshots to every 8 hours. Depends on how much churn there is in your mail spool I suppose. This system should move into production later this month. We have 70,000 accounts that we'll begin a long and slow migration from our UW-IMAP pool of servers. We have an existing Perdition proxy server setup, which will allow us to migrate users transparently. Hopefully I'll have more good things to say about it sometime thereafter. Cyrus Home Page: http://cyrusimap.web.cmu.edu/ Cyrus Wiki/FAQ: http://cyrusimap.web.cmu.edu/twiki List Archives/Info: http://asg.web.cmu.edu/cyrus/mailing-list.html
Re: ZFS for Cyrus IMAP storage
Gary Mills wrote: There was a question earlier regarding ZFS for Cyrus IMAP storage. We recently converted to that filesystem here. I'm extremely pleased with it. Our server has about 30,000 users with over 200,000 mailboxes. It peaks at about 1900 IMAP sessions. It currently has 1 TB of storage with about 300 GB in use. The server is a Sun T2000 with 16 GB of memory, running Solaris 10. It uses iSCSI to provide two 500 GB devices from our Netapp filer. Disk redundancy is all on the Netapp because it's currently superior to that provided by ZFS. Our T2000 have 8 GB for the cluster we are building here at University of California Davis. Would you expect we will be having to upgrade RAM, or is it just gravy? Cyrus Home Page: http://cyrusimap.web.cmu.edu/ Cyrus Wiki/FAQ: http://cyrusimap.web.cmu.edu/twiki List Archives/Info: http://asg.web.cmu.edu/cyrus/mailing-list.html
Re: ZFS for Cyrus IMAP storage
On Sat, May 05, 2007 at 10:14:07AM -0700, Vincent Fox wrote: Gary Mills wrote: There was a question earlier regarding ZFS for Cyrus IMAP storage. We recently converted to that filesystem here. I'm extremely pleased with it. Our server has about 30,000 users with over 200,000 mailboxes. It peaks at about 1900 IMAP sessions. It currently has 1 TB of storage with about 300 GB in use. The server is a Sun T2000 with 16 GB of memory, running Solaris 10. It uses iSCSI to provide two 500 GB devices from our Netapp filer. Disk redundancy is all on the Netapp because it's currently superior to that provided by ZFS. Our T2000 have 8 GB for the cluster we are building here at University of California Davis. Would you expect we will be having to upgrade RAM, or is it just gravy? I really don't know. IMAP sessions need lots of memory. ZFS will use all it can get for its ARC cache. For a one-off server, I find it cheaper to overdesign it than to devote a lot of time to full load testing. -- -Gary Mills--Unix Support--U of M Academic Computing and Networking- Cyrus Home Page: http://cyrusimap.web.cmu.edu/ Cyrus Wiki/FAQ: http://cyrusimap.web.cmu.edu/twiki List Archives/Info: http://asg.web.cmu.edu/cyrus/mailing-list.html
Re: ZFS for Cyrus IMAP storage
Vincent Fox wrote: I originally brought up the ZFS question. We seem to have arrived at a similar solution after much experimentation. Meaning using ZFS for the things it does well already, and leveraging proven hardware to fill in the weak spots. I have a pair of Sun 3510FC arrays we have exported 2 RAID5 LUNs (5 disks each) with one on primary and the other on second controller. This is to exploit the active-active controller feature of the 3510FC. We are also doing multipathing through a pair of SAN fabric switches. On top of that we then use ZFS to join a LUN from each array into a mirror pair, and then add the other pair as well. I guess you could call it RAID 5+1+0. This architecture allows us to add more storage to the pool while online, by adding more 3510FC array pairs to the pool. Performance in benchmarking (Bonnie++ etc.) has shown to be little different from turning them into JBOD and doing everything with ZFS. Behavior is more predictable to me since I know that the 3510 firmware knows how to rebuild a RAID5 set using the assigned spare drive in that array. With ZFS I see no way to specify which disk is assigned as spare to a particular set of disks, which could mean a spare is pulled from another array. It's pretty nifty to be able to walk into the machine room and flip off the power to an entire array and things keep working without a blip. It's not the most efficient usage of disk space but with performance safety this promising for an EMAIL SERVER it will definitely be welcome. I dread the idea of silent data corruption or long fsck time on a 1+ TB mail spool which ZFS should save us from. I have atime=off and compression=on. Our setup is slightly different from yours in that we are clustering 2 T2000 with 8GB RAM each, and we are currently setting up Solaris Cluster 3.2 software in failover configuration so we can patch without downtime. Thanks for the idea about daily snapshots for recovering recent data, I like that idea a lot. I'll tinker around with it I wonder if there'd be much penalty to upping the snapshots to every 8 hours. Depends on how much churn there is in your mail spool I suppose. This system should move into production later this month. We have 70,000 accounts that we'll begin a long and slow migration from our UW-IMAP pool of servers. We have an existing Perdition proxy server setup, which will allow us to migrate users transparently. Hopefully I'll have more good things to say about it sometime thereafter. Are you going to do this with 1 perdition server? Make sure you have compiled perdition with /dev/urandom, or an other sort of non blocking entropy providing device :) Rudy Cyrus Home Page: http://cyrusimap.web.cmu.edu/ Cyrus Wiki/FAQ: http://cyrusimap.web.cmu.edu/twiki List Archives/Info: http://asg.web.cmu.edu/cyrus/mailing-list.html
ZFS for Cyrus IMAP storage
There was a question earlier regarding ZFS for Cyrus IMAP storage. We recently converted to that filesystem here. I'm extremely pleased with it. Our server has about 30,000 users with over 200,000 mailboxes. It peaks at about 1900 IMAP sessions. It currently has 1 TB of storage with about 300 GB in use. The server is a Sun T2000 with 16 GB of memory, running Solaris 10. It uses iSCSI to provide two 500 GB devices from our Netapp filer. Disk redundancy is all on the Netapp because it's currently superior to that provided by ZFS. ZFS is fully POSIX compliant; Cyrus is quite happy with that. Performance is excellent. The full backups go much faster than they did when we were using UFS on a pair of locally-attached RAID-5 arrays. The ability to expand the storage pool simply by adding more devices from the Netapp is very appealing. No longer will we have to move mailboxes from one partition to another. The snapshots are wonderful. We keep 14 daily snapshots, instead of doing daily incremental backups. I can certainly recommend ZFS for Cyrus. -- -Gary Mills--Unix Support--U of M Academic Computing and Networking- Cyrus Home Page: http://cyrusimap.web.cmu.edu/ Cyrus Wiki/FAQ: http://cyrusimap.web.cmu.edu/twiki List Archives/Info: http://asg.web.cmu.edu/cyrus/mailing-list.html