+1 on both. Mostly SmartOS, some FreeNAS (which is FreeBSD underneath). -r
Ryan Brooks <r...@hack.net> writes: > Zfs on BSD or a Solaris like OS > > >> On Dec 11, 2014, at 10:06 AM, Bacon Zombie <baconzom...@gmail.com> wrote: >> >> Are you running ZFS and RAIDZ on Linux or BSD? >>> On 10 Dec 2014 23:21, "Javier J" <jav...@advancedmachines.us> wrote: >>> >>> I'm just going to chime in here since I recently had to deal with bit-rot >>> affecting a 6TB linux raid5 setup using mdadm (6x 1TB disks) >>> >>> We couldn't rebuild because of 5 URE sectors on one of the other disks in >>> the array after a power / ups issue rebooted our storage box. >>> >>> We are now using ZFS RAIDZ and the question I ask myself is, why wasn't I >>> using ZFS years ago? >>> >>> +1 for ZFS and RAIDZ >>> >>> >>> >>>> On Wed, Dec 10, 2014 at 8:40 AM, Rob Seastrom <r...@seastrom.com> wrote: >>>> >>>> >>>> The subject is drifting a bit but I'm going with the flow here: >>>> >>>> Seth Mos <seth....@dds.nl> writes: >>>> >>>>> Raid10 is the only valid raid format these days. With the disks as big >>>>> as they get these days it's possible for silent corruption. >>>> >>>> How do you detect it? A man with two watches is never sure what time it >>>> is. >>>> >>>> Unless you have a filesystem that detects and corrects silent >>>> corruption, you're still hosed, you just don't know it yet. RAID10 >>>> between the disks in and of itself doesn't help. >>>> >>>>> And with 4TB+ disks that is a real thing. Raid 6 is ok, if you accept >>>>> rebuilds that take a week, literally. Although the rebuild rate on our >>>>> 11 disk raid 6 SSD array (2TB) is less then a day. >>>> >>>> I did a rebuild on a RAIDZ2 vdev recently (made out of 4tb WD reds). >>>> It took nowhere near a day let alone a week. Theoretically takes 8-11 >>>> hours if the vdev is completely full, proportionately less if it's >>>> not, and I was at about 2/3 in use. >>>> >>>> -r >>>