Re: [zfs-discuss] ZFS and spindle speed (7.2k / 10k / 15k)

2011-02-06 Thread Brandon High
On Sat, Feb 5, 2011 at 3:34 PM, Roy Sigurd Karlsbakk r...@karlsbakk.net wrote: so as not to exceed the channel bandwidth. When they need to get higher disk capacity, they add more platters. May this mean those drives are more robust in terms of reliability, since the leaks between sectors is

Re: [zfs-discuss] Drive id confusion

2011-02-06 Thread Chris Ridd
On 6 Feb 2011, at 03:14, David Dyer-Bennet wrote: I'm thinking either Solaris' appalling mess of device files is somehow scrod, or else ZFS is confused in its reporting (perhaps because of cache file contents?). Is there anything I can do about either of these? Does devfsadm really

Re: [zfs-discuss] ZFS/Drobo (Newbie) Question

2011-02-06 Thread Orvar Korvar
Yes, you create three groups as you described and insert them into your zpool (the zfs raid). So you have only one ZFS raid, consisting of three groups. You dont have three different ZFS raids (unless you configure that). You can also later, swap one disk to a larger and repair the group. Then

Re: [zfs-discuss] ZFS and TRIM - No need for TRIM

2011-02-06 Thread Orvar Korvar
Ok, so can we say that the conclusion for a home user is: 1) Using SSD without TRIM is acceptable. The only drawback is that without TRIM, the SSD will write much more, which effects life time. Because when the SSD has written enough, it will break. I dont have high demands for my OS disk, so

Re: [zfs-discuss] Identifying drives (SATA)

2011-02-06 Thread Orvar Korvar
Will this not ruin the zpool? If you overwrite one of discs in the zpool won't the zpool go broke, so you need to repair it? -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org

Re: [zfs-discuss] Identifying drives (SATA)

2011-02-06 Thread Roy Sigurd Karlsbakk
Will this not ruin the zpool? If you overwrite one of discs in the zpool won't the zpool go broke, so you need to repair it? As suggested, dd if=/dev/rdsk/c8t3d0s0 of=/dev/null bs=4k count=10, that will do its best to overwrite /dev/null, which the system is likely to allow :P Vennlige

Re: [zfs-discuss] Identifying drives (SATA)

2011-02-06 Thread David Dyer-Bennet
On 2011-02-06 05:58, Orvar Korvar wrote: Will this not ruin the zpool? If you overwrite one of discs in the zpool won't the zpool go broke, so you need to repair it? Without quoting I can't tell what you think you're responding to, but from my memory of this thread, I THINK you're forgetting

[zfs-discuss] Replace block devices to increase pool size

2011-02-06 Thread Achim Wolpers
Hi! I have a zpool biult up from two vdrives (one mirror and one raidz). The raidz is built up from 4x1TB HDs. When I successively replace each 1TB drive with a 2TB drive will the capacity of the raidz double after the last block device is replaced? Achim signature.asc Description: OpenPGP

Re: [zfs-discuss] Replace block devices to increase pool size

2011-02-06 Thread taemun
If autoexpand = on, then yes. zpool get autoexpand pool zpool set autoexpand=on pool The expansion is vdev specific, so if you replaced the mirror first, you'd get that much (the extra 2TB) without touching the raidz. Cheers, On 7 February 2011 01:41, Achim Wolpers achim...@googlemail.com

Re: [zfs-discuss] ZFS and TRIM - No need for TRIM

2011-02-06 Thread Roy Sigurd Karlsbakk
2) And later, when Solaris gets TRIM support, should I reformat or is there no need to reformat? I mean, maybe I must format and reinstall to get TRIM all over the disk. Or will TRIM immediately start to do it's magic? Trim works on the device level, so a reformat won't be necessary Vennlige

Re: [zfs-discuss] Identifying drives (SATA), question about hot spare allocation

2011-02-06 Thread David Dyer-Bennet
Following up to myself, I think I've got things sorted, mostly. 1. The thing I was most sure of, I was wrong about. Some years back, I must have split the mirrors so that they used different brand disks. I probably did this, maybe even accidentally, when I had to restore from backups at

Re: [zfs-discuss] ZFS and TRIM - No need for TRIM

2011-02-06 Thread Erik Trimble
On 2/6/2011 3:51 AM, Orvar Korvar wrote: Ok, so can we say that the conclusion for a home user is: 1) Using SSD without TRIM is acceptable. The only drawback is that without TRIM, the SSD will write much more, which effects life time. Because when the SSD has written enough, it will break. I

Re: [zfs-discuss] ZFS and TRIM - No need for TRIM

2011-02-06 Thread Erik Trimble
On 2/6/2011 3:51 AM, Orvar Korvar wrote: Ok, so can we say that the conclusion for a home user is: 1) Using SSD without TRIM is acceptable. The only drawback is that without TRIM, the SSD will write much more, which effects life time. Because when the SSD has written enough, it will break. I

Re: [zfs-discuss] zfs-discuss Digest, Vol 64, Issue 13

2011-02-06 Thread Michael Armstrong
Additionally, the way I do it is to draw a diagram of the drives in the system, labelled with the drive serial numbers. Then when a drive fails, I can find out from smartctl which drive it is and remove/replace without trial and error. On 5 Feb 2011, at 21:54,

Re: [zfs-discuss] Identifying drives (SATA)

2011-02-06 Thread Orvar Korvar
Heh. My bad. Didnt read the command. Yes, that should be safe. -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Re: [zfs-discuss] Identifying drives (SATA)

2011-02-06 Thread Orvar Korvar
Roy, I read your question on OpenIndiana mail lists: how can you rebalance your huge raid, without implementing block pointer rewrite? You have an old vdev full of data, and now you have added a new vdev - and you want the data to be evenly spread out to all vdevs. I answer here beceause it is

Re: [zfs-discuss] kernel messages question

2011-02-06 Thread Krunal Desai
On Sat, Feb 5, 2011 at 5:44 PM, Roy Sigurd Karlsbakk r...@karlsbakk.net wrote: Hi I keep getting these messages on this one box. There are issues with at least one of the drives in it, but since there are some 80 drives in it, that's not really an issue. I just want to know, if anyone

[zfs-discuss] RAID Failure Calculator (for 8x 2TB RAIDZ)

2011-02-06 Thread Matthew Angelo
I require a new high capacity 8 disk zpool.  The disks I will be purchasing (Samsung or Hitachi) have an Error Rate (non-recoverable, bits read) of 1 in 10^14 and will be 2TB.  I'm staying clear of WD because they have the new 2048b sectors which don't play nice with ZFS at the moment. My

Re: [zfs-discuss] Repairing Faulted ZFS pool when zbd doesn't recognize the pool as existing

2011-02-06 Thread George Wilson
Chris, I might be able to help you recover the pool but will need access to your system. If you think this is possible just ping me off list and let me know. Thanks, George On Sun, Feb 6, 2011 at 4:56 PM, Chris Forgeron cforge...@acsi.ca wrote: Hello all, Long time reader, first time

Re: [zfs-discuss] RAID Failure Calculator (for 8x 2TB RAIDZ)

2011-02-06 Thread Ian Collins
On 02/ 7/11 03:45 PM, Matthew Angelo wrote: I require a new high capacity 8 disk zpool. The disks I will be purchasing (Samsung or Hitachi) have an Error Rate (non-recoverable, bits read) of 1 in 10^14 and will be 2TB. I'm staying clear of WD because they have the new 2048b sectors which

Re: [zfs-discuss] RAID Failure Calculator (for 8x 2TB RAIDZ)

2011-02-06 Thread Edward Ned Harvey
From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss- boun...@opensolaris.org] On Behalf Of Matthew Angelo My question is, how do I determine which of the following zpool and vdev configuration I should run to maximize space whilst mitigating rebuild failure risk? 1. 2x

Re: [zfs-discuss] Understanding directio, O_DSYNC and zfs_nocacheflush on ZFS

2011-02-06 Thread Richard Elling
On Feb 5, 2011, at 8:10 AM, Yi Zhang wrote: Hi all, I'm trying to achieve the same effect of UFS directio on ZFS and here is what I did: Solaris UFS directio has three functions: 1. improved async code path 2. multiple concurrent writers 3. no buffering Of the

Re: [zfs-discuss] kernel messages question

2011-02-06 Thread Richard Elling
On Feb 5, 2011, at 2:44 PM, Roy Sigurd Karlsbakk wrote: Hi I keep getting these messages on this one box. There are issues with at least one of the drives in it, but since there are some 80 drives in it, that's not really an issue. I just want to know, if anyone knows, what this kernel

Re: [zfs-discuss] RAID Failure Calculator (for 8x 2TB RAIDZ)

2011-02-06 Thread Matthew Angelo
Yes I did mean 6+2, Thank you for fixing the typo. I'm actually more leaning towards running a simple 7+1 RAIDZ1. Running this with 1TB is not a problem but I just wanted to investigate at what TB size the scales would tip. I understand RAIDZ2 protects against failures during a rebuild process.

Re: [zfs-discuss] RAID Failure Calculator (for 8x 2TB RAIDZ)

2011-02-06 Thread Richard Elling
On Feb 6, 2011, at 6:45 PM, Matthew Angelo wrote: I require a new high capacity 8 disk zpool. The disks I will be purchasing (Samsung or Hitachi) have an Error Rate (non-recoverable, bits read) of 1 in 10^14 and will be 2TB. I'm staying clear of WD because they have the new 2048b sectors