Re: [zfs-discuss] Sun Flash Accelerator F20 numbers

2010-03-31 Thread Neil Perrin
On 03/30/10 20:00, Bob Friesenhahn wrote: On Tue, 30 Mar 2010, Edward Ned Harvey wrote: But the speedup of disabling the ZIL altogether is appealing (and would probably be acceptable in this environment). Just to make sure you know ... if you disable the ZIL altogether, and you have a power

Re: [zfs-discuss] Sun Flash Accelerator F20 numbers

2010-04-03 Thread Neil Perrin
On 04/02/10 08:24, Edward Ned Harvey wrote: The purpose of the ZIL is to act like a fast "log" for synchronous writes. It allows the system to quickly confirm a synchronous write request with the minimum amount of work. Bob and Casper and some others clearly know a lot here. But I'm he

Re: [zfs-discuss] Removing SSDs from pool

2010-04-05 Thread Neil Perrin
On 04/05/10 11:43, Andreas Höschler wrote: Hi Khyron, No, he did *not* say that a mirrored SLOG has no benefit, redundancy-wise. He said that YOU do *not* have a mirrored SLOG. You have 2 SLOG devices which are striped. And if this machine is running Solaris 10, then you cannot remove a lo

Re: [zfs-discuss] Sun Flash Accelerator F20 numbers

2010-04-07 Thread Neil Perrin
On 04/07/10 09:19, Bob Friesenhahn wrote: On Wed, 7 Apr 2010, Robert Milkowski wrote: it is only read at boot if there are uncomitted data on it - during normal reboots zfs won't read data from slog. How does zfs know if there is uncomitted data on the slog device without reading it? The m

Re: [zfs-discuss] Sun Flash Accelerator F20 numbers

2010-04-07 Thread Neil Perrin
On 04/07/10 10:18, Edward Ned Harvey wrote: From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss- boun...@opensolaris.org] On Behalf Of Bob Friesenhahn It is also worth pointing out that in normal operation the slog is essentially a write-only device which is only read at boot time. Th

Re: [zfs-discuss] Sync Write - ZIL log performance - Feedback for ZFS developers?

2010-04-10 Thread Neil Perrin
On 04/10/10 09:28, Edward Ned Harvey wrote: Neil or somebody? Actual ZFS developers? Taking feedback here? ;-) While I was putting my poor little server through cruel and unusual punishment as described in my post a moment ago, I noticed something unexpected: I expected that while

Re: [zfs-discuss] Sync Write - ZIL log performance - Feedback for ZFS developers?

2010-04-10 Thread Neil Perrin
On 04/10/10 14:55, Daniel Carosone wrote: On Sat, Apr 10, 2010 at 11:50:05AM -0500, Bob Friesenhahn wrote: Huge synchronous bulk writes are pretty rare since usually the bottleneck is elsewhere, such as the ethernet. Also, large writes can go straight to the pool, and the zil only logs

Re: [zfs-discuss] creating a fast ZIL device for $200

2010-05-26 Thread Neil Perrin
On 05/26/10 07:10, sensille wrote: Recently, I've been reading through the ZIL/slog discussion and have the impression that a lot of folks here are (like me) interested in getting a viable solution for a cheap, fast and reliable ZIL device. I think I can provide such a solution for about $200, bu

Re: [zfs-discuss] Zpool import not working

2010-06-11 Thread Neil Perrin
On 06/11/10 22:07, zfsnoob4 wrote: Hey, I'm running some test right now before setting up my server. I'm running Nexenta Core 3.02 (RC2, based on opensolaris build 134 I believe) in Virtualbox. To do the test, I'm creating three empty files and then making a raidz mirror: mkfile -n 1g /foo mkf

Re: [zfs-discuss] Zpool import not working

2010-06-12 Thread Neil Perrin
On 06/12/10 17:13, zfsnoob4 wrote: Thanks. As I discovered from that post, VB does not have cache flush enabled by default. Ignoreflush must be explicitly turned off. VBoxManage setextradata VMNAME "VBoxInternal/Devices/piix3ide/0/LUN#[x]/Config/IgnoreFlush" 0 where VMNAME is the name of your

Re: [zfs-discuss] size of slog device

2010-06-14 Thread Neil Perrin
On 06/14/10 12:29, Bob Friesenhahn wrote: On Mon, 14 Jun 2010, Roy Sigurd Karlsbakk wrote: It is good to keep in mind that only small writes go to the dedicated slog. Large writes to to main store. A succession of that many small writes (to fill RAM/2) is highly unlikely. Also, that the zil is

Re: [zfs-discuss] size of slog device

2010-06-14 Thread Neil Perrin
On 06/14/10 19:35, Erik Trimble wrote: On 6/14/2010 12:10 PM, Neil Perrin wrote: On 06/14/10 12:29, Bob Friesenhahn wrote: On Mon, 14 Jun 2010, Roy Sigurd Karlsbakk wrote: It is good to keep in mind that only small writes go to the dedicated slog. Large writes to to main store. A succession

Re: [zfs-discuss] Dedup RAM requirements, vs. L2ARC?

2010-07-01 Thread Neil Perrin
On 07/01/10 22:33, Erik Trimble wrote: On 7/1/2010 9:23 PM, Geoff Nordli wrote: Hi Erik. Are you saying the DDT will automatically look to be stored in an L2ARC device if one exists in the pool, instead of using ARC? Or is there some sort of memory pressure point where the DDT gets moved fr

Re: [zfs-discuss] Dedup RAM requirements, vs. L2ARC?

2010-07-02 Thread Neil Perrin
On 07/02/10 00:57, Erik Trimble wrote: On 7/1/2010 10:17 PM, Neil Perrin wrote: On 07/01/10 22:33, Erik Trimble wrote: On 7/1/2010 9:23 PM, Geoff Nordli wrote: Hi Erik. Are you saying the DDT will automatically look to be stored in an L2ARC device if one exists in the pool, instead of using

Re: [zfs-discuss] Dedup RAM requirements, vs. L2ARC?

2010-07-02 Thread Neil Perrin
On 07/02/10 11:14, Erik Trimble wrote: On 7/2/2010 6:30 AM, Neil Perrin wrote: On 07/02/10 00:57, Erik Trimble wrote: That's what I assumed. One further thought, though. Is the DDT is treated as a single entity - so it's *all* either in the ARC or in the L2ARC? Or does it move on

Re: [zfs-discuss] Debunking the dedup memory myth

2010-07-09 Thread Neil Perrin
On 07/09/10 19:40, Erik Trimble wrote: On 7/9/2010 5:18 PM, Brandon High wrote: On Fri, Jul 9, 2010 at 5:00 PM, Edward Ned Harvey mailto:solar...@nedharvey.com>> wrote: The default ZFS block size is 128K. If you have a filesystem with 128G used, that means you are consuming 1,048,576

Re: [zfs-discuss] Does ZFS work with SAN-attached devices?

2009-10-13 Thread Neil Perrin
Also, ZFS does things like putting the ZIL data (when not on a dedicated device) at the outer edge of disks, that being faster. No, ZFS does not do that. It will chain the intent log from blocks allocated from the same metaslabs that the pool is allocating from. This actually works out well be

Re: [zfs-discuss] ZFS and NFS

2009-11-20 Thread Neil Perrin
On 11/18/09 12:21, Joe Cicardo wrote: Hi, My customer says: Application has NFS directories with millions of files in a directory, and this can't changed. We are having issues with the EMC appliance and RPC timeouts on the NFS lookup. I am looking doing

Re: [zfs-discuss] Is write(2) made durable atomically?

2009-11-30 Thread Neil Perrin
Under the hood in ZFS, writes are committed using either shadow paging or logging, as I understand it. So I believe that I mean to ask whether a write(2), pushed to ZPL, and pushed on down the stack, can be split into multiple transactions? Or, instead, is it guaranteed to be committed in a si

Re: [zfs-discuss] Separate Zil on HDD ?

2009-12-03 Thread Neil Perrin
On 12/03/09 09:21, mbr wrote: Hello, Bob Friesenhahn wrote: On Thu, 3 Dec 2009, mbr wrote: What about the data that were on the ZILlog SSD at the time of failure, is a copy of the data still in the machines memory from where it can be used to put the transaction to the stable storage poo

Re: [zfs-discuss] zfs on ssd

2009-12-05 Thread Neil Perrin
On 12/05/09 01:36, anu...@kqinfotech.com wrote: Hi, What you say is probably right with respect to L2ARC, but logging (ZIL or database log) is required for consistency purpose. No, the ZIL is not required for consistency. The pool is fully consistent without the ZIL. See http://blogs.sun.

Re: [zfs-discuss] [zfs-code] Transaction consistency of ZFS

2009-12-06 Thread Neil Perrin
On 12/06/09 10:11, Anurag Agarwal wrote: Hi, My reading of write code of ZFS (zfs_write in zfs_vnops.c), is that all the writes in zfs are logged in the ZIL. Each write gets recorded in memory in case it needs to be forced out later (eg fsync()), but is not written to the on-disk log until

Re: [zfs-discuss] Transaction consistency of ZFS

2009-12-06 Thread Neil Perrin
I'll try to find out whether ZFS binding the same file always to the same opening transaction group. Not sure what you mean by this. Transactions (eg writes) will go into the current open transaction group (txg). Subsequent writes may enter the same or a future txg. Txgs are obviously committe

Re: [zfs-discuss] Planed ZFS-Features - Is there a List or something else

2009-12-09 Thread Neil Perrin
On 12/09/09 13:52, Glenn Lagasse wrote: * R.G. Keen (k...@geofex.com) wrote: I didn't see "remove a simple device" anywhere in there. Is it: too hard to even contemplate doing, or too silly a thing to do to even consider letting that happen or too stupid a question to even consider or too

Re: [zfs-discuss] zfs on ssd

2009-12-11 Thread Neil Perrin
On 12/11/09 14:56, Bill Sommerfeld wrote: On Fri, 2009-12-11 at 13:49 -0500, Miles Nordin wrote: "sh" == Seth Heeren writes: sh> If you don't want/need log or cache, disable these? You might sh> want to run your ZIL (slog) on ramdisk. seems quite silly. why would you do that instea

Re: [zfs-discuss] New ZFS Intent Log (ZIL) device available - Beta program now open!

2010-01-13 Thread Neil Perrin
Hi Adam, So was FW aware of this or in contact with these guys? Also are you requesting/ordering any of these cards to evaluate? The device seems kind of small at 4GB, and uses a double wide PCI Express slot. Neil. On 01/13/10 12:27, Adam Leventhal wrote: Hey Chris, The DDRdrive X1 OpenSola

Re: [zfs-discuss] ZIL to disk

2010-01-15 Thread Neil Perrin
On 01/15/10 12:59, Jeffry Molanus wrote: Sometimes people get confused about the ZIL and separate logs. For sizing purposes, the ZIL is a write-only workload. Data which is written to the ZIL is later asynchronously written to the pool when the txg is committed. Right; the tgx needs time t

Re: [zfs-discuss] Intrusion Detection - powered by ZFS Checksumming ?

2010-02-09 Thread Neil Perrin
On 02/09/10 08:18, Kjetil Torgrim Homme wrote: Richard Elling writes: On Feb 8, 2010, at 9:10 PM, Damon Atkins wrote: I would have thought that if I write 1k then ZFS txg times out in 30secs, then the 1k will be written to disk in a 1k record block, and then if I write 4k then 30secs latte

Re: [zfs-discuss] ZFS performance benchmarks in various configurations

2010-02-19 Thread Neil Perrin
If I understand correctly, ZFS now adays will only flush data to non volatile storage (such as a RAID controller NVRAM), and not all the way out to disks. (To solve performance problems with some storage systems, and I believe that it also is the right thing to do under normal circumstances.) D

Re: [zfs-discuss] Log disk with all ssd pool?

2011-10-27 Thread Neil Perrin
On 10/28/11 00:04, Mark Wolek wrote: Still kicking around this idea and didn’t see it addressed in any of the threads before the forum closed.    If one made an all ssd pool, would a log/cache drive just slow you down?  Would zil slow you down?  Thinking rotate MLC drives w

Re: [zfs-discuss] Log disk with all ssd pool?

2011-10-28 Thread Neil Perrin
On 10/28/11 00:54, Neil Perrin wrote: On 10/28/11 00:04, Mark Wolek wrote: Still kicking around this idea and didn’t see it addressed in any of the threads before the forum closed.    If one made an all ssd pool, would a log/cache drive just slow you

Re: [zfs-discuss] Log disk with all ssd pool?

2011-10-28 Thread Neil Perrin
Of Neil Perrin Sent: Friday, October 28, 2011 11:38 AM To: zfs-discuss@opensolaris.org Subject: Re: [zfs-discuss] Log disk with all ssd pool?   On 10/28/11 00:54, Neil Perrin wrote: On 10/28/11 00:04, Mark Wolek wrote: Still kicking around this idea and didn’t see it addressed

Re: [zfs-discuss] what have you been buying for slog and l2arc?

2012-08-03 Thread Neil Perrin
On 08/03/12 19:39, Bob Friesenhahn wrote: On Fri, 3 Aug 2012, Karl Rossing wrote: I'm looking at http://www.intel.com/content/www/us/en/solid-state-drives/solid-state-drives-ssd.html wondering what I should get. Are people getting intel 330's for l2arc and 520's for slog? For the slog, you

Re: [zfs-discuss] Making ZIL faster

2012-10-04 Thread Neil Perrin
On 10/04/12 05:30, Schweiss, Chip wrote: Thanks for all the input.  It seems information on the performance of the ZIL is sparse and scattered.   I've spent significant time researching this the past day.  I'll summarize what I've found.   Please correct me if I'm w

Re: [zfs-discuss] Making ZIL faster

2012-10-04 Thread Neil Perrin
On 10/04/12 15:59, Edward Ned Harvey (opensolarisisdeadlongliveopensolaris) wrote: From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss- boun...@opensolaris.org] On Behalf Of Neil Perrin The ZIL code chains blocks together and these are allocated round robin among slogs or if they

Re: [zfs-discuss] SSD ZIL/L2ARC partitioning

2012-11-14 Thread Neil Perrin
On 11/14/12 03:24, Sašo Kiselkov wrote: On 11/14/2012 11:14 AM, Michel Jansens wrote: Hi, I've ordered a new server with: - 4x600GB Toshiba 10K SAS2 Disks - 2x100GB OCZ DENEVA 2R SYNC eMLC SATA (no expander so I hope no SAS/SATA problems). Specs: http://www.oczenterprise.com/ssd-products/deneva

Re: [zfs-discuss] ZFS offline ZIL corruption not detected

2010-08-23 Thread Neil Perrin
This is a consequence of the design for performance of the ZIL code. Intent log blocks are dynamically allocated and chained together. When reading the intent log we read each block and checksum it with the embedded checksum within the same block. If we can't read a block due to an IO error then t

Re: [zfs-discuss] ZFS offline ZIL corruption not detected

2010-08-23 Thread Neil Perrin
replay the following entries in case the log records in the missing log block were important (eg create file). Mirroring the slogs is recommended to minimise concerns about slogs corruption. Â Regards, Markus Neil Perrin hat am 23. August 2010 um 19:44 geschrieben: This is a consequence of

Re: [zfs-discuss] ZFS offline ZIL corruption not detected

2010-08-25 Thread Neil Perrin
On 08/25/10 20:33, Edward Ned Harvey wrote: From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss- boun...@opensolaris.org] On Behalf Of Neil Perrin This is a consequence of the design for performance of the ZIL code. Intent log blocks are dynamically allocated and chained together

Re: [zfs-discuss] NFS performance near zero on a very full pool

2010-09-09 Thread Neil Perrin
Arne, NFS often demands it's transactions are stable before returning. This forces ZFS to do the system call synchronously. Usually the ZIL (code) allocates and writes a new block in the intent log chain to achieve this. If ever it fails to allocate a block (of the size requested) it it forced

Re: [zfs-discuss] NFS performance near zero on a very full pool

2010-09-09 Thread Neil Perrin
I should also have mentioned that if the pool has a separate log device then this shouldn't happen.Assuming the slog is big enough then it it should have enough blocks to not be forced into using main pool device blocks. Neil. On 09/09/10 10:36, Neil Perrin wrote: Arne, NFS often de

Re: [zfs-discuss] Best practice for Sol10U9 ZIL -- mirrored or not?

2010-09-17 Thread Neil Perrin
On 09/17/10 06:24, Edward Ned Harvey wrote: From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss- boun...@opensolaris.org] On Behalf Of Bryan Horstmann-Allen The ability to remove the slogs isn't really the win here, it's import -F. The Disagree. Although I agree the -F is impor

Re: [zfs-discuss] Best practice for Sol10U9 ZIL -- mirrored or not?

2010-09-17 Thread Neil Perrin
On 09/17/10 18:32, Edward Ned Harvey wrote: From: Neil Perrin [mailto:neil.per...@oracle.com] you lose information. Not your whole pool. You lose up to 30 sec of writes The default is now 5 seconds (zfs_txg_timeout). When did that become default? It was changed more

Re: [zfs-discuss] Best practice for Sol10U9 ZIL -- mirrored or not?

2010-09-17 Thread Neil Perrin
On 09/17/10 23:31, Ian Collins wrote: On 09/18/10 04:46 PM, Neil Perrin wrote: On 09/17/10 18:32, Edward Ned Harvey wrote: From: Neil Perrin [mailto:neil.per...@oracle.com] you lose information. Not your whole pool. You lose up to 30 sec of writes The default is now 5

Re: [zfs-discuss] ZFS COW and simultaneous read & write of files

2010-09-22 Thread Neil Perrin
On 09/22/10 11:22, Moazam Raja wrote: Hi all, I have a ZFS question related to COW and scope. If user A is reading a file while user B is writing to the same file, when do the changes introduced by user B become visible to everyone? Is there a block level scope, or file level, or something else?

Re: [zfs-discuss] What is l2cache setting?

2010-09-22 Thread Neil Perrin
On 09/22/10 11:23, Peter Taps wrote: Folks, While going through zpool source code, I see a configuration option called l2cache. What is this option for? It doesn't seem to be documented. Thank you in advance for your help. Regards, Peter man zpool under "Cache Devices" section ___

Re: [zfs-discuss] What is l2cache setting?

2010-09-22 Thread Neil Perrin
On 09/22/10 13:40, Peter Taps wrote: Neil, Thank you for your help. However, I don't see anything about l2cache under "Cache devices" man pages. To be clear, there are two different vdev types defined in zfs source code - "cache" and "l2cache." I am familiar with "cache" devices. I am curious

Re: [zfs-discuss] What is "dedupditto" property on zpool?

2010-09-24 Thread Neil Perrin
On 09/24/10 11:26, Peter Taps wrote: Folks, One of the zpool properties that is reported is "dedupditto." However, there is no documentation available, either in man pages or anywhere else on the Internet. What exactly is this property? Thank you in advance for your help. Regards, Peter

Re: [zfs-discuss] How does dedup work over iSCSI?

2010-10-22 Thread Neil Perrin
On 10/22/10 15:34, Peter Taps wrote: Folks, Let's say I have a volume being shared over iSCSI. The dedup has been turned on. Let's say I copy the same file twice under different names at the initiator end. Let's say each file ends up taking 5 blocks. For dedupe to work, each block for a file

Re: [zfs-discuss] How does dedup work over iSCSI?

2010-10-22 Thread Neil Perrin
On 10/22/10 17:28, Peter Taps wrote: Hi Neil, if the file offset does not match, the chances that the checksum would match, especially sha256, is almost 0. May be I am missing something. Let's say I have a file that contains 11 letters - ABCDEFGHIJK. Let's say the block size is 5. For the fi

Re: [zfs-discuss] ashift and vdevs

2010-12-01 Thread Neil Perrin
On 12/01/10 22:14, Miles Nordin wrote: Also did anyone ever clarify whether the slog has an ashift? or is it forced-512? or derived from whatever vdev will eventually contain the separately-logged data? I would expect generalized immediate Caring about that since no slogs except ACARD and DDRD

Re: [zfs-discuss] BOOT, ZIL, L2ARC one one SSD?

2011-01-04 Thread Neil Perrin
On 12/25/10 19:32, Bill Werner wrote: Understood Edward, and if this was a production data center, I wouldn't be doing it this way. This is for my home lab, so spending hundreds of dollars on SSD devices isn't practical. Can several datasets share a single ZIL and a single L2ARC, or much must

Re: [zfs-discuss] Cannot remove zil device

2011-03-31 Thread Neil Perrin
On 03/31/11 12:28, Roy Sigurd Karlsbakk wrote: http://pastebin.com/nD2r2qmh Here is zpool status and zpool version The only thing I wonder about here, is why you have two striped log devices. I didn't even know that was supported. Yes it's supported. ZFS will round robin writes to

Re: [zfs-discuss] Dedup and L2ARC memory requirements (again)

2011-04-25 Thread Neil Perrin
On 04/25/11 11:55, Erik Trimble wrote: On 4/25/2011 8:20 AM, Edward Ned Harvey wrote: And one more comment: Based on what's below, it seems that the DDT gets stored on the cache device and also in RAM. Is that correct? What if you didn't have a cache device? Shouldn't it *always* be in r

Re: [zfs-discuss] Dedup and L2ARC memory requirements (again)

2011-04-28 Thread Neil Perrin
On 4/28/11 12:45 PM, Edward Ned Harvey wrote: From: Erik Trimble [mailto:erik.trim...@oracle.com] OK, I just re-looked at a couple of things, and here's what I /think/ is the correct numbers. I just checked, and the current size of this structure is 0x178, or 376 bytes. Each ARC entry, which p

Re: [zfs-discuss] Dedup and L2ARC memory requirements (again)

2011-04-30 Thread Neil Perrin
On 04/30/11 01:41, Sean Sprague wrote: : xvm-4200m2-02 ; I can do the echo | mdb -k. But what is that : xvm-4200 command? My guess is that is a very odd shell prompt ;-) - Indeed ':' means what follows a comment (at least to /bin/ksh) 'xvm-4200m2-02' is the comment - actua

Re: [zfs-discuss] ls reports incorrect file size

2011-05-02 Thread Neil Perrin
On 05/02/11 14:02, Nico Williams wrote: Also, sparseness need not be apparent to applications. Until recent improvements to lseek(2) to expose hole/non-hole offsets, the only way to know about sparseness was to notice that a file's reported size is more than the file's reported filesystem blocks

Re: [zfs-discuss] Summary: Dedup and L2ARC memory requirements

2011-05-08 Thread Neil Perrin
On 05/08/11 09:22, Andrew Gabriel wrote: Toby Thain wrote: On 08/05/11 10:31 AM, Edward Ned Harvey wrote: ... Incidentally, does fsync() and sync return instantly or wait? Cuz "time sync" might product 0 sec every time even if there were something waiting to be flushed to disk. The

Re: [zfs-discuss] write cache partial-disk pools (was Server with 4 drives, how to configure ZFS?)

2011-06-16 Thread Neil Perrin
On 06/16/11 20:26, Daniel Carosone wrote: On Thu, Jun 16, 2011 at 09:15:44PM -0400, Edward Ned Harvey wrote: My personal preference, assuming 4 disks, since the OS is mostly reads and only a little bit of writes, is to create a 4-way mirrored 100G partition for the OS, and the remaining 900G

Re: [zfs-discuss] ZFS Fragmentation issue - examining the ZIL

2011-08-01 Thread Neil Perrin
In general the blogs conclusion is correct . When file systems get full there is fragmentation (happens to all file systems) and for ZFS the pool uses gang blocks of smaller blocks when there are insufficient large blocks. However, the ZIL never allocates or uses gang blocks. It directly allocate

Re: [zfs-discuss] Advice with SSD, ZIL and L2ARC

2011-08-30 Thread Neil Perrin
On 08/30/11 08:31, Edward Ned Harvey wrote: From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss- boun...@opensolaris.org] On Behalf Of Jesus Cea 10. What happens if my 1GB of ZIL is too optimistic?. Will ZFS use the disks or it will stop writers until flushing ZIL to the HDs?.

Re: [zfs-discuss] Advice with SSD, ZIL and L2ARC

2011-09-19 Thread Neil Perrin
On 9/19/11 11:45 AM, Jesus Cea wrote: -BEGIN PGP SIGNED MESSAGE- Hash: SHA1 I have a new answer: interaction between dataset encryption and L2ARC and ZIL. 1. I am pretty sure (but not completely sure) that data stored in the ZIL is encrypted, if the destination dataset uses encryption.

Re: [zfs-discuss] Supermicro AOC-SAT2-MV8 hang when drive removed

2008-07-30 Thread Neil Perrin
Peter Cudhea wrote: > Your point is well taken that ZFS should not duplicate functionality > that is already or should be available at the device driver level.In > this case, I think it misses the point of what ZFS should be doing that > it is not. > > ZFS does its own periodic commits to

Re: [zfs-discuss] Disabling disks' write-cache in J4200 with ZFS?

2008-08-02 Thread Neil Perrin
Carson Gaspar wrote: > Todd E. Moore wrote: >> I'm working with a group that wants to commit all the way to disk every >> single write - flushing or bypassing all the caches each time. The >> fsync() call will flush the ZIL. As for the disk's cache, if given the >> entire disk, ZFS enables its ca

Re: [zfs-discuss] Zpool import not working - I broke my pool...

2008-08-04 Thread Neil Perrin
Richard Elling wrote: > Ross wrote: >> I'm trying to import a pool I just exported but I can't, even -f doesn't >> help. Every time I try I'm getting an error: >> "cannot import 'rc-pool': one or more devices is currently unavailable" >> >> Now I suspect the reason it's not happy is that the p

Re: [zfs-discuss] Zpool import not working - I broke my pool...

2008-08-06 Thread Neil Perrin
Ross, Thanks, I have updated the bug with this info. Neil. Ross Smith wrote: > Hmm... got a bit more information for you to add to that bug I think. > > Zpool import also doesn't work if you have mirrored log devices and > either one of them is offline. > > I created two ramdisks with: > #

Re: [zfs-discuss] zfs crash CR6727355 marked incomplete

2008-08-06 Thread Neil Perrin
Michael Hale wrote: > A bug report I've submitted for a zfs-related kernel crash has been > marked incomplete and I've been asked to provide more information. > > This CR has been marked as "incomplete" by > for the reason "Need More Info". Please update the CR > providing the information re

Re: [zfs-discuss] ZIL & NVRAM partitioning?

2008-09-05 Thread Neil Perrin
On 09/05/08 14:42, Narayan Venkat wrote: > I understand that if you want to use ZIL, then the requirement is one or more > ZILs per pool. A little clarification of ZFS terms may help here. The term ZIL is somewhat overloaded. I think what you mean here is a separate log device (slog), because in

Re: [zfs-discuss] zfs cp hangs when the mirrors are removed ..

2008-10-15 Thread Neil Perrin
Karthik, The pool failmode property as implemented governs the behaviour when all the devices needed are unavailable. The default behaviour is to wait (block) until the IO can continue - perhaps by re-enabling the device(s). The behaviour you expected can be achieved by "zpool set failmode=continu

Re: [zfs-discuss] zfs cp hangs when the mirrors are removed ..

2008-10-15 Thread Neil Perrin
vices. This should prevent the hang. It isn't used to recover from it. If you did do that then it seems like a bug somewhere in ZFS or the IO stack below it. In which case you should file a bug. Neil. > > thanks and regards, > Karthik > > On 10/15/08 22:03, Neil Perrin wrot

Re: [zfs-discuss] Disabling COMMIT at NFS level, or disabling ZIL on a per-filesystem basis

2008-10-22 Thread Neil Perrin
On 10/22/08 10:26, Constantin Gonzalez wrote: > Hi, > > On a busy NFS server, performance tends to be very modest for large amounts > of small files due to the well known effects of ZFS and ZIL honoring the > NFS COMMIT operation[1]. > > For the mature sysadmin who knows what (s)he does, there

Re: [zfs-discuss] Disabling COMMIT at NFS level, or disabling ZIL on a per-filesystem basis

2008-10-22 Thread Neil Perrin
> But the slog is the ZIL. formaly a *separate* intent log. No the slog is not the ZIL! Here's the definition of the terms as we've been trying to use them: ZIL: The body of code the supports synchronous requests, which writes out to the Intent Logs Intent Log: A stable s

Re: [zfs-discuss] Disabling COMMIT at NFS level, or disabling ZIL on a per-filesystem basis

2008-10-22 Thread Neil Perrin
On 10/22/08 13:56, Marcelo Leal wrote: >>> But the slog is the ZIL. formaly a *separate* >> intent log. >> >> No the slog is not the ZIL! > Ok, when you did write this: > "I've been slogging for a while on support for separate intent logs (slogs) > for ZFS. > Without slogs, the ZIL is allocated

Re: [zfs-discuss] Cannot remove slog device from zpool

2008-10-28 Thread Neil Perrin
Ethan, It is still not possible to remove a slog from a pool. This is bug: 6574286 removing a slog doesn't work The error message: "cannot remove c4t15d0p0: only inactive hot spares or cache devices can be removed" is correct and this is the same as documented in the zpool man page: zpo

Re: [zfs-discuss] DNLC and ARC

2008-10-30 Thread Neil Perrin
Leal, ZFS uses the DNLC. It still provides the fastest lookup of to vnode. The DNLC is kind of LRU. An async process will use a rotor to move through the hash chains and select the LRU entry but will select first negative cache entries and vnodes only referenced by the DNLC. Underlying this ZFS

Re: [zfs-discuss] DNLC and ARC

2008-10-30 Thread Neil Perrin
On 10/30/08 11:00, Marcelo Leal wrote: > Hello Neil, > >> Leal, >> >> ZFS uses the DNLC. It still provides the fastest >> lookup of to vnode. > > Ok, so the whole concept remains true? We can tune the DNLC and expect the > same behaviour on ZFS? Yes. > >> The DNLC is kind of LRU. An async

Re: [zfs-discuss] s10u6--will using disk slices for zfs logs improve nfs performance?

2008-11-13 Thread Neil Perrin
I wouldn't expect any improvement using a separate disk slice for the Intent Log unless that disk was much faster and was otherwise largely idle. If it was heavily used then I'd expect quite the performance degradation as the disk head bounces around between slices. Separate intent logs are really

Re: [zfs-discuss] ZIL performance on "traditional" HDDs

2008-11-20 Thread Neil Perrin
On 11/20/08 12:52, Danilo Poccia wrote: > Hi, > > I was wondering is there is a performance gain for an OLTP-like workload > in putting the ZFS Intent Log (ZIL) on "traditional" HDDs. > It's probably always best to benchmark it yourself, but my experience has shown that it's better to only ha

Re: [zfs-discuss] zfs znode changes getting lost

2008-11-26 Thread Neil Perrin
I suspect ZFS is unaware that anything has changed in the z_phys so it never gets written out. You probably need to create a dmu transaction and call dmu_buf_will_dirty(zp->z_dbuf, tx); Neil. On 11/26/08 03:36, shelly wrote: > In place of padding in zfs znode i added a new field. stored an intege

Re: [zfs-discuss] zfs_nocacheflush, nvram, and root pools

2008-12-03 Thread Neil Perrin
On 12/02/08 03:47, River Tarnell wrote: > hi, > > i have a system connected to an external DAS (SCSI) array, using ZFS. the > array has an nvram write cache, but it honours SCSI cache flush commands by > flushing the nvram to disk. the array has no way to disable this behaviour. > a > well-kno

Re: [zfs-discuss] What will happen when write a block of 8k if the recordsize is 128k. Will 128k be written instead of 8k?

2008-12-24 Thread Neil Perrin
The default recordsize is 128K. So you are correct, for random reads performance will be bad as excess data is read. For Oracle it is recommended to set the recordsize to 8k. This can be done when creating the filesystem using 'zfs create -o recordsize=8k '. If the fs has already been created th

Re: [zfs-discuss] Problems at 90% zpool capacity 2008.05

2009-01-06 Thread Neil Perrin
On 01/06/09 21:25, Nicholas Lee wrote: > Since zfs is so smart is other areas is there a particular reason why a > high water mark is not calculated and the available space not reset to this? > > I'd far rather have a zpool of 1000GB that said it only had 900GB but > did not have corruption as

Re: [zfs-discuss] Several questions concerning ZFS

2009-01-12 Thread Neil Perrin
On 01/12/09 20:45, Simon wrote: > Hi Experts, > > IHAC who using Solaris 10 + ZFS,two questions they're concerned: > > - ZIL(zfs intent log) is enabled by default for ZFS,there are varied > storage purchased by customer(such as EMC CX/DMX series,HDS AMS/USP > series and etc),customer wonder whe

Re: [zfs-discuss] zfs panic

2009-01-12 Thread Neil Perrin
This is a known bug: 6678070 Panic from vdev_mirror_map_alloc() http://bugs.opensolaris.org/view_bug.do?bug_id=6678070 Neil. On 01/12/09 21:12, Krzys wrote: > any idea what could cause my system to panic? I get my system rebooted daily > at > various times. very strange, but its pointing to zf

Re: [zfs-discuss] zfs panic

2009-01-13 Thread Neil Perrin
and I do not > care about it to that extend but if I had other system that has much greater > importance and I get such situation its quite scary... :( > > On Mon, 12 Jan 2009, Neil Perrin wrote: > >> This is a known bug: >> >> 6678070 Panic from vdev_mirror_ma

Re: [zfs-discuss] Lackluster ZFS performance trials using various ZIL and L2ARC configurations...

2009-01-16 Thread Neil Perrin
I don't believe that iozone does any synchronous calls (fsync/O_DSYNC/O_SYNC), so the ZIL and separate logs (slogs) would be unused. I'd recommend performance testing by configuring filebench to do synchronous writes: http://opensolaris.org/os/community/performance/filebench/ Neil. On 01/15/09

Re: [zfs-discuss] write cache and cache flush

2009-01-29 Thread Neil Perrin
On 01/29/09 21:32, Greg Mason wrote: > This problem only manifests itself when dealing with many small files > over NFS. There is no throughput problem with the network. > > I've run tests with the write cache disabled on all disks, and the cache > flush disabled. I'm using two Intel SSDs for

Re: [zfs-discuss] ZFS core contributor nominations

2009-02-02 Thread Neil Perrin
Looks reasonable +1 Neil. On 02/02/09 08:55, Mark Shellenbaum wrote: > The time has come to review the current Contributor and Core contributor > grants for ZFS. Since all of the ZFS core contributors grants are set > to expire on 02-24-2009 we need to renew the members that are still > contr

Re: [zfs-discuss] Max size of log device?

2009-02-08 Thread Neil Perrin
On 02/08/09 11:50, Vincent Fox wrote: > So I have read in the ZFS Wiki: > > # The minimum size of a log device is the same as the minimum size of device > in > pool, which is 64 Mbytes. The amount of in-play data that might be stored on > a log > device is relatively small. Log blocks are fre

Re: [zfs-discuss] Solaris and zfs versions

2009-02-12 Thread Neil Perrin
Mark, I believe creating a older version pool is supported: zpool create -o version= whirl c0t0d0 I'm not sure what version of ZFS in Solaris 10 you are running. Try running "zpool upgrade" and replacing above with that version number. Neil. : trasimene ; zpool create -o version=11 whirl c0t

Re: [zfs-discuss] ZFS: unreliable for professional usage?

2009-02-13 Thread Neil Perrin
Having a separate intent log on good hardware will not prevent corruption on a pool with bad hardware. By "good" I mean hardware that correctly flush their write caches when requested. Note, a pool is always consistent (again when using good hardware). The function of the intent log is not to pro

Re: [zfs-discuss] ZFS and SNDR..., now I'm confused.

2009-03-06 Thread Neil Perrin
I'd like to correct a few misconceptions about the ZIL here. On 03/06/09 06:01, Jim Dunham wrote: ZFS the filesystem is always on disk consistent, and ZFS does maintain filesystem consistency through coordination between the ZPL (ZFS POSIX Layer) and the ZIL (ZFS Intent Log). Pool and file sy

Re: [zfs-discuss] ZFS and SNDR..., now I'm confused.

2009-03-06 Thread Neil Perrin
On 03/06/09 08:10, Jim Dunham wrote: Andrew, Jim Dunham wrote: ZFS the filesystem is always on disk consistent, and ZFS does maintain filesystem consistency through coordination between the ZPL (ZFS POSIX Layer) and the ZIL (ZFS Intent Log). Unfortunately for SNDR, ZFS caches a lot of an a

Re: [zfs-discuss] ZFS and SNDR..., now I'm confused.

2009-03-06 Thread Neil Perrin
On 03/06/09 14:51, Miles Nordin wrote: "np" == Neil Perrin writes: np> Alternatively, a lockfs will flush just a file system to np> stable storage but in this case just the intent log is np> written. (Then later when the txg commits those intent log np>

Re: [zfs-discuss] ZIL SSD performance testing... -IOzone works great, others not so great

2009-04-09 Thread Neil Perrin
Patrick, The ZIL is only used for synchronous requests like O_DSYNC/O_SYNC and fsync(). Your iozone command must be doing some synchronous writes. All the other tests (dd, cat, cp, ...) do everything asynchronously. That is they do not require the data to be on stable storage on return from the w

Re: [zfs-discuss] ZIL SSD performance testing... -IOzone works great, others not so great

2009-04-10 Thread Neil Perrin
On 04/10/09 20:15, Toby Thain wrote: On 10-Apr-09, at 5:05 PM, Mark J Musante wrote: On Fri, 10 Apr 2009, Patrick Skerrett wrote: degradation) when these write bursts come in, and if I could buffer them even for 60 seconds, it would make everything much smoother. ZFS already batches up w

Re: [zfs-discuss] Degraded log device in "zpool status" output

2009-04-18 Thread Neil Perrin
Will, This is bug: 6710376 log device can show incorrect status when other parts of pool are degraded This is just an error in the reporting. There was nothing actually wrong with the log device. It is picking up the degraded status from the rest of the pool. The bug was fixed only yesterday a

Re: [zfs-discuss] zpool iostat and iostat discrepancy

2009-06-20 Thread Neil Perrin
On 06/20/09 11:14, tester wrote: Hi, Does anyone know the difference between zpool iostat and iostat? dd if=/dev/zero of=/test/test1/trash count=1 bs=1024k;sync pool only shows 236K IO and 13 write ops. whereas iostat shows a correctly meg of activity. The zfs numbers are per second as we

Re: [zfs-discuss] How Virtual Box handles the IO

2009-07-31 Thread Neil Perrin
I understand > that the ZILs are allocated out of the general pool. There is one intent log chain per dataset (file system or zvol). The head of each log the log is kept in the main pool. Without slog(s) we allocate (and chain) blocks from the main pool. If separate intent log(s) exist then bloc

Re: [zfs-discuss] zfs fragmentation

2009-08-07 Thread Neil Perrin
On 08/07/09 10:54, Scott Meilicke wrote: ZFS absolutely observes synchronous write requests (e.g. by NFS or a database). The synchronous write requests do not benefit from the long write aggregation delay so the result may not be written as ideally as ordinary write requests. Recently zfs has

  1   2   3   >