[zfs-discuss] Pools inside pools

2010-09-22 Thread Markus Kovero
Hi, I'm asking for opinions here, any possible disaster happening or performance issues related in setup described below. Point being to create large pool and smaller pools within where you can monitor easily iops and bandwidth usage without using dtrace or similar techniques. 1. Create pool #

Re: [zfs-discuss] Has anyone seen zpool corruption with VirtualBox shared folders?

2010-09-22 Thread Orvar Korvar
Now this is a testament to the power of ZFS. Only ZFS is so sensitive it observed these errors to you. Had you run another filesystem, you would never got a notice that your data is slowly being corrupted by some faulty hardware. :o) -- This message posted from opensolaris.org

Re: [zfs-discuss] Please warn a home user against OpenSolaris under VirtualBox under WinXP ; )

2010-09-22 Thread Orvar Korvar
There was a guy doing that: Windows as host and OpenSolaris as guest with raw access to his disks. He lost his 12 TB data. It turned out that VirtualBox dont honor the write flush flag (or something similar). In other words, I would never ever do that. Your data is safer with Windows only and

Re: [zfs-discuss] space_map again nuked!!

2010-09-22 Thread Orvar Korvar
Did you see this thread? http://opensolaris.org/jive/thread.jspa?messageID=500659#500659 He had problems with ZFS. It turned out to be faulty RAM. ZFS is so sensitive it detects and reports problems to you. No other filesystem does that, so you think ZFS is problematic and switch. But the other

Re: [zfs-discuss] Please warn a home user against OpenSolaris under VirtualBox under WinXP ; )

2010-09-22 Thread Nicolas Williams
On Wed, Sep 22, 2010 at 07:14:43AM -0700, Orvar Korvar wrote: There was a guy doing that: Windows as host and OpenSolaris as guest with raw access to his disks. He lost his 12 TB data. It turned out that VirtualBox dont honor the write flush flag (or something similar). VirtualBox has an

Re: [zfs-discuss] Pools inside pools

2010-09-22 Thread Pawel Jakub Dawidek
On Wed, Sep 22, 2010 at 02:06:27PM +, Markus Kovero wrote: Hi, I'm asking for opinions here, any possible disaster happening or performance issues related in setup described below. Point being to create large pool and smaller pools within where you can monitor easily iops and bandwidth

[zfs-discuss] ZFS COW and simultaneous read write of files

2010-09-22 Thread Moazam Raja
Hi all, I have a ZFS question related to COW and scope. If user A is reading a file while user B is writing to the same file, when do the changes introduced by user B become visible to everyone? Is there a block level scope, or file level, or something else? Thanks!

[zfs-discuss] What is l2cache setting?

2010-09-22 Thread Peter Taps
Folks, While going through zpool source code, I see a configuration option called l2cache. What is this option for? It doesn't seem to be documented. Thank you in advance for your help. Regards, Peter -- This message posted from opensolaris.org ___

Re: [zfs-discuss] Pools inside pools

2010-09-22 Thread Markus Kovero
Such configuration was known to cause deadlocks. Even if it works now (which I don't expect to be the case) it will make your data to be cached twice. The CPU utilization will also be much higher, etc. All in all I strongly recommend against such setup. -- Pawel Jakub Dawidek

Re: [zfs-discuss] ZFS COW and simultaneous read write of files

2010-09-22 Thread Neil Perrin
On 09/22/10 11:22, Moazam Raja wrote: Hi all, I have a ZFS question related to COW and scope. If user A is reading a file while user B is writing to the same file, when do the changes introduced by user B become visible to everyone? Is there a block level scope, or file level, or something

Re: [zfs-discuss] What is l2cache setting?

2010-09-22 Thread Neil Perrin
On 09/22/10 11:23, Peter Taps wrote: Folks, While going through zpool source code, I see a configuration option called l2cache. What is this option for? It doesn't seem to be documented. Thank you in advance for your help. Regards, Peter man zpool under Cache Devices section

Re: [zfs-discuss] Pools inside pools

2010-09-22 Thread Erik Trimble
On 9/22/2010 11:15 AM, Markus Kovero wrote: Such configuration was known to cause deadlocks. Even if it works now (which I don't expect to be the case) it will make your data to be cached twice. The CPU utilization will also be much higher, etc. All in all I strongly recommend against such

Re: [zfs-discuss] ZFS COW and simultaneous read write of files

2010-09-22 Thread Nicolas Williams
On Wed, Sep 22, 2010 at 12:30:58PM -0600, Neil Perrin wrote: On 09/22/10 11:22, Moazam Raja wrote: Hi all, I have a ZFS question related to COW and scope. If user A is reading a file while user B is writing to the same file, when do the changes introduced by user B become visible to

Re: [zfs-discuss] Pools inside pools

2010-09-22 Thread Markus Kovero
Actually, the mechanics of local pools inside pools is significantly different than using remote volumes (potentially exported ZFS volumes) to build a local pool from. I don't see how, I'm referring to method where hostA shares local iscsi volume to hostB where volume is being mirrored

Re: [zfs-discuss] What is l2cache setting?

2010-09-22 Thread Peter Taps
Neil, Thank you for your help. However, I don't see anything about l2cache under Cache devices man pages. To be clear, there are two different vdev types defined in zfs source code - cache and l2cache. I am familiar with cache devices. I am curious about l2cache devices. Regards, Peter --

Re: [zfs-discuss] Pools inside pools

2010-09-22 Thread Mattias Pantzare
On Wed, Sep 22, 2010 at 20:15, Markus Kovero markus.kov...@nebula.fi wrote: Such configuration was known to cause deadlocks. Even if it works now (which I don't expect to be the case) it will make your data to be cached twice. The CPU utilization will also be much higher, etc. All in all

Re: [zfs-discuss] What is l2cache setting?

2010-09-22 Thread Cindy Swearingen
On 9/22/10 1:40 PM, Peter Taps wrote: Neil, Thank you for your help. However, I don't see anything about l2cache under Cache devices man pages. To be clear, there are two different vdev types defined in zfs source code - cache and l2cache. I am familiar with cache devices. I am curious

[zfs-discuss] Feature differences between Solaris 10 9/10 and build 147

2010-09-22 Thread Peter Taps
Folks, Here is the list of ZFS enhancements as mentioned for the latest Solaris 10 update: * ZFS device replacement enhancements - namely autoexpand * some changes to the zpool list command * Holding ZFS snapshots * Triple parity RAID-Z (raidz3) * The logbias property *

[zfs-discuss] scrub: resilver in progress for 0h38m, 0.00% done, 1131207h51m to go

2010-09-22 Thread LIC mesh
What options are there to turn off or reduce the priority of a resilver? This is on a 400TB iSCSI based zpool (8 LUNs per raidz2 vdev, 4 LUNs per shelf, 6 drives per LUN - 16 shelves total) - my client has gotten to the point that they just want to get their data off, but this resilver won't

Re: [zfs-discuss] space_map again nuked!!

2010-09-22 Thread Stephan Ferraro
Did you see this thread? http://opensolaris.org/jive/thread.jspa?messageID=5006 59#500659 I get on this link: Tomcat http error 500: The server encountered an internal error () that prevented it from fulfilling this request. He had problems with ZFS. It turned out to be faulty RAM. ZFS

Re: [zfs-discuss] Pools inside pools

2010-09-22 Thread Maurice Volaski
If you write to a zvol on a different host (via iSCSI) those writes use memory in a different memory pool (on the other computer). No deadlock. I would expect in a usual configuration that one side of a mirrored iSCSI-based pool would be on the same host as it's underlying zvol's pool. --

Re: [zfs-discuss] space_map again nuked!!

2010-09-22 Thread Victor Latushkin
On Sep 23, 2010, at 1:11 AM, Stephan Ferraro wrote: He had problems with ZFS. It turned out to be faulty RAM. ZFS is so sensitive it detects and reports problems to you. No other filesystem does that, so you think ZFS is problematic and switch. But the other filesystems is slowly

Re: [zfs-discuss] Feature differences between Solaris 10 9/10 and build 147

2010-09-22 Thread Richard Elling
On Sep 22, 2010, at 1:43 PM, Peter Taps wrote: Folks, Here is the list of ZFS enhancements as mentioned for the latest Solaris 10 update: * ZFS device replacement enhancements - namely autoexpand * some changes to the zpool list command * Holding ZFS snapshots * Triple

Re: [zfs-discuss] scrub: resilver in progress for 0h38m, 0.00% done, 1131207h51m to go

2010-09-22 Thread Richard Elling
On Sep 22, 2010, at 1:46 PM, LIC mesh wrote: What options are there to turn off or reduce the priority of a resilver? This is on a 400TB iSCSI based zpool (8 LUNs per raidz2 vdev, 4 LUNs per shelf, 6 drives per LUN - 16 shelves total) - my client has gotten to the point that they just

[zfs-discuss] ZFS Tutorial at USENIX LISA '10 conference

2010-09-22 Thread Richard Elling
Back by popular demand! The USENIX LISA conference will be hosting a full day ZFS Tutorial on Monday, November 8, 2010. A lot has changed since last year's LISA conference and the new, up-to-date tutorial will surely be a session of extreme gratification. The conference home page is

Re: [zfs-discuss] What is l2cache setting?

2010-09-22 Thread Neil Perrin
On 09/22/10 13:40, Peter Taps wrote: Neil, Thank you for your help. However, I don't see anything about l2cache under Cache devices man pages. To be clear, there are two different vdev types defined in zfs source code - cache and l2cache. I am familiar with cache devices. I am curious about

[zfs-discuss] Growing a root ZFS mirror on b134?

2010-09-22 Thread Carl Brewer
G'day, My OpenSolaris (b134) box is low on space and has a ZFS mirror for root : uname -a SunOS wattage 5.11 snv_134 i86pc i386 i86pc rpool 696G 639G 56.7G91% 1.09x ONLINE - It's currently a pair of 750GB drives. In my bag I have a pair of brand spanking new 2TB seagates that

Re: [zfs-discuss] Pools inside pools

2010-09-22 Thread Markus Kovero
If you write to a zvol on a different host (via iSCSI) those writes use memory in a different memory pool (on the other computer). No deadlock. I would expect in a usual configuration that one side of a mirrored iSCSI-based pool would be on the same host as it's underlying zvol's pool.