Re: [zfs-discuss] HAMMER
Hello Dave, Tuesday, October 16, 2007, 9:17:30 PM, you wrote: DJ you mean c9n ? ;) DJ does anyone actually *use* compression ? i'd like to see a poll on how many DJ people are using (or would use) compression on production systems that are DJ larger than your little department catch-all dumping ground server. i mean, DJ unless you had some NDMP interface directly to ZFS, daily tape backups for DJ any large system will likely be an excersize in futility unless the systems DJ are largely just archive servers, at which point it's probably smarter to DJ perform backups less often, coinciding with the workflow of migrating DJ archive data to it. otherwise wouldn't the system just plain get pounded? LDAP servers with several dozen millions accounts? Why? First you get about 2:1 compression ratio with lzjb, and you also get better performance. -- Best regards, Robert Milkowskimailto:[EMAIL PROTECTED] http://milek.blogspot.com ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] practicality of zfs send/receive for failover
Hello Matthew, Wednesday, October 17, 2007, 1:46:02 AM, you wrote: MA Richard Elling wrote: Paul B. Henson wrote: On Fri, 12 Oct 2007, Paul B. Henson wrote: I've read a number of threads and blog posts discussing zfs send/receive and its applicability is such an implementation, but I'm curious if anyone has actually done something like that in practice, and if so how well it worked. So I didn't hear from anyone on this thread actually running such an implementation in production? Could someone maybe comment on a theoretical level :) whether this would be realistic for multiple terabytes, or if I should just give up on it? It should be more reasonable to use ZFS send/recv than a dumb volume block copy. It should be on the same order of goodness as rsync-style copying. I use send/recv quite often, but my wife doesn't have a TByte of pictures (yet :-) MA Incremental zfs send/recv is actually orders of magnitude more goodness MA than rsync (due to much faster finding of changed files). MA I know of customers who are using send|ssh|recv to replicate entire thumpers MA across the country, in production. I'm sure they'll speak up here if/when MA they find this thread... I know such environment too, however just across a server room :) Is it perfect? No... but still comparing to legacy backup it's much better in terms of performance and much worse in terms of manageability. -- Best regards, Robert Milkowskimailto:[EMAIL PROTECTED] http://milek.blogspot.com ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] nfs-ownership
Is the mount using NFSv4? If so, there is likely a midguided mapping of the user/groups between the client and server. While not including BSD info, there is a little bit on NFSv4 user/group mappings at this blog: http://blogs.sun.com/nfsv4 It defaults to nfs ver. 3. As a sidenote samba is running on this machine as well, and the windows-share is able to read and write to the sub-partitons (under home/user). -- regards Claus When lenity and cruelty play for a kingdom, the gentlest gamester is the soonest winner. Shakespeare ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] HAMMER
From: Robert Milkowski [EMAIL PROTECTED] LDAP servers with several dozen millions accounts? Why? First you get about 2:1 compression ratio with lzjb, and you also get better performance. a busy ldap server certainly seems a good fit for compression but when i said large i meant, as in bytes and numbers of files :) seriously, is anyone out there using zfs for large storage servers? you know, the same usage that 90% of the storage sold in the world is used for ? (yes, i pulled that figure out of my *ss ;) are my concerns invalid with the current implementation of zfs with compression ? is the compression so lightweight that it can be decompressed as fast as the disks can stream uncompressed backup data to tape while the server is still servicing clients ? the days of nightly backups seem long gone in the space I've been working in the last several years... backups run almost 'round the clock it seems on our biggest systems (15-30Tb and 150-300mil files , which may be small by the standard of others of you out there.) what really got my eyes rolling about c9n and prompted my question was all this talk about gzip compression and other even heavierweight compression algor's. lzjb is relatively lightweight but i could still see it being a bottleneck in a 'weekly full backups' scenario unless you had a very new system with kilowatts of cpu to spare. gzip ? pulease. bzip and lzma someone has *got* to be joking ? i see these as ideal candiates for AVS scenarios where the aplication never requires full dumps to tape, but on a typical storage server ? the compression would be ideal but would also make it impossible to backup in any reasonable window. back to my postulation, if it is correct, what about some NDMP interface to ZFS ? it seems a more than natural candidate. in this scenario, compression would be a boon since the blocks would already be in a compressed state. I'd imagine this fitting into the 'zfs send' codebase somewhere. thoughts (on either c9n and/or 'zfs send ndmp') ? -=dave - Original Message - From: Robert Milkowski [EMAIL PROTECTED] To: Dave Johnson [EMAIL PROTECTED] Cc: roland [EMAIL PROTECTED]; zfs-discuss@opensolaris.org Sent: Wednesday, October 17, 2007 2:35 AM Subject: Re[2]: [zfs-discuss] HAMMER Hello Dave, Tuesday, October 16, 2007, 9:17:30 PM, you wrote: DJ you mean c9n ? ;) DJ does anyone actually *use* compression ? i'd like to see a poll on how many DJ people are using (or would use) compression on production systems that are DJ larger than your little department catch-all dumping ground server. i mean, DJ unless you had some NDMP interface directly to ZFS, daily tape backups for DJ any large system will likely be an excersize in futility unless the systems DJ are largely just archive servers, at which point it's probably smarter to DJ perform backups less often, coinciding with the workflow of migrating DJ archive data to it. otherwise wouldn't the system just plain get pounded? LDAP servers with several dozen millions accounts? Why? First you get about 2:1 compression ratio with lzjb, and you also get better performance. -- Best regards, Robert Milkowskimailto:[EMAIL PROTECTED] http://milek.blogspot.com ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Upgrade from B62 ZFS Boot/Root to B70b
How painful is this going to be? Completely? -brian -- Perl can be fast and elegant as much as J2EE can be fast and elegant. In the hands of a skilled artisan, it can and does happen; it's just that most of the shit out there is built by people who'd be better suited to making sure that my burger is cooked thoroughly. -- Jonathan Patschke ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] nfs-ownership
On 10/16/07, Claus Guttesen [EMAIL PROTECTED] wrote: I have created some zfs-partitions. First I create the home/user-partitions. Beneath that I create additional partitions. Then I have do a chown -R for that user. These partitions are shared using the sharenfs=on. The owner- and group-id is 1009. These partitions are visible as the user assigned above. But when I mount the home/user partition from a FreeBSD-client, only the top-partiton has the proper uid- and guid-assignment. The partitons beneath are assigned to the root/wheel (uid 0 and gid 0 on FreeBSD). Am I doing something wrong Did you mount both the parent and all the children on the client ? -- Paul Kraus ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] nfs-ownership
I have created some zfs-partitions. First I create the home/user-partitions. Beneath that I create additional partitions. Then I have do a chown -R for that user. These partitions are shared using the sharenfs=on. The owner- and group-id is 1009. These partitions are visible as the user assigned above. But when I mount the home/user partition from a FreeBSD-client, only the top-partiton has the proper uid- and guid-assignment. The partitons beneath are assigned to the root/wheel (uid 0 and gid 0 on FreeBSD). Am I doing something wrong Did you mount both the parent and all the children on the client ? No, I just assumed that the sub-partitions would inherit the same uid/gid as the parent. I have done a chown -R. That would be a neat feature because I have some fairly large partitions hosted on another server (solaris 9 on sparc and vxfs) which shares disk via nfs. Everytime I create a new partition I must create and mount the partitons on each webserver. Not that I have many webservers, but it would be nice to create a /data-partition, then a image-partition below that and then a, b, c, d, e, f etc. (/data/image/a). Then all I had to do was to mount the image-partiton and I wouldn't have to worry about mounting other than the image-partition on the webserver. -- regards Claus When lenity and cruelty play for a kingdom, the gentlest gamester is the soonest winner. Shakespeare ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Home fileserver with solaris 10 and zfs
hi I am currently running a linux box as my fileserver at home. It's got eight 250 gig sata2 drives connected to two sata pci controllers and configured as one big raid5 with linux software raid. Linux is (and solaris will be) installed on two separate mirrored disks. I've been playing around with solaris 10 and zfs at work and I'm pretty exited about it and now I'd like to migrate my fileserver to solaris and zfs. Now I just wanted to hear your opinion on how to configure these eight disks. Should I use raidz or raidz2 and how many pools, block size and so on? Most of my files are between 200 and 800 megs and a total of 100.000 files. The new solaris box will have 2 gigs of mem and a AMD dualcore cpu around 2ghz. Any suggestions, tips or experiences ? Thanks in advance regards This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] nfs-ownership
On 10/17/07, Claus Guttesen [EMAIL PROTECTED] wrote: Did you mount both the parent and all the children on the client ? No, I just assumed that the sub-partitions would inherit the same uid/gid as the parent. I have done a chown -R. Ahhh, the issue is not permissions, but how the NFS server sees the various directories to share. Each dataset in the zpool is seen as a separate FS from the OS perspective; each is a separate NFS share. In which case each has to be mounted separately on the NFS client. -- Paul Kraus ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] nfs-ownership
Did you mount both the parent and all the children on the client ? No, I just assumed that the sub-partitions would inherit the same uid/gid as the parent. I have done a chown -R. Ahhh, the issue is not permissions, but how the NFS server sees the various directories to share. Each dataset in the zpool is seen as a separate FS from the OS perspective; each is a separate NFS share. In which case each has to be mounted separately on the NFS client. Thank you for the clarification. When mounting the same partitions from a windows-client I get r/w access to both the parent- and child-partition. Will it be possible to implement such a feature in nfs? -- regards Claus When lenity and cruelty play for a kingdom, the gentlest gamester is the soonest winner. Shakespeare ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] HAMMER
Dave Johnson wrote: From: Robert Milkowski [EMAIL PROTECTED] LDAP servers with several dozen millions accounts? Why? First you get about 2:1 compression ratio with lzjb, and you also get better performance. a busy ldap server certainly seems a good fit for compression but when i said large i meant, as in bytes and numbers of files :) seriously, is anyone out there using zfs for large storage servers? you know, the same usage that 90% of the storage sold in the world is used for ? (yes, i pulled that figure out of my *ss ;) We're using ZFS compression on Netbackup Disk Cache Media Servers. I have 3 media servers with 42TB usable each, with compression enabled. I had to wait for Sol10 U4 to run compression because these are T2000's and there was a problem that zfs was using only 1 compression thread per pool which made it too slow. But after U4, I have no problem handling bursts of nearly 2Gbit/s of backup streams in over the network while still spooling to a pair of 30MByte/s tape drives on each server. -Andy ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] HAMMER
We are using zfs compression across 5 zpools, about 45TB of data on iSCSI storage. I/O is very fast, with small fractional CPU usage (seat of the pants metrics here, sorry). We have one other large 10TB volume for nearline Networker backups, and that one isn't compressed. We already compress these data on the backup client, and there wasn't any more compression to be had on the zpool, so it isn't worth it there. There's no doubt that heavier weight compression would be a problem as you say. One thing that would be ultra cool on the backup pool would be to have post write compression. After backups are done, the backup server sits more or less idle. It would be cool to do a compress on scrub operation that cold do some real high level compression. Then we could zfssend | ssh-remote | zfsreceive to an off site location with far less less network bandwidth, not to mention the remote storage could be really small. Datadomain (www.datadomain.com) does block level checksumming to save files as link lists of common blocks. They get very high compression ratios (in our tests about 6/1, but with more frequent full backups, more like 20/1). Then off site transfers go that much faster. Jon Dave Johnson wrote: From: Robert Milkowski [EMAIL PROTECTED] LDAP servers with several dozen millions accounts? Why? First you get about 2:1 compression ratio with lzjb, and you also get better performance. a busy ldap server certainly seems a good fit for compression but when i said large i meant, as in bytes and numbers of files :) seriously, is anyone out there using zfs for large storage servers? you know, the same usage that 90% of the storage sold in the world is used for ? (yes, i pulled that figure out of my *ss ;) are my concerns invalid with the current implementation of zfs with compression ? is the compression so lightweight that it can be decompressed as fast as the disks can stream uncompressed backup data to tape while the server is still servicing clients ? the days of nightly backups seem long gone in the space I've been working in the last several years... backups run almost 'round the clock it seems on our biggest systems (15-30Tb and 150-300mil files , which may be small by the standard of others of you out there.) what really got my eyes rolling about c9n and prompted my question was all this talk about gzip compression and other even heavierweight compression algor's. lzjb is relatively lightweight but i could still see it being a bottleneck in a 'weekly full backups' scenario unless you had a very new system with kilowatts of cpu to spare. gzip ? pulease. bzip and lzma someone has *got* to be joking ? i see these as ideal candiates for AVS scenarios where the aplication never requires full dumps to tape, but on a typical storage server ? the compression would be ideal but would also make it impossible to backup in any reasonable window. back to my postulation, if it is correct, what about some NDMP interface to ZFS ? it seems a more than natural candidate. in this scenario, compression would be a boon since the blocks would already be in a compressed state. I'd imagine this fitting into the 'zfs send' codebase somewhere. thoughts (on either c9n and/or 'zfs send ndmp') ? -=dave - Original Message - From: Robert Milkowski [EMAIL PROTECTED] To: Dave Johnson [EMAIL PROTECTED] Cc: roland [EMAIL PROTECTED]; zfs-discuss@opensolaris.org Sent: Wednesday, October 17, 2007 2:35 AM Subject: Re[2]: [zfs-discuss] HAMMER Hello Dave, Tuesday, October 16, 2007, 9:17:30 PM, you wrote: DJ you mean c9n ? ;) DJ does anyone actually *use* compression ? i'd like to see a poll on how many DJ people are using (or would use) compression on production systems that are DJ larger than your little department catch-all dumping ground server. i mean, DJ unless you had some NDMP interface directly to ZFS, daily tape backups for DJ any large system will likely be an excersize in futility unless the systems DJ are largely just archive servers, at which point it's probably smarter to DJ perform backups less often, coinciding with the workflow of migrating DJ archive data to it. otherwise wouldn't the system just plain get pounded? LDAP servers with several dozen millions accounts? Why? First you get about 2:1 compression ratio with lzjb, and you also get better performance. -- Best regards, Robert Milkowskimailto:[EMAIL PROTECTED] http://milek.blogspot.com ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss -- - _/ _/ / - Jonathan Loran - - -/ / /IT Manager - - _ / _ / / Space Sciences Laboratory, UC Berkeley -/
Re: [zfs-discuss] HAMMER
Jonathan Loran wrote: We are using zfs compression across 5 zpools, about 45TB of data on iSCSI storage. I/O is very fast, with small fractional CPU usage (seat of the pants metrics here, sorry). We have one other large 10TB volume for nearline Networker backups, and that one isn't compressed. We already compress these data on the backup client, and there wasn't any more compression to be had on the zpool, so it isn't worth it there. cool. There's no doubt that heavier weight compression would be a problem as you say. One thing that would be ultra cool on the backup pool would be to have post write compression. After backups are done, the backup server sits more or less idle. It would be cool to do a compress on scrub operation that cold do some real high level compression. Then we could zfssend | ssh-remote | zfsreceive to an off site location with far less less network bandwidth, not to mention the remote storage could be really small. Datadomain (www.datadomain.com http://www.datadomain.com) does block level checksumming to save files as link lists of common blocks. They get very high compression ratios (in our tests about 6/1, but with more frequent full backups, more like 20/1). Then off site transfers go that much faster. Do not assume that a compressed file system will send compressed. IIRC, it does not. But since UNIX is a land of pipe dreams, you can always compress anyway :-) zfs send ... | compress | ssh ... | uncompress | zfs receive ... -- richard ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] HAMMER
Richard Elling wrote: Jonathan Loran wrote: snip... Do not assume that a compressed file system will send compressed. IIRC, it does not. Let's say, if it were possible to detect the remote compression support, couldn't we send it compressed? With higher compression rates, wouldn't that be smart? The Internet is not the land of infinite bandwidth that we often think it is. But since UNIX is a land of pipe dreams, you can always compress anyway :-) zfs send ... | compress | ssh ... | uncompress | zfs receive ... Gosh, how obvious is that, eh? Thanks Richard. -- richard Jon -- - _/ _/ / - Jonathan Loran - - -/ / /IT Manager - - _ / _ / / Space Sciences Laboratory, UC Berkeley -/ / / (510) 643-5146 [EMAIL PROTECTED] - __/__/__/ AST:7731^29u18e3 ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] practicality of zfs send/receive for failover
On Tue, 16 Oct 2007, Matthew Ahrens wrote: I know of customers who are using send|ssh|recv to replicate entire thumpers across the country, in production. I'm sure they'll speak up here if/when they find this thread... Ah, that's who I'd like to hear from :)... Thanks for the secondhand information though... -- Paul B. Henson | (909) 979-6361 | http://www.csupomona.edu/~henson/ Operating Systems and Network Analyst | [EMAIL PROTECTED] California State Polytechnic University | Pomona CA 91768 ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] HAMMER
Jonathan Loran wrote: Richard Elling wrote: Jonathan Loran wrote: snip... Do not assume that a compressed file system will send compressed. IIRC, it does not. Let's say, if it were possible to detect the remote compression support, couldn't we send it compressed? With higher compression rates, wouldn't that be smart? The Internet is not the land of infinite bandwidth that we often think it is. But since UNIX is a land of pipe dreams, you can always compress anyway :-) zfs send ... | compress | ssh ... | uncompress | zfs receive ... Gosh, how obvious is that, eh? Thanks Richard. -- richard Jon even better is the ability of ssh to compress the stream for you via ssh -C :) ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Lack of physical memory evidences
In pre-ZFS era, we had observable parameters like scan rate and anonymous page-in/-out counters to discover situations when a system experiences a lack of physical memory. With ZFS, it's difficult to use mentioned parameters to figure out situations like that. Has someone any idea what we can use for the same purpose now ? Thanks in advance, Dmeetry This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Adding my own compression to zfs
being at $300 now - a friend of mine just adding another $100 This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] df command in ZFS?
I was presenting to a customer at the EBC yesterday, and one of the people at the meeting said using df in ZFS really drives him crazy (no, that's all the detail I have). Any ideas/suggestions? -- David Runyon Disk Sales Specialist Sun Microsystems, Inc. 4040 Palm Drive Santa Clara, CA 95054 US Mobile 925 323-1211 Email [EMAIL PROTECTED] ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] df command in ZFS?
I asked this recently, but haven't done anything else about it: http://www.opensolaris.org/jive/thread.jspa?messageID=155583#155583 This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] df command in ZFS?
On 10/17/07, David Runyon [EMAIL PROTECTED] wrote: I was presenting to a customer at the EBC yesterday, and one of the people at the meeting said using df in ZFS really drives him crazy (no, that's all the detail I have). Any ideas/suggestions? I suspect that this is related to the notion that file systems are cheap and the traditional notion of quotas is replaced by cheap file systems. This makes it so that a system with 1000 users that previously had a small number of file systems now has over 1000 file systems. What used to be relatively simple output from df now turns into 40+ screens[1] on the default sized terminal window. 1. If you are in this situation, there is a good chance that the formatting of df cause line folding or wrapping that doubles the number of lines to 80+ screens of df output. -- Mike Gerdts http://mgerdts.blogspot.com/ ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] GRUB + zpool version mismatches
Apparently with zfs boot, if the zpool is a version grub doesn't recognize, it merely ignores any zfs entries in menu.lst, and apparently instead boots the first entry it thinks it can boot. I ran into this myself due to some boneheaded mistakes while doing a very manual zfs / install at the summit. Shouldn't it at least spit out a warning? If so, I have no issues filing a bug, but wanted to bounce it off those more knowledgeable in this area than I am. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Fracture Clone Into FS
Hey Guys, Its not possible yet to fracture a snapshot or clone into a self-standing filesystem is it? Basically, I'd like to fracture a snapshot/clone into is own FS so I can rollback past that snapshot in the original filesystem and still keep that data. Thank you in advance. Best Regards, Jason ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Home fileserver with solaris 10 and zfs
Sandro wrote: hi I am currently running a linux box as my fileserver at home. It's got eight 250 gig sata2 drives connected to two sata pci controllers and configured as one big raid5 with linux software raid. Linux is (and solaris will be) installed on two separate mirrored disks. I've been playing around with solaris 10 and zfs at work and I'm pretty exited about it and now I'd like to migrate my fileserver to solaris and zfs. Now I just wanted to hear your opinion on how to configure these eight disks. Should I use raidz or raidz2 and how many pools, block size and so on? Most of my files are between 200 and 800 megs and a total of 100.000 files. The new solaris box will have 2 gigs of mem and a AMD dualcore cpu around 2ghz. Any suggestions, tips or experiences ? Have a look back the the many similar threads to this in this list, the general answer is it depends. There are many possibilities with 8 drives, each offering different performance/reliability trade offs. Ian ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS+NFS on storedge 6120 (sun t4)
Ok...got a break from the 25xx release... Trying to catch up so...sorry for the late response... The 6120 firmware does not support the Cache Sync command at all... You could try using a smaller blocksize setting on the array to attempt to reduce the number of read/modify/writes that you will incur... It also can be important to understand how zfs attempts to make aligned transactions as wellsince a single 128k write that starts on the beginning of a RAID stripe is guaranteed to do a full-stripe write v/s 2 read/modify/write stripes I have considered making an unsupported firmware that turns it into a caching JBOD...I just have not had any infinite spare time -Joel This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] characterizing I/O on a per zvol basis.
Hey all - Time for my silly question of the day, and before I bust out vi and dtrace... If there a simple, existing way I can observe the read / write / IOPS on a per-zvol basis? If not, is there interest in having one? Cheers! Nathan. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss