[zfs-discuss] How to enforce probing of all disks?

2013-03-22 Thread Jim Klimov
Hello all, I have a kind of lame question here: how can I force the system (OI) to probe all the HDD controllers and disks that it can find, and be certain that it has searched everywhere for disks? My remotely supported home-NAS PC was unavailable for a while, and a friend rebooted it for

Re: [zfs-discuss] SSD for L2arc

2013-03-21 Thread Jim Klimov
On 2013-03-21 16:24, Ram Chander wrote: Hi, Can I know how to configure a SSD to be used for L2arc ? Basically I want to improve read performance. The man zpool page is quite informative on theory and concepts ;) If your pool already exists, you can prepare the SSD (partition/slice it) and:

Re: [zfs-discuss] System started crashing hard after zpool reconfigure and OI upgrade

2013-03-20 Thread Jim Klimov
On 2013-03-20 17:15, Peter Wood wrote: I'm going to need some help with the crash dumps. I'm not very familiar with Solaris. Do I have to enable something to get the crash dumps? Where should I look for them? Typically the kernel crash dumps are created as a result of kernel panic; also they

Re: [zfs-discuss] partioned cache devices

2013-03-19 Thread Jim Klimov
for some dual-booted OSes, but repurposed as L2ARC and ZIL devices for the storage pool on other disks, when I played with that technology. Didn't gain much with a single spindle ;) HTH, //Jim Klimov ___ zfs-discuss mailing list zfs-discuss@opensolaris.org

Re: [zfs-discuss] partioned cache devices

2013-03-19 Thread Jim Klimov
On 2013-03-19 22:07, Andrew Gabriel wrote: The GPT partitioning spec requires the disk to be FDISK partitioned with just one single FDISK partition of type EFI, so that tools which predate GPT partitioning will still see such a GPT disk as fully assigned to FDISK partitions, and therefore less

Re: [zfs-discuss] [zfs] Petabyte pool?

2013-03-16 Thread Jim Klimov
the boxes' individual power sources can do. Conveniently, they also allow to do a remote hard-reset of hung boxes without walking to the server room ;) My 2c, //Jim Klimov ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http

Re: [zfs-discuss] [zfs] Petabyte pool?

2013-03-16 Thread Jim Klimov
On 2013-03-16 15:20, Bob Friesenhahn wrote: On Sat, 16 Mar 2013, Kristoffer Sheather @ CloudCentral wrote: Well, off the top of my head: 2 x Storage Heads, 4 x 10G, 256G RAM, 2 x Intel E5 CPU's 8 x 60-Bay JBOD's with 60 x 4TB SAS drives RAIDZ2 stripe over the 8 x JBOD's That should fit

Re: [zfs-discuss] Sun X4200 Question...

2013-03-14 Thread Jim Klimov
On 2013-03-11 21:50, Bob Friesenhahn wrote: On Mon, 11 Mar 2013, Tiernan OToole wrote: I know this might be the wrong place to ask, but hopefully someone can point me in the right direction... I got my hands on a Sun x4200. Its the original one, not the M2, and has 2 single core Opterons, 4Gb

Re: [zfs-discuss] Sun X4200 Question...

2013-03-14 Thread Jim Klimov
On 2013-03-15 01:58, Gary Driggs wrote: On Mar 14, 2013, at 5:55 PM, Jim Klimov jimkli...@cos.ru wrote: However, recently the VM virtual hardware clocks became way slow. Does NTP help correct the guest's clock? Unfortunately no, neither guest NTP, ntpdate or rdate in crontabs, nor

Re: [zfs-discuss] SVM ZFS

2013-02-26 Thread Jim Klimov
might be able to store these pieces in different pools (i.e. SSDs for some data and HDDs for other - though most list members would rightfully argue in favor of L2ARC on the SSDs). HTH, //Jim Klimov ___ zfs-discuss mailing list zfs-discuss@opensolaris.org

Re: [zfs-discuss] SVM ZFS

2013-02-26 Thread Jim Klimov
Ah, I forgot to mention - ufsdump|ufsrestore was at some time also a recommended way of such transition ;) I think it should be aware of all intimacies of the FS, including sparse files which reportedly may puzzle some other archivers. Although with any sort of ZFS compression (including

Re: [zfs-discuss] ZFS Distro Advice

2013-02-26 Thread Jim Klimov
in the directory during the process). If you use rsync over network to back up stuff, here's an example of SMF wrapper for rsyncd, and a config sample to make a snapshot after completion of the rsync session. http://wiki.openindiana.org/oi/rsync+daemon+service+on+OpenIndiana HTH, //Jim Klimov

Re: [zfs-discuss] Feature Request for zfs pool/filesystem protection?

2013-02-21 Thread Jim Klimov
On 2013-02-20 23:49, Markus Grundmann wrote: add an pool / filesystem property as an additional security layer for administrators. Whenever I modify zfs pools or filesystems it's possible to destroy [on a bad day :-)] my data. A new property protected=on|off in the pool and/or filesystem can

Re: [zfs-discuss] Feature Request for zfs pool/filesystem protection?

2013-02-21 Thread Jim Klimov
On 2013-02-21 16:54, Markus Grundmann wrote: It's anyone here on the list that's have some tips for me what files are to modify ? :-) In my current source tree now is a new property PROTECTED available both for pool- und zfs-objects. I have also two functions added to get and set the property

Re: [zfs-discuss] cannot destroy, volume is busy

2013-02-21 Thread Jim Klimov
On 2013-02-21 17:02, John D Groenveld wrote: # zfs list -t vol NAME USED AVAIL REFER MOUNTPOINT rpool/dump4.00G 99.9G 4.00G - rpool/foo128 66.2M 100G16K - rpool/swap4.00G 99.9G 4.00G - # zfs destroy rpool/foo128 cannot destroy 'rpool/foo128': volume is busy

Re: [zfs-discuss] zfs raid1 error resilvering and mount

2013-02-19 Thread Jim Klimov
On 2013-02-19 12:39, Konstantin Kuklin wrote: i did`t replace disk, after reboot system not started (zfs installed as default root system) and i boot from another system(from flash) and resilvering has auto start and show me warnings with freeze progress(dead on checking zroot/var/crash )

Re: [zfs-discuss] zfs raid1 error resilvering and mount

2013-02-19 Thread Jim Klimov
On 2013-02-19 14:24, Konstantin Kuklin wrote: zfs set canmount=off zroot/var/crash i can`t do this, because zfs list empty I'd argue that in your case it might be desirable to evacuate data and reinstall the OS - just to be certain that ZFS on-disk structures on new installation have no

Re: [zfs-discuss] zfs raid1 error resilvering and mount

2013-02-19 Thread Jim Klimov
On 2013-02-19 17:02, Victor Latushkin wrote: On 2/19/13 6:32 AM, Jim Klimov wrote: On 2013-02-19 14:24, Konstantin Kuklin wrote: zfs set canmount=off zroot/var/crash i can`t do this, because zfs list empty I'd argue that in your case it might be desirable to evacuate data and reinstall

Re: [zfs-discuss] zfs raid1 error resilvering and mount

2013-02-17 Thread Jim Klimov
delete or drop it fs zroot/var/crash (1m-10m size i didn`t remember) and mount other zfs points with my data -- С уважением Куклин Константин. Good luck, //Jim Klimov ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org

Re: [zfs-discuss] zfs raid1 error resilvering and mount

2013-02-17 Thread Jim Klimov
Also, adding to my recent post: instead of resilvering, try to run zpool scrub first - it should verify all checksums and repair whatever it can via redundancy (for metadata - extra copies). Resilver is similar to scrub, but it has its other goals and implementation, and might be not so

Re: [zfs-discuss] HELP! RPool problem

2013-02-16 Thread Jim Klimov
On 2013-02-16 21:49, John D Groenveld wrote: By the way, whatever the error message is when booting, it disapears so quickly I can't read it, so I am only guessing that this is the reason. Boot with kernel debugger so you can see the panic. And that would be so: 1) In the boot loader (GRUB)

Re: [zfs-discuss] zfs-discuss mailing list opensolaris EOL

2013-02-16 Thread Jim Klimov
Hello Cindy, Are there any plans to preserve the official mailing lists' archives, or will they go the way of Jive forums and the future digs for bits of knowledge would rely on alternate mirrors and caches? I understand that Oracle has some business priorities, but retiring hardware causes

Re: [zfs-discuss] Slow zfs writes

2013-02-12 Thread Jim Klimov
On 2013-02-12 10:32, Ian Collins wrote: Ram Chander wrote: Hi Roy, You are right. So it looks like re-distribution issue. Initially there were two Vdev with 24 disks ( disk 0-23 ) for close to year. After which which we added 24 more disks and created additional vdevs. The initial vdevs are

Re: [zfs-discuss] ZFS monitoring

2013-02-11 Thread Jim Klimov
it gets its numbers from... and perhaps make and RTI relevant kstats, if they aren't yet there ;) On the other hand, I am not certain how Solaris-based kstats interact or correspond to structures in FreeBSD (or Linux for that matter)?.. HTH, //Jim Klimov

Re: [zfs-discuss] Freeing unused space in thin provisioned zvols

2013-02-10 Thread Jim Klimov
on the original dataset. HTH, //Jim Klimov ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Re: [zfs-discuss] how to know available disk space

2013-02-08 Thread Jim Klimov
On 2013-02-08 22:47, Edward Ned Harvey (opensolarisisdeadlongliveopensolaris) wrote: Maybe this isn't exactly what you need, but maybe: for fs in `zfs list -H -o name` ; do echo $fs ; zfs get reservation,refreservation,usedbyrefreservation $fs ; done What is the sacramental purpose of such

Re: [zfs-discuss] Scrub performance

2013-02-04 Thread Jim Klimov
On 2013-02-04 15:52, Edward Ned Harvey (opensolarisisdeadlongliveopensolaris) wrote: I noticed that sometimes I had terrible rates with 10MB/sec. Then later it rose up to 70MB/sec. Are you talking about scrub rates for the complete scrub? Because if you sit there and watch it, from minute

Re: [zfs-discuss] Scrub performance

2013-02-04 Thread Jim Klimov
On 2013-02-04 17:10, Karl Wagner wrote: OK then, I guess my next question would be what's the best way to undedupe the data I have? Would it work for me to zfs send/receive on the same pool (with dedup off), deleting the old datasets once they have been 'copied'? I think I remember reading

Re: [zfs-discuss] RFE: Un-dedup for unique blocks

2013-01-24 Thread Jim Klimov
On 2013-01-24 11:06, Darren J Moffat wrote: On 01/24/13 00:04, Matthew Ahrens wrote: On Tue, Jan 22, 2013 at 5:29 AM, Darren J Moffat darr...@opensolaris.org mailto:darr...@opensolaris.org wrote: Preallocated ZVOLs - for swap/dump. Darren, good to hear about the cool stuff in S11.

Re: [zfs-discuss] RFE: Un-dedup for unique blocks

2013-01-23 Thread Jim Klimov
On 2013-01-23 09:41, casper@oracle.com wrote: Yes and no: the system reserves a lot of additional memory (Solaris doesn't over-commits swap) and swap is needed to support those reservations. Also, some pages are dirtied early on and never touched again; those pages should not be kept in

Re: [zfs-discuss] RFE: Un-dedup for unique blocks

2013-01-22 Thread Jim Klimov
On 2013-01-22 14:29, Darren J Moffat wrote: Preallocated ZVOLs - for swap/dump. Sounds like something I proposed on these lists, too ;) Does this preallocation only mean filling an otherwise ordinary ZVOL with zeroes (or some other pattern) - if so, to what effect? Or is it also supported to

Re: [zfs-discuss] RFE: Un-dedup for unique blocks

2013-01-22 Thread Jim Klimov
On 2013-01-22 23:03, Sašo Kiselkov wrote: On 01/22/2013 10:45 PM, Jim Klimov wrote: On 2013-01-22 14:29, Darren J Moffat wrote: Preallocated ZVOLs - for swap/dump. Or is it also supported to disable COW for such datasets, so that the preallocated swap/dump zvols might remain contiguous

Re: [zfs-discuss] RFE: Un-dedup for unique blocks

2013-01-22 Thread Jim Klimov
On 2013-01-22 23:32, Nico Williams wrote: IIRC dump is special. As for swap... really, you don't want to swap. If you're swapping you have problems. Any swap space you have is to help you detect those problems and correct them before apps start getting ENOMEM. There *are* exceptions to this,

Re: [zfs-discuss] RFE: Un-dedup for unique blocks

2013-01-22 Thread Jim Klimov
The discussion gets suddenly hot and interesting - albeit quite diverged from the original topic ;) First of all, as a disclaimer, when I have earlier proposed such changes to datasets for swap (and maybe dump) use, I've explicitly proposed that this be a new dataset type - compared to zvol and

Re: [zfs-discuss] Resilver w/o errors vs. scrub with errors

2013-01-21 Thread Jim Klimov
On 2013-01-21 07:06, Stephan Budach wrote: Are there switch stats on whether it has seen media errors? Has anybody gotton QLogic's SanSurfer to work with anything newer than Java 1.4.2? ;) I checked the logs on my switches and they don't seem to indicate such issues, but I am lacking the

Re: [zfs-discuss] RFE: Un-dedup for unique blocks

2013-01-20 Thread Jim Klimov
On 2013-01-20 19:55, Tomas Forsman wrote: On 19 January, 2013 - Jim Klimov sent me these 2,0K bytes: Hello all, While revising my home NAS which had dedup enabled before I gathered that its RAM capacity was too puny for the task, I found that there is some deduplication among the data bits

Re: [zfs-discuss] Resilver w/o errors vs. scrub with errors

2013-01-20 Thread Jim Klimov
On 2013-01-20 16:56, Edward Ned Harvey (opensolarisisdeadlongliveopensolaris) wrote: From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss- boun...@opensolaris.org] On Behalf Of Jim Klimov And regarding the considerable activity - AFAIK there is little way for ZFS to reliably read

Re: [zfs-discuss] Resilver w/o errors vs. scrub with errors

2013-01-20 Thread Jim Klimov
Did you try replacing the patch-cables and/or SFPs on the path between servers and disks, or at least cleaning them? A speck of dust (or, God forbid, a pixel of body fat from a fingerprint) caught between the two optic cable cutoffs might cause any kind of signal weirdness from time to time...

Re: [zfs-discuss] RFE: Un-dedup for unique blocks

2013-01-20 Thread Jim Klimov
On 2013-01-20 17:16, Edward Harvey wrote: But, by talking about it, we're just smoking pipe dreams. Cuz we all know zfs is developmentally challenged now. But one can dream... I beg to disagree. While most of my contribution was so far about learning stuff and sharing with others, as well

[zfs-discuss] RFE: Un-dedup for unique blocks

2013-01-19 Thread Jim Klimov
Hello all, While revising my home NAS which had dedup enabled before I gathered that its RAM capacity was too puny for the task, I found that there is some deduplication among the data bits I uploaded there (makes sense, since it holds backups of many of the computers I've worked on - some of

Re: [zfs-discuss] Resilver w/o errors vs. scrub with errors

2013-01-19 Thread Jim Klimov
On 2013-01-19 18:17, Bob Friesenhahn wrote: Resilver may in fact be just verifying that the pool disks are coherent via metadata. This might happen if the fiber channel is flapping. Correction: that (verification) would be scrubbing ;) The way I get it, resilvering is related to scrubbing

Re: [zfs-discuss] Resilver w/o errors vs. scrub with errors

2013-01-19 Thread Jim Klimov
On 2013-01-19 20:08, Bob Friesenhahn wrote: On Sat, 19 Jan 2013, Jim Klimov wrote: On 2013-01-19 18:17, Bob Friesenhahn wrote: Resilver may in fact be just verifying that the pool disks are coherent via metadata. This might happen if the fiber channel is flapping. Correction

Re: [zfs-discuss] Resilver w/o errors vs. scrub with errors

2013-01-19 Thread Jim Klimov
On 2013-01-19 20:23, Jim Klimov wrote: On 2013-01-19 20:08, Bob Friesenhahn wrote: On Sat, 19 Jan 2013, Jim Klimov wrote: On 2013-01-19 18:17, Bob Friesenhahn wrote: Resilver may in fact be just verifying that the pool disks are coherent via metadata. This might happen if the fiber channel

Re: [zfs-discuss] iSCSI access patterns and possible improvements?

2013-01-19 Thread Jim Klimov
On 2013-01-19 23:39, Richard Elling wrote: This is not quite true for raidz. If there is a 4k write to a raidz comprised of 4k sector disks, then there will be one data and one parity block. There will not be 4 data + 1 parity with 75% space wastage. Rather, the space allocation more closely

Re: [zfs-discuss] iSCSI access patterns and possible improvements?

2013-01-18 Thread Jim Klimov
On 2013-01-18 06:35, Thomas Nau wrote: If almost all of the I/Os are 4K, maybe your ZVOLs should use a volblocksize of 4K? This seems like the most obvious improvement. 4k might be a little small. 8k will have less metadata overhead. In some cases we've seen good performance on these

Re: [zfs-discuss] iSCSI access patterns and possible improvements?

2013-01-17 Thread Jim Klimov
On 2013-01-17 16:04, Bob Friesenhahn wrote: If almost all of the I/Os are 4K, maybe your ZVOLs should use a volblocksize of 4K? This seems like the most obvious improvement. Matching the volume block size to what the clients are actually using (due to their filesystem configuration) should

Re: [zfs-discuss] Heavy write IO for no apparent reason

2013-01-17 Thread Jim Klimov
. Moving such temporary works to a different dataset with a different snapshot schedule and/or to a different pool (to keep related fragmentation constrained) may prove useful. HTH, //Jim Klimov ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http

Re: [zfs-discuss] help zfs pool with duplicated and missing entry of hdd

2013-01-10 Thread Jim Klimov
On 2013-01-10 08:51, Jason wrote: Hi, One of my server's zfs faulted and it shows following: NAMESTATE READ WRITE CKSUM backup UNAVAIL 0 0 0 insufficient replicas raidz2-0 UNAVAIL 0 0 0 insufficient replicas c4t0d0

Re: [zfs-discuss] Pool performance when nearly full

2012-12-20 Thread Jim Klimov
or hardlinks, they won't actually be released when you delete a particular copy of the file. HTH, //Jim Klimov ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Re: [zfs-discuss] The format command crashes on 3TB disk but zpool create ok

2012-12-15 Thread Jim Klimov
no such problems with newer build of parted as in OI, so that disk was in fact labeled for SXCE while the box was booted with OI LiveCD. HTH, //Jim Klimov ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman

Re: [zfs-discuss] Digging in the bowels of ZFS

2012-12-11 Thread Jim Klimov
On 2012-12-02 05:42, Jim Klimov wrote: My plan is to dig out the needed sectors of the broken block from each of the 6 disks and try any and all reasonable recombinations of redundancy and data sectors to try and match the checksum - this should be my definite answer on whether ZFS

Re: [zfs-discuss] Digging in the bowels of ZFS

2012-12-11 Thread Jim Klimov
On 2012-12-11 16:44, Jim Klimov wrote: For single-break-per-row tests based on hypotheses from P parities, D data disks and R broken rows, we need to checksum P*(D^R) userdata recombinations in order to determine that we can't recover the block. A small maths correction: the formula above

Re: [zfs-discuss] Digging in the bowels of ZFS

2012-12-10 Thread Jim Klimov
and the software further along the road). I am not sure which one of these *couldn't* issue (or be interpreted to issue) a number of weird identical writes to different disks at same offsets. Everyone is a suspect :( Thanks, //Jim Klimov ___ zfs-discuss mailing list

Re: [zfs-discuss] Digging in the bowels of ZFS

2012-12-09 Thread Jim Klimov
more below... On 2012-12-06 03:06, Jim Klimov wrote: It also happens that on disks 1,2,3 the first row's sectors (d0, d2, d3) are botched - ranges from 0x9C0 to 0xFFF (end of 4KB sector) are zeroes. The neighboring blocks, located a few sectors away from this one, also have compressed data

[zfs-discuss] Zpool error in metadata:0x0

2012-12-08 Thread Jim Klimov
an error in metadata:0x0? Thanks, //Jim Klimov ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Re: [zfs-discuss] Userdata physical allocation in rows vs. columns WAS Digging in the bowels of ZFS

2012-12-08 Thread Jim Klimov
and their ordering don't even matter to ZFS layers until decompression - parities and checksums just apply to prepared bulk data... //Jim Klimov On 2012-12-06 02:08, Jim Klimov wrote: On 2012-12-05 05:52, Jim Klimov wrote: For undersized allocations, i.e. of compressed data, it is possible

Re: [zfs-discuss] Remove disk

2012-12-06 Thread Jim Klimov
On 2012-12-06 09:35, Albert Shih wrote: 1) add a 5th top-level vdev (eg. another set of 12 disks) That's not a problem. That IS a problem if you're going to ultimately remove an enclosure - once added, you won't be able to remove the extra top-level VDEV from your ZFS pool. 2) replace the

Re: [zfs-discuss] ZFS QoS and priorities

2012-12-05 Thread Jim Klimov
On 2012-12-05 04:11, Richard Elling wrote: On Nov 29, 2012, at 1:56 AM, Jim Klimov jimkli...@cos.ru mailto:jimkli...@cos.ru wrote: I've heard a claim that ZFS relies too much on RAM caching, but implements no sort of priorities (indeed, I've seen no knobs to tune those) - so

Re: [zfs-discuss] ZFS QoS and priorities

2012-12-05 Thread Jim Klimov
On 2012-11-29 10:56, Jim Klimov wrote: For example, I might want to have corporate webshop-related databases and appservers to be the fastest storage citizens, then some corporate CRM and email, then various lower priority zones and VMs, and at the bottom of the list - backups. On a side note

Re: [zfs-discuss] zvol access rights - chown zvol on reboot / startup / boot

2012-12-05 Thread Jim Klimov
/before zpool exports), but brave souls can feel free to try it out and comment. Presence of the service didn't cause any noticeable troubles on my test boxen over the past couple of weeks. http://vboxsvc.svn.sourceforge.net/viewvc/vboxsvc/lib/svc/method/zfs-zvolrights HTH, //Jim Klimov

Re: [zfs-discuss] VXFS to ZFS

2012-12-05 Thread Jim Klimov
On 2012-12-05 23:11, Morris Hooten wrote: Is there a documented way or suggestion on how to migrate data from VXFS to ZFS? Off the top of my head, I think this would go like any other migration - create the new pool on new disks and use rsync for simplicity (if your VxFS setup does not utilize

Re: [zfs-discuss] Digging in the bowels of ZFS

2012-12-05 Thread Jim Klimov
On 2012-12-05 05:52, Jim Klimov wrote: For undersized allocations, i.e. of compressed data, it is possible to see P-sizes not divisible by 4 (disks) in 4KB sectors, however, some sectors do apparently get wasted because the A-size in the DVA is divisible by 6*4KB. With columnar allocation

Re: [zfs-discuss] Digging in the bowels of ZFS

2012-12-05 Thread Jim Klimov
more below... On 2012-12-05 23:16, Timothy Coalson wrote: On Tue, Dec 4, 2012 at 10:52 PM, Jim Klimov jimkli...@cos.ru mailto:jimkli...@cos.ru wrote: On 2012-12-03 18:23, Jim Klimov wrote: On 2012-12-02 05:42, Jim Klimov wrote: 4) Where are the redundancy algorithms

Re: [zfs-discuss] Digging in the bowels of ZFS

2012-12-04 Thread Jim Klimov
On 2012-12-03 18:23, Jim Klimov wrote: On 2012-12-02 05:42, Jim Klimov wrote: So... here are some applied questions: Well, I am ready to reply a few of my own questions now :) Continuing the desecration of my deceased files' resting grounds... 2) Do I understand correctly

Re: [zfs-discuss] zpool rpool rename offline

2012-12-03 Thread Jim Klimov
and use autoexpansion (or manual expansion) to let your VM's rpool capture the whole increased virtual disk. If automagic doesn't work, I posted about a month ago about the manual procedure on this list: http://mail.opensolaris.org/pipermail/zfs-discuss/2012-November/052712.html HTH, //Jim Klimov

[zfs-discuss] Eradicating full-disk pool headers safely

2012-12-03 Thread Jim Klimov
Hello all, When I started with my old test box (the 6-disk raidz2 pool), I had first created the pool on partitions (i.e. c7t1d0p0 or physical paths like /pci@0,0/pci1043,81ec@1f,2/disk@1,0:q), but I've soon destroyed it and recreated (with the same name pool) in slices (i.e. c7t0d0s0 or

Re: [zfs-discuss] Digging in the bowels of ZFS

2012-12-03 Thread Jim Klimov
On 2012-12-02 05:42, Jim Klimov wrote: So... here are some applied questions: Well, I am ready to reply a few of my own questions now :) I've staged an experiment by taking a 128Kb block from that file and appending it to a new file in a test dataset, where I changed the compression settings

Re: [zfs-discuss] zpool rpool rename offline

2012-12-03 Thread Jim Klimov
On 2012-12-03 20:35, Heiko L. wrote: I've already tested: beadm create -p $dstpool $bename beadm list zpool set bootfs=$dstpool/ROOT/$bename $dstpool beadm activate $bename beadm list init 6 - result: root@opensolaris:~# init 6 updating //platform/i86pc/boot_archive updating

Re: [zfs-discuss] zpool rpool rename offline

2012-12-03 Thread Jim Klimov
On 2012-12-03 20:51, Heiko L. wrote: jimklimov wrote: In general, I'd do the renaming with a different bootable media, including a LiveCD/LiveUSB, another distro that can import and rename this pool version, etc. - as long as booting does not involve use of the old rpool. Thank you. I will

Re: [zfs-discuss] 6Tb Database with ZFS

2012-12-01 Thread Jim Klimov
(as is the default), while the randomly accessed tablespaces might be or not be good candidates for such caching - however you can test this setting change on the fly. I believe, you must allow caching userdata for a dataset in RAM if you want to let it spill over onto L2ARC. HTH, //Jim Klimov

[zfs-discuss] Digging in the bowels of ZFS

2012-12-01 Thread Jim Klimov
for any info, ideas, insights, and just for reading this long post to the end ;) //Jim Klimov ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Re: [zfs-discuss] Remove disk

2012-11-30 Thread Jim Klimov
On 2012-11-30 15:52, Tomas Forsman wrote: On 30 November, 2012 - Albert Shih sent me these 0,8K bytes: Hi all, I would like to knwon if with ZFS it's possible to do something like that : http://tldp.org/HOWTO/LVM-HOWTO/removeadisk.html Removing a disk - no, one still can not reduce

[zfs-discuss] ZFS QoS and priorities

2012-11-29 Thread Jim Klimov
I've heard a claim that ZFS relies too much on RAM caching, but implements no sort of priorities (indeed, I've seen no knobs to tune those) - so that if the storage box receives many different types of IO requests with different administrative weights in the view of admins, it can not really

Re: [zfs-discuss] zfs on SunFire X2100M2 with hybrid pools

2012-11-28 Thread Jim Klimov
Edward Ned Harvey (opensolarisisdeadlongliveopensolaris) wrote: There are very few situations where (gzip) option is better than the default lzjb. Well, for the most part my question regarded the slowness (or lack of) gzip DEcompression as compared to lz* algorithms. If there are files and

Re: [zfs-discuss] zfs on SunFire X2100M2 with hybrid pools

2012-11-27 Thread Jim Klimov
, or in a separate dataset, cloned and delegated into LZs (if JDK customizations are further needed by some - but not all - local zones, i.e. timezone updates, trusted CA certs, etc.). HTH, //Jim Klimov ___ zfs-discuss mailing list zfs-discuss@opensolaris.org

Re: [zfs-discuss] zfs on SunFire X2100M2 with hybrid pools

2012-11-27 Thread Jim Klimov
blocks for OS image files would be the same and dedupable. HTH, //Jim Klimov ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Re: [zfs-discuss] Directory is not accessible

2012-11-26 Thread Jim Klimov
possible that the block would go away (if it is not referenced also by snapshots/clones/dedup), and such drastic measures won't be needed. HTH, //Jim Klimov ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo

[zfs-discuss] ZFS Appliance as a general-purpose server question

2012-11-22 Thread Jim Klimov
? Is it possible to run VirtualBoxes in the ZFS-SA OS, dare I ask? ;) Thanks, //Jim Klimov ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Re: [zfs-discuss] ZFS Appliance as a general-purpose server question

2012-11-22 Thread Jim Klimov
On 2012-11-22 17:31, Darren J Moffat wrote: Is it possible to use the ZFS Storage appliances in a similar way, and fire up a Solaris zone (or a few) directly on the box for general-purpose software; or to shell-script administrative tasks such as the backup archive management in the global zone

Re: [zfs-discuss] mixing WD20EFRX and WD2002FYPS in one pool

2012-11-21 Thread Jim Klimov
the found discrepancies and those sectors it's going to use to recover a block, at least of the kernel knows it is on non-ECC RAM (if it does), but I don't know if it really does that. (Worthy RFE if not). HTH, //Jim Klimov ___ zfs-discuss mailing list

Re: [zfs-discuss] Intel DC S3700

2012-11-21 Thread Jim Klimov
On 2012-11-21 21:55, Ian Collins wrote: I can't help thinking these drives would be overkill for an ARC device. All of the expensive controller hardware is geared to boosting random write IOPs, which somewhat wasted on a write slowly, read often device. The enhancements would be good for a ZIL,

Re: [zfs-discuss] zvol wrapped in a vmdk by Virtual Box and double writes?

2012-11-20 Thread Jim Klimov
On 2012-11-21 03:21, nathan wrote: Overall, the pain of the doubling of bandwidth requirements seems like a big downer for *my* configuration, as I have just the one SSD, but I'll persist and see what I can get out of it. I might also speculate that for each rewritten block of userdata in the

Re: [zfs-discuss] Repairing corrupted ZFS pool

2012-11-19 Thread Jim Klimov
. Perhaps, if the zfs diff does perform reasonably for you, you can feed its output as the list of objects to replicate in rsync's input and save many cycles this way. Good luck, //Jim Klimov ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http

Re: [zfs-discuss] Repairing corrupted ZFS pool

2012-11-19 Thread Jim Klimov
On 2012-11-19 20:58, Mark Shellenbaum wrote: There is probably nothing wrong with the snapshots. This is a bug in ZFS diff. The ZPL parent pointer is only guaranteed to be correct for directory objects. What you probably have is a file that was hard linked multiple times and the parent

Re: [zfs-discuss] Repairing corrupted ZFS pool

2012-11-19 Thread Jim Klimov
Oh, and one more thing: rsync is only good if your filesystems don't really rely on ZFS/NFSv4-style ACLs. If you need those, you are stuck with Solaris tar or Solaris cpio to carry the files over, or you have to script up replication of ACLs after rsync somehow. You should also replicate the

Re: [zfs-discuss] Repairing corrupted ZFS pool

2012-11-19 Thread Jim Klimov
On 2012-11-19 22:38, Mark Shellenbaum wrote: The parent pointer is a single 64 bit quantity that can't track all the possible parents a hard linked file could have. I believe it is inode number of the parent, or similar to that - and an available inode number can get recycled and used by newer

Re: [zfs-discuss] zvol access rights - chown zvol on reboot / startup / boot

2012-11-16 Thread Jim Klimov
On 2012-11-15 21:43, Geoff Nordli wrote: Instead of using vdi, I use comstar targets and then use vbox built-in scsi initiator. Out of curiosity: in this case are there any devices whose ownership might get similarly botched, or you've tested that this approach also works well for non-root

Re: [zfs-discuss] zvol access rights - chown zvol on reboot / startup / boot

2012-11-16 Thread Jim Klimov
On 2012-11-16 12:43, Robert Milkowski wrote: No, there isn’t other way to do it currently. SMF approach is probably the best option for the time being. I think that there should be couple of other properties for zvol where permissions could be stated. +1 :) Well, when the subject was

Re: [zfs-discuss] zvol access rights - chown zvol on reboot / startup / boot

2012-11-16 Thread Jim Klimov
Well, as a simple stone-age solution (to simplify your SMF approach), you can define custom attributes on dataset, zvols included. I think a custom attr must include a colon : in the name, and values can be multiline if needed. Simple example follows: # zfs set owner:user=jim pool/rsvd # zfs set

Re: [zfs-discuss] zvol access rights - chown zvol on reboot / startup / boot

2012-11-16 Thread Jim Klimov
On 2012-11-16 14:45, Jim Klimov wrote: Well, as a simple stone-age solution (to simplify your SMF approach), you can define custom attributes on dataset, zvols included. I think a custom attr must include a colon : in the name, and values can be multiline if needed. Simple example follows

Re: [zfs-discuss] Intel DC S3700

2012-11-14 Thread Jim Klimov
On 2012-11-14 18:05, Eric D. Mudama wrote: On Wed, Nov 14 at 0:28, Jim Klimov wrote: All in all, I can't come up with anything offensive against it quickly ;) One possible nit regards the ratings being geared towards 4KB block (which is not unusual with SSDs), so it may be further from

Re: [zfs-discuss] Intel DC S3700

2012-11-13 Thread Jim Klimov
On 2012-11-13 22:56, Mauricio Tavares wrote: Trying again: Intel just released those drives. Any thoughts on how nicely they will play in a zfs/hardware raid setup? Seems interesting - fast, assumed reliable and consistent in its IOPS (according to marketing talk), addresses power loss

Re: [zfs-discuss] Dedicated server running ESXi with no RAID card, ZFS for storage?

2012-11-13 Thread Jim Klimov
On 2012-11-14 03:20, Dan Swartzendruber wrote: Well, I think I give up for now. I spent quite a few hours over the last couple of days trying to get gnome desktop working on bare-metal OI, followed by virtualbox. Supposedly that works in headless mode with RDP for management, but nothing but

[zfs-discuss] Expanding a ZFS pool disk in Solaris 10 on VMWare (or other expandable storage technology)

2012-11-11 Thread Jim Klimov
on pool! ;) HTH, //Jim Klimov ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Re: [zfs-discuss] cannot replace X with Y: devices have different sector alignment

2012-11-10 Thread Jim Klimov
process partial overwrites of a 4KB sector with 512b pieces of data - would other bytes remain intact or not?.. Before trying to fool a production system this way, if at all, I believe some stress-tests with small blocks are due on some other system. My 2c, //Jim Klimov

Re: [zfs-discuss] Dedicated server running ESXi with no RAID card, ZFS for storage?

2012-11-09 Thread Jim Klimov
On 2012-11-09 16:14, Edward Ned Harvey (opensolarisisdeadlongliveopensolaris) wrote: From: Karl Wagner [mailto:k...@mouse-hole.com] If I was doing this now, I would probably use the ZFS aware OS bare metal, but I still think I would use iSCSI to export the ZVols (mainly due to the ability to

Re: [zfs-discuss] Dedicated server running ESXi with no RAID card, ZFS for storage?

2012-11-09 Thread Jim Klimov
/vbox-svc.xml http://vboxsvc.svn.sourceforge.net/viewvc/vboxsvc/usr/share/doc/vboxsvc/README-vboxsvc.txt See you in the VirtualBox forum thread if you do have questions :) //Jim Klimov ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http

[zfs-discuss] Forcing ZFS options

2012-11-09 Thread Jim Klimov
into a livecd/failsafe, just so that the needed datasets or paths won't be busy and so I can set, verify and apply these mountpoint values. This is not a convenient way to do things :) Thanks, //Jim Klimov ___ zfs-discuss mailing list zfs-discuss@opensolaris.org

Re: [zfs-discuss] Forcing ZFS options

2012-11-09 Thread Jim Klimov
On 2012-11-09 18:06, Gregg Wonderly wrote: Do you move the pools between machines, or just on the same physical machine? Could you just use symlinks from the new root to the old root so that the names work until you can reboot? It might be more practical to always use symlinks if you do a

Re: [zfs-discuss] Dedicated server running ESXi with no RAID card, ZFS for storage?

2012-11-08 Thread Jim Klimov
. HTH, //Jim Klimov ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

  1   2   3   4   5   6   >