Re: [zfs-discuss] Re: ZFS uses 1.1GB more space, reports conflicting information...
On 9/6/06, UNIX admin <[EMAIL PROTECTED]> wrote: Yes, the man page says that. However, it is possible to mix disks of different sizes in a RAIDZ, and this works. Why does it work? Because RAIDZ stripes are dynamic in size. From that I infer that disks can be any size because the stripes can be any size. Should one or more of the smaller disks become full, the stripe information is written across the remaining space on the rest of the bigger disks. Or should be anyway. If you think about it, it makes sense. You are right that it can be done in theory. The practical difficulty is to do it efficiently. My raidz pool tend to have a strong preference to utilize all the disks so disks with different sizes might result in smaller column widths later. I'm not sure if that spells any performance issues but at least the capacity difference between the 2 largest disk will not be usable. Is maximising use of capacity of all available disk in the pipeline for RAID-Z or is it to be avoided intentionally? -- Just me, Wire ... ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Overview (rollup) of recent activity on zfs-discuss
For background on what this is, see: http://www.opensolaris.org/jive/message.jspa?messageID=24416#24416 http://www.opensolaris.org/jive/message.jspa?messageID=25200#25200 = zfs-discuss 08/16 - 08/31 = Size of all threads during period: Thread size Topic --- - 22 ZFS + rsync, backup on steroids. 13 SCSI synchronize cache cmd 13 Niagara and ZFS compression? 12 ZFS Boot Disk 10 zpool status panics server 10 ZFS & se6920 9 Oracle on ZFS 8 zpool core dumped 8 in-kernel gzip compression 8 Significant "pauses" during zfs writes 7 Home Server with ZFS 7 3510 - some new tests 6 unaccounted for daily growth in ZFS disk space usage 6 ZFS write performance problem with compression set to ON 6 Why is ZFS raidz slower than simple ufs mount? 6 Tape backup 6 SPEC SFS97 benchmark of ZFS,UFS,VxFS 6 Proposal: user-defined properties 6 Porting ZFS file system to FreeBSD. 6 Need Help: didn't create the pool as radiz but stripes 6 Information on ZFS API? 5 zpool import: snv_33 to S10 6/06 5 multi-layer ZFS filesystems and exporting: my stupid question for the day 5 commercial backup software and zfs 5 ZFS compression 5 ZFS Filesytem Corrpution 5 Issue with zfs snapshot replication from version2 to version3 pool. 5 File level compression 4 zpool import dropped a core. 4 zpool import - cannot mount [...] directory is not empty 4 zpool hangs 4 destroyed pools signatures 4 [fbsd] Porting ZFS file system to FreeBSD. 4 ZFS with expanding LUNs 4 ZFS web admin - No items found. 4 ZFS and very large directories 4 ZFS Load-balancing over vdevs vs. real disks? 4 Storage Compatibilty list 4 Find the difference between two snapshots 4 Encryption on ZFS / Disk Usage 4 Compression with fixed record sizes 3 zpool iostat, scrubbing increases used disk space 3 zpool import/export 3 system unresponsive after issuing a zpool attach 3 problem with zfs receive -i 3 does zfs can be failover between two mechine? 3 does anybody port the zfs webadmin to webmin? 3 can't create snapshot 3 ZFS questions with mirrors 3 ZFS compression / space efficiency 3 ZFS Performance compared to UFS & VxFS 3 Want to try ZFS, but "format" got message 3 Newbie questions about drive problems 3 Can a zfs storage pool be imported readonly? 2 zfs and vmware 2 new ZFS links page 2 neopath vs automount 2 migrating data across boxes 2 libzfs question 2 delete acl not working on zfs.v3? 2 ZFS 'quot' command 2 Tunable parameter to zfs memory use 2 Solaris 6/06 ZFS and OpenSolaris ZFS 2 Interesting zfs destroy failure 1 zpool status inconsistent after user error? 1 zfs list - unexpected error 17 at line 1684 1 space accounting with RAID-Z 1 pool ID 1 fdatasync 1 ZFS Performance compared to UFS & VxFS - offtopic 1 ZFS ACL: append_data didn't do what I expected 1 Sol 10 x86_64 intermittent SATA device locks up server 1 Question on Zones and memory usage (65120349) 1 Query on ZFS 1 Equivalent of command quot 1 Apple Time Machine Posting activity by person for period: # of posts By -- -- 35 rmilkowski at task.gda.pl (robert milkowski) 17 ahrens at eng.sun.com (matthew ahrens) 15 eric.schrock at sun.com (eric schrock) 11 roch.bourbonnais at sun.com (roch) 11 fcusack at fcusack.com (frank cusack) 10 matthew.ahrens at sun.com (matthew ahrens) 9 jamesd.wi at gmail.com (james dickens) 8 george.wilson at sun.com (george wilson) 7 rasputnik at gmail.com (dick davies) 6 mgerdts at gmail.com (mike gerdts) 6 jmlittle at gmail.com (joe little) 6 anantha.srirama at cdc.hhs.gov (anantha n. srirama) 5 torrey.mcmahon at sun.com (torrey mcmahon) 5 richard.elling at sun.com (richard elling - pae) 5 nicolas.williams at sun.com (nicolas williams) 5 michael.schuster at sun.com (michael schuster - sun microsystems)
Re: [zfs-discuss] Re: Re: ZFS forces system to paging to the point it is
Robert Milkowski wrote: On Wed, 6 Sep 2006, Mark Maybee wrote: Robert Milkowski wrote: ::dnlc!wc 1048545 3145811 76522461 Well, that explains half your problem... and maybe all of it: After I reduced vdev prefetch from 64K to 8K for last few hours system is working properly without workaround and free memory stays at about 1GB. Reducing vdev prefetch to 8K alse reduced read thruoutput 10x. I belive this is somehow related - maybe vdev cache was so aggressive (I got 40-100MB/s of reads) and consuming memory so fast that thread which is supposed to regain some memory couldn't keep up? I suppose, although the data volume doesn't seem that high... maybe you are just operating at the hairy edge here. Anyway, I have filed a bug to track this issue: 6467963 do_dnlc_reduce_cache() can be blocked by ZFS_OBJ_HOLD_ENTER() -Mark ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Re: Recommendation ZFS on StorEdge 3320
Roch - PAE wrote: Thinking some more about this. If your requirements does mandate some form of mirroring, then it truly seems that ZFS should take that in charge if only because of the self-healing characteristics. So I feel the storage array's job is to export low latency Luns to ZFS. The hard part is getting a set of simple requirements. As you go into more complex data center environments you get hit with older Solaris revs, other OSs, SOX compliance issues, etc. etc. etc. The world where most of us seem to be playing with ZFS is on the lower end of the complexity scale. Sure, throw your desktop some fast SATA drives. No problem. Oh wait, you've got ten Oracle DBs on three E25Ks that need to be backed up every other blue moon ... I agree with the general idea that an array, be it one disk or some raid combination, should simply export low latency LUNs. However, its the features offered by the array - Like site to site replication - used to meet more complex requirements that literally slow things down. In many cases you'll see years old operational procedures causing those low latency LUNs to slow down even more. Something really hard to get a customer to undo because a new fangled file system is out. ;) I'd be happy to live with those simple Luns but I guess some storage will just refuse to export non-protected luns. Now we can definitively take advantage of the Array's capability of exporting highly resilient Luns; RAID-5 seems to fit the bill rather well here. Even an 9+1 luns will be quite resilient and have a low block overhead. I think 99x0 used to do 3+1 only. Now it's 7+1 if I recall. Close enough I suppose. So we benefit from the arrays resiliency as well as it's low latency characteristics. And we mirror data at the ZFS level which means great performance and great data integrity and great availability. Note that ZFS write characteristics (all sequential) means that we will commonly be filling full stripes on the luns thus avoiding the partial stripe performance pitfall. One thing comes to mind in that case. Many arrays do sequential detect on the blocks that come in to the front end ports. If things get split up to much or out of order or array characteristic here> then you could induce more latency as the array does cartwheels trying to figure out whats going on. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Re: Re: ZFS forces system to paging to the point it is
Robert Milkowski wrote: ::dnlc!wc 1048545 3145811 76522461 Well, that explains half your problem... and maybe all of it: We have a thread that *should* be trying to free up these entries in the DNLC, however it appears to be blocked: stack pointer for thread 2a10014fcc0: 2a10014edd1 [ 02a10014edd1 turnstile_block+0x5e8() ] 02a10014ee81 mutex_vector_enter+0x424(181bad0, fffe9bac45d5c95c, 6000e685bc8, 30001119340, 30001119340, 0) 02a10014ef31 zfs_zinactive+0x24(300a9196f00, 6000e685bc8, 6000e685a58, 6000e685980, 300a9196f28, 300c11d7b40) 02a10014efe1 zfs_inactive+0x168(6000e6859d8, 60001001ee8, 2a10014f948, 2, 0, 300a9196f00) 02a10014f091 fop_inactive+0x50(300c11d7b40, 60001001ee8, 2000, 60004511f00 , 1, 7b763864) 02a10014f151 do_dnlc_reduce_cache+0x210(0, 1853da0, 1863c70, 6000175c868, 18ab838, 60cbd) 02a10014f201 taskq_d_thread+0x88(60003f9f4a0, 32878c0, 6000100b520, 0 , 1636a70535248, 60003f9f4d0) 02a10014f2d1 thread_start+4(60003f9f4a0, 0, 6057b48, ffbffc6f , 4558505f5355, 5252) We are trying to obtain a mutex that is currently held by another thread trying to get memory. I suspect that the rest of the active vnodes are probably being held by the arc, as a side-effect of the fact that its holding onto the associated dnodes (and its holding onto these dnodes because they are in the same block as some still-dnlc-referenced vnode/dnode). Note that almost all of the arc memory is tied up in the MRU cache: > ARC_mru::print { list = 0x80 lsize = 0x200 size = 0x8a030400 hits = 0x16dcadb mtx = { _opaque = [ 0 ] } } Almost none of this is freeable, so the arc cannot shrink in size. -Mark ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Need input on implementing a ZFS layout
Darren Dunham wrote: Let's say the devices are named thus (and I'm making this up): /devices/../../SUNW,[EMAIL PROTECTED]/[EMAIL PROTECTED],0/WWN:sliceno [EMAIL PROTECTED] denotes the FLX380 frame, [0-6] [EMAIL PROTECTED],n denotes the virtual disk,LUN, [0-19],[0-3] How do I know that my stripes are rotated among [EMAIL PROTECTED], [EMAIL PROTECTED], ... [EMAIL PROTECTED] Today, you'd have to create each of the VDEVs to explicitly use one LUN from each array. There's no parameter for ZFS to pick them automatically. yep, something like: # zpool create mybigzpool \ raidz2 c10t0d0 c11t0d0 c12t0d0 c13t0d0 c14t0d0 c15t0d0 c16t0d0 \ raidz2 c10t0d1 c11t0d1 c12t0d1 c13t0d1 c14t0d1 c15t0d1 c16t0d1 \ ... raidz2 c10t0dN c11t0dN c12t0dN c13t0dN c14t0dN c15t0dN c16t0dN Obviously the c#t#d# would need to match your hardware, but you should be able to see the pattern. Later, you could add: # zpool add mybigzpool \ raidz2 c10t0dM c11t0dM c12t0dM c13t0dM c14t0dM c15t0dM c16t0dM -- richard ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Need input on implementing a ZFS layout
> Let's say the devices are named thus (and I'm making this up): > > /devices/../../SUNW,[EMAIL PROTECTED]/[EMAIL PROTECTED],0/WWN:sliceno > > [EMAIL PROTECTED] denotes the FLX380 frame, [0-6] > [EMAIL PROTECTED],n denotes the virtual disk,LUN, [0-19],[0-3] > > How do I know that my stripes are rotated among [EMAIL PROTECTED], [EMAIL > PROTECTED], > ... [EMAIL PROTECTED] Today, you'd have to create each of the VDEVs to explicitly use one LUN from each array. There's no parameter for ZFS to pick them automatically. -- Darren Dunham [EMAIL PROTECTED] Senior Technical Consultant TAOShttp://www.taos.com/ Got some Dr Pepper? San Francisco, CA bay area < This line left intentionally blank to confuse you. > ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Re: Re: ZFS forces system to paging to the point it is
> ::dnlc!wc 1048545 3145811 76522461 > This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] ZFS automatic snapshots SMF service updated
Hi All, I just posted version 0.6 of the automatic snapshots prototype on my web log. The new features are: * ZFS send/receive support * Multiple schedules per filesystem More at: http://blogs.sun.com/timf/entry/zfs_automatic_snapshots_now_with Note, this is just something I've been messing about with in my free time - it's not an official ZFS project and may not be bullet-proof enough for production. It works for me, but bug reports would be appreciated ;-) cheers, tim Tim Foster, Sun Microsystems Inc, Solaris Engineering Ops http://blogs.sun.com/timf ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Need input on implementing a ZFS layout
This is a most interesting thread. I'm a little be-fuddled, though. How will ZFS know to select the RAID-Z2 stripes from each FLX380, because if it stripes the (5+2) from the LUNS within one FLX380, this will not help if one frame goes irreplaceably out of service. Let's say the devices are named thus (and I'm making this up): /devices/../../SUNW,[EMAIL PROTECTED]/[EMAIL PROTECTED],0/WWN:sliceno [EMAIL PROTECTED] denotes the FLX380 frame, [0-6] [EMAIL PROTECTED],n denotes the virtual disk,LUN, [0-19],[0-3] How do I know that my stripes are rotated among [EMAIL PROTECTED], [EMAIL PROTECTED], ... [EMAIL PROTECTED] When I make pools I don't give the raw device name, and ZFS may not know it has selected its (5+2) stripes from one frame. This placement is for redundancy, but then will I be wasting the other 79 spindles in each frame? It's not just 7 giant disk. If I needed to see this for myself, or show it to a customer, what test may I set up to observe RAID-Z2 in action, so that I/O are observed to be spread among the 7 frames? I'm not yet comfortable with giving ZFS entire control over my disks without verification. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Re: Need input on implementing a ZFS layout
However performance will be much worse as data will be striped to only those mirrors already available. However is performance isn't an issue it could be interesting. This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Re: ZFS forces system to paging to the point it is
Hmmm, interesting data. See comments in-line: Robert Milkowski wrote: Yes, server has 8GB of RAM. Most of the time there's about 1GB of free RAM. bash-3.00# mdb 0 Loading modules: [ unix krtld genunix dtrace specfs ufs sd md ip sctp usba fcp fctl qlc ssd lofs zfs random logindmux ptm cpc nfs ipc ] arc::print { anon = ARC_anon mru = ARC_mru mru_ghost = ARC_mru_ghost mfu = ARC_mfu mfu_ghost = ARC_mfu_ghost size = 0x8b72ae00 We are referencing about 2.2GB of data from the ARC. p = 0xfe41b00 c = 0xfe51b00 We are trying to get down to our minimum target size of 16MB. So we are obviously feeling memory pressure and trying to react. c_min = 0xfe51b00 c_max = 0x1bca36000 ... >::kmastat cachebufbufbufmemory alloc alloc namesize in use totalin use succeed fail - -- -- -- - - - ... vn_cache 240 2400324 2507745 662691840 6307891 0 This is very interesting: 2.4 million vnodes are "active". ... zio_buf_512 512 2388292 2388330 1304346624 176134688 0 zio_buf_10241024 18 96 98304 17058709 0 zio_buf_15361536 0 30 49152 2791254 0 zio_buf_20482048 0 20 40960 1051435 0 zio_buf_25602560 0 33 90112 1716360 0 zio_buf_30723072 0 40122880 1902497 0 zio_buf_35843584 0225819200 3918593 0 zio_buf_40964096 3 34139264 20336550 0 zio_buf_51205120 0144737280 8932632 0 zio_buf_61446144 0 36221184 5274922 0 zio_buf_71687168 0 16114688 3350804 0 zio_buf_81928192 0 11 90112 9131264 0 zio_buf_10240 10240 0 12122880 2268700 0 zio_buf_12288 12288 0 8 98304 3258896 0 zio_buf_14336 14336 0 60860160 15853089 0 zio_buf_16384 16384 142762 142793 2339520512 74889652 0 zio_buf_20480 20480 0 6122880 1299564 0 zio_buf_24576 24576 0 5122880 1063597 0 zio_buf_28672 28672 0 6172032712545 0 zio_buf_32768 32768 0 4131072 1339604 0 zio_buf_40960 40960 0 6245760 1736172 0 zio_buf_49152 49152 0 4196608609853 0 zio_buf_57344 57344 0 5286720428139 0 zio_buf_65536 65536520522 34209792 8839788 0 zio_buf_73728 73728 0 5368640284979 0 zio_buf_81920 81920 0 5409600133392 0 zio_buf_90112 90112 0 6540672 96787 0 zio_buf_98304 98304 0 4393216133942 0 zio_buf_106496106496 0 5532480 91769 0 zio_buf_114688114688 0 5573440 72130 0 zio_buf_122880122880 0 5614400 52151 0 zio_buf_131072131072100107 14024704 7326248 0 dmu_buf_impl_t 328 2531066 2531232 863993856 237052643 0 dnode_t 648 2395209 2395212 1635131392 83304588 0 arc_buf_hdr_t128 142786 390852 50823168 155745359 0 arc_buf_t 40 142786 347333 14016512 160502001 0 zil_lwb_cache208 28468 98304 30507668 0 zfs_znode_cache 192 2388224 2388246 465821696 83149771 0 ... Because of all of those vnodes, we are seeing a lot of extra memory being used by ZFS: - about 1.5GB for the dnodes - another 800MB for dbufs - plus 1.3GB for the "bonus buffers" (not accounted for in the arc) - plus about 400MB for znodes This totals to another 4GB + .6GB held in vnodes The question is who is holding these vnodes in memory... Could you do a >::dnlc!wc and let me know what it comes back with? -Mark ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Need input on implementing a ZFS layout
+5 I've been saving my +1s for a few weeks now. ;) Richard Elling - PAE wrote: There is another option. I'll call it "grow into your storage." Pre-ZFS, for most systems you would need to allocate the storage well in advance of its use. For the 7xFLX380 case using SVM and UFS, you would typically setup the FLX380 LUNs, merge them together using SVM, and newfs. Growing is somewhat difficult for that size of systems because UFS has some smallish limits (16 TBytes per file system, less for older Solaris releases). Planning this in advance is challenging and the process for growing existing file systems or adding new file systems would need careful attention. By contrast, with ZFS we can add vdevs to the zpool on the fly to an existing zpool and the file systems can immediately use the new space. The reliability of devices is measured in operational hours. So, for a fixed reliability metric one way to improve your real-life happiness is to reduce the operational hours. Putting these together, it makes sense to only add disks as you need the space. Keep the disks turned off, until needed, to lengthen their life. In other words, grow into your storage. This doesn't work for everyone, or every situation, but ZFS makes it an easy, viable option to consider. -- richard ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Re: zoned datasets in zfs list
As user properties are coming then maybe when a fs is mounted in a local zone a user property would be set, like zone_mounted=test1. Perhaps during each mount such property would be created. In case with several zones just put names after , or something similar. ?? This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Re: How to destroy a pool wich you can't import because it is in faulted st
On September 6, 2006 7:19:32 AM -0700 Lieven De Geyndt <[EMAIL PROTECTED]> wrote: sorry guys ...RTF did the job [b]Legacy Mount Points[/b] That just means filesystems in the pool won't get mounted, not that the pool won't be imported. -frank ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Need input on implementing a ZFS layout
There is another option. I'll call it "grow into your storage." Pre-ZFS, for most systems you would need to allocate the storage well in advance of its use. For the 7xFLX380 case using SVM and UFS, you would typically setup the FLX380 LUNs, merge them together using SVM, and newfs. Growing is somewhat difficult for that size of systems because UFS has some smallish limits (16 TBytes per file system, less for older Solaris releases). Planning this in advance is challenging and the process for growing existing file systems or adding new file systems would need careful attention. By contrast, with ZFS we can add vdevs to the zpool on the fly to an existing zpool and the file systems can immediately use the new space. The reliability of devices is measured in operational hours. So, for a fixed reliability metric one way to improve your real-life happiness is to reduce the operational hours. Putting these together, it makes sense to only add disks as you need the space. Keep the disks turned off, until needed, to lengthen their life. In other words, grow into your storage. This doesn't work for everyone, or every situation, but ZFS makes it an easy, viable option to consider. -- richard ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Re: ZFS issues too much IOs
Lowering from default 64K to 16K turned into about 10x less read throutput! And similar factor for latency for nfs clients. For now I'll probably leave it as it is and later will do some comparisons with different settings. ps. very big thanks to Roch! I owe you! This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zoned datasets in zfs list
On Wed, Sep 06, 2006 at 04:52:48PM +0100, Dick Davies wrote: > > Oh God no. That's exactly what I wanted to avoid. > Why wouldn't you want it stored in the dataset, out of interest? There are a couple of reasons: - We don't want to re-create the same information in multiple places. Keeping both the zone configuration file and the on-disk state in sync would be difficult. - The association of a dataset to a zone is a property of the zone, not the dataset. This is particularly important if you want to migrate zones with delegated datasets between machines, or export and import your pool on a machine with a different zone configuration. - Unlike NFS options or mountpoints, the dataset is part of a larger zone configuration, and cannot replace the zone configuration file entirely. Ideally, we would have liked to not have the 'zoned' property at all, but it was necessary to enforce sane behavior between the global and local zone. > > I'm reading that as 'the only real use case for 1 dataset in multiple zones' > (sorry if I'm misunderstanding you)? > Yep. - Eric -- Eric Schrock, Solaris Kernel Development http://blogs.sun.com/eschrock ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Re: Re: Re: datasets,zones and mounts
No, remove all other datasets from zone config and just put: add dataset set name=telecom/oracle/production end and that's it. That way you will see all filesystem beneath tyelecom/oracle/production. Additionally in a zone production you will be able to create more file systems inside without changing zone configuration. Additionally snapshots and clones will work inside zone. You are not only not loosing granulity you are adding flexibity that way. Sorry for not being clear. This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zoned datasets in zfs list
On 06/09/06, Eric Schrock <[EMAIL PROTECTED]> wrote: On Wed, Sep 06, 2006 at 03:53:52PM +0100, Dick Davies wrote: > That's a bit nicer, thanks. > Still not that clear which zone they belong to though - would > it be an idea to add a 'zone' property be a string == zonename ? Yes, this is possible, but it's annoying because the actual owning zone isn't stored with the dataset (nor should it be). We'd have to grovel around every zone's configuration file, which is certainly doable, just annoying. Oh God no. That's exactly what I wanted to avoid. Why wouldn't you want it stored in the dataset, out of interest? In addition, it's possible (though not recommended) to have a single dataset in multiple zones. Ah Ok, that explains why a single string wouldn't cut it (although it sounds insane to me)! The only real use case would be a read-only, unmounted dataset whose snapshots could serve as a clone source for other delegated datasets. I'm reading that as 'the only real use case for 1 dataset in multiple zones' (sorry if I'm misunderstanding you)? -- Rasputin :: Jack of All Trades - Master of Nuns http://number9.hellooperator.net/ ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zoned datasets in zfs list
On Wed, Sep 06, 2006 at 08:34:26AM -0700, Eric Schrock wrote: > > Feel free to file an RFE. > Oops, found one already: 6313352 'zpool list' & 'zfs list' should add '-z' & '-Z' to identifier a zone - Eric -- Eric Schrock, Solaris Kernel Development http://blogs.sun.com/eschrock ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] creating zvols in a non-global zone (or 'Doctor, it hurts when I do this')
On Wed, Sep 06, 2006 at 04:23:32PM +0100, Dick Davies wrote: > > a) prevent attempts to create zvols in non-global zones > b) somehow allow it (?) or > c) Don't do That > > I vote for a) myself - should I raise an RFE? Yes, that was _supposed_ to be the original behavior, and I thought we had it working that way at one point. Apparently I'm imagining things, or it got broken somewhere along the way. Please file a bug. - Eric -- Eric Schrock, Solaris Kernel Development http://blogs.sun.com/eschrock ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Re: datasets,zones and mounts
On Wed, Sep 06, 2006 at 09:01:00AM -0400, Kenneth Mikelinich wrote: > Hi Robert -- Here are the outputs. I cannot seem to see the last isapps > dataset via zfs list. The non-global zone will be used to host a 10G > Oracle. Yes, this is definitely a bug somewhere. I'll try to reproduce this on a test machine and see what's going on. - Eric -- Eric Schrock, Solaris Kernel Development http://blogs.sun.com/eschrock ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zoned datasets in zfs list
On Wed, Sep 06, 2006 at 03:53:52PM +0100, Dick Davies wrote: > That's a bit nicer, thanks. > Still not that clear which zone they belong to though - would > it be an idea to add a 'zone' property be a string == zonename ? Yes, this is possible, but it's annoying because the actual owning zone isn't stored with the dataset (nor should it be). We'd have to grovel around every zone's configuration file, which is certainly doable, just annoying. In addition, it's possible (though not recommended) to have a single dataset in multiple zones. The only real use case would be a read-only, unmounted dataset whose snapshots could serve as a clone source for other delegated datasets. Since this is extremely rare, it would probably suffice to have a special string like "(shared)" to indicate that it is being shared between multiple zones. Feel free to file an RFE. - Eric -- Eric Schrock, Solaris Kernel Development http://blogs.sun.com/eschrock ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] creating zvols in a non-global zone (or 'Doctor, it hurts when I do this')
A colleague just asked if zfs delegation worked with zvols too. Thought I'd give it a go and got myself in a mess (tank/linkfixer is the delegated dataset): [EMAIL PROTECTED] / # zfs create -V 500M tank/linkfixer/foo cannot create device links for 'tank/linkfixer/foo': permission denied cannot create 'tank/linkfixer/foo': permission denied Ok, so we'll try a normal filesystem: [EMAIL PROTECTED] / # zfs create tank/linkfixer/foo cannot create 'tank/linkfixer/foo': dataset already exists [EMAIL PROTECTED] / # zfs list NAME USED AVAIL REFER MOUNTPOINT tank 2.09G 33.8G 24.5K legacy tank/linkfixer36.3M 9.96G 24.5K legacy tank/linkfixer/foo22.5K 9.96G 22.5K - [EMAIL PROTECTED] / # zfs destroy -f tank/linkfixer/foo cannot remove device links for 'tank/linkfixer/foo': permission denied [EMAIL PROTECTED] / # zfs list NAME USED AVAIL REFER MOUNTPOINT tank 2.09G 33.8G 24.5K legacy tank/linkfixer36.3M 9.96G 24.5K legacy tank/linkfixer/foo22.5K 9.96G 22.5K - I can destroy it ok from the global zone, and I know I could just create a top-level zvol and grant the zone access. Not sure if the 'fix' is : a) prevent attempts to create zvols in non-global zones b) somehow allow it (?) or c) Don't do That I vote for a) myself - should I raise an RFE? -- Rasputin :: Jack of All Trades - Master of Nuns http://number9.hellooperator.net/ ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zoned datasets in zfs list
That's a bit nicer, thanks. Still not that clear which zone they belong to though - would it be an idea to add a 'zone' property be a string == zonename ? On 06/09/06, Kenneth Mikelinich <[EMAIL PROTECTED]> wrote: zfs mount should show where all your datasets are mounted. I too was confused with the zfs list readout. On Wed, 2006-09-06 at 07:37, Dick Davies wrote: > Just did my first dataset delegation, so be gentle :) > > Was initially terrified to see that changes to the mountpoint in the non-global > zone were visible in the global zone. > > Then I realised it wasn't actually mounted (except in the delegated zone). > But I couldn't see any obvious indication that the dataset was delegated to > another zone in zfs list. > Eventually I found the 'zoned' property. Couple of thoughts: > > 1) would it be worth changing 'zfs list' to clarify where a dataset >is actually mounted? > 2) Is there any way to indicate _what_ zone a dataset is mounted in > (other than greppping the zones configuration)? > > -- > Rasputin :: Jack of All Trades - Master of Nuns > http://number9.hellooperator.net/ > ___ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss -- Rasputin :: Jack of All Trades - Master of Nuns http://number9.hellooperator.net/ ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Need input on implementing a ZFS layout
Oatway, Ted wrote: Thanks for the response Richard. Forgive my ignorance but the following questions come to mind as I read your response. I would then have to create 80 RAIDz(6+1) Volumes and the process of creating these Volumes can be scripted. But - 1) I would then have to create 80 mount points to mount each of these Volumes (?) No. In ZFS, you create a zpool which has the devices and RAID configurations. The file systems (plural) are then put in the zpool. You could have one file system, or thousands. Each file system, by default, will be in the heirarchy under the zpool name, or you can change it as you need. 2) I would have no load balancing across mount points and I would have to specifically direct the files to a mount point using an algorithm of some design ZFS will dynamically stripe across the sets. In traditional RAID terms, this is like RAID-1+0, RAID-5+0, or RAID-6+0. 3) A file landing on any one mount point would be constrained to the I/O of the underlying disk which would represent 1/80th of the potential available It would be spread across the 80 sets. 4) Expansion of the architecture, by adding in another single disk array, would be difficult and would probably be some form of data migration (?). For 800TB of data that would be unacceptable. It depends on how you do this. There are techniques for balancing which might work, but they have availability trade-offs because you are decreasing your diversity. I'm encouraged by the fact that they are planning ahead :-). Also, unlike a traditional disk array or LVM software, ZFS will only copy the data. For example, in SVM, if you replace a whole disk, the resync will copy the "data" for the whole disk. For ZFS, it knows what data is valid, and will only copy the valid data. Thus the resync time is based upon the size of the data, not the size of the disk. There are more nuances here, but that covers it to the first order. -- richard ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Need input on implementing a ZFS layout
> I would then have to create 80 RAIDz(6+1) Volumes and the process of > creating these Volumes can be scripted. But - > > 1) I would then have to create 80 mount points to mount each of these > Volumes (?) No. Each of the RAIDZs that you create can be combined into a single pool. Data written to the pool will stripe across all the RAIDz devices. > 2) I would have no load balancing across mount points and I would have > to specifically direct the files to a mount point using an algorithm of > some design > > 3) A file landing on any one mount point would be constrained to the I/O > of the underlying disk which would represent 1/80th of the potential > available See #1. > 4) Expansion of the architecture, by adding in another single disk > array, would be difficult and would probably be some form of data > migration (?). For 800TB of data that would be unacceptable. Today, you wouldn't be able to do it easily. In the future, you may be able to expand the RAIDz device. (or if you could remove a VDEV from a pool, you could rotate through and remove each of the RAIDz devices followed by an addition of a new (8-column) RAIDz). -- Darren Dunham [EMAIL PROTECTED] Senior Technical Consultant TAOShttp://www.taos.com/ Got some Dr Pepper? San Francisco, CA bay area < This line left intentionally blank to confuse you. > ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Re: How to destroy a pool wich you can't import because it is in faulted st
sorry guys ...RTF did the job [b]Legacy Mount Points[/b] You can manage ZFS file systems with legacy tools by setting the mountpoint property to legacy. Legacy file systems must be managed through the mount and umount commands and the /etc/vfstab file. ZFS does not automatically mount legacy file systems on boot, and the ZFS mount and umount command do not operate on datasets of this type. The following examples show how to set up and manage a ZFS dataset in legacy mode: # zfs set mountpoint=legacy tank/home/eschrock # mount -F zfs tank/home/eschrock /mnt In particular, if you have set up separate ZFS /usr or /var file systems, you must indicate that theyare legacy file systems. In addition, you must mount them by creating entries in the /etc/vfstab file. Otherwise, the system/filesystem/local service enters maintenance mode when the system boots. To automatically mount a legacy file system on boot, you must add an entry to the /etc/vfstab file. The following example shows what the entry in the /etc/vfstab file might look like: #device device mount FSfsck mount mount #to mount to fsckpoint type pass at boot options # tank/home/eschrock - /mnt zfs - yes - Note that the device to fsck and fsck pass entries are set to -. This syntax is because the fsck command is not applicable to ZFS file systems. For more information regarding data integrity and the lack of need for fsck in ZFS This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Re: Re: How to destroy a pool wich you can't import
This could still corrupt the pool. Probably the customer has to write its own tool to import a pool using libzfs and not creating zpool.cache. Eventually just after pool is imported remove zpool.cache - I'm not sure but it should work. This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
RE: [zfs-discuss] Re: Re: datasets,zones and mounts
Thanks. I will try this out Ken Mikelinich Computer Operations Manager Telecommunications and Client Services University of New Hampshire 603.862.4220 -Original Message- From: Dick Davies [mailto:[EMAIL PROTECTED] Sent: Wednesday, September 06, 2006 9:54 AM To: Mikelinich, Ken Cc: zfs-discuss@opensolaris.org Subject: Re: [zfs-discuss] Re: Re: datasets,zones and mounts On 06/09/06, Kenneth Mikelinich <[EMAIL PROTECTED]> wrote: > Are you suggesting that I not get too granular with datasets and use a > higher level one versus several? I tihnk what he's saying is you should only have to delegate one dataset (telecom/oracle/production, for example), and all the 'child' datasets can be created/administered/snapshotted etc. in the non -global zone itself. -- Rasputin :: Jack of All Trades - Master of Nuns http://number9.hellooperator.net/ ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Re: Re: datasets,zones and mounts
On 06/09/06, Kenneth Mikelinich <[EMAIL PROTECTED]> wrote: Are you suggesting that I not get too granular with datasets and use a higher level one versus several? I tihnk what he's saying is you should only have to delegate one dataset (telecom/oracle/production, for example), and all the 'child' datasets can be created/administered/snapshotted etc. in the non -global zone itself. -- Rasputin :: Jack of All Trades - Master of Nuns http://number9.hellooperator.net/ ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Re: How to destroy a pool wich you can't import because it is in faulted st
Lieven De Geyndt wrote: zpool create -R did his job . Thanks for the tip . Is ther a way to disable the auto mount when you boot a system ? The customer has some kind of poor mans cluster . 2 systems has access to a SE3510 with ZFS . System A was powered-off as test , system B did an import of the pools . When system A rebooted , this system tries to import his pools , so 2 systems are accessing the same pool . Probably this caused a corruption in his pool . So how to disable automount of zfs pools ? Oh heck PMC 0.0.0alpha again :( How about # zfs set mountpoint=none fsname James C. McPherson ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Re: Re: datasets,zones and mounts
Hmmm. I thought I was doing this via zonecfg -z production, which zonecfg is run from the global zone. add dataset set name=telecom/oracle/production/oraapp end ... repeat add dataset set name=telecom/oracle/production/isapps end commit exit The zone took all the datasets (shown in the earlier export), yet the zone only reveals/mounts the first seven when it runs. Are you suggesting that I not get too granular with datasets and use a higher level one versus several? thanks ken On Wed, 2006-09-06 at 09:21, Robert Milkowski wrote: > Well, that's interesting. Looks like some limit/bug here. However whatever > the limit is have you considered to add dataset into a zone? That way you can > actually create new file systems as needed inside a zone without changing > zone configuration, etc. You can also utilize snapshots, clones inside a > zone. That's what I did. That way DB admins can create file systems, > snapshots, etc. > > add dataset > set name=pool/dataset > end > > > This message posted from opensolaris.org > ___ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Re: Re: datasets,zones and mounts
Hmmm. I thought I was doing this via zonecfg -z production, which zonecfg is run from the global zone. add dataset set name=telecom/oracle/production/oraapp end ... repeat add dataset set name=telecom/oracle/production/isapps end commit exit The zone took all the datasets (shown in the earlier export), yet the zone only reveals/mounts the first seven when it runs. Are you suggesting that I not get too granular with datasets and use a higher level one versus several? thanks ken On Wed, 2006-09-06 at 09:21, Robert Milkowski wrote: > Well, that's interesting. Looks like some limit/bug here. However whatever > the limit is have you considered to add dataset into a zone? That way you can > actually create new file systems as needed inside a zone without changing > zone configuration, etc. You can also utilize snapshots, clones inside a > zone. That's what I did. That way DB admins can create file systems, > snapshots, etc. > > add dataset > set name=pool/dataset > end > > > This message posted from opensolaris.org > ___ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Re: Recommendation ZFS on StorEdge 3320
Wee Yeh Tan writes: > On 9/5/06, Torrey McMahon <[EMAIL PROTECTED]> wrote: > > This is simply not true. ZFS would protect against the same type of > > errors seen on an individual drive as it would on a pool made of HW raid > > LUN(s). It might be overkill to layer ZFS on top of a LUN that is > > already protected in some way by the devices internal RAID code but it > > does not "make your data susceptible to HW errors caused by the storage > > subsystem's RAID algorithm, and slow down the I/O". > > & Roch's recommendation to leave at least 1 layer of redundancy to ZFS > allows the extension of ZFS's own redundancy features for some truely > remarkable data reliability. > > Perhaps, the question should be how one could mix them to get the best > of both worlds instead of going to either extreme. > > > True, ZFS can't manage past the LUN into the array. Guess what? ZFS > > can't get past the disk drive firmware eitherand thats a good thing > > for all parties involved. > > > -- > Just me, > Wire ... > ___ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss Thinking some more about this. If your requirements does mandate some form of mirroring, then it truly seems that ZFS should take that in charge if only because of the self-healing characteristics. So I feel the storage array's job is to export low latency Luns to ZFS. I'd be happy to live with those simple Luns but I guess some storage will just refuse to export non-protected luns. Now we can definitively take advantage of the Array's capability of exporting highly resilient Luns; RAID-5 seems to fit the bill rather well here. Even an 9+1 luns will be quite resilient and have a low block overhead. So we benefit from the arrays resiliency as well as it's low latency characteristics. And we mirror data at the ZFS level which means great performance and great data integrity and great availability. Note that ZFS write characteristics (all sequential) means that we will commonly be filling full stripes on the luns thus avoiding the partial stripe performance pitfall. If you must shy away from any form of mirroring, then it's either stripe your raid-5 luns (performance edge for those who live dangerously) or raid-z around those raid-5 luns (lower cost, survives lun failures). -r ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Re: How to destroy a pool wich you can't import because it is in faulted st
zpool create -R did his job . Thanks for the tip . Is ther a way to disable the auto mount when you boot a system ? The customer has some kind of poor mans cluster . 2 systems has access to a SE3510 with ZFS . System A was powered-off as test , system B did an import of the pools . When system A rebooted , this system tries to import his pools , so 2 systems are accessing the same pool . Probably this caused a corruption in his pool . So how to disable automount of zfs pools ? This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Re: Re: datasets,zones and mounts
Well, that's interesting. Looks like some limit/bug here. However whatever the limit is have you considered to add dataset into a zone? That way you can actually create new file systems as needed inside a zone without changing zone configuration, etc. You can also utilize snapshots, clones inside a zone. That's what I did. That way DB admins can create file systems, snapshots, etc. add dataset set name=pool/dataset end This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Re: datasets,zones and mounts
Hi Robert -- Here are the outputs. I cannot seem to see the last isapps dataset via zfs list. The non-global zone will be used to host a 10G Oracle. //** From the global zone ** //** Pool is named telecom # zonecfg -z production export create -b set zonepath=/zones/production set autoboot=true add inherit-pkg-dir set dir=/lib end add inherit-pkg-dir set dir=/platform end add inherit-pkg-dir set dir=/sbin end add inherit-pkg-dir set dir=/usr end add dataset set name=telecom/oracle/production/backup end add dataset set name=telecom/oracle/production/logs end add dataset set name=telecom/oracle/production/oraapp end add dataset set name=telecom/oracle/production/oradata end add dataset set name=telecom/oracle/production/oradmp end add dataset set name=telecom/oracle/production/oramirror1 end add dataset set name=telecom/oracle/production/oramirror2 end add dataset set name=telecom/oracle/production/isapps end //**from my non-global zone called production /** I have used zfs set mountpoint= to adjust the mounts where i need them /** Telecom and telecom/oracle remain accessible only to the global as desired # zfs list NAME USED AVAIL REFER MOUNTPOINT telecom799K 669G51K /telecom telecom/oracle 490K 669G49K /telecom/oracle telecom/oracle/production 441K 669G49K /telecom/oracle/production telecom/oracle/production/backup49K 20.0G49K /backup telecom/oracle/production/logs49K 5.00G49K /logs telecom/oracle/production/oraapp49K 15.0G49K /oraapp telecom/oracle/production/oradata49K 20.0G49K /oradata telecom/oracle/production/oradmp49K 2.00G49K /oradmp telecom/oracle/production/oramirror149K 5.00G49K /oramirror1 telecom/oracle/production/oramirror249K 5.00G49K /oramirror2 If I do a zfs mount in either the global zone or the non-global zone, I do not see telecom/oracle/production/isapps. The zoned flag is set. Regards and Thanks for your help Ken University of New Hampshire On Wed, 2006-09-06 at 08:36, Robert Milkowski wrote: > Hi. > > Can you post zonecfg -z export and zfs list in that XXX zone? > > > This message posted from opensolaris.org > ___ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Re: datasets,zones and mounts
Hi. Can you post zonecfg -z export and zfs list in that XXX zone? This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Re: ZFS forces system to paging to the point it is unresponsive
It looks like I discovered a workaround. I've got another zpool within rg in SC. The other zpool does not have production data (yet) so I can switch it between nodes freely. By doing this every 3 minutes I can stay safe on free memory, at least so far. I guess it frees some ARC cache. What is not that clear however is that when I created another pool and just did export/import sometimes I get back some memory sometimes not. At the end it works worse than switching entire RG which actually means not only exporting zpool but also restarting nfsd. Looks like those two helps. This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] datasets,zones and mounts
Hi this question is along the lines of datasets and number of zfs file systems available per zone. I suspect I am missing something obvious. We have added 8 datasets to one non-global Zone. While logged in and doing a ZFS list in that zone, I am only able to see the first 7 available ZFS file systems. The 8th one is not available; however a zonecfg:info from the Global Zone shows it is indeed added to the zone. How do I go about making a Zone see more (eg mounting) more than 7 ZFS file systems? Note my datasets all are originating from one pool sitting on top of a StorEdge 3320 RAID. Regards Ken ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zoned datasets in zfs list
zfs mount should show where all your datasets are mounted. I too was confused with the zfs list readout. On Wed, 2006-09-06 at 07:37, Dick Davies wrote: > Just did my first dataset delegation, so be gentle :) > > Was initially terrified to see that changes to the mountpoint in the > non-global > zone were visible in the global zone. > > Then I realised it wasn't actually mounted (except in the delegated zone). > But I couldn't see any obvious indication that the dataset was delegated to > another zone in zfs list. > Eventually I found the 'zoned' property. Couple of thoughts: > > 1) would it be worth changing 'zfs list' to clarify where a dataset >is actually mounted? > 2) Is there any way to indicate _what_ zone a dataset is mounted in > (other than greppping the zones configuration)? > > -- > Rasputin :: Jack of All Trades - Master of Nuns > http://number9.hellooperator.net/ > ___ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] How to destroy a pool wich you can't import because it is in faulted state
Lieven De Geyndt wrote: When a pool is in a faulted state , you can't import it . Even -f fails . When you to decide to recreate the pool , you cannot execute zpool destroy , because it is not imported . Also -f does not work . Any idea how to get out of this situation ? try something like zpool create -R /alternate_root newpoolname vdevlist You might need to add a "-f", but try it without "-f" first. cheers, James C. McPherson ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Re: How to destroy a pool wich you can't import because it is in faulted st
Hi. Just re-create it or create new pool with disks from the old one and use -f flag. This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] zoned datasets in zfs list
Just did my first dataset delegation, so be gentle :) Was initially terrified to see that changes to the mountpoint in the non-global zone were visible in the global zone. Then I realised it wasn't actually mounted (except in the delegated zone). But I couldn't see any obvious indication that the dataset was delegated to another zone in zfs list. Eventually I found the 'zoned' property. Couple of thoughts: 1) would it be worth changing 'zfs list' to clarify where a dataset is actually mounted? 2) Is there any way to indicate _what_ zone a dataset is mounted in (other than greppping the zones configuration)? -- Rasputin :: Jack of All Trades - Master of Nuns http://number9.hellooperator.net/ ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] How to destroy a pool wich you can't import because it is in faulted state
When a pool is in a faulted state , you can't import it . Even -f fails . When you to decide to recreate the pool , you cannot execute zpool destroy , because it is not imported . Also -f does not work . Any idea how to get out of this situation ? This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] How to flush ARC cache?
Hi. Is there a waf to safely flush ARC cache (get most of ZFS memory back to system)? ps. not by export/import This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss