Re: [zfs-discuss] Re: ZFS uses 1.1GB more space, reports conflicting information...

2006-09-06 Thread Wee Yeh Tan

On 9/6/06, UNIX admin <[EMAIL PROTECTED]> wrote:

Yes, the man page says that. However, it is possible to mix disks of different 
sizes in a RAIDZ, and this works. Why does it work? Because RAIDZ stripes are 
dynamic in size. From that I infer that disks can be any size because the 
stripes can be any size. Should one or more of the smaller disks become full, 
the stripe information is written across the remaining space on the rest of the 
bigger disks. Or should be anyway. If you think about it, it makes sense.


You are right that it can be done in theory.  The practical difficulty
is to do it efficiently.  My raidz pool tend to have a strong
preference to utilize all the disks so disks with different sizes
might result in smaller column widths later.  I'm not sure if that
spells any performance issues but at least the capacity difference
between the 2 largest disk will not be usable.

Is maximising use of capacity of all available disk in the pipeline
for RAID-Z or is it to be avoided intentionally?


--
Just me,
Wire ...
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Overview (rollup) of recent activity on zfs-discuss

2006-09-06 Thread Eric Boutilier

For background on what this is, see:

http://www.opensolaris.org/jive/message.jspa?messageID=24416#24416
http://www.opensolaris.org/jive/message.jspa?messageID=25200#25200

=
zfs-discuss 08/16 - 08/31
=

Size of all threads during period:

Thread size Topic
--- -
 22   ZFS + rsync, backup on steroids.
 13   SCSI synchronize cache cmd
 13   Niagara and ZFS compression?
 12   ZFS Boot Disk
 10   zpool status panics server
 10   ZFS & se6920
  9   Oracle on ZFS
  8   zpool core dumped
  8   in-kernel gzip compression
  8   Significant "pauses" during zfs writes
  7   Home Server with ZFS
  7   3510 - some new tests
  6   unaccounted for daily growth in ZFS disk space usage
  6   ZFS write performance problem with compression set to ON
  6   Why is ZFS raidz slower than simple ufs mount?
  6   Tape backup
  6   SPEC SFS97 benchmark of ZFS,UFS,VxFS
  6   Proposal: user-defined properties
  6   Porting ZFS file system to FreeBSD.
  6   Need Help: didn't create the pool as radiz but stripes
  6   Information on ZFS API?
  5   zpool import: snv_33 to S10 6/06
  5   multi-layer ZFS filesystems and exporting: my stupid question for 
the day
  5   commercial backup software and zfs
  5   ZFS compression
  5   ZFS Filesytem Corrpution
  5   Issue with zfs snapshot replication from version2 to version3 
pool.
  5   File level compression
  4   zpool import dropped a core.
  4   zpool import - cannot mount [...] directory is not empty
  4   zpool hangs
  4   destroyed pools signatures
  4   [fbsd] Porting ZFS file system to FreeBSD.
  4   ZFS with expanding LUNs
  4   ZFS web admin - No items found.
  4   ZFS and very large directories
  4   ZFS Load-balancing over vdevs vs. real disks?
  4   Storage Compatibilty list
  4   Find the difference between two snapshots
  4   Encryption on ZFS / Disk Usage
  4   Compression with fixed record sizes
  3   zpool iostat, scrubbing increases used disk space
  3   zpool import/export
  3   system unresponsive after issuing a zpool attach
  3   problem with zfs receive -i
  3   does zfs can be failover between two mechine?
  3   does anybody port the zfs webadmin to webmin?
  3   can't create snapshot
  3   ZFS questions with mirrors
  3   ZFS compression / space efficiency
  3   ZFS Performance compared to UFS & VxFS
  3   Want to try ZFS, but "format" got  message
  3   Newbie questions about drive problems
  3   Can a zfs storage pool be imported readonly?
  2   zfs and vmware
  2   new ZFS links page
  2   neopath vs automount
  2   migrating data across boxes
  2   libzfs question
  2   delete acl not working on zfs.v3?
  2   ZFS 'quot' command
  2   Tunable parameter to zfs memory use
  2   Solaris 6/06 ZFS and OpenSolaris ZFS
  2   Interesting zfs destroy failure
  1   zpool status inconsistent after user error?
  1   zfs list - unexpected error 17 at line 1684
  1   space accounting with RAID-Z
  1   pool ID
  1   fdatasync
  1   ZFS Performance compared to UFS & VxFS - offtopic
  1   ZFS ACL: append_data didn't do what I expected
  1   Sol 10 x86_64 intermittent SATA device locks up server
  1   Question on Zones and memory usage (65120349)
  1   Query on ZFS
  1   Equivalent of command quot
  1   Apple Time Machine


Posting activity by person for period:

# of posts  By
--   --
 35   rmilkowski at task.gda.pl (robert milkowski)
 17   ahrens at eng.sun.com (matthew ahrens)
 15   eric.schrock at sun.com (eric schrock)
 11   roch.bourbonnais at sun.com (roch)
 11   fcusack at fcusack.com (frank cusack)
 10   matthew.ahrens at sun.com (matthew ahrens)
  9   jamesd.wi at gmail.com (james dickens)
  8   george.wilson at sun.com (george wilson)
  7   rasputnik at gmail.com (dick davies)
  6   mgerdts at gmail.com (mike gerdts)
  6   jmlittle at gmail.com (joe little)
  6   anantha.srirama at cdc.hhs.gov (anantha n. srirama)
  5   torrey.mcmahon at sun.com (torrey mcmahon)
  5   richard.elling at sun.com (richard elling - pae)
  5   nicolas.williams at sun.com (nicolas williams)
  5   michael.schuster at sun.com (michael schuster - sun microsystems)
  

Re: [zfs-discuss] Re: Re: ZFS forces system to paging to the point it is

2006-09-06 Thread Mark Maybee

Robert Milkowski wrote:



On Wed, 6 Sep 2006, Mark Maybee wrote:


Robert Milkowski wrote:


::dnlc!wc



 1048545 3145811 76522461


Well, that explains half your problem... and maybe all of it:




After I reduced vdev prefetch from 64K to 8K for last few hours system 
is working properly without workaround and free memory stays at about 1GB.


Reducing vdev prefetch to 8K alse reduced read thruoutput 10x.

I belive this is somehow related - maybe vdev cache was so aggressive (I 
got 40-100MB/s of reads) and consuming memory so fast that thread which 
is supposed to regain some memory couldn't keep up?


I suppose, although the data volume doesn't seem that high... maybe you
are just operating at the hairy edge here.  Anyway, I have filed a bug
to track this issue:

6467963 do_dnlc_reduce_cache() can be blocked by ZFS_OBJ_HOLD_ENTER()

-Mark
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re: Recommendation ZFS on StorEdge 3320

2006-09-06 Thread Torrey McMahon

Roch - PAE wrote:

Thinking some more about this. If your requirements does
mandate some form of mirroring, then it truly seems that ZFS 
should take that in charge if only because of the

self-healing characteristics. So I feel the storage array's
job is to export low latency Luns to ZFS.
  



The hard part is getting a set of simple requirements. As you go into 
more complex data center environments you get hit with older Solaris 
revs, other OSs, SOX compliance issues, etc. etc. etc. The world where 
most of us seem to be playing with ZFS is on the lower end of the 
complexity scale. Sure, throw your desktop some fast SATA drives. No 
problem. Oh wait, you've got ten Oracle DBs on three E25Ks that need to 
be backed up every other blue moon ...


I agree with the general idea that an array, be it one disk or some raid 
combination, should simply export low latency LUNs. However, its the 
features offered by the array - Like site to site replication - used to 
meet more complex requirements that literally slow things down. In many 
cases you'll see years old operational procedures causing those low 
latency LUNs to slow down even more. Something really hard to get a 
customer to undo because a new fangled file system is out. ;)



I'd be happy to live with those simple Luns but I guess some
storage will just  refuse to export non-protected  luns. Now
we can definitively take advantage of the Array's capability
of exporting highly resilient Luns;  RAID-5 seems to fit the
bill  rather   well here. Even  an 9+1   luns will  be quite
resilient and have a low block overhead.
  



I think 99x0 used to do 3+1 only. Now it's 7+1 if I recall. Close enough 
I suppose.

So we benefit from the arrays resiliency as well as it's low
latency characteristics. And we mirror data at the ZFS level 
which means great performance and great data integrity and

great availability.

Note that ZFS  write characteristics (all  sequential) means
that  we will commonly be filling  full  stripes on the luns
thus avoiding the partial stripe performance pitfall.



One thing comes to mind in that case. Many arrays do sequential detect 
on the blocks that come in to the front end ports.
If things get split up to much or out of order or array characteristic here> then you could induce more latency as the 
array does cartwheels trying to figure out whats going on.



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re: Re: ZFS forces system to paging to the point it is

2006-09-06 Thread Mark Maybee

Robert Milkowski wrote:

::dnlc!wc


 1048545 3145811 76522461


Well, that explains half your problem... and maybe all of it:

We have a thread that *should* be trying to free up these entries
in the DNLC, however it appears to be blocked:

stack pointer for thread 2a10014fcc0: 2a10014edd1
[ 02a10014edd1 turnstile_block+0x5e8() ]
  02a10014ee81 mutex_vector_enter+0x424(181bad0, fffe9bac45d5c95c,
  6000e685bc8, 30001119340, 30001119340, 0)
  02a10014ef31 zfs_zinactive+0x24(300a9196f00, 6000e685bc8, 
6000e685a58,

  6000e685980, 300a9196f28, 300c11d7b40)
  02a10014efe1 zfs_inactive+0x168(6000e6859d8, 60001001ee8, 
2a10014f948, 2,

  0, 300a9196f00)
  02a10014f091 fop_inactive+0x50(300c11d7b40, 60001001ee8, 2000, 
60004511f00  , 1, 7b763864)
  02a10014f151 do_dnlc_reduce_cache+0x210(0, 1853da0, 1863c70, 
6000175c868,

  18ab838, 60cbd)
  02a10014f201 taskq_d_thread+0x88(60003f9f4a0, 32878c0, 
6000100b520, 0

  , 1636a70535248, 60003f9f4d0)
  02a10014f2d1 thread_start+4(60003f9f4a0, 0, 6057b48, 
ffbffc6f

  , 4558505f5355, 5252)

We are trying to obtain a mutex that is currently held by another
thread trying to get memory.

I suspect that the rest of the active vnodes are probably being held
by the arc, as a side-effect of the fact that its holding onto the
associated dnodes (and its holding onto these dnodes because they are
in the same block as some still-dnlc-referenced vnode/dnode).  Note
that almost all of the arc memory is tied up in the MRU cache:

> ARC_mru::print
{
list = 0x80
lsize = 0x200
size = 0x8a030400
hits = 0x16dcadb
mtx = {
_opaque = [ 0 ]
}
}
Almost none of this is freeable, so the arc cannot shrink in size.

-Mark
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Need input on implementing a ZFS layout

2006-09-06 Thread Richard Elling - PAE

Darren Dunham wrote:

Let's say the devices are named thus (and I'm making this up):

/devices/../../SUNW,[EMAIL PROTECTED]/[EMAIL PROTECTED],0/WWN:sliceno

[EMAIL PROTECTED] denotes the FLX380 frame, [0-6]
[EMAIL PROTECTED],n denotes the virtual disk,LUN, [0-19],[0-3]

How do I know that my stripes are rotated among [EMAIL PROTECTED], [EMAIL 
PROTECTED],
... [EMAIL PROTECTED] 


Today, you'd have to create each of the VDEVs to explicitly use one LUN
from each array.  There's no parameter for ZFS to pick them
automatically.


yep, something like:
# zpool create mybigzpool \
  raidz2 c10t0d0 c11t0d0 c12t0d0 c13t0d0 c14t0d0 c15t0d0 c16t0d0 \
  raidz2 c10t0d1 c11t0d1 c12t0d1 c13t0d1 c14t0d1 c15t0d1 c16t0d1 \
...
  raidz2 c10t0dN c11t0dN c12t0dN c13t0dN c14t0dN c15t0dN c16t0dN

Obviously the c#t#d# would need to match your hardware, but you should
be able to see the pattern.

Later, you could add:
# zpool add mybigzpool \
  raidz2 c10t0dM c11t0dM c12t0dM c13t0dM c14t0dM c15t0dM c16t0dM

 -- richard
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Need input on implementing a ZFS layout

2006-09-06 Thread Darren Dunham
> Let's say the devices are named thus (and I'm making this up):
> 
> /devices/../../SUNW,[EMAIL PROTECTED]/[EMAIL PROTECTED],0/WWN:sliceno
> 
> [EMAIL PROTECTED] denotes the FLX380 frame, [0-6]
> [EMAIL PROTECTED],n denotes the virtual disk,LUN, [0-19],[0-3]
> 
> How do I know that my stripes are rotated among [EMAIL PROTECTED], [EMAIL 
> PROTECTED],
> ... [EMAIL PROTECTED] 

Today, you'd have to create each of the VDEVs to explicitly use one LUN
from each array.  There's no parameter for ZFS to pick them
automatically.

-- 
Darren Dunham   [EMAIL PROTECTED]
Senior Technical Consultant TAOShttp://www.taos.com/
Got some Dr Pepper?   San Francisco, CA bay area
 < This line left intentionally blank to confuse you. >
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Re: Re: ZFS forces system to paging to the point it is

2006-09-06 Thread Robert Milkowski
> ::dnlc!wc
 1048545 3145811 76522461
>
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] ZFS automatic snapshots SMF service updated

2006-09-06 Thread Tim Foster
Hi All,

I just posted version 0.6 of the automatic snapshots prototype on my web
log. The new features are:

 * ZFS send/receive support
 * Multiple schedules per filesystem

More at:
http://blogs.sun.com/timf/entry/zfs_automatic_snapshots_now_with


Note, this is just something I've been messing about with in my free
time - it's not an official ZFS project and may not be bullet-proof
enough for production.

It works for me, but bug reports would be appreciated ;-)

cheers,
tim

Tim Foster, Sun Microsystems Inc, Solaris Engineering Ops
http://blogs.sun.com/timf

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Need input on implementing a ZFS layout

2006-09-06 Thread Christine Tran
This is a most interesting thread.  I'm a little be-fuddled, though. 
How will ZFS know to select the RAID-Z2 stripes from each FLX380, 
because if it stripes the (5+2) from the LUNS within one FLX380, this 
will not help if one frame goes irreplaceably out of service.


Let's say the devices are named thus (and I'm making this up):

/devices/../../SUNW,[EMAIL PROTECTED]/[EMAIL PROTECTED],0/WWN:sliceno

[EMAIL PROTECTED] denotes the FLX380 frame, [0-6]
[EMAIL PROTECTED],n denotes the virtual disk,LUN, [0-19],[0-3]

How do I know that my stripes are rotated among [EMAIL PROTECTED], [EMAIL PROTECTED], ... [EMAIL PROTECTED] 
 When I make pools I don't give the raw device name, and ZFS may not 
know it has selected its (5+2) stripes from one frame.  This placement 
is for redundancy, but then will I be wasting the other 79 spindles in 
each frame?  It's not just 7 giant disk.


If I needed to see this for myself, or show it to a customer, what test 
may I set up to observe RAID-Z2 in action, so that I/O are observed to 
be spread among the 7 frames?  I'm not yet comfortable with giving ZFS 
entire control over my disks without verification.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Re: Need input on implementing a ZFS layout

2006-09-06 Thread Robert Milkowski
However performance will be much worse as data will be striped to only those 
mirrors already available.

However is performance isn't an issue it could be interesting.
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re: ZFS forces system to paging to the point it is

2006-09-06 Thread Mark Maybee

Hmmm, interesting data.  See comments in-line:

Robert Milkowski wrote:

Yes, server has 8GB of RAM.
Most of the time there's about 1GB of free RAM.

bash-3.00# mdb 0
Loading modules: [ unix krtld genunix dtrace specfs ufs sd md ip sctp usba fcp 
fctl qlc ssd lofs zfs random logindmux ptm cpc nfs ipc ]


arc::print


{
anon = ARC_anon
mru = ARC_mru
mru_ghost = ARC_mru_ghost
mfu = ARC_mfu
mfu_ghost = ARC_mfu_ghost
size = 0x8b72ae00


We are referencing about 2.2GB of data from the ARC.


p = 0xfe41b00
c = 0xfe51b00


We are trying to get down to our minimum target size of 16MB.
So we are obviously feeling memory pressure and trying to react.


c_min = 0xfe51b00
c_max = 0x1bca36000

...

>::kmastat

cachebufbufbufmemory alloc alloc
namesize in use  totalin use   succeed  fail

- -- -- -- - - -
...

vn_cache 240 2400324 2507745 662691840   6307891 0


This is very interesting: 2.4 million vnodes are "active".

...

zio_buf_512  512 2388292 2388330 1304346624 176134688 0
zio_buf_10241024 18 96 98304  17058709 0
zio_buf_15361536  0 30 49152   2791254 0
zio_buf_20482048  0 20 40960   1051435 0
zio_buf_25602560  0 33 90112   1716360 0
zio_buf_30723072  0 40122880   1902497 0
zio_buf_35843584  0225819200   3918593 0
zio_buf_40964096  3 34139264  20336550 0
zio_buf_51205120  0144737280   8932632 0
zio_buf_61446144  0 36221184   5274922 0
zio_buf_71687168  0 16114688   3350804 0
zio_buf_81928192  0 11 90112   9131264 0
zio_buf_10240  10240  0 12122880   2268700 0
zio_buf_12288  12288  0  8 98304   3258896 0
zio_buf_14336  14336  0 60860160  15853089 0
zio_buf_16384  16384 142762 142793 2339520512  74889652 0
zio_buf_20480  20480  0  6122880   1299564 0
zio_buf_24576  24576  0  5122880   1063597 0
zio_buf_28672  28672  0  6172032712545 0
zio_buf_32768  32768  0  4131072   1339604 0
zio_buf_40960  40960  0  6245760   1736172 0
zio_buf_49152  49152  0  4196608609853 0
zio_buf_57344  57344  0  5286720428139 0
zio_buf_65536  65536520522  34209792   8839788 0
zio_buf_73728  73728  0  5368640284979 0
zio_buf_81920  81920  0  5409600133392 0
zio_buf_90112  90112  0  6540672 96787 0
zio_buf_98304  98304  0  4393216133942 0
zio_buf_106496106496  0  5532480 91769 0
zio_buf_114688114688  0  5573440 72130 0
zio_buf_122880122880  0  5614400 52151 0
zio_buf_131072131072100107  14024704   7326248 0
dmu_buf_impl_t   328 2531066 2531232 863993856 237052643 0
dnode_t  648 2395209 2395212 1635131392  83304588 0
arc_buf_hdr_t128 142786 390852  50823168 155745359 0
arc_buf_t 40 142786 347333  14016512 160502001 0
zil_lwb_cache208 28468 98304  30507668 0
zfs_znode_cache  192 2388224 2388246 465821696  83149771 0

...
Because of all of those vnodes, we are seeing a lot of extra memory
being used by ZFS:
- about 1.5GB for the dnodes
- another 800MB for dbufs
- plus 1.3GB for the "bonus buffers" (not accounted for in the arc)
- plus about 400MB for znodes

This totals to another 4GB + .6GB held in vnodes

The question is who is holding these vnodes in memory... Could you do a
>::dnlc!wc
and let me know what it comes back with?

-Mark
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Need input on implementing a ZFS layout

2006-09-06 Thread Torrey McMahon

+5

I've been saving my +1s for a few weeks now. ;)

Richard Elling - PAE wrote:

There is another option.  I'll call it "grow into your storage."
Pre-ZFS, for most systems you would need to allocate the storage well
in advance of its use.  For the 7xFLX380 case using SVM and UFS, you
would typically setup the FLX380 LUNs, merge them together using SVM,
and newfs.  Growing is somewhat difficult for that size of systems
because UFS has some smallish limits (16 TBytes per file system, less
for older Solaris releases).  Planning this in advance is challenging
and the process for growing existing file systems or adding new file
systems would need careful attention.

By contrast, with ZFS we can add vdevs to the zpool on the fly to an
existing zpool and the file systems can immediately use the new space.

The reliability of devices is measured in operational hours.  So, for
a fixed reliability metric one way to improve your real-life happiness
is to reduce the operational hours.

Putting these together, it makes sense to only add disks as you need
the space.  Keep the disks turned off, until needed, to lengthen their
life.  In other words, grow into your storage.  This doesn't work for
everyone, or every situation, but ZFS makes it an easy, viable option
to consider.
 -- richard
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Re: zoned datasets in zfs list

2006-09-06 Thread Robert Milkowski
As user properties are coming then maybe when a fs is mounted in a local zone a 
user property would be set, like zone_mounted=test1. Perhaps during each mount 
such property would be created. In case with several zones just put names after 
, or something similar.

??
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re: How to destroy a pool wich you can't import because it is in faulted st

2006-09-06 Thread Frank Cusack
On September 6, 2006 7:19:32 AM -0700 Lieven De Geyndt <[EMAIL PROTECTED]> 
wrote:

sorry guys ...RTF did the job

[b]Legacy Mount Points[/b]


That just means filesystems in the pool won't get mounted, not that the
pool won't be imported.

-frank

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Need input on implementing a ZFS layout

2006-09-06 Thread Richard Elling - PAE

There is another option.  I'll call it "grow into your storage."
Pre-ZFS, for most systems you would need to allocate the storage well
in advance of its use.  For the 7xFLX380 case using SVM and UFS, you
would typically setup the FLX380 LUNs, merge them together using SVM,
and newfs.  Growing is somewhat difficult for that size of systems
because UFS has some smallish limits (16 TBytes per file system, less
for older Solaris releases).  Planning this in advance is challenging
and the process for growing existing file systems or adding new file
systems would need careful attention.

By contrast, with ZFS we can add vdevs to the zpool on the fly to an
existing zpool and the file systems can immediately use the new space.

The reliability of devices is measured in operational hours.  So, for
a fixed reliability metric one way to improve your real-life happiness
is to reduce the operational hours.

Putting these together, it makes sense to only add disks as you need
the space.  Keep the disks turned off, until needed, to lengthen their
life.  In other words, grow into your storage.  This doesn't work for
everyone, or every situation, but ZFS makes it an easy, viable option
to consider.
 -- richard
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Re: ZFS issues too much IOs

2006-09-06 Thread Robert Milkowski
Lowering from default 64K to 16K turned into about 10x less read throutput! And 
similar factor for latency for nfs clients. For now I'll probably leave it as 
it is and later will do some comparisons with different settings.

ps. very big thanks to Roch! I owe you!
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zoned datasets in zfs list

2006-09-06 Thread Eric Schrock
On Wed, Sep 06, 2006 at 04:52:48PM +0100, Dick Davies wrote:
> 
> Oh God no. That's exactly what I wanted to avoid.
> Why wouldn't you want it stored in the dataset, out of interest?

There are a couple of reasons:

- We don't want to re-create the same information in multiple places.
  Keeping both the zone configuration file and the on-disk state in
  sync would be difficult.
  
- The association of a dataset to a zone is a property of the
  zone, not the dataset.  This is particularly important if you want to
  migrate zones with delegated datasets between machines, or export
  and import your pool on a machine with a different zone configuration.

- Unlike NFS options or mountpoints, the dataset is part of a larger
  zone configuration, and cannot replace the zone configuration file
  entirely.

Ideally, we would have liked to not have the 'zoned' property at all,
but it was necessary to enforce sane behavior between the global and
local zone.

> 
> I'm reading that as 'the only real use case for 1 dataset in multiple zones'
> (sorry if I'm misunderstanding you)?
> 

Yep.

- Eric

--
Eric Schrock, Solaris Kernel Development   http://blogs.sun.com/eschrock
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Re: Re: Re: datasets,zones and mounts

2006-09-06 Thread Robert Milkowski
No, remove all other datasets from zone config and just put:

add dataset
set name=telecom/oracle/production
end

and that's it.

That way you will see all filesystem beneath tyelecom/oracle/production.
Additionally in a zone production you will be able to create more file systems 
inside without changing zone configuration.
Additionally snapshots and clones will work inside zone.

You are not only not loosing granulity you are adding flexibity that way.

Sorry for not being clear.
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zoned datasets in zfs list

2006-09-06 Thread Dick Davies

On 06/09/06, Eric Schrock <[EMAIL PROTECTED]> wrote:

On Wed, Sep 06, 2006 at 03:53:52PM +0100, Dick Davies wrote:
> That's a bit nicer, thanks.
> Still not that clear which zone they belong to though - would
> it be an idea to add a 'zone' property be a string == zonename ?

Yes, this is possible, but it's annoying because the actual owning zone
isn't stored with the dataset (nor should it be).   We'd have to grovel
around every zone's configuration file, which is certainly doable, just
annoying.


Oh God no. That's exactly what I wanted to avoid.
Why wouldn't you want it stored in the dataset, out of interest?


In addition, it's possible (though not recommended) to have a
single dataset in multiple zones.


Ah Ok, that explains why a single string wouldn't cut it
(although it sounds insane to me)!


The only real use case would be a
read-only, unmounted dataset whose snapshots could serve as a clone
source for other delegated datasets.


I'm reading that as 'the only real use case for 1 dataset in multiple zones'
(sorry if I'm misunderstanding you)?

--
Rasputin :: Jack of All Trades - Master of Nuns
http://number9.hellooperator.net/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zoned datasets in zfs list

2006-09-06 Thread Eric Schrock
On Wed, Sep 06, 2006 at 08:34:26AM -0700, Eric Schrock wrote:
> 
> Feel free to file an RFE.
> 

Oops, found one already:

6313352 'zpool list' & 'zfs list' should add '-z' & '-Z' to identifier a zone

- Eric

--
Eric Schrock, Solaris Kernel Development   http://blogs.sun.com/eschrock
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] creating zvols in a non-global zone (or 'Doctor, it hurts when I do this')

2006-09-06 Thread Eric Schrock
On Wed, Sep 06, 2006 at 04:23:32PM +0100, Dick Davies wrote:
> 
> a) prevent attempts to create zvols in non-global zones
> b) somehow allow it (?) or
> c) Don't do That
> 
> I vote for a) myself - should I raise an RFE?

Yes, that was _supposed_ to be the original behavior, and I thought we
had it working that way at one point.  Apparently I'm imagining things,
or it got broken somewhere along the way.  Please file a bug.

- Eric

--
Eric Schrock, Solaris Kernel Development   http://blogs.sun.com/eschrock
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re: datasets,zones and mounts

2006-09-06 Thread Eric Schrock
On Wed, Sep 06, 2006 at 09:01:00AM -0400, Kenneth Mikelinich wrote:
> Hi Robert -- Here are the outputs.  I cannot seem to see the last isapps
> dataset via zfs list. The non-global zone will be used to host a 10G
> Oracle.

Yes, this is definitely a bug somewhere.  I'll try to reproduce this on
a test machine and see what's going on.

- Eric

--
Eric Schrock, Solaris Kernel Development   http://blogs.sun.com/eschrock
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zoned datasets in zfs list

2006-09-06 Thread Eric Schrock
On Wed, Sep 06, 2006 at 03:53:52PM +0100, Dick Davies wrote:
> That's a bit nicer, thanks.
> Still not that clear which zone they belong to though - would
> it be an idea to add a 'zone' property be a string == zonename ?

Yes, this is possible, but it's annoying because the actual owning zone
isn't stored with the dataset (nor should it be).   We'd have to grovel
around every zone's configuration file, which is certainly doable, just
annoying. In addition, it's possible (though not recommended) to have a
single dataset in multiple zones.  The only real use case would be a
read-only, unmounted dataset whose snapshots could serve as a clone
source for other delegated datasets.  Since this is extremely rare, it
would probably suffice to have a special string like "(shared)" to
indicate that it is being shared between multiple zones.

Feel free to file an RFE.

- Eric

--
Eric Schrock, Solaris Kernel Development   http://blogs.sun.com/eschrock
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] creating zvols in a non-global zone (or 'Doctor, it hurts when I do this')

2006-09-06 Thread Dick Davies

A colleague just asked if zfs delegation worked with zvols too.
Thought I'd give it a go and got myself in a mess
(tank/linkfixer is the delegated dataset):

[EMAIL PROTECTED] / # zfs create -V 500M tank/linkfixer/foo
cannot create device links for 'tank/linkfixer/foo': permission denied
cannot create 'tank/linkfixer/foo': permission denied

Ok, so we'll try a normal filesystem:

[EMAIL PROTECTED] / # zfs create  tank/linkfixer/foo
cannot create 'tank/linkfixer/foo': dataset already exists
[EMAIL PROTECTED] / # zfs list
NAME   USED  AVAIL  REFER  MOUNTPOINT
tank  2.09G  33.8G  24.5K  legacy
tank/linkfixer36.3M  9.96G  24.5K  legacy
tank/linkfixer/foo22.5K  9.96G  22.5K  -
[EMAIL PROTECTED] / # zfs destroy -f  tank/linkfixer/foo
cannot remove device links for 'tank/linkfixer/foo': permission denied
[EMAIL PROTECTED] / # zfs list
NAME   USED  AVAIL  REFER  MOUNTPOINT
tank  2.09G  33.8G  24.5K  legacy
tank/linkfixer36.3M  9.96G  24.5K  legacy
tank/linkfixer/foo22.5K  9.96G  22.5K  -

I can destroy it ok from the global zone, and I know I could just
create a top-level zvol and grant the zone access.
Not sure if the 'fix' is :

a) prevent attempts to create zvols in non-global zones
b) somehow allow it (?) or
c) Don't do That

I vote for a) myself - should I raise an RFE?

--
Rasputin :: Jack of All Trades - Master of Nuns
http://number9.hellooperator.net/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zoned datasets in zfs list

2006-09-06 Thread Dick Davies

That's a bit nicer, thanks.
Still not that clear which zone they belong to though - would
it be an idea to add a 'zone' property be a string == zonename ?

On 06/09/06, Kenneth Mikelinich <[EMAIL PROTECTED]> wrote:

zfs mount

should show where all your datasets are mounted.

I too was confused with the zfs list readout.



On Wed, 2006-09-06 at 07:37, Dick Davies wrote:
> Just did my first dataset delegation, so be gentle :)
>
> Was initially terrified to see that changes to the mountpoint in the 
non-global
> zone were visible in the global zone.
>
> Then I realised it wasn't actually mounted (except in the delegated zone).
> But I couldn't see any obvious indication that the dataset was delegated to
> another zone in zfs list.
> Eventually I found the 'zoned' property. Couple of thoughts:
>
> 1) would it be worth changing 'zfs list' to clarify where a dataset
>is actually mounted?
> 2) Is there any way to indicate _what_ zone a dataset is mounted in
>   (other than greppping the zones configuration)?
>
> --
> Rasputin :: Jack of All Trades - Master of Nuns
> http://number9.hellooperator.net/
> ___
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss




--
Rasputin :: Jack of All Trades - Master of Nuns
http://number9.hellooperator.net/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Need input on implementing a ZFS layout

2006-09-06 Thread Richard Elling - PAE

Oatway, Ted wrote:

Thanks for the response Richard. Forgive my ignorance but the following
questions come to mind as I read your response.

I would then have to create 80 RAIDz(6+1) Volumes and the process of
creating these Volumes can be scripted. But -

1) I would then have to create 80 mount points to mount each of these
Volumes (?)


No.  In ZFS, you create a zpool which has the devices and RAID configurations.
The file systems (plural) are then put in the zpool.  You could have one
file system, or thousands.  Each file system, by default, will be in the
heirarchy under the zpool name, or you can change it as you need.


2) I would have no load balancing across mount points and I would have
to specifically direct the files to a mount point using an algorithm of
some design


ZFS will dynamically stripe across the sets.  In traditional RAID terms,
this is like RAID-1+0, RAID-5+0, or RAID-6+0.


3) A file landing on any one mount point would be constrained to the I/O
of the underlying disk which would represent 1/80th of the potential
available


It would be spread across the 80 sets.


4) Expansion of the architecture, by adding in another single disk
array, would be difficult and would probably be some form of data
migration (?). For 800TB of data that would be unacceptable.


It depends on how you do this.  There are techniques for balancing which
might work, but they have availability trade-offs because you are decreasing
your diversity.  I'm encouraged by the fact that they are planning ahead :-).

Also, unlike a traditional disk array or LVM software, ZFS will only copy
the data.  For example, in SVM, if you replace a whole disk, the resync will
copy the "data" for the whole disk.  For ZFS, it knows what data is valid,
and will only copy the valid data.  Thus the resync time is based upon the
size of the data, not the size of the disk.  There are more nuances here,
but that covers it to the first order.
 -- richard
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Need input on implementing a ZFS layout

2006-09-06 Thread Darren Dunham
> I would then have to create 80 RAIDz(6+1) Volumes and the process of
> creating these Volumes can be scripted. But -
> 
> 1) I would then have to create 80 mount points to mount each of these
> Volumes (?)

No.  Each of the RAIDZs that you create can be combined into a single
pool.  Data written to the pool will stripe across all the RAIDz
devices.

> 2) I would have no load balancing across mount points and I would have
> to specifically direct the files to a mount point using an algorithm of
> some design
>
> 3) A file landing on any one mount point would be constrained to the I/O
> of the underlying disk which would represent 1/80th of the potential
> available

See #1.

> 4) Expansion of the architecture, by adding in another single disk
> array, would be difficult and would probably be some form of data
> migration (?). For 800TB of data that would be unacceptable.

Today, you wouldn't be able to do it easily.  In the future, you may be
able to expand the RAIDz device.  (or if you could remove a VDEV from a
pool, you could rotate through and remove each of the RAIDz devices
followed by an addition of a new (8-column) RAIDz).

-- 
Darren Dunham   [EMAIL PROTECTED]
Senior Technical Consultant TAOShttp://www.taos.com/
Got some Dr Pepper?   San Francisco, CA bay area
 < This line left intentionally blank to confuse you. >
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Re: How to destroy a pool wich you can't import because it is in faulted st

2006-09-06 Thread Lieven De Geyndt
sorry guys ...RTF did the job 

[b]Legacy Mount Points[/b]
You can manage ZFS file systems with legacy tools by setting the mountpoint 
property to legacy.
Legacy file systems must be managed through the mount and umount commands and the
/etc/vfstab file. ZFS does not automatically mount legacy file systems on boot, 
and the ZFS mount and umount command do not operate on datasets of this type. 
The following examples show how to
set up and manage a ZFS dataset in legacy mode:
# zfs set mountpoint=legacy tank/home/eschrock
# mount -F zfs tank/home/eschrock /mnt
In particular, if you have set up separate ZFS /usr or /var file systems, you 
must indicate that theyare legacy file systems. In addition, you must mount them 
by creating entries in the /etc/vfstab file.
Otherwise, the system/filesystem/local service enters maintenance mode when the 
system boots.
To automatically mount a legacy file system on boot, you must add an entry to 
the /etc/vfstab file.
The following example shows what the entry in the /etc/vfstab file might look 
like:
#device device mount   FSfsck mount 
mount
#to mount   to fsckpoint   type  pass at 
boot options
#
tank/home/eschrock - /mnt  zfs   - yes  
   -

Note that the device to fsck and fsck pass entries are set to -. This syntax is 
because the fsck command is not applicable to ZFS file systems. For more 
information regarding data integrity and the lack of need for fsck in ZFS
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Re: Re: How to destroy a pool wich you can't import

2006-09-06 Thread Robert Milkowski
This could still corrupt the pool.
Probably the customer has to write its own tool to import a pool using libzfs 
and not creating zpool.cache.

Eventually just after pool is imported remove zpool.cache - I'm not sure but it 
should work.
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


RE: [zfs-discuss] Re: Re: datasets,zones and mounts

2006-09-06 Thread Mikelinich, Ken

Thanks.
I will try this out


Ken Mikelinich
Computer Operations Manager
Telecommunications and Client Services
University of New Hampshire
603.862.4220

-Original Message-
From: Dick Davies [mailto:[EMAIL PROTECTED] 
Sent: Wednesday, September 06, 2006 9:54 AM
To: Mikelinich, Ken
Cc: zfs-discuss@opensolaris.org
Subject: Re: [zfs-discuss] Re: Re: datasets,zones and mounts

On 06/09/06, Kenneth Mikelinich <[EMAIL PROTECTED]> wrote:

> Are you suggesting that I not get too granular with datasets and use a
> higher level one versus several?

I tihnk what he's saying is you should only have to
delegate one dataset (telecom/oracle/production, for example),
and all the 'child' datasets can be created/administered/snapshotted
etc.
in the non -global zone itself.
-- 
Rasputin :: Jack of All Trades - Master of Nuns
http://number9.hellooperator.net/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re: Re: datasets,zones and mounts

2006-09-06 Thread Dick Davies

On 06/09/06, Kenneth Mikelinich <[EMAIL PROTECTED]> wrote:


Are you suggesting that I not get too granular with datasets and use a
higher level one versus several?


I tihnk what he's saying is you should only have to
delegate one dataset (telecom/oracle/production, for example),
and all the 'child' datasets can be created/administered/snapshotted etc.
in the non -global zone itself.
--
Rasputin :: Jack of All Trades - Master of Nuns
http://number9.hellooperator.net/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re: How to destroy a pool wich you can't import because it is in faulted st

2006-09-06 Thread James C. McPherson

Lieven De Geyndt wrote:

zpool create -R did his job . Thanks for the tip .

Is ther a way to disable the auto mount when you boot a system ?
The customer has some kind of poor mans cluster . 
2 systems has access to a SE3510 with ZFS . 
System A was powered-off as test , system B did an import of the pools .
When system A rebooted , this system tries to import his pools , so 2 
systems are accessing the same pool . Probably this caused a corruption in his pool .

So how to disable automount of zfs pools ?


Oh heck  PMC 0.0.0alpha again :(

How about

# zfs set mountpoint=none fsname


James C. McPherson
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re: Re: datasets,zones and mounts

2006-09-06 Thread Kenneth Mikelinich
Hmmm.

I thought I was doing this via zonecfg -z production, which zonecfg is
run from the global zone.

add dataset
set name=telecom/oracle/production/oraapp
end

... repeat

add dataset
set name=telecom/oracle/production/isapps
end
commit
exit

The zone took all the datasets (shown in the earlier export), yet the
zone only reveals/mounts the first seven when it runs.  

Are you suggesting that I not get too granular with datasets and use a
higher level one versus several?

thanks
ken






On Wed, 2006-09-06 at 09:21, Robert Milkowski wrote:
> Well, that's interesting. Looks like some limit/bug here. However whatever 
> the limit is have you considered to add dataset into a zone? That way you can 
> actually create new file systems as needed inside a zone without changing 
> zone configuration, etc. You can also utilize snapshots, clones inside a 
> zone. That's what I did. That way DB admins can create file systems, 
> snapshots, etc.
> 
> add dataset
> set name=pool/dataset
> end
>  
> 
> This message posted from opensolaris.org
> ___
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re: Re: datasets,zones and mounts

2006-09-06 Thread Kenneth Mikelinich
Hmmm.

I thought I was doing this via zonecfg -z production, which zonecfg is
run from the global zone.

add dataset
set name=telecom/oracle/production/oraapp
end

... repeat

add dataset
set name=telecom/oracle/production/isapps
end
commit
exit

The zone took all the datasets (shown in the earlier export), yet the
zone only reveals/mounts the first seven when it runs.  

Are you suggesting that I not get too granular with datasets and use a
higher level one versus several?

thanks
ken






On Wed, 2006-09-06 at 09:21, Robert Milkowski wrote:
> Well, that's interesting. Looks like some limit/bug here. However whatever 
> the limit is have you considered to add dataset into a zone? That way you can 
> actually create new file systems as needed inside a zone without changing 
> zone configuration, etc. You can also utilize snapshots, clones inside a 
> zone. That's what I did. That way DB admins can create file systems, 
> snapshots, etc.
> 
> add dataset
> set name=pool/dataset
> end
>  
> 
> This message posted from opensolaris.org
> ___
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re: Recommendation ZFS on StorEdge 3320

2006-09-06 Thread Roch - PAE
Wee Yeh Tan writes:
 > On 9/5/06, Torrey McMahon <[EMAIL PROTECTED]> wrote:
 > > This is simply not true. ZFS would protect against the same type of
 > > errors seen on an individual drive as it would on a pool made of HW raid
 > > LUN(s). It might be overkill to layer ZFS on top of a LUN that is
 > > already protected in some way by the devices internal RAID code but it
 > > does not "make your data susceptible to HW errors caused by the storage
 > > subsystem's RAID algorithm, and slow down the I/O".
 > 
 > & Roch's recommendation to leave at least 1 layer of redundancy to ZFS
 > allows the extension of ZFS's own redundancy features for some truely
 > remarkable data reliability.
 > 
 > Perhaps, the question should be how one could mix them to get the best
 > of both worlds instead of going to either extreme.
 > 
 > > True, ZFS can't manage past the LUN into the array. Guess what? ZFS
 > > can't get past the disk drive firmware eitherand thats a good thing
 > > for all parties involved.
 > 
 > 
 > -- 
 > Just me,
 > Wire ...
 > ___
 > zfs-discuss mailing list
 > zfs-discuss@opensolaris.org
 > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Thinking some more about this. If your requirements does
mandate some form of mirroring, then it truly seems that ZFS 
should take that in charge if only because of the
self-healing characteristics. So I feel the storage array's
job is to export low latency Luns to ZFS.

I'd be happy to live with those simple Luns but I guess some
storage will just  refuse to export non-protected  luns. Now
we can definitively take advantage of the Array's capability
of exporting highly resilient Luns;  RAID-5 seems to fit the
bill  rather   well here. Even  an 9+1   luns will  be quite
resilient and have a low block overhead.

So we benefit from the arrays resiliency as well as it's low
latency characteristics. And we mirror data at the ZFS level 
which means great performance and great data integrity and
great availability.

Note that ZFS  write characteristics (all  sequential) means
that  we will commonly be filling  full  stripes on the luns
thus avoiding the partial stripe performance pitfall.

If you must shy away from any form of mirroring, then it's
either stripe your raid-5 luns (performance edge for those
who live dangerously) or raid-z around those raid-5 luns
(lower cost, survives lun failures).

-r

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Re: How to destroy a pool wich you can't import because it is in faulted st

2006-09-06 Thread Lieven De Geyndt
zpool create -R did his job . Thanks for the tip .

Is ther a way to disable the auto mount when you boot a system ?
The customer has some kind of poor mans cluster . 
2 systems has access to a SE3510 with ZFS . 
System A was powered-off as test , system B did an import of the pools .
When system A rebooted , this system tries to import his pools , so 2 systems 
are accessing the same pool . Probably this caused a corruption in his pool .
So how to disable automount of zfs pools ?
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Re: Re: datasets,zones and mounts

2006-09-06 Thread Robert Milkowski
Well, that's interesting. Looks like some limit/bug here. However whatever the 
limit is have you considered to add dataset into a zone? That way you can 
actually create new file systems as needed inside a zone without changing zone 
configuration, etc. You can also utilize snapshots, clones inside a zone. 
That's what I did. That way DB admins can create file systems, snapshots, etc.

add dataset
set name=pool/dataset
end
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re: datasets,zones and mounts

2006-09-06 Thread Kenneth Mikelinich
Hi Robert -- Here are the outputs.  I cannot seem to see the last isapps
dataset via zfs list. The non-global zone will be used to host a 10G
Oracle.

//** From the global zone **
//** Pool is named telecom

# zonecfg -z production export
create -b
set zonepath=/zones/production
set autoboot=true
add inherit-pkg-dir
set dir=/lib
end
add inherit-pkg-dir
set dir=/platform
end
add inherit-pkg-dir
set dir=/sbin
end
add inherit-pkg-dir
set dir=/usr
end
add dataset
set name=telecom/oracle/production/backup
end
add dataset
set name=telecom/oracle/production/logs
end
add dataset
set name=telecom/oracle/production/oraapp
end
add dataset
set name=telecom/oracle/production/oradata
end
add dataset
set name=telecom/oracle/production/oradmp
end
add dataset
set name=telecom/oracle/production/oramirror1
end
add dataset
set name=telecom/oracle/production/oramirror2
end
add dataset
set name=telecom/oracle/production/isapps
end

//**from my non-global zone called production
/** I have used zfs set mountpoint=  to adjust the mounts where i need
them
/** Telecom and telecom/oracle remain accessible only to the global as
desired

# zfs list
NAME   USED  AVAIL  REFER  MOUNTPOINT
telecom799K   669G51K  /telecom
telecom/oracle 490K   669G49K  /telecom/oracle
telecom/oracle/production   441K   669G49K 
/telecom/oracle/production
telecom/oracle/production/backup49K  20.0G49K  /backup
telecom/oracle/production/logs49K  5.00G49K  /logs
telecom/oracle/production/oraapp49K  15.0G49K  /oraapp
telecom/oracle/production/oradata49K  20.0G49K  /oradata
telecom/oracle/production/oradmp49K  2.00G49K  /oradmp
telecom/oracle/production/oramirror149K  5.00G49K  /oramirror1
telecom/oracle/production/oramirror249K  5.00G49K  /oramirror2


If I do a zfs mount in either the  global zone or the non-global zone, I
do not see telecom/oracle/production/isapps.  The zoned flag is set.

Regards and Thanks for your help
Ken
University of New Hampshire



On Wed, 2006-09-06 at 08:36, Robert Milkowski wrote:
> Hi.
> 
>   Can you post zonecfg -z  export and zfs list in that XXX zone?
>  
> 
> This message posted from opensolaris.org
> ___
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Re: datasets,zones and mounts

2006-09-06 Thread Robert Milkowski
Hi.

  Can you post zonecfg -z  export and zfs list in that XXX zone?
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Re: ZFS forces system to paging to the point it is unresponsive

2006-09-06 Thread Robert Milkowski
It looks like I discovered a workaround. I've got another zpool within rg in 
SC. The other zpool does not have production data (yet) so I can switch it 
between nodes freely. By doing this every 3 minutes I can stay safe on free 
memory, at least so far.

I guess it frees some ARC cache. What is not that clear however is that when I 
created another pool and just did export/import sometimes I get back some 
memory sometimes not. At the end it works worse than switching entire RG which 
actually means not only exporting zpool but also restarting nfsd. Looks like 
those two helps.
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] datasets,zones and mounts

2006-09-06 Thread Kenneth Mikelinich
Hi this question is along the lines of datasets and number of zfs file
systems available per zone.  I suspect I am missing something obvious.

We have added 8 datasets to one non-global Zone. While logged in and
doing a ZFS list in that zone, I am only able to see the first 7
available ZFS file systems.  The 8th one is not available; however a
zonecfg:info from the Global Zone shows it is indeed added to the zone. 
How do I go about making a Zone see more (eg mounting) more than 7 ZFS
file systems?

Note my datasets all are originating from one pool sitting on top of a
StorEdge 3320 RAID.

Regards
Ken
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zoned datasets in zfs list

2006-09-06 Thread Kenneth Mikelinich
zfs mount

should show where all your datasets are mounted.

I too was confused with the zfs list readout.



On Wed, 2006-09-06 at 07:37, Dick Davies wrote:
> Just did my first dataset delegation, so be gentle :)
> 
> Was initially terrified to see that changes to the mountpoint in the 
> non-global
> zone were visible in the global zone.
> 
> Then I realised it wasn't actually mounted (except in the delegated zone).
> But I couldn't see any obvious indication that the dataset was delegated to
> another zone in zfs list.
> Eventually I found the 'zoned' property. Couple of thoughts:
> 
> 1) would it be worth changing 'zfs list' to clarify where a dataset
>is actually mounted?
> 2) Is there any way to indicate _what_ zone a dataset is mounted in
>   (other than greppping the zones configuration)?
> 
> -- 
> Rasputin :: Jack of All Trades - Master of Nuns
> http://number9.hellooperator.net/
> ___
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] How to destroy a pool wich you can't import because it is in faulted state

2006-09-06 Thread James C. McPherson

Lieven De Geyndt wrote:

When a pool is in a faulted state , you can't import it . Even -f fails .
When you to decide to recreate the pool , you cannot execute zpool destroy , 
because it is not imported . Also -f does not work .
Any idea how to get out of this situation ?



try something like

zpool create -R /alternate_root newpoolname vdevlist


You might need to add a "-f", but try it without "-f" first.


cheers,
James C. McPherson

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Re: How to destroy a pool wich you can't import because it is in faulted st

2006-09-06 Thread Robert Milkowski
Hi.

  Just re-create it or create new pool with disks from the old one and use -f 
flag.
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] zoned datasets in zfs list

2006-09-06 Thread Dick Davies

Just did my first dataset delegation, so be gentle :)

Was initially terrified to see that changes to the mountpoint in the non-global
zone were visible in the global zone.

Then I realised it wasn't actually mounted (except in the delegated zone).
But I couldn't see any obvious indication that the dataset was delegated to
another zone in zfs list.
Eventually I found the 'zoned' property. Couple of thoughts:

1) would it be worth changing 'zfs list' to clarify where a dataset
  is actually mounted?
2) Is there any way to indicate _what_ zone a dataset is mounted in
 (other than greppping the zones configuration)?

--
Rasputin :: Jack of All Trades - Master of Nuns
http://number9.hellooperator.net/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] How to destroy a pool wich you can't import because it is in faulted state

2006-09-06 Thread Lieven De Geyndt
When a pool is in a faulted state , you can't import it . Even -f fails .
When you to decide to recreate the pool , you cannot execute zpool destroy , 
because it is not imported . Also -f does not work .
Any idea how to get out of this situation ?
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] How to flush ARC cache?

2006-09-06 Thread Robert Milkowski
Hi.

  Is there a waf to safely flush ARC cache (get most of ZFS memory back to 
system)?

ps. not by export/import
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss