Re: [zfs-discuss] zpool split problem?

2010-04-01 Thread Damon Atkins
You mean /usr/sbin/sys-unconfig? No, it does not reset a system back far enough. You still left with the orginal path_to_inst and the device tree. e.g. take a disk to a different system and the first disk might end up being sd10 and c15t0d0s0 instead of sd0 and c0 without cleaning up the system

Re: [zfs-discuss] zpool split problem?

2010-03-31 Thread Damon Atkins
Why do we still need /etc/zfs/zpool.cache file??? (I could understand it was useful when zfs import was slow) zpool import is now multi-threaded (http://bugs.opensolaris.org/bugdatabase/view_bug.do?bug_id=6844191), hence a lot faster, each disk contains the hostname

Re: [zfs-discuss] zpool split problem?

2010-03-31 Thread Damon Atkins
I assume the swap, dumpadm, grub is because the pool has a different name now, but is it still a problem if you take it to a *different system* boot off a CD change it back to rpool. (which is most likley unsupported, ie no help to get it working) Over 10 years ago (way before flash archive

[zfs-discuss] zfs send and ARC

2010-03-25 Thread Damon Atkins
In the Thoughts on ZFS Pool Backup Strategies thread it was stated that zfs send, sends uncompress data and uses the ARC. If zfs send sends uncompress data which has already been compress this is not very efficient, and it would be *nice* to see it send the original compress data. (or an

Re: [zfs-discuss] ZFS file system confusion

2010-03-25 Thread Damon Atkins
NFSv4 has a concept of a root of the overall exported filesystem (Pseudofilesystem). FileHandle 0 in terms of Linux it is setting fsid=0 when exporting. Which would explain why someone said Linux (NFSv4) automounts an exported filesystem under another exported filesystem ie mount servername:/

Re: [zfs-discuss] CR 6880994 and pkg fix

2010-03-24 Thread Damon Atkins
You could try copying the file to /tmp (ie swap/ram) and do a continues loop of checksums e.g. while [ ! -f ibdlpi.so.1.x ] ; do sleep 1; cp libdlpi.so.1 libdlpi.so.1.x ; A=`sha512sum -b libdlpi.so.1.x` ; [ $A == what it should be libdlpi.so.1.x ] rm libdlpi.so.1.x ; done ; date Assume the

Re: [zfs-discuss] CR 6880994 and pkg fix

2010-03-24 Thread Damon Atkins
you could also use psradm to take a CPU off-line. At boot I would ??assume?? the system boots the same way every time unless something changes, so you could be hiting the came CPU core every time or the same bit of RAM until booted fully. Or even run SunVTS Validation Test Suite which I belive

Re: [zfs-discuss] Thoughts on ZFS Pool Backup Strategies

2010-03-18 Thread Damon Atkins
A system with 100TB of data its 80% full and the a user ask their local system admin to restore a directory with large files, as it was 30days ago with all Windows/CIFS ACLS and NFSv4/ACLS etc. If we used zfs send, we need to go back to a zfs send some 30days ago, and find 80TB of disk space

Re: [zfs-discuss] Thoughts on ZFS Pool Backup Strategies

2010-03-17 Thread Damon Atkins
I vote for zfs needing a backup and restore command against a snapshot. backup command should output on stderr at least Full_Filename SizeBytes Modification_Date_1970secSigned so backup software can build indexes and stdout contains the data. The advantage of zfs providing the command is that

Re: [zfs-discuss] sharenfs option rw,root=host1 don't take effect

2010-03-11 Thread Damon Atkins
pantzer5 wrote: These days I am a fan for forward check access lists, because any one who owns a DNS server can say that for IPAddressX returns aserver.google.com. They can not set the forward lookup outside of their domain but they can setup a reverse lookup. The other advantage is

Re: [zfs-discuss] Should ZFS write data out when disk are idle

2010-03-10 Thread Damon Atkins
For a RaidZ, when data is written to a disk, are individual 32k join together to the same disk and written out as a single I/O to the disk? I/Os can be coalesced, but there is no restriction as to what can be coalesced. In other words, subsequent writes can also be coalesced if they

Re: [zfs-discuss] sharenfs option rw,root=host1 don't take effect

2010-03-10 Thread Damon Atkins
In /etc/hosts for the format is IP FQDN Alias... Which would means 1.1.1.1 aserver.google.com aserver aserver-le0 I have seen a lot of sysadmins do the following: 1.1.1.1 aserver aserver.google.com which means the host file (or NIS) does not match DNS As the first entry is FQDN it is then name

Re: [zfs-discuss] Should ZFS write data out when disk are idle

2010-03-09 Thread Damon Atkins
I am talking about having a write queue, which points to ready to write, full stripes. Ready to write full stripes would be *The last byte of the full stripe has been updated. *The file has been closed for writing. (Exception to the above rule) I believe there is now a scheduler for ZFS, to

Re: [zfs-discuss] Should ZFS write data out when disk are idle

2010-03-09 Thread Damon Atkins
Sorry, Full Stripe on a RaidZ is the recordsize ie if the record size is 128k on a RaidZ and its made up of 5 disks, then 128k is spread across 4 disks with the calc parity on the 5 disk, which means the writes are 32k to each disk. For a RaidZ, when data is written to a disk, are individual

Re: [zfs-discuss] Should ZFS write data out when disk are idle

2010-03-07 Thread Damon Atkins
I think ZFS should look for more opportunities to write to disk rather than leaving it to the last second (5seconds) as it appears it does. e.g. if a file has record size worth of data outstanding it should be queued within ZFS to be written out. If the record is updated again before a txg,

Re: [zfs-discuss] Help with corrupted pool

2010-02-17 Thread Damon Atkins
Create a new empty pool on the solaris system, let it format the disks etc ie used the disk names cXtXd0 This should put the EFI label on the disks and then setup the partitions for you. Just encase here is an example. Go back to the Linux box, and see if you can use tools to see the same

Re: [zfs-discuss] Intrusion Detection - powered by ZFS Checksumming ?

2010-02-08 Thread Damon Atkins
May be look at rsync and rsync lib (http://librsync.sourceforge.net/) code to see if a ZFS API could be design to help rsync/librsync in the future as well as diff. It might be a good idea for POSIX to have a single checksum and a multi-checksum interface. One problem could be block sizes,

Re: [zfs-discuss] Intrusion Detection - powered by ZFS Checksumming ?

2010-02-08 Thread Damon Atkins
I would have thought that if I write 1k then ZFS txg times out in 30secs, then the 1k will be written to disk in a 1k record block, and then if I write 4k then 30secs latter txg happen another 4k record size block will be written, and then if I write 130k a 128k and 2k record block will be

[zfs-discuss] send/received inherited bug?, received overrides parent, snv_130 6920906

2010-01-29 Thread Damon Atkins
Here is the output -bash-4.0# uname -a SunOS 5.11 snv_130 i86pc i386 i86pc -bash-4.0# zfs get -r -o all compression mainfs01 | egrep -v \@ NAMEPROPERTY VALUE RECEIVED SOURCE mainfs01 compression gzip-3- local mainfs01/home compression

Re: [zfs-discuss] zpool fragmentation issues? (dovecot)

2010-01-16 Thread Damon Atkins
In my previous post I was refering more to mdbox (Multi-dbox) rather than dbox, however I beleive the meta data is store with the mail msg in version 1.x and 2.x meta is not updated within the msg which would be better for ZFS. What I am saying is msg per file which is not updated is better for

Re: [zfs-discuss] (Practical) limit on the number of snapshots?

2010-01-11 Thread Damon Atkins
One thing which may help is the zfs import was single threaded, ie it open every disk one disk (maybe slice) at a time and processed it, as of 128b it is multi-threaded, ie it opens N disks/slices at once and process N disks/slices at once. When N is the number of threads it decides to use.

Re: [zfs-discuss] Import a SAN cloned disk

2009-12-16 Thread Damon Atkins
Before Veritas VM had support for this, you needed to use a different server to import a disk group. You could use a different server for ZFS, which will also take the backup load off the Server? Cheers -- This message posted from opensolaris.org ___

Re: [zfs-discuss] Transaction consistency of ZFS

2009-12-07 Thread Damon Atkins
Because ZFS is transaction, (effectively preserves order), the rename trick will work. If you find the .filename delete create a new .filename and when finish writing rename it to filename. If filename exists you no all writes were completed. If you have a batch system which looks for the file

Re: [zfs-discuss] Accidentally added disk instead of attaching

2009-12-07 Thread Damon Atkins
What about removing attach/deattach and replace it with zpool add [-fn] 'pool' submirror 'device/mirrorname' 'new_device' e.g. NAMESTATE READ WRITE CKSUM rpoolONLINE 0 0 0 mirror-01 ONLINE 0 0 0 c4d0s0 ONLINE

Re: [zfs-discuss] Transaction consistency of ZFS

2009-12-05 Thread Damon Atkins
If power failure happens you will lose anything in cache. So you could lose the entire file on power failure if the system is not busy (ie ZFS does delay writes, unless you do a fsync before closing the file). I would still like to see a file system option sync on close or even wait for txg on

Re: [zfs-discuss] Zpools on USB zpool.cache zpool import

2009-03-24 Thread Damon Atkins
The zpool.cache file makes clustering complex. {Assume the man page is still correct} From the zpool man page: cachefile=path | none Controls the location of where the pool configuration is cached. Discovering all pools on system startup requires a cached copy of the configuration data that

Re: [zfs-discuss] Zpools on USB zpool.cache

2009-03-23 Thread Damon Atkins
Do we still need the zpool.cache still. I believe early versions of zpool used the cache to remember what zpools to import at boot. I understand newer versions of zfs still use the cache but also check to see if the pool contains the correct host name of the server, and will only import if the

Re: [zfs-discuss] Zpools on USB zpool.cache zpool import

2009-03-23 Thread Damon Atkins
The zpool.cache file makes clustering complex. {Assume the man page is still correct} From the zpool man page: cachefile=path | none Controls the location of where the pool configuration is cached. Discovering all pools on system startup requires a cached copy of the configuration data that

Re: [zfs-discuss] device alias

2007-09-26 Thread Damon Atkins
ZFS should allow 31+NULL chars for a comment against each disk. This would work well with the host name string (I assume is max_hostname 255+NULL) If a disk fails it should report c6t4908029d0 failed comment from disk, it should also remember the comment until reboot This would be useful for

Re: [zfs-discuss] Wish List

2007-08-17 Thread Damon Atkins
Close Sync on file systems option (ie when the app calls close the file is flushed, including mmap, no data loss of closed files on system crash) Atomic/Locked operations across all pools e.g. snapshot all or selected pools at the same time. Allowance for offline files, eg. first part of a

Re: [zfs-discuss] ZFS with HDS TrueCopy and EMC SRDF

2007-08-03 Thread Damon Atkins
Date: Thu, 26 Jul 2007 20:39:09 PDT From: Anton B. Rang That said, I?m not sure exactly what this buys you for disk replication. What?s special about files which have been closed? Is the point that applications might close a file and then notify some other process of the file?s availability

Re: [zfs-discuss] ZFS with HDS TrueCopy and EMC SRDF [CloseSync]

2007-07-27 Thread Damon Atkins
Date: Thu, 26 Jul 2007 20:39:09 PDT From: Anton B. Rang [EMAIL PROTECTED] That said, I?m not sure exactly what this buys you for disk replication. What?s special about files which have been closed? Is the point that applications might close a file and then notify some other process of the

[zfs-discuss] ZFS with HDS TrueCopy and EMC SRDF

2007-07-26 Thread Damon Atkins
Guys, What is the best way to ask for a feature enhancement to ZFS. To allow ZFS to be usefull for DR disk replication, we need to be able set an option against the pool or file system or both, called close sync. ie When a programme closes a file any outstanding writes are flush to disk,