You mean /usr/sbin/sys-unconfig?
No, it does not reset a system back far enough.
You still left with the orginal path_to_inst and the device tree.
e.g. take a disk to a different system and the first disk might end up
being sd10 and c15t0d0s0 instead of sd0 and c0 without cleaning up the system
Why do we still need /etc/zfs/zpool.cache file???
(I could understand it was useful when zfs import was slow)
zpool import is now multi-threaded
(http://bugs.opensolaris.org/bugdatabase/view_bug.do?bug_id=6844191), hence a
lot faster, each disk contains the hostname
I assume the swap, dumpadm, grub is because the pool has a different name now,
but is it still a problem if you take it to a *different system* boot off a CD
change it back to rpool. (which is most likley unsupported, ie no help to get
it working)
Over 10 years ago (way before flash archive
In the Thoughts on ZFS Pool Backup Strategies thread it was stated that zfs
send, sends uncompress data and uses the ARC.
If zfs send sends uncompress data which has already been compress this is not
very efficient, and it would be *nice* to see it send the original compress
data. (or an
NFSv4 has a concept of a root of the overall exported filesystem
(Pseudofilesystem).
FileHandle 0 in terms of Linux it is setting fsid=0 when exporting.
Which would explain why someone said Linux (NFSv4) automounts an exported
filesystem under another exported filesystem
ie mount servername:/
You could try copying the file to /tmp (ie swap/ram) and do a continues loop of
checksums e.g.
while [ ! -f ibdlpi.so.1.x ] ; do sleep 1; cp libdlpi.so.1 libdlpi.so.1.x ;
A=`sha512sum -b libdlpi.so.1.x` ; [ $A == what it should be
libdlpi.so.1.x ] rm libdlpi.so.1.x ; done ; date
Assume the
you could also use psradm to take a CPU off-line.
At boot I would ??assume?? the system boots the same way every time unless
something changes, so you could be hiting the came CPU core every time or the
same bit of RAM until booted fully.
Or even run SunVTS Validation Test Suite which I belive
A system with 100TB of data its 80% full and the a user ask their local system
admin to restore a directory with large files, as it was 30days ago with all
Windows/CIFS ACLS and NFSv4/ACLS etc.
If we used zfs send, we need to go back to a zfs send some 30days ago, and find
80TB of disk space
I vote for zfs needing a backup and restore command against a snapshot.
backup command should output on stderr at least
Full_Filename SizeBytes Modification_Date_1970secSigned
so backup software can build indexes and stdout contains the data.
The advantage of zfs providing the command is that
pantzer5 wrote:
These days I am a fan for forward check access
lists, because any one who
owns a DNS server can say that for IPAddressX
returns aserver.google.com.
They can not set the forward lookup outside of
their domain but they can
setup a reverse lookup. The other advantage is
For a RaidZ, when data is written to a disk, are
individual 32k join together to the same disk and
written out as a single I/O to the disk?
I/Os can be coalesced, but there is no restriction as
to what can be coalesced.
In other words, subsequent writes can also be
coalesced if they
In /etc/hosts for the format is
IP FQDN Alias...
Which would means 1.1.1.1 aserver.google.com aserver aserver-le0
I have seen a lot of sysadmins do the following:
1.1.1.1 aserver aserver.google.com
which means the host file (or NIS) does not match DNS
As the first entry is FQDN it is then name
I am talking about having a write queue, which points to ready to write, full
stripes.
Ready to write full stripes would be
*The last byte of the full stripe has been updated.
*The file has been closed for writing. (Exception to the above rule)
I believe there is now a scheduler for ZFS, to
Sorry, Full Stripe on a RaidZ is the recordsize ie if the record size is 128k
on a RaidZ and its made up of 5 disks, then 128k is spread across 4 disks with
the calc parity on the 5 disk, which means the writes are 32k to each disk.
For a RaidZ, when data is written to a disk, are individual
I think ZFS should look for more opportunities to write to disk rather than
leaving it to the last second (5seconds) as it appears it does. e.g.
if a file has record size worth of data outstanding it should be queued within
ZFS to be written out. If the record is updated again before a txg,
Create a new empty pool on the solaris system, let it format the disks etc
ie used the disk names cXtXd0 This should put the EFI label on the disks and
then setup the partitions for you. Just encase here is an example.
Go back to the Linux box, and see if you can use tools to see the same
May be look at rsync and rsync lib (http://librsync.sourceforge.net/) code to
see if a ZFS API could be design to help rsync/librsync in the future as well
as diff.
It might be a good idea for POSIX to have a single checksum and a
multi-checksum interface.
One problem could be block sizes,
I would have thought that if I write 1k then ZFS txg times out in 30secs, then
the 1k will be written to disk in a 1k record block, and then if I write 4k
then 30secs latter txg happen another 4k record size block will be written, and
then if I write 130k a 128k and 2k record block will be
Here is the output
-bash-4.0# uname -a
SunOS 5.11 snv_130 i86pc i386 i86pc
-bash-4.0# zfs get -r -o all compression mainfs01 | egrep -v \@
NAMEPROPERTY VALUE RECEIVED SOURCE
mainfs01 compression gzip-3- local
mainfs01/home compression
In my previous post I was refering more to mdbox (Multi-dbox) rather than dbox,
however I beleive the meta data is store with the mail msg in version 1.x and
2.x meta is not updated within the msg which would be better for ZFS.
What I am saying is msg per file which is not updated is better for
One thing which may help is the zfs import was single threaded, ie it open
every disk one disk (maybe slice) at a time and processed it, as of 128b it is
multi-threaded, ie it opens N disks/slices at once and process N disks/slices
at once. When N is the number of threads it decides to use.
Before Veritas VM had support for this, you needed to use a different server to
import a disk group. You could use a different server for ZFS, which will also
take the backup load off the Server?
Cheers
--
This message posted from opensolaris.org
___
Because ZFS is transaction, (effectively preserves order), the rename trick
will work.
If you find the .filename delete create a new .filename and when finish
writing rename it to filename. If filename exists you no all writes were
completed. If you have a batch system which looks for the file
What about removing attach/deattach and replace it with
zpool add [-fn] 'pool' submirror 'device/mirrorname' 'new_device'
e.g.
NAMESTATE READ WRITE CKSUM
rpoolONLINE 0 0 0
mirror-01 ONLINE 0 0 0
c4d0s0 ONLINE
If power failure happens you will lose anything in cache. So you could lose the
entire file on power failure if the system is not busy (ie ZFS does delay
writes, unless you do a fsync before closing the file). I would still like to
see a file system option sync on close or even wait for txg on
The zpool.cache file makes clustering complex. {Assume the man page is
still correct}
From the zpool man page:
cachefile=path | none
Controls the location of where the pool configuration is cached.
Discovering all pools on system startup requires a cached copy of the
configuration data that
Do we still need the zpool.cache still. I believe early versions of
zpool used the cache to remember what zpools to import at boot.
I understand newer versions of zfs still use the cache but also check to
see if the pool contains the correct host name of the server, and will
only import if the
The zpool.cache file makes clustering complex. {Assume the man page is
still correct}
From the zpool man page:
cachefile=path | none
Controls the location of where the pool configuration is cached.
Discovering all pools on system startup requires a cached copy of the
configuration data that
ZFS should allow 31+NULL chars for a comment against each disk.
This would work well with the host name string (I assume is max_hostname
255+NULL)
If a disk fails it should report c6t4908029d0 failed comment from
disk, it should also remember the comment until reboot
This would be useful for
Close Sync on file systems option (ie when the app calls close the file
is flushed, including mmap, no data loss of closed files on system
crash) Atomic/Locked operations across all pools e.g. snapshot all or
selected pools at the same time.
Allowance for offline files, eg. first part of a
Date: Thu, 26 Jul 2007 20:39:09 PDT
From: Anton B. Rang
That said, I?m not sure exactly what this buys you for disk replication.
What?s special about files which have been closed? Is the point that
applications might close a file and then notify some other process of the
file?s availability
Date: Thu, 26 Jul 2007 20:39:09 PDT
From: Anton B. Rang [EMAIL PROTECTED]
That said, I?m not sure exactly what this buys you for disk replication.
What?s special about files which have been closed? Is the point that
applications might close a file and then notify some other process of the
Guys,
What is the best way to ask for a feature enhancement to ZFS.
To allow ZFS to be usefull for DR disk replication, we need to be able
set an option against the pool or file system or both, called close
sync. ie When a programme closes a file any outstanding writes are flush
to disk,
33 matches
Mail list logo