Brian Wilson wrote:
On Jul 16, 2007, at 6:06 PM, Torrey McMahon wrote:
Darren Dunham wrote:
My previous experience with powerpath was that it rode below the
Solaris
device layer. So you couldn't cause trespass by using the wrong
device. It would just go to powerpath which would
Hello Łukasz,
Monday, July 23, 2007, 1:19:16 PM, you wrote:
Ł ZFS send is very slow.
Ł dmu_sendbackup function is traversing dataset in one thread and in
Ł traverse callback function ( backup_cb ) we are waiting for data in
Ł arc_read called with ARC_WAIT flag.
Ł I want to parallize zfs send
The Sharemgr test suite is available on OpenSolaris.org.
The source tarball, binary package and baseline can be downloaded from the test
consolidation download center at:
http://dlc.sun.com/osol/test/downloads/current
The source code can be viewed in the Solaris Test Collection (STC) 2.0
Hi
I'm trying to setup a new NFS server, and wish to use Solaris and ZFS. I have a
ZFS filesystem set up to handle the users home directories and setup sharing
# zfs list
NAME USED AVAIL REFER MOUNTPOINT
data 896K 9.75G 35.3K /data
data/home
Robert Milkowski wrote:
Hello Łukasz,
Monday, July 23, 2007, 1:19:16 PM, you wrote:
Ł ZFS send is very slow.
Ł dmu_sendbackup function is traversing dataset in one thread and in
Ł traverse callback function ( backup_cb ) we are waiting for data in
Ł arc_read called with ARC_WAIT flag.
On Jul 22, 2007, at 7:39 PM, JS wrote:
There a way to take advantage of this in Sol10/u03?
sorry, variable 'zfs_vdev_cache_max' is not defined in the 'zfs'
module
That tunable/hack will be available in s10u4:
http://bugs.opensolaris.org/view_bug.do?bug_id=6472021
wait about a month and
Scott Adair wrote:
Hi
I'm trying to setup a new NFS server, and wish to use Solaris and ZFS. I have
a ZFS filesystem set up to handle the users home directories and setup sharing
# zfs list
NAME USED AVAIL REFER MOUNTPOINT
data 896K 9.75G
Richard Elling wrote:
Haudy Kazemi wrote:
How would one calculate system reliability estimates here? One is a
RAIDZ set of 6 disks, the other a set of 8. The reliability of each
RAIDZ set by itself isn't too hard to calculate, but put together,
especially since they're different sizes, I
Hi Louwtjie,
(CC'd to the list as an FYI to others)
The biggest gotcha is the SE6140's have a 12 byte SCSI control data
block, and thus can only do 2TB LUNs out to the host. That's not an
issue with ZFS however since you can just tack them together and grow
your pool that way. See the