Note the you only need SUNWsan if you're running Solaris < 10.
Why one would run Solaris < 10 these days is beyond me, but...
/dale
On Mar 20, 2008, at 3:40 PM, Kim Kimball wrote:
Thanks, Jason.
Is the hardware the same as what you tested last year?
Kim
Jason Edgecombe wrote:
Is this what you need?
PKGINST: SUNWsan
NAME: SAN Foundation Kit
CATEGORY: system
ARCH: sparc
VERSION: 1.0
BASEDIR: /
VENDOR: Sun Microsystems, Inc.
DESC: This package provides a support for the SAN Foundation
Kit.
PSTAMP: sanserve-a20031029172438
INSTDATE: Jan 15 2008 10:37
HOTLINE: Please contact your local service provider
STATUS: completely installed
FILES: 22 installed pathnames
4 shared pathnames
1 linked files
11 directories
2 executables
239 blocks used (approx)
Running Solaris 9 09/05HW Sparc with Sun SAN foundation.
Jason
Kim Kimball wrote:
Hi Jason,
Thanks!
Can you tell me which flavor of SAN you're using?
Kim
Jason Edgecombe wrote:
Robert Banz wrote:
AFS can't really cause "san issues" in that it's just another
application using your filesystem. In some cases, it can be
quite a heavy user of such, but since its only interacting
through the fs, its not going to know anything about your
underlying storage fabric, or have any way of targeting it for
any more badness than any other filesystem user.
One of the big differences that would effect the filesystem IO
load that occurred between 1.4.1 & 1.4.6 was the removal
functions that made copious fsync operations. These operations
were called in fileserver/volserver functions that modified
various in-volume structures, specifically file creations and
deletions, and would lead to rather underwhelming performance
when doing vos restores, deleting, or copying large file trees.
In many configurations, this causes the OS to pass on a call to
the underlying storage to verify that all changes written have
been written to *disk*, causing the storage controller to flush
its write cache. Since this defeats many of the benefits (wrt I/
O scheduling) on your storage hardware of having a cache, this
could lead to overloaded storage.
Some storage devices have the option to ignore these calls from
devices, assuming your write cache is reliable.
Under UFS, I would suggest that you'd be running in 'logging'
mode when using the namei fileserver on Solaris, as yes, fsck is
rather horrible to run. Performance on reasonably recent
versions of ZFS were quite acceptable as well.
I can confirm Robert's observations. I recently tested openafs
1.4.1 inode vs 1.4.6 namei on solaris 9 sparc with a Sun Storedge
3511 Expansion tray fibre channel device. The difference is
stagerring with vos move and such. We have been using the 1.4.6
namei config on a SAN for a few months now with no issues.
Jason
_______________________________________________
OpenAFS-info mailing list
[email protected]
https://lists.openafs.org/mailman/listinfo/openafs-info
_______________________________________________
OpenAFS-info mailing list
[email protected]
https://lists.openafs.org/mailman/listinfo/openafs-info
_______________________________________________
OpenAFS-info mailing list
[email protected]
https://lists.openafs.org/mailman/listinfo/openafs-info
_______________________________________________
OpenAFS-info mailing list
[email protected]
https://lists.openafs.org/mailman/listinfo/openafs-info
_______________________________________________
OpenAFS-info mailing list
[email protected]
https://lists.openafs.org/mailman/listinfo/openafs-info