I've seen the problems with bug 6343667, but I haven't seen the problem
I have at the the moment.
I started a scrub of a b72 system that doesn't have any recent snapshots
(none since the last scrub) and the % complete is cycling:
scrub: scrub in progress, 69.08% done, 0h13m to go
scrub: scrub
On Dec 14, 2007 1:12 AM, can you guess?
[EMAIL PROTECTED] wrote:
yes. far rarer and yet home users still see
them.
I'd need to see evidence of that for current
hardware.
What would constitute evidence? Do anecdotal tales
from home users
qualify? I have two disks (and one
...
though I'm not familiar with any recent examples in normal desktop environments
One example found during early use of zfs in Solaris engineering was
a system with a flaky power supply.
It seemed to work just fine with ufs but when zfs was installed the
sata drives started to shows many
Hello every ZFS gurus
I've been using a ZFS server for about one year now (for rsync-based disk
backup purpose).
The process is quite simple :
I backup each fs using rsync.
After each filesystem backup, I take a zfs snapshot to freeze read-only the
saved data.
So I end up with a zfs snapshot
On Dec 14, 2007, at 12:27 AM, Jorgen Lundman wrote:
Shawn Ferry wrote:
Jorgen,
You may want to try running 'bootadm update-archive'
Assuming that your boot-archive problem is an out of date boot-
archive
message at boot and/or doing a clean reboot to let the system try to
write an up
Hi all,
we are using the following setup as file server:
---
# uname -a
SunOS troubadix 5.10 Generic_120011-14 sun4u sparc SUNW,Sun-Fire-280R
# prtconf -D
System Configuration: Sun Microsystems sun4u
Memory size: 2048 Megabytes
System Peripherals (Software Nodes):
SUNW,Sun-Fire-280R (driver
Steve,
I have a couple of questions and concerns about using ZFS in an
environment where the underlying LUNs are replicated at a block
level using products like HDS TrueCopy or EMC SRDF. Apologies in
advance for the length, but I wanted the explanation to be clear.
(I do realise that
Hello,
We have a StorageTek FLX280 (very similar to a 6140) with 16 750 GB SATA drives
that we would like to use for disk-based backups. I am trying to make an
(educated) guess at what the best configuration for the LUN's on the FLX280
might be.
I've read, or at least skimmed, most of the
Frank Penczek wrote:
The performance is slightly disappointing. Does anyone have
a similar setup and can anyone share some figures?
Any pointers to possible improvements are greatly appreciated.
Use a faster processor or change to a mirrored configuration.
raidz2 can become processor
The throughput when writing from a local disk to the
zpool is around 30MB/s, when writing from a client
Err.. sorry, the internal storage would be good old 1Gbit FCAL disks @
10K rpm. Still, not the fastest around ;)
___
zfs-discuss mailing list
So does anyone have any insight on BugID 6535160?
We have verified on a similar system, that ZFS shows big latency in filebench
varmail test.
We formatted the same LUN with UFS and latency went down from 300 ms to 1-2 ms.
http://sunsolve.sun.com/search/document.do?assetkey=1-1-6535160-1
We
On Fri, 14 Dec 2007, Andrew Chace wrote:
[ reformatted ]
Hello,
We have a StorageTek FLX280 (very similar to a 6140) with 16 750 GB
SATA drives that we would like to use for disk-based backups. I am
trying to make an (educated) guess at what the best configuration
for the LUN's
Vincent Fox wrote:
So does anyone have any insight on BugID 6535160?
We have verified on a similar system, that ZFS shows big latency in filebench
varmail test.
We formatted the same LUN with UFS and latency went down from 300 ms to 1-2
ms.
This is such a big difference it makes me
Vincent Fox wrote:
So does anyone have any insight on BugID 6535160?
We have verified on a similar system, that ZFS shows big latency in filebench
varmail test.
We formatted the same LUN with UFS and latency went down from 300 ms to 1-2
ms.
This is such a big difference it makes me think
...
though I'm not familiar with any recent examples in
normal desktop environments
One example found during early use of zfs in Solaris
engineering was
a system with a flaky power supply.
It seemed to work just fine with ufs but when zfs was
installed the
sata drives started to
) The write cache is non volatile, but ZFS hasn't
been configured
to stop flushing it (set zfs:zfs_nocacheflush =
1).
These are a pair of 2540 with dual-controllers, definitely non-volatile cache.
We set the zfs_nocacheflush=1 and that improved things considerably.
ZFS filesystem (2540
On Dec 14, 2007 4:23 AM, can you guess? [EMAIL PROTECTED] wrote:
I assume that you're referring to ZFS checksum errors rather than to transfer
errors caught by the CRC resulting in retries.
Correct.
If so, then the next obvious question is, what is causing the ZFS checksum
errors? And
I'm testing an Iscsi multipath configuration on a T2000 with two disk
devices provided by a Netapp filer. Both the T2000 and the Netapp
have two ethernet interfaces for Iscsi, going to separate switches on
separate private networks. The scsi_vhci devices look like this in
`format':
1.
Hi Folks,
Begin forwarded message:
From: Edward Irvine [EMAIL PROTECTED]
Date: 12 December 2007 8:44:57 AM
To: [EMAIL PROTECTED]
Subject: Fwd: [zfs-discuss] zpool kernel panics.
FYI ...
Begin forwarded message:
From: James C. McPherson [EMAIL PROTECTED]
Date: 12 December 2007 8:06:51
the next obvious question is, what is
causing the ZFS checksum errors? And (possibly of
some help in answering that question) is the disk
seeing CRC transfer errors (which show up in its
SMART data)?
The memory is ECC in this machine, and Memtest passed
it for five
days. The disk was
This is the same configuration we use on 4 separate servers (T2000, two
X4100, and a V215). We do use a different iSCSI solution, but we have
the same multi path config setup with scsi_vhci. Dual GigE switches on
separate NICs both server and iSCSI node side. We suffered from the
e1000g
Use a faster processor or change to a mirrored configuration.
raidz2 can become processor bound in the Reed-Soloman calculations
for the 2nd parity set. You should be able to see this in mpstat, and to
a coarser grain in vmstat.
Hmm. Is the OP's hardware *that* slow? (I don't know enough
22 matches
Mail list logo