fyi to everyone, the Asus P5W64 motherboard previously in my opensolaris machine
was the culprit, and not the general mpt issues. At the time the motherboard
was
originally put in that machine, there was not enough zfs i/o load to trigger the
problem which led to the false impression the
Thanks for the info on the yukon driver. I realize too many variables makes
things impossible to determine, but I had made these hardware changes awhile
back, and they seemed to work fine at the time. Since they aren't now, even
in the older OpenSolaris (i've tried 2009.06 and 2008.11 now), the
Hi all,
Unfortunately for me, there does seem to be a hardware component to my problem.
Although my rsync copied almost 4TB of
data with no iostat errors after going back to OpenSolaris 2009.06, I/O on one
of my mpt cards did eventually hang, with
6 disk lights on and 2 off, until rebooting.
I eventually performed a few more tests, adjusting some zfs tuning options
which had no effect, and trying the
itmpt driver which someone had said would work, and regardless my system would
always freeze quite rapidly in
snv 127 and 128a. Just to double check my hardware, I went back to the
I don't think the hardware has any problems, it only started having errors when
I upgraded OpenSolaris.
It's still working fine again now after a reboot. Actually, I reread one of
your earlier messages,
and I didn't realize at first when you said non-Sun JBOD that this didn't
apply to me (in
Well, ok, the msi=0 thing didn't help after all. A few minutes after my last
message a few errors showed
up in iostat, and then in a few minutes more the machine was locked up hard...
Maybe I will try just
doing a scrub instead of my rsync process and see how that does.
Chad
On Tue, Dec 01,
This is basically just a me too. I'm using different hardware but essentially
the same problems. The relevant hardware I have is:
---
SuperMicro MBD-H8Di3+-F-O motherboard with LSI 1068E onboard
SuperMicro SC846E2-R900B 4U chassis with two LSI SASx36 expander chips on the
backplane
24 Western
Chad Cantwell wrote:
Hi,
I was using for quite awhile OpenSolaris 2009.06
with the opensolaris-provided mpt driver to operate a zfs raidz2 pool of
about ~20T and this worked perfectly fine (no issues or device errors
logged for several months, no hanging). A few days ago I decided to
Mark Johnson wrote:
Chad Cantwell wrote:
Hi,
I was using for quite awhile OpenSolaris 2009.06
with the opensolaris-provided mpt driver to operate a zfs raidz2 pool of
about ~20T and this worked perfectly fine (no issues or device errors
logged for several months, no hanging). A few days
To: zfs-discuss@opensolaris.org
Subject: Re: [zfs-discuss] mpt errors on snv 127
Mark Johnson wrote:
Chad Cantwell wrote:
Hi,
I was using for quite awhile OpenSolaris 2009.06
with the opensolaris-provided mpt driver to operate a zfs raidz2 pool of
about ~20T and this worked perfectly fine
What's the earliest build someone has seen this
problem? i.e. if we binary chop, has anyone seen it
in
b118?
We have used every stable build from b118 up, as b118 was the first reliable
one that could be used is a CIFS-heavy environment. The problem occurs on all
of them.
- Adam
--
If someone from Sun will confirm that it should work to use the mpt driver from
2009.06, I'd be willing to set up a BE and try it. I still have the snapshot
from my 2009.06 install, so I should be able to mount that and grab the files
easily enough.
--
This message posted from opensolaris.org
Travis Tabbal wrote:
If someone from Sun will confirm that it should work to use the mpt
driver from 2009.06, I'd be willing to set up a BE and try it. I
still have the snapshot from my 2009.06 install, so I should be able
to mount that and grab the files easily enough.
I tried, it doesn't
First I tried just upgrading to b127, that had a few issues besides the mpt
driver. After that
I did a clean install of b127, but no I don't have my osol2009.06 root still
there. I wasn't
sure how to install another copy and leave it there (I suspect it is possible,
since I saw
when doing
To update everyone, I did a complete zfs scrub, and it it generated no errors
in iostat, and I have 4.8T of
data on the filesystem so it was a fairly lengthy test. The machine also has
exhibited no evidence of
instability. If I were to start copying a lot of data to the filesystem again
Hi,
Sorry for not replying to one of the already open threads on this topic;
I've just joined the list for the purposes of this discussion and have
nothing in my client to reply to yet.
I have an x86_64 opensolaris machine running on a Core 2 Quad Q9650
platform with two LSI SAS3081E-R PCI-E 8
Chad Cantwell wrote:
Hi,
Sorry for not replying to one of the already open threads on this topic;
I've just joined the list for the purposes of this discussion and have
nothing in my client to reply to yet.
I have an x86_64 opensolaris machine running on a Core 2 Quad Q9650
platform with two
Hi,
Replied to your previous general query already, but in summary, they are in the
server chassis. It's a Chenbro 16 hotswap bay case. It has 4 mini backplanes
that each connect via an SFF-8087 cable (1m) to my LSI cards (2 cables / 8
drives
per card).
Chad
On Tue, Dec 01, 2009 at
Chad Cantwell wrote:
Hi,
Replied to your previous general query already, but in summary, they are in the
server chassis. It's a Chenbro 16 hotswap bay case. It has 4 mini backplanes
that each connect via an SFF-8087 cable (1m) to my LSI cards (2 cables / 8
drives
per card).
Hi Chad,
thanks
Hi,
The Chenbro chassis contains everything - the motherboard/CPU, and the disks.
As far as
I know the chenbro backplanes are basically electrical jumpers that the LSI
cards shouldn't
be aware of. They pass through the SATA signals directly from SFF-8087 cables
to the
disks.
Thanks,
Chad
After another crash I checked the syslog and there were some different errors
than the ones
I saw previously during operation:
Nov 30 20:26:11 the-vault scsi: [ID 107833 kern.warning] WARNING:
/p...@0,0/pci8086,2...@3/pci111d,8...@0/pci111d,8...@1/pci1000,3...@0 (mpt1):
Nov 30 20:26:11
Chad Cantwell wrote:
After another crash I checked the syslog and there were some different errors
than the ones
I saw previously during operation:
...
Nov 30 20:59:13 the-vault LSI PCI device (1000,) not supported.
...
Nov 30 20:59:13 the-vault mpt_config_space_init failed
22 matches
Mail list logo