I turned off softupdates, and had the same thing happen. Is this a bogus test? Is it reasonable to expect that a scsi drive failure should of been tolerated w/o crashing?
(bunch of scsi msgs to console) sub-disk down plex degraded g_access failed:6
trap 12
page fault while in kernel mode
cpuid=1 apic id=01
fault virtual address =0x18c
fault code supervisor write, page not present
instruction pointer =0x8:0xc043d72c
stack pointer =0x10:cbb17bf0
code segment =base 0x0, limit 0xfff, type 0x1b
=DPL0, pres1,def32,gran1
Processor flags interupt enabled, resume,IOPL=0
current process 22(irq11:ahc1)
Matthias Schuendehuette wrote:
gvinum> start <plexname>
This (as far as I investigated :-)
a) initializes a newly created RAID5-plex or
b) recalculates parity informations on a degraded RAID5-plex with a new replaced subdisk.
So, a 'gvinum start raid5.p0' initializes my RAID5-plex if newly created. You can monitor the initialization process with subsequent 'gvinum list' commands.
If you degrade a RAID5-plex with 'camcontrol stop <diskname>' (in case of SCSI-Disks) and 'repair' it afterwards with 'camcontrol start <diskname>', the 'gvinum start raid5.p0' (my volume here is called 'raid5') command recalculates the parity and revives the subdisk which was on disk <diskname>.
_______________________________________________ [EMAIL PROTECTED] mailing list http://lists.freebsd.org/mailman/listinfo/freebsd-stable To unsubscribe, send any mail to "[EMAIL PROTECTED]"
