Nate,
Thanks for investigating this. Sounds like ZFS is either conflicting
with the Linux partition or running off the end of its partition in the
VMware configuration you set up. The result is the CKSUM errors you
are observing. This could well lead to errors when we try to pagefault
in the iso image blocks at mount.
There is still a bug here I think in ZFS in the way it is handling these
pagefault errors. We should not be panicing.
Given your analysis, I don't think I need your crash dump.
Thanks for using (and finding a bug in) ZFS!
-Mark
Nathanael Burton wrote:
Mark,
I might know a little bit more about what's causing this particular panic. I'm
currently running OpenSolaris as a guest OS under VMware Server RC1 on a CentOS
4.3 host OS. I have 3 - 300GB (~280GB usable) SATA disks in the server that
are all formatted under CentOS like so:
[EMAIL PROTECTED] sdb]# fdisk -l /dev/sda
Disk /dev/sda: 300.0 GB, 300069052416 bytes
255 heads, 63 sectors/track, 36481 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Device Boot Start End Blocks Id System
/dev/sda1 * 1 3187 25599546 fd Linux raid autodetect
/dev/sda2 3188 36481 267434055 bf Solaris[/b]
So I use the first ~25GB/disk in a Linux software RAID 5, the rest of the disk ~240GB (usable) is given to OpenSolaris (via VMware) as a raw physical disk partition. OpenSolaris still thinks that those disks that it's been given are the full size (~280GB) -- PROBLEM 1.
Next I can create a simple ZFS pool using one of the SATA disks like so:
[b]bash-3.00# zpool create sata c2t0d0[/b]
Then I copy an iso file from my OpenBSD file server via ftp... As soon as data
starts writing into the ZFS file system I notice zpool CKSUM errors -- PROBLEM
2. The first time I saw this problem occur I never checked the output of zpool
status, and I believe I must have had a bunch of CKSUM errors then too.
Current info:
[b]bash-3.00# pwd
/data
bash-3.00# ls -al
total 1423398
drwxr-xr-x 2 root sys 3 Jun 15 20:55 .
drwxr-xr-x 43 root root 1024 Jun 15 20:57 ..
-rw-r--r-- 1 root root 728190976 Sep 23 2005
KNOPPIX_V4.0.2CD-2005-09-23-EN.iso
bash-3.00# zpool status
pool: sata
state: ONLINE
status: One or more devices has experienced an unrecoverable error. An
attempt was made to correct the error. Applications are unaffected.
action: Determine if the device needs to be replaced, and clear the errors
using 'zpool clear' or replace the device with 'zpool replace'.
see: http://www.sun.com/msg/ZFS-8000-9P
scrub: none requested
config:
NAME STATE READ WRITE CKSUM
sata ONLINE 0 0 20
c2t0d0 ONLINE 0 0 20
errors: No known data errors
bash-3.00# zfs list
NAME USED AVAIL REFER MOUNTPOINT
sata 695M 273G 24.5K /sata
sata/data 695M 273G 695M /data
sata/mp3s 24.5K 273G 24.5K /mp3s[/b]
Now, I attempt to mount the iso file via lofiadm and the panic occurs:
[b]bash-3.00# mount -F hsfs `lofiadm -a
/data/KNOPPIX_V4.0.2CD-2005-09-23-EN.iso` /tmp/test[/b]
I have also tested the above scenario but instead of giving OpenSolaris the SATA disk via raw physical disk access I create a VMware vmdk disk image file on the SATA disk and give that to OpenSolaris. In this case I can successfully create a ZFS file system, copy the same iso to it, and mount it via lofiadm.
So I have a new panic/crash dump -- it's absolutely huge, ~400MB after tar and bzip. If you still want it I can upload it to sunsolve as you requested. Or if there is a way to make it smaller let me know.
Thanks,
Nate
This message posted from opensolaris.org
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss