Hello, I built one of my servers (V120) with a zfs root to see how it works, now when I try to rebuild and re-use the two internal disks using slices/sds and ufs I can still see that there are references to the old zpool called rpool. When I try to destroy the old pool I get core dups. Does anyone have any idea how I resolve this?
Many thanks, John ok boot rootmirror Resetti LOM event: +6d+5h2m18s host reset ng ... ChassisSerialNumber UNDEFINED Sun Fire V120 (UltraSPARC-IIe 648MHz), No Keyboard OpenBoot 4.0, 3072 MB memory installed, Serial #55470796. Ethernet address 0:3:ba:4e:6a:cc, Host ID: 834e6acc. Executing last command: boot rootmirror Boot device: /p...@1f,0/p...@1/s...@8/d...@1,0 File and args: SunOS Release 5.10 Version Generic_137137-09 64-bit Copyright 1983-2008 Sun Microsystems, Inc. All rights reserved. Use is subject to license terms. Hostname: ted SUNW,eri0 : 100 Mbps full duplex link up /dev/md/rdsk/d40 is clean /dev/md/rdsk/d50 is clean Reading ZFS config: done. ted console login: root Password: Nov 9 16:56:49 ted login: ROOT LOGIN /dev/console Last login: Sun Nov 9 16:14:35 on console Sun Microsystems Inc. SunOS 5.10 Generic January 2005 # zpool list NAME SIZE USED AVAIL CAP HEALTH ALTROOT rpool - - - - FAULTED - # zpool status rpool pool: rpool state: UNAVAIL scrub: none requested config: NAME STATE READ WRITE CKSUM rpool UNAVAIL 0 0 0 insufficient replicas mirror UNAVAIL 0 0 0 corrupted data c1t0d0s0 ONLINE 0 0 0 c1t1d0s0 ONLINE 0 0 0 # df -h Filesystem size used avail capacity Mounted on /dev/md/dsk/d10 9.8G 4.3G 5.5G 44% / /devices 0K 0K 0K 0% /devices ctfs 0K 0K 0K 0% /system/contract proc 0K 0K 0K 0% /proc mnttab 0K 0K 0K 0% /etc/mnttab swap 4.2G 1.5M 4.2G 1% /etc/svc/volatile objfs 0K 0K 0K 0% /system/object sharefs 0K 0K 0K 0% /etc/dfs/sharetab fd 0K 0K 0K 0% /dev/fd swap 4.2G 32K 4.2G 1% /tmp swap 4.2G 48K 4.2G 1% /var/run /dev/md/dsk/d50 480M 1.0M 431M 1% /globaldevices /dev/md/dsk/d40 11G 11M 11G 1% /export/home # metastat -c d50 m 512MB d51 d51 s 512MB c1t1d0s5 d40 m 11GB d41 d41 s 11GB c1t1d0s4 d30 m 10GB d31 d31 s 10GB c1t1d0s3 d20 m 2.0GB d21 d21 s 2.0GB c1t1d0s1 d10 m 10GB d11 d11 s 10GB c1t1d0s0 # zpool export rpool internal error: Invalid argument Abort - core dumped # zfs destroy -f rpool internal error: Invalid argument Abort - core dumped # -- This message posted from opensolaris.org _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss